Lindenstrauss

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 321 Experts worldwide ranked by ideXlab platform

Jelani Nelson - One of the best experts on this subject based on the ideXlab platform.

  • Optimality of the Johnson-Lindenstrauss Lemma
    2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS), 2017
    Co-Authors: Kasper Green Larsen, Jelani Nelson
    Abstract:

    For any d, n ≥ 2 and 1/(min{n, d})0.4999

  • optimality of the johnson Lindenstrauss lemma
    arXiv: Information Theory, 2016
    Co-Authors: Kasper Green Larsen, Jelani Nelson
    Abstract:

    For any integers $d, n \geq 2$ and $1/({\min\{n,d\}})^{0.4999} < \varepsilon<1$, we show the existence of a set of $n$ vectors $X\subset \mathbb{R}^d$ such that any embedding $f:X\rightarrow \mathbb{R}^m$ satisfying $$ \forall x,y\in X,\ (1-\varepsilon)\|x-y\|_2^2\le \|f(x)-f(y)\|_2^2 \le (1+\varepsilon)\|x-y\|_2^2 $$ must have $$ m = \Omega(\varepsilon^{-2} \lg n). $$ This lower bound matches the upper bound given by the Johnson-Lindenstrauss lemma [JL84]. Furthermore, our lower bound holds for nearly the full range of $\varepsilon$ of interest, since there is always an isometric embedding into dimension $\min\{d, n\}$ (either the identity map, or projection onto $\mathop{span}(X)$). Previously such a lower bound was only known to hold against linear maps $f$, and not for such a wide range of parameters $\varepsilon, n, d$ [LN16]. The best previously known lower bound for general $f$ was $m = \Omega(\varepsilon^{-2}\lg n/\lg(1/\varepsilon))$ [Wel74, Lev83, Alo03], which is suboptimal for any $\varepsilon = o(1)$.

  • Optimality of the Johnson-Lindenstrauss Lemma
    arXiv: Information Theory, 2016
    Co-Authors: Kasper Green Larsen, Jelani Nelson
    Abstract:

    For any integers $d, n \geq 2$ and $1/({\min\{n,d\}})^{0.4999} < \varepsilon

  • Toward a unified theory of sparse dimensionality reduction in Euclidean space
    Geometric and Functional Analysis, 2015
    Co-Authors: Jean Bourgain, Sjoerd Dirksen, Jelani Nelson
    Abstract:

    Let $${\Phi\in\mathbb{R}^{m\times n}}$$ Φ ∈ R m × n be a sparse Johnson–Lindenstrauss transform (Kane and Nelson in J ACM 61(1):4, 2014 ) with s non-zeroes per column. For a subset T of the unit sphere, $${\varepsilon\in(0,1/2)}$$ ε ∈ ( 0 , 1 / 2 ) given, we study settings for m , s required to ensure $$\mathop{\mathbb{E}}_\Phi \sup_{x\in T}\left|\|\Phi x\|_2^2 - 1\right| < \varepsilon,$$ E Φ sup x ∈ T ‖ Φ x ‖ 2 2 - 1 < ε , i.e. so that $${\Phi}$$ Φ preserves the norm of every $${x\in T}$$ x ∈ T simultaneously and multiplicatively up to $${1+\varepsilon}$$ 1 + ε . We introduce a new complexity parameter, which depends on the geometry of T , and show that it suffices to choose s and m such that this parameter is small. Our result is a sparse analog of Gordon’s theorem, which was concerned with a dense $${\Phi}$$ Φ having i.i.d. Gaussian entries. We qualitatively unify several results related to the Johnson–Lindenstrauss lemma, subspace embeddings, and Fourier-based restricted isometries. Our work also implies new results in using the sparse Johnson–Lindenstrauss transform in numerical linear algebra, classical and model-based compressed sensing, manifold learning, and constrained least squares problems such as the Lasso.

  • the johnson Lindenstrauss lemma is optimal for linear dimensionality reduction
    arXiv: Information Theory, 2014
    Co-Authors: Kasper Green Larsen, Jelani Nelson
    Abstract:

    For any $n>1$ and $0<\varepsilon<1/2$, we show the existence of an $n^{O(1)}$-point subset $X$ of $\mathbb{R}^n$ such that any linear map from $(X,\ell_2)$ to $\ell_2^m$ with distortion at most $1+\varepsilon$ must have $m = \Omega(\min\{n, \varepsilon^{-2}\log n\})$. Our lower bound matches the upper bounds provided by the identity matrix and the Johnson-Lindenstrauss lemma, improving the previous lower bound of Alon by a $\log(1/\varepsilon)$ factor.

Kasper Green Larsen - One of the best experts on this subject based on the ideXlab platform.

  • on using toeplitz and circulant matrices for johnson Lindenstrauss transforms
    Algorithmica, 2020
    Co-Authors: Casper Benjamin Freksen, Kasper Green Larsen
    Abstract:

    The Johnson–Lindenstrauss lemma is one of the cornerstone results in dimensionality reduction. A common formulation of it, is that there exists a random linear mapping \(f : {\mathbb {R}}^n \rightarrow {\mathbb {R}}^m\) such that for any vector \(x \in {\mathbb {R}}^n\), f preserves its norm to within \((1 {\pm } \varepsilon )\) with probability \(1 - \delta \) if \(m = \varTheta (\varepsilon ^{-2} \lg (1/\delta ))\). Much effort has gone into developing fast embedding algorithms, with the Fast Johnson–Lindenstrauss transform of Ailon and Chazelle being one of the most well-known techniques. The current fastest algorithm that yields the optimal \(m = {\mathcal {O}}(\varepsilon ^{-2}\lg (1/\delta ))\) dimensions has an embedding time of \({\mathcal {O}}(n \lg n + \varepsilon ^{-2} \lg ^3 (1/\delta ))\). An exciting approach towards improving this, due to Hinrichs and Vybiral, is to use a random \(m \times n\) Toeplitz matrix for the embedding. Using Fast Fourier Transform, the embedding of a vector can then be computed in \({\mathcal {O}}(n \lg m)\) time. The big question is of course whether \(m = {\mathcal {O}}(\varepsilon ^{-2} \lg (1/\delta ))\) dimensions suffice for this technique. If so, this would end a decades long quest to obtain faster and faster Johnson–Lindenstrauss transforms. The current best analysis of the embedding of Hinrichs and Vybiral shows that \(m = {\mathcal {O}}(\varepsilon ^{-2}\lg ^2 (1/\delta ))\) dimensions suffice. The main result of this paper, is a proof that this analysis unfortunately cannot be tightened any further, i.e., there exist vectors requiring \(m = \varOmega (\varepsilon ^{-2} \lg ^2 (1/\delta ))\) for the Toeplitz approach to work.

  • On Using Toeplitz and Circulant Matrices for Johnson–Lindenstrauss Transforms
    Algorithmica, 2019
    Co-Authors: Casper Benjamin Freksen, Kasper Green Larsen
    Abstract:

    The Johnson–Lindenstrauss lemma is one of the cornerstone results in dimensionality reduction. A common formulation of it, is that there exists a random linear mapping $$f : {\mathbb {R}}^n \rightarrow {\mathbb {R}}^m$$ f : R n → R m such that for any vector $$x \in {\mathbb {R}}^n$$ x ∈ R n , f preserves its norm to within $$(1 {\pm } \varepsilon )$$ ( 1 ± ε ) with probability $$1 - \delta $$ 1 - δ if $$m = \varTheta (\varepsilon ^{-2} \lg (1/\delta ))$$ m = Θ ( ε - 2 lg ( 1 / δ ) ) . Much effort has gone into developing fast embedding algorithms, with the Fast Johnson–Lindenstrauss transform of Ailon and Chazelle being one of the most well-known techniques. The current fastest algorithm that yields the optimal $$m = {\mathcal {O}}(\varepsilon ^{-2}\lg (1/\delta ))$$ m = O ( ε - 2 lg ( 1 / δ ) ) dimensions has an embedding time of $${\mathcal {O}}(n \lg n + \varepsilon ^{-2} \lg ^3 (1/\delta ))$$ O ( n lg n + ε - 2 lg 3 ( 1 / δ ) ) . An exciting approach towards improving this, due to Hinrichs and Vybíral, is to use a random $$m \times n$$ m × n Toeplitz matrix for the embedding. Using Fast Fourier Transform, the embedding of a vector can then be computed in $${\mathcal {O}}(n \lg m)$$ O ( n lg m ) time. The big question is of course whether $$m = {\mathcal {O}}(\varepsilon ^{-2} \lg (1/\delta ))$$ m = O ( ε - 2 lg ( 1 / δ ) ) dimensions suffice for this technique. If so, this would end a decades long quest to obtain faster and faster Johnson–Lindenstrauss transforms. The current best analysis of the embedding of Hinrichs and Vybíral shows that $$m = {\mathcal {O}}(\varepsilon ^{-2}\lg ^2 (1/\delta ))$$ m = O ( ε - 2 lg 2 ( 1 / δ ) ) dimensions suffice. The main result of this paper, is a proof that this analysis unfortunately cannot be tightened any further, i.e., there exist vectors requiring $$m = \varOmega (\varepsilon ^{-2} \lg ^2 (1/\delta ))$$ m = Ω ( ε - 2 lg 2 ( 1 / δ ) ) for the Toeplitz approach to work.

  • on using toeplitz and circulant matrices for johnson Lindenstrauss transforms
    arXiv: Functional Analysis, 2017
    Co-Authors: Casper Benjamin Freksen, Kasper Green Larsen
    Abstract:

    The Johnson-Lindenstrauss lemma is one of the corner stone results in dimensionality reduction. It says that given $N$, for any set of $N$ vectors $X \subset \mathbb{R}^n$, there exists a mapping $f : X \to \mathbb{R}^m$ such that $f(X)$ preserves all pairwise distances between vectors in $X$ to within $(1 \pm \varepsilon)$ if $m = O(\varepsilon^{-2} \lg N)$. Much effort has gone into developing fast embedding algorithms, with the Fast Johnson-Lindenstrauss transform of Ailon and Chazelle being one of the most well-known techniques. The current fastest algorithm that yields the optimal $m = O(\varepsilon^{-2}\lg N)$ dimensions has an embedding time of $O(n \lg n + \varepsilon^{-2} \lg^3 N)$. An exciting approach towards improving this, due to Hinrichs and Vybiral, is to use a random $m \times n$ Toeplitz matrix for the embedding. Using Fast Fourier Transform, the embedding of a vector can then be computed in $O(n \lg m)$ time. The big question is of course whether $m = O(\varepsilon^{-2} \lg N)$ dimensions suffice for this technique. If so, this would end a decades long quest to obtain faster and faster Johnson-Lindenstrauss transforms. The current best analysis of the embedding of Hinrichs and Vybiral shows that $m = O(\varepsilon^{-2}\lg^2 N)$ dimensions suffices. The main result of this paper, is a proof that this analysis unfortunately cannot be tightened any further, i.e., there exists a set of $N$ vectors requiring $m = \Omega(\varepsilon^{-2} \lg^2 N)$ for the Toeplitz approach to work.

  • Optimality of the Johnson-Lindenstrauss Lemma
    2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS), 2017
    Co-Authors: Kasper Green Larsen, Jelani Nelson
    Abstract:

    For any d, n ≥ 2 and 1/(min{n, d})0.4999

  • optimality of the johnson Lindenstrauss lemma
    arXiv: Information Theory, 2016
    Co-Authors: Kasper Green Larsen, Jelani Nelson
    Abstract:

    For any integers $d, n \geq 2$ and $1/({\min\{n,d\}})^{0.4999} < \varepsilon<1$, we show the existence of a set of $n$ vectors $X\subset \mathbb{R}^d$ such that any embedding $f:X\rightarrow \mathbb{R}^m$ satisfying $$ \forall x,y\in X,\ (1-\varepsilon)\|x-y\|_2^2\le \|f(x)-f(y)\|_2^2 \le (1+\varepsilon)\|x-y\|_2^2 $$ must have $$ m = \Omega(\varepsilon^{-2} \lg n). $$ This lower bound matches the upper bound given by the Johnson-Lindenstrauss lemma [JL84]. Furthermore, our lower bound holds for nearly the full range of $\varepsilon$ of interest, since there is always an isometric embedding into dimension $\min\{d, n\}$ (either the identity map, or projection onto $\mathop{span}(X)$). Previously such a lower bound was only known to hold against linear maps $f$, and not for such a wide range of parameters $\varepsilon, n, d$ [LN16]. The best previously known lower bound for general $f$ was $m = \Omega(\varepsilon^{-2}\lg n/\lg(1/\varepsilon))$ [Wel74, Lev83, Alo03], which is suboptimal for any $\varepsilon = o(1)$.

Edo Liberty - One of the best experts on this subject based on the ideXlab platform.

  • an almost optimal unrestricted fast johnson Lindenstrauss transform
    Symposium on Discrete Algorithms, 2011
    Co-Authors: Nir Ailon, Edo Liberty
    Abstract:

    The problems of random projections and sparse reconstruction have much in common and individually received much attention. Surprisingly, until now they progressed in parallel and remained mostly separate. Here, we employ new tools from probability in Banach spaces that were successfully used in the context of sparse reconstruction to advance on an open problem in random pojection. In particular, we generalize and use an intricate result by Rudelson and Veshynin [2008] for sparse reconstruction which uses Dudley’s theorem for bounding Gaussian processes. Our main result states that any set of N = exp(O(n)) real vectors in n dimensional space can be linearly mapped to a space of dimension k = O(log N polylog(n)), while (1) preserving the pairwise distances among the vectors to within any constant distortion and (2) being able to apply the transformation in time O(n log n) on each vector. This improves on the best known bound N = exp(O(n1/2)) achieved by Ailon and Liberty [2009] and N = exp(O(n1/3)) by Ailon and Chazelle [2010]. The dependence in the distortion constant however is suboptimal, and since the publication of an early version of the work, the gap between upper and lower bounds has been considerably tightened obtained by Krahmer and Ward [2011]. For constant distortion, this settles the open question posed by these authors up to a polylog(n) factor while considerably simplifying their constructions.

  • almost optimal unrestricted fast johnson Lindenstrauss transform
    arXiv: Data Structures and Algorithms, 2010
    Co-Authors: Nir Ailon, Edo Liberty
    Abstract:

    The problems of random projections and sparse reconstruction have much in common and individually received much attention. Surprisingly, until now they progressed in parallel and remained mostly separate. Here, we employ new tools from probability in Banach spaces that were successfully used in the context of sparse reconstruction to advance on an open problem in random pojection. In particular, we generalize and use an intricate result by Rudelson and Vershynin for sparse reconstruction which uses Dudley's theorem for bounding Gaussian processes. Our main result states that any set of $N = \exp(\tilde{O}(n))$ real vectors in $n$ dimensional space can be linearly mapped to a space of dimension $k=O(\log N\polylog(n))$, while (1) preserving the pairwise distances among the vectors to within any constant distortion and (2) being able to apply the transformation in time $O(n\log n)$ on each vector. This improves on the best known $N = \exp(\tilde{O}(n^{1/2}))$ achieved by Ailon and Liberty and $N = \exp(\tilde{O}(n^{1/3}))$ by Ailon and Chazelle. The dependence in the distortion constant however is believed to be suboptimal and subject to further investigation. For constant distortion, this settles the open question posed by these authors up to a $\polylog(n)$ factor while considerably simplifying their constructions.

Daniel M Kane - One of the best experts on this subject based on the ideXlab platform.

  • sparser johnson Lindenstrauss transforms
    Journal of the ACM, 2014
    Co-Authors: Daniel M Kane, Jelani Nelson
    Abstract:

    We give two different and simple constructions for dimensionality reduction in e2 via linear mappings that are sparse: only an O(e)-fraction of entries in each column of our embedding matrices are non-zero to achieve distortion 1 + e with high probability, while still achieving the asymptotically optimal number of rows. These are the first constructions to provide subconstant sparsity for all values of parameters, improving upon previous works of Achlioptas [2003] and Dasgupta et al. [2010]. Such distributions can be used to speed up applications where e2 dimensionality reduction is used.

  • sparser johnson Lindenstrauss transforms
    Symposium on Discrete Algorithms, 2012
    Co-Authors: Daniel M Kane, Jelani Nelson
    Abstract:

    We give two different Johnson-Lindenstrauss distributions, each with column sparsity s = Θ(e−1 log(1/δ)) and embedding into optimal dimension k = O(e−2 log(1/δ)) to achieve distortion 1±e with probability 1−δ. That is, only an O(e)-fraction of entries are non-zero in each embedding matrix in the supports of our distributions. These are the first distributions to provide o(k) sparsity for all values of e,δ. Previously the best known construction obtained s = Θ(e-1 log2 (1/δ))1 [Dasgupta-Kumar-Sarlos, STOC 2010]2. In addition, one of our distributions can be sampled from a seed of O(log(1/δ) log d) uniform random bits. Some applications that use Johnson-Lindenstrauss embeddings as a black box, such as those in approximate numerical linear algebra ([Sarlos, FOCS 2006], [Clarkson-Woodruff, STOC 2009]), require exponentially small δ. Our linear dependence on log(1/δ) in the sparsity is thus crucial in these applications to obtain speedup.

  • SODA - Sparser Johnson-Lindenstrauss transforms
    2012
    Co-Authors: Daniel M Kane, Jelani Nelson
    Abstract:

    We give two different Johnson-Lindenstrauss distributions, each with column sparsity s = Θ(e−1 log(1/δ)) and embedding into optimal dimension k = O(e−2 log(1/δ)) to achieve distortion 1±e with probability 1−δ. That is, only an O(e)-fraction of entries are non-zero in each embedding matrix in the supports of our distributions. These are the first distributions to provide o(k) sparsity for all values of e,δ. Previously the best known construction obtained s = Θ(e-1 log2 (1/δ))1 [Dasgupta-Kumar-Sarlos, STOC 2010]2. In addition, one of our distributions can be sampled from a seed of O(log(1/δ) log d) uniform random bits. Some applications that use Johnson-Lindenstrauss embeddings as a black box, such as those in approximate numerical linear algebra ([Sarlos, FOCS 2006], [Clarkson-Woodruff, STOC 2009]), require exponentially small δ. Our linear dependence on log(1/δ) in the sparsity is thus crucial in these applications to obtain speedup.

  • almost optimal explicit johnson Lindenstrauss families
    International Workshop and International Workshop on Approximation Randomization and Combinatorial Optimization. Algorithms and Techniques, 2011
    Co-Authors: Daniel M Kane, Raghu Meka, Jelani Nelson
    Abstract:

    The Johnson-Lindenstrauss lemma is a fundamental result in probability with several applications in the design and analysis of algorithms. Constructions of linear embeddings satisfying the Johnson-Lindenstrauss property necessarily involve randomness and much attention has been given to obtain explicit constructions minimizing the number of random bits used. In this work we give explicit constructions with an almost optimal use of randomness: For 0 e] ≤ δ, with seed-length r = O(log d + log(1/δ) ċ log (log(1/δ/e)). In particular, for δ = 1/ poly(d) and fixed e > 0, we obtain seed-length O((log d)(log log d)). Previous constructions required Ω(log2 d) random bits to obtain polynomially small error. We also give a new elementary proof of the optimality of the JL lemma showing a lower bound of Ω(log(1/δ)/e2) on the embedding dimension. Previously, Jayram and Woodruff [10] used communication complexity techniques to show a similar bound.

  • APPROX-RANDOM - Almost optimal explicit Johnson-Lindenstrauss families
    Approximation Randomization and Combinatorial Optimization. Algorithms and Techniques, 2011
    Co-Authors: Daniel M Kane, Raghu Meka, Jelani Nelson
    Abstract:

    The Johnson-Lindenstrauss lemma is a fundamental result in probability with several applications in the design and analysis of algorithms. Constructions of linear embeddings satisfying the Johnson-Lindenstrauss property necessarily involve randomness and much attention has been given to obtain explicit constructions minimizing the number of random bits used. In this work we give explicit constructions with an almost optimal use of randomness: For 0 e] ≤ δ, with seed-length r = O(log d + log(1/δ) ċ log (log(1/δ/e)). In particular, for δ = 1/ poly(d) and fixed e > 0, we obtain seed-length O((log d)(log log d)). Previous constructions required Ω(log2 d) random bits to obtain polynomially small error. We also give a new elementary proof of the optimality of the JL lemma showing a lower bound of Ω(log(1/δ)/e2) on the embedding dimension. Previously, Jayram and Woodruff [10] used communication complexity techniques to show a similar bound.

Nir Ailon - One of the best experts on this subject based on the ideXlab platform.

  • an almost optimal unrestricted fast johnson Lindenstrauss transform
    Symposium on Discrete Algorithms, 2011
    Co-Authors: Nir Ailon, Edo Liberty
    Abstract:

    The problems of random projections and sparse reconstruction have much in common and individually received much attention. Surprisingly, until now they progressed in parallel and remained mostly separate. Here, we employ new tools from probability in Banach spaces that were successfully used in the context of sparse reconstruction to advance on an open problem in random pojection. In particular, we generalize and use an intricate result by Rudelson and Veshynin [2008] for sparse reconstruction which uses Dudley’s theorem for bounding Gaussian processes. Our main result states that any set of N = exp(O(n)) real vectors in n dimensional space can be linearly mapped to a space of dimension k = O(log N polylog(n)), while (1) preserving the pairwise distances among the vectors to within any constant distortion and (2) being able to apply the transformation in time O(n log n) on each vector. This improves on the best known bound N = exp(O(n1/2)) achieved by Ailon and Liberty [2009] and N = exp(O(n1/3)) by Ailon and Chazelle [2010]. The dependence in the distortion constant however is suboptimal, and since the publication of an early version of the work, the gap between upper and lower bounds has been considerably tightened obtained by Krahmer and Ward [2011]. For constant distortion, this settles the open question posed by these authors up to a polylog(n) factor while considerably simplifying their constructions.

  • almost optimal unrestricted fast johnson Lindenstrauss transform
    arXiv: Data Structures and Algorithms, 2010
    Co-Authors: Nir Ailon, Edo Liberty
    Abstract:

    The problems of random projections and sparse reconstruction have much in common and individually received much attention. Surprisingly, until now they progressed in parallel and remained mostly separate. Here, we employ new tools from probability in Banach spaces that were successfully used in the context of sparse reconstruction to advance on an open problem in random pojection. In particular, we generalize and use an intricate result by Rudelson and Vershynin for sparse reconstruction which uses Dudley's theorem for bounding Gaussian processes. Our main result states that any set of $N = \exp(\tilde{O}(n))$ real vectors in $n$ dimensional space can be linearly mapped to a space of dimension $k=O(\log N\polylog(n))$, while (1) preserving the pairwise distances among the vectors to within any constant distortion and (2) being able to apply the transformation in time $O(n\log n)$ on each vector. This improves on the best known $N = \exp(\tilde{O}(n^{1/2}))$ achieved by Ailon and Liberty and $N = \exp(\tilde{O}(n^{1/3}))$ by Ailon and Chazelle. The dependence in the distortion constant however is believed to be suboptimal and subject to further investigation. For constant distortion, this settles the open question posed by these authors up to a $\polylog(n)$ factor while considerably simplifying their constructions.