No Arabic abstract
Random subspaces $X$ of $mathbb{R}^n$ of dimension proportional to $n$ are, with high probability, well-spread with respect to the $ell_p$-norm (for $p in [1,2]$). Namely, every nonzero $x in X$ is robustly non-sparse in the following sense: $x$ is $varepsilon |x|_p$-far in $ell_p$-distance from all $delta n$-sparse vectors, for positive constants $varepsilon, delta$ bounded away from $0$. This $ell_p$-spread property is the natural counterpart, for subspaces over the reals, of the minimum distance of linear codes over finite fields, and, for $p = 2$, corresponds to $X$ being a Euclidean section of the $ell_1$ unit ball. Explicit $ell_p$-spread subspaces of dimension $Omega(n)$, however, are not known except for $p=1$. The construction for $p=1$, as well as the best known constructions for $p in (1,2]$ (which achieve weaker spread properties), are analogs of low density parity check (LDPC) codes over the reals, i.e., they are kernels of sparse matrices. We study the spread properties of the kernels of sparse random matrices. Rather surprisingly, we prove that with high probability such subspaces contain vectors $x$ that are $o(1)cdot |x|_2$-close to $o(n)$-sparse with respect to the $ell_2$-norm, and in particular are not $ell_2$-spread. On the other hand, for $p < 2$ we prove that such subspaces are $ell_p$-spread with high probability. Moreover, we show that a random sparse matrix has the stronger restricted isometry property (RIP) with respect to the $ell_p$ norm, and this follows solely from the unique expansion of a random biregular graph, yielding a somewhat unexpected generalization of a similar result for the $ell_1$ norm [BGI+08]. Instantiating this with explicit expanders, we obtain the first explicit constructions of $ell_p$-spread subspaces and $ell_p$-RIP matrices for $1 leq p < p_0$, where $1 < p_0 < 2$ is an absolute constant.
We determine the rank of a random matrix over an arbitrary field with prescribed numbers of non-zero entries in each row and column. As an application we obtain a formula for the rate of low-density parity check codes. This formula vindicates a conjecture of Lelarge (2013). The proofs are based on coupling arguments and a novel random perturbation, applicable to any matrix, that diminishes the number of short linear relations.
We characterize the approximate monomial complexity, sign monomial complexity , and the approximate L 1 norm of symmetric functions in terms of simple combinatorial measures of the functions. Our characterization of the approximate L 1 norm solves the main conjecture in [AFH12]. As an application of the characterization of the sign monomial complexity, we prove a conjecture in [ZS09] and provide a characterization for the unbounded-error communication complexity of symmetric-xor functions.
This paper was removed due to an error in the proof (Claim 4.12 as stated is not true). The authors would like to thank Ilya Volkovich for pointing out a counterexample to this papers main result in positive characteristic: If $F$ is a field with prime characteristic $p$, then the polynomial $x_1^p + x_2^p + ldots + x^n^p$ has the following factor: $(x_1+x_2+ ldots + x_n)^{p-1}$, which has sparsity $n^p$.
Principal Component Analysis (PCA) is a powerful tool in statistics and machine learning. While existing study of PCA focuses on the recovery of principal components and their associated eigenvalues, there are few precise characterizations of individual principal component scores that yield low-dimensional embedding of samples. That hinders the analysis of various spectral methods. In this paper, we first develop an $ell_p$ perturbation theory for a hollowed version of PCA in Hilbert spaces which provably improves upon the vanilla PCA in the presence of heteroscedastic noises. Through a novel $ell_p$ analysis of eigenvectors, we investigate entrywise behaviors of principal component score vectors and show that they can be approximated by linear functionals of the Gram matrix in $ell_p$ norm, which includes $ell_2$ and $ell_infty$ as special examples. For sub-Gaussian mixture models, the choice of $p$ giving optimal bounds depends on the signal-to-noise ratio, which further yields optimality guarantees for spectral clustering. For contextual community detection, the $ell_p$ theory leads to a simple spectral algorithm that achieves the information threshold for exact recovery. These also provide optimal recovery results for Gaussian mixture and stochastic block models as special cases.
Accurate estimation of tail probabilities of projections of high-dimensional probability measures is of relevance in high-dimensional statistics and asymptotic geometric analysis. For fixed $p in (1,infty)$, let $(X^{(n,p)})$ and $(theta^n)$ be independent sequences of random vectors with $theta^n$ distributed according to the normalized cone measure on the unit $ell_2^n$ sphere, and $X^{(n,p)}$ distributed according to the normalized cone measure on the unit $ell_p^n$ sphere. For almost every sequence of projection directions $(theta^n)$, (quenched) sharp large deviation estimates are established for suitably normalized (scalar) projections of $X^{n,p}$ onto $theta^n$, that are asymptotically exact (as the dimension $n$ tends to infinity). Furthermore, the case when $(X^{(n,p)})$ is replaced with $(mathscr{X}^{(n,p)})$, where $mathscr{X}^{(n,p)}$ is distributed according to the uniform (or normalized volume) measure on the unit $ell_p^n$ ball, is also considered. In both cases, in contrast to the (quenched) large deviation rate function, the prefactor exhibits a dependence on the projection directions $(theta^n)$ that encodes geometric information. Moreover, although the (quenched) large deviation rate functions for the sequences of random projections of $(X^{(n,p)})$ and $(mathscr{X}^{(n,p)})$ are known to coincide, it is shown that the prefactor distinguishes between these two cases. The results on the one hand provide quantitative estimates of tail probabilities of random projections of $ell_p^n$ balls and spheres, valid for finite $n$, generalizing previous results due to Gantert, Kim and Ramanan, and on the other hand, generalize classical sharp large deviation estimates in the spirit of Bahadur and Ranga Rao to a geometric setting.