Abstract
Recovery procedures in Data Science are often based on stable point separation. In its simplest form, stable point separation implies that if f is “far away” from 0, and one is given a random sample \((f(Z_i))_{i=1}^m\) where a proportional number of the sample points may be corrupted by noise—even maliciously, that information is still enough to exhibit that f is far from 0.
Stable point separation is well understood in the context of iid sampling, and to explore it for general sampling methods we introduce a new notion—the coordinate small-ball of a random vector X. Roughly put, this feature captures the number of “relatively large coordinates” of \((|\left \langle TX,u_i \right \rangle |)_{i=1}^m\), where \(T:\mathbb {R}^n \to \mathbb {R}^m\) is an arbitrary linear operator and \((u_i)_{i=1}^m\) is any fixed orthonormal basis of \(\mathbb {R}^m\).
We show that under the bare-minimum assumptions on X, and with high probability, many of the values \(|\left \langle TX,u_i \right \rangle |\) are at least of the order \(\|T\|{ }_{S_2}/\sqrt {m}\). As a result, the “coordinate structure” of TX exhibits the typical Euclidean norm of TX and does so in a stable way.
One outcome of our analysis is that random sub-sampled convolutions satisfy stable point separation under minimal assumptions on the generating random vector—a fact that was known previously only in a highly restrictive setup, namely, for random vectors with iid subgaussian coordinates. As an application we address the problem of sparse signal recovery using a circulant matrix when a proportion of the given sample is corrupted by malicious noise.
Supported by NSF grant DMS-1812240 and Simons Fellows in Mathematics Award 823432.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Recall that X is log-concave if it has a density \( f_{X}\) that satisfies that for every \( x, y\) in the support of \( f_{X}\) and every \(0 \leq \lambda \leq 1\), \(f_{X} ((1- \lambda ) x+ \lambda y )\geq f_{X}^{(1-\lambda )} (x) f_{X}^{\lambda } (y)\).
- 2.
A centred random variable x is L-subgaussian if for every \(p \geq 2\), \(\|x\|{ }_{L_p} \leq L \sqrt {p}\|x\|{ }_{L_2}\).
- 3.
We say that a vector is s-sparse if it is supported on a set of cardinality no more than s.
References
K. Ball, Logarithmically concave functions and sections of convex sets in \({\mathbf {R}}^n\). Stud. Math. 88(1), 69–84 (1988)
J. Bourgain, On the distribution of polynomials on high-dimensional convex sets, in Geometric Aspects of Functional Analysis (1989–90). Lecture Notes in Mathematics, vol. 1469 (Springer, Berlin, 1991), pp. 127–137
J. Bourgain, On the isotropy-constant problem for “PSI-2”-bodies, in Geometric Aspects of Functional Analysis. Lecture Notes in Mathematics, vol. 1807 (Springer, Berlin, 2003), pp. 114–121
S. Brazitikos, A. Giannopoulos, P. Valettas, B.-H. Vritsiou, Geometry of Isotropic Convex Bodies. Mathematical Surveys and Monographs, vol. 196 (American Mathematical Society, Providence, 2014)
D. Cordero-Erausquin, M. Fradelizi, B. Maurey, The (B) conjecture for the Gaussian measure of dilates of symmetric convex sets and related problems. J. Funct. Anal. 214(2), 410–427 (2004)
S. Foucart, H. Rauhut, A Mathematical Introduction to Compressive Sensing. Applied and Numerical Harmonic Analysis (Birkhäuser/Springer, New York, 2013)
B. Klartag, On convex perturbations with a bounded isotropic constant. Geom. Funct. Anal. 16(6), 1274–1290 (2006)
B. Klartag, R. Vershynin, Small ball probability and Dvoretzky’s theorem. Isr. J. Math. 157, 193–207 (2007)
R. Latal, K. Oleszkiewicz, Small ball probability estimates in terms of widths. Stud. Math. 169(3), 305–314 (2005)
R. Latala, P. Mankiewicz, K. Oleszkiewicz, N. Tomczak-Jaegermann, Banach-Mazur distances and projections on random subgaussian polytopes. Discrete Comput. Geom. 38(1), 29–50 (2007)
G. Livshyts, G. Paouris, P. Pivovarov, On sharp bounds for marginal densities of product measures. Isr. J. Math. 216(2), 877–889 (2016)
G. Lugosi, S. Mendelson, Near-optimal mean estimators with respect to general norms. Probab. Theory Relat. Fields 175(3–4), 957–973 (2019)
G. Lugosi, S. Mendelson, Sub-Gaussian estimators of the mean of a random vector. Ann. Stat. 47(2), 783–794 (2019)
G. Lugosi, S. Mendelson, Risk minimization by median-of-means tournaments. J. Eur. Math. Soc. 22(3), 925–965 (2020)
S. Mendelson, Learning without concentration. J. ACM 62(3), Article 21, 25 (2015)
S. Mendelson, An unrestricted learning procedure. J. ACM 66(6), Article 42, 42 (2019)
S. Mendelson, Approximating the covariance ellipsoid. Commun. Contemp. Math. 22(8), 1950089, 24 (2020)
S. Mendelson, N. Zhivotovskiy, Robust covariance estimation under \(L_4\)-\(L_2\) norm equivalence. Ann. Stat. 48(3), 1648–1664 (2020)
S. Mendelson, H. Rauhut, R. Ward, Improved bounds for sparse recovery from subsampled random convolutions. Ann. Appl. Probab. 28(6), 3491–3527 (2018)
S. Mendelson, E. Milman, G. Paouris, Generalized dual Sudakov minoration via dimension-reduction—a program. Stud. Math. 244(2), 159–202 (2019)
A. Naor, P. Youssef, Restricted invertibility revisited, in A Journey Through Discrete Mathematics (Springer, Cham, 2017), pp. 657–691
G. Paouris, On the isotropic constant of marginals. Stud. Math. 212(3), 219–236 (2012)
G. Paouris. Small ball probability estimates for log-concave measures. Trans. Am. Math. Soc. 364(1), 287–308 (2012)
G. Paouris, P. Pivovarov, Randomized isoperimetric inequalities, in Convexity and Concentration. IMA Vol. Mathematics Applied, vol. 161 (Springer, New York, 2017), pp. 391–425
J. Romberg, Compressive sensing by random convolution. SIAM J. Imaging Sci. 2(4), 1098–1128 (2009)
M. Rudelson, R. Vershynin, Small ball probabilities for linear images of high-dimensional distributions. Int. Math. Res. Not. IMRN (19), 9594–9617 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendices
A Examples of Vectors That Satisfy The SBA
Here we give examples of several generic random vectors that satisfy the SBA. This is far from being an exhaustive list and should be viewed only as an indication to the fact that the SBA is a property shared by many natural random vectors.
-
(1)
Let \(X= (\xi _{1}, \cdots , \xi _{n})\) where the \( \xi _{i}\)’s are independent random variables with densities bounded by \( {\mathcal {L}}\). It was shown in [26] that X satisfies the SBA with constant \( c{\mathcal {L}}\), where \( c>0\) is an absolute constant.
This fact was further extended in [11, 24]; most notably, it was shown in [24] that if the coordinates of \(X=(\xi _i)_{i=1}^n\) are independent random variables with densities bounded by 1 and the coordinates of \(Y=(\eta _i)_{i=1}^n\) are uniformly distributed in \([-\frac {1}{2} , \frac {1}{2}]\), then for every semi-norm \( \| \cdot \|\) and \( t>0\),
$$\displaystyle \begin{aligned} {} \mathbb P ( \|X\|\leq t ) \leq \mathbb P ( \| Y \|\leq t ). \end{aligned} $$(A.1)In particular, among all such vectors the ‘worse’ small-ball behaviour—with respect to any semi-norm—is exhibited by the uniform measure on the cube \([-\frac {1}{2} , \frac {1}{2}]^n\).
Observe that for the Euclidean norm, the small-ball behaviour of Y and of the standard Gaussian vector G is the same up to absolute constants.
-
(2)
Perturbations: It is standard to verify that if X satisfies the SBA with a constant \(\mathcal {L}\) and W is an arbitrary random vector that is independent of X, then \( W+\delta X\) satisfies SBA with a constant depending on \( \delta \) and \(\mathcal {L}\).
-
(3)
The question of whether there is a constant \(\mathcal {L}\) such that any isotropic log-concave random vector satisfies the SBA with constant \(\mathcal {L}\) is equivalent to Bourgain’s celebrated Hyperplane Conjecture (see [2] and the discussion in [7] and [4])).
Thanks to the extensive study of log-concave measures and the connection the SBA has with the Hyperplane conjecture for such measures, there are some important examples of isotropic, log-concave random vectors that are known to satisfy the SBA with an absolute constant:
-
If X is also 1-unconditional (see [20], section 8.2);
-
If X is also supergaussian (this follows from results of [22]).
B Proof of Remark 1.14
The proof requires some additional notation. Let X be a random vector in \( \mathbb {R}^{n}\) and let \(p\geq 1\). The \(Z_{p}\) body of X is defined as the (centrally-symmetric) convex body whose support function is
It is straightforward to verify that if \( T :\mathbb {R}^{n} \to \mathbb {R}^{m}\) is a linear operator then
Lemma B.1
There are absolute constants \(c_1\) and \(c_2\) for which the following holds. Let \( X\) be a centred log-concave random vector in \(\mathbb {R}^{n}\) that satisfies the SBA with constant \( {\mathcal {L}}\) . For any \( T\in GL_{n}\) and \( F\in \mathcal {G}_{n,k}\) one has
where the left-hand side holds true under the additional assumption that X is isotropic.
The proof of Lemma B.1 is based on two facts. The first is a standard observation from linear algebra: let \(T: \mathbb {R}^{n} \to \mathbb {R}^{k}\), set \(E= \mathrm {ker}(T)^{\perp }= \mathrm {im} (T^{\ast } ) \) and denote by \( T|{ }_{E}\) the restriction of T to E. Then for any compact set \(K \subset \mathbb {R}^n\),
The second observation is Proposition 3.7 from [23]: If X is a centred, log-concave random vector then
Proof of Lemma B.1
By the Prekopá-Leindler inequality, for every linear operator S, the random vector SX is also log-concave and centred. Hence, using (B.2), (B.5) and (B.4), it is evident that
Clearly, \( f_{P_{E}X}^{\frac {1}{k}}(0) \leq {\mathcal {L}}\), which proved the right-hand side inequality in (B.3). Moreover if X is an isotropic log-concave random vector in \(\mathbb {R}^{n}\) then \(f_{X}^{\frac {1}{n}} (0 ) \geq c \), where c is an absolute constant (see, e.g. [1]). And since \( P_{F} X\) is also isotropic when X is, the left-hand side inequality in (B.3) follows. □
Combining (B.3) and (2.1) it is evident that:
Proposition B.2
There are absolute constants \(c_1\) and \(c_2\) for which the following holds. Let X be an isotropic log-concave random vector in \(\mathbb {R}^{n}\) that satisfies the SBA with constant \( {\mathcal {L}}\) and let \( T: \mathbb {R}^{n} \to \mathbb {R}^{m}\) be a linear operator. Then
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Mendelson, S., Paouris, G. (2023). Stable Recovery and the Coordinate Small-Ball Behaviour of Random Vectors. In: Eldan, R., Klartag, B., Litvak, A., Milman, E. (eds) Geometric Aspects of Functional Analysis. Lecture Notes in Mathematics, vol 2327. Springer, Cham. https://doi.org/10.1007/978-3-031-26300-2_8
Download citation
DOI: https://doi.org/10.1007/978-3-031-26300-2_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-26299-9
Online ISBN: 978-3-031-26300-2
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)