Restricted isometry properties, a.k.a. uniform uncertainty principles (E. Candès and Tao 2005; E. J. Candès, Romberg, and Tao 2006), mutual incoherence (David L. Donoho 2006; D. L. Donoho, Elad, and Temlyakov 2006), irrepresentability conditions (Zhao and Yu 2006)…

This is mostly notes while I learn some definitions; expect no actual thoughts.

Recoverability conditions, as seen in sparse regression, sparse basis dictionaries, function approximation, compressed sensing etc. If you squint right you could imagine these uncertainty principles for a sparse world, or a the foundations of a particular type of sampling theory.

Terry Tao mentions the various related conditions for the compressed sensing problem, and which types of random matrices satisfy them.

## Restricted Isometry

The compressed sensing formulation.

The chatty lecture notes on uniform uncertainty look fun.

The restricted isometry constant of a matrix \(A\), is the smallest constant \(\delta_s\) \((1-\delta_s(A))\|x\|_2^2\leq \|Ax\|_2^2\leq (1+\delta_s(A))\|x\|_2^2\) for all \(s\)-sparse \(x\). That is, the measurement matrix does not change the norm of sparse signals “too much”, and in particular, does not null them when \(\delta_s < 1.\)

## Irrepresentability

The set up is a little different for regression-type problems,
which is where “representability” comes from.
Here we care also about the *design*, roughly, the dependence of covariates we actually observe, and the noise distribution.

Zhao and Yu (2006) present an abstract condition called strong irrepresentability,
which guarantees asymptotic sign consistency of selection.
See also Meinshausen and Bühlmann (2006), who call this *neighborhood stability*,
which is even less catchy.

More recently Meinshausen and Yu (2009) extend this (and explain the original irrepresentability more clearly IMO):

Here we examine the behavior of the Lasso estimators if the irrepresentable condition is relaxed. Even though the Lasso cannot recover the correct sparsity pattern, we show that the estimator is still consistent in the l2-norm sense for fixed designs under conditions on (a) the number \(s_n\) of nonzero components of the vector \(\beta_n\) and (b) the minimal singular values of design matrices that are induced by selecting small subsets of variables. Furthermore, a rate of convergence result is obtained on the l2 error with an appropriate choice of the smoothing parameter.

They do a good gob of uniting prediction-error and model-selection consistency
approaches. In fact, I will base *everything* off Meinshausen and Yu (2009), since not only is
the prose lucid,
it gives the background to the design assumptions and relaxation of coherence.

TBC.

## Incoherence

A Basis-Pursuit noise-free setting.

D. L. Donoho, Elad, and Temlyakov (2006):

We can think of the atoms in our dictionary as columns in a matrix \(\Phi\), so that \(\Phi\) is \(n\) by \(m\) and \(m > n.\). A representation of \(y\in\mathbb{R}^n\) can be thought of as a vector \(\alpha\in\mathbb{R}^m\) satisfying \(y=\Phi\alpha.\)

The concept of mutual coherence of the dictionary […] is defined, assuming that the columns of are normalized to unit \(\ell^2\)-norm, in terms of the Gram matrix \(G=\Phi^T\Phi\). With \(G(k,j)\) denoting entries of this matrix, the mutual coherence is

\[ M(\Phi) = \max_{1\leq k, j\leq m, k\neq j} |G(k,j)| \]

A dictionary is

incoherentif \(M\) is small.

## Frame theory

See frames.

## References

*Compressed Sensing and Its Applications: MATHEON Workshop 2013*, edited by Holger Boche, Robert Calderbank, Gitta Kutyniok, and Jan Vybíral, 143–67. Applied and Numerical Harmonic Analysis. Cham: Springer International Publishing.

*IEEE Transactions on Information Theory*56 (4): 1982–2001.

*Constructive Approximation*28 (3): 253–63.

*The Annals of Statistics*36 (1): 64–94.

*arXiv:0805.0149 [Cs]*, May.

*Applied and Computational Harmonic Analysis*6 (2): 197–218.

*Applied and Computational Harmonic Analysis*31 (1): 59–73.

*Communications on Pure and Applied Mathematics*59 (8): 1207–23.

*IEEE Transactions on Information Theory*52 (12): 5406–25.

*IEEE Transactions on Information Theory*51 (12): 4203–15.

*An Introduction to Frames and Riesz Bases*. Second edtion. Applied and Numerical Harmonic Analysis. Cham: Springer International Publishing.

*IEEE Transactions on Information Theory*36 (5): 961–1005.

*Ten lectures on wavelets*. Philadelphia, Pa: Society for Industrial and Applied Mathematics (SIAM, 3600 Market Street, Floor 6, Philadelphia, PA 19104).

*Communications on Pure and Applied Mathematics*63 (1): 1–38.

*IEEE Transactions on Information Theory*52 (1): 6–18.

*IEEE Transactions on Information Theory*52 (4): 1289–1306.

*Proceedings of the National Academy of Sciences*100 (5): 2197–2202.

*Transactions of the American Mathematical Society*72 (2): 341–66.

*New Journal of Physics*14 (9): 095022.

*arXiv:1108.0373 [Math, Stat]*, August.

*arXiv:1403.7023 [Math, Stat]*. Vol. 131.

*IEEE Transactions on Information Theory*58 (12): 7204–14.

*SIAM Review*31 (4): 628–66.

*arXiv:1704.02107 [Stat]*, April.

*An Introduction to Frames*. Vol. 2.

*Computational Statistics*29 (3-4): 547–68.

*arXiv:2004.11554 [Stat]*, April.

*The Annals of Statistics*34 (3): 1436–62.

*The Annals of Statistics*37 (1): 246–70.

*Foundations of Computational Mathematics*9 (3): 317–34.

*2*11 (Aug): 2241–59.

*Electronic Journal of Statistics*5: 935–80.

*Compressed Sensing & Sparse Filtering*, edited by Avishy Y. Carmi, Lyudmila Mihaylova, and Simon J. Godsill, 77–93. Signals and Communication Technology. Springer Berlin Heidelberg.

*arXiv:1702.05181 [Cs, Stat]*, February.

*IEEE Signal Processing Letters*20 (4): 403–6.

*An Introduction to Non-Harmonic Fourier Series, Revised Edition, 93*. Academic Press.

*The Annals of Statistics*38 (2): 894–942.

*Journal of Machine Learning Research*7 (Nov): 2541–63.

## No comments yet. Why not leave one?