<< Chapter < Page Chapter >> Page >
In this module we provide a proof that sub-Gaussian matrices satisfy the restricted isometry property.

We now show how to exploit the concentration of measure properties of sub-Gaussian distributions to provide a simple proof that sub-Gaussian matrices satisfy the restricted isometry property (RIP). Specifically, we wish to show that by constructing an M × N matrix Φ at random with M sufficiently large, then with high probability there exists a δ K ( 0 , 1 ) such that

( 1 - δ K ) x 2 2 Φ x 2 2 ( 1 + δ K ) x 2 2

holds for all x Σ K (where Σ K denotes the set of all signals x with at most K nonzeros).

We begin by observing that if all we require is that δ 2 K > 0 , then we may set M = 2 K and draw a Φ according to a Gaussian distribution, or indeed any continuous univariate distribution. In this case, with probability 1, any subset of 2 K columns will be linearly independent, and hence all subsets of 2 K columns will be bounded below by 1 - δ 2 K where δ 2 K > 0 . However, suppose we wish to know the constant δ 2 K . In order to find the value of the constant we must consider all possible N K K -dimensional subspaces of R N . From a computational perspective, this is impossible for any realistic values of N and K . Moreover, in light of the lower bounds we described earlier in this course, the actual value of δ 2 K in this case is likely to be very close to 1. Thus, we focus instead on the problem of achieving the RIP of order 2 K for a specified constant δ 2 K .

To ensure that the matrix will satisfy the RIP, we will impose two conditions on the random distribution. First, we require that the distribution is sub-Gaussian. In order to simplify our argument, we will use the simpler results stated in Corollary 2 from "Concentration of measure for sub-Gaussian random variables" , which we briefly recall.

Suppose that Φ is an M × N matrix whose entries φ i j are i.i.d. with φ i j SSub ( 1 / M ) . Let Y = Φ x for x R N . Then for any ϵ > 0 , and any x R N ,

E Y 2 2 = x 2 2

and

P Y 2 2 - x 2 2 ϵ x 2 2 2 exp - M ϵ 2 κ *

with κ * = 2 / ( 1 - log ( 2 ) ) 6 . 52 .

By exploiting this theorem, we assume that the distribution used to construct Φ is strictly sub-Gaussian. This is done simply to yield more concrete constants. The argument could easily be modified to establish a similar result for general sub-Gaussian distributions by instead using Theorem 2 from "Concentration of measure for sub-Gaussian random variables" .

Our second condition is that we require that the distribution yield a matrix that is approximately norm-preserving, which will require that

E ( φ i j 2 ) = 1 M ,

and hence the variance is 1 / M .

We shall now show how the concentration of measure inequality in [link] can be used together with covering arguments to prove the RIP for sub-Gaussian random matrices. Our general approach will be to construct nets of points in each K -dimensional subspace, apply [link] to all of these points through a union bound, and then extend the result from our finite set of points to all possible K -dimensional signals. Thus, in order to prove the result, we will require the following upper bound on the number of points required to construct the nets of points. (For an overview of results similar to [link] and of various related concentration of measure results, we refer the reader to the excellent introduction of  [link] .)

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, An introduction to compressive sensing. OpenStax CNX. Apr 02, 2011 Download for free at http://legacy.cnx.org/content/col11133/1.5
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'An introduction to compressive sensing' conversation and receive update notifications?

Ask