<< Chapter < Page Chapter >> Page >
In this module we introduce coherence, which provides a more computationally friendly alternative to conditions such as the spark, NSP, or RIP. We briefly describe the theoretical relationship between these conditions.

While the spark , null space property (NSP), and restricted isometry property (RIP) all provide guarantees for the recovery of sparse signals, verifying that a general matrix Φ satisfies any of these properties has a combinatorial computational complexity, since in each case one must essentially consider N K submatrices. In many settings it is preferable to use properties of Φ that are easily computable to provide more concrete recovery guarantees. The coherence of a matrix is one such property  [link] , [link] .

The coherence of a matrix Φ , μ ( Φ ) , is the largest absolute inner product between any two columns φ i , φ j of Φ :

μ ( Φ ) = max 1 i < j N φ i , φ j φ i 2 φ j 2 .

It is possible to show that the coherence of a matrix is always in the range μ ( Φ ) N - M M ( N - 1 ) , 1 ; the lower bound is known as the Welch bound  [link] , [link] , [link] . Note that when N M , the lower bound is approximately μ ( Φ ) 1 / M .

One can sometimes relate coherence to the spark, NSP, and RIP. For example, the coherence and spark properties of a matrix can be related by employing the Gershgorin circle theorem  [link] , [link] .

(theorem 2 of [link] )

The eigenvalues of an N × N matrix M with entries m i j , 1 i , j N , lie in the union of N discs d i = d i ( c i , r i ) , 1 i N , centered at c i = m i i and with radius r i = j i | m i j | .

Applying this theorem on the Gram matrix G = Φ Λ T Φ Λ leads to the following straightforward result.

For any matrix Φ ,

spark ( Φ ) 1 + 1 μ ( Φ ) .

Since spark ( Φ ) does not depend on the scaling of the columns, we can assume without loss of generality that Φ has unit-norm columns. Let Λ { 1 , ... , N } with | Λ | = p determine a set of indices. We consider the restricted Gram matrix G = Φ Λ T Φ Λ , which satisfies the following properties:

  • g i i = 1 , 1 i p ;
  • | g i j | μ ( Φ ) , 1 i , j p , i j .

From [link] , if j i | g i j | < | g i i | then the matrix G is positive definite, so that the columns of Φ Λ are linearly independent. Thus, the spark condition implies ( p - 1 ) μ ( Φ ) < 1 or, equivalently, p < 1 + 1 / μ ( Φ ) for all p < spark ( Φ ) , yielding spark ( Φ ) 1 + 1 / μ ( Φ ) .

By merging Theorem 1 from "Null space conditions" with [link] , we can pose the following condition on Φ that guarantees uniqueness.

(theorem 12 of [link] )

If

K < 1 2 1 + 1 μ ( Φ ) ,

then for each measurement vector y R M there exists at most one signal x Σ K such that y = Φ x .

[link] , together with the Welch bound, provides an upper bound on the level of sparsity K that guarantees uniqueness using coherence: K = O ( M ) . Another straightforward application of the Gershgorin circle theorem ( [link] ) connects the RIP to the coherence property.

If Φ has unit-norm columns and coherence μ = μ ( Φ ) , then Φ satisfies the RIP of order K with δ = ( K - 1 ) μ for all K < 1 / μ .

The proof of this lemma is similar to that of [link] .

The results given here emphasize the need for small coherence μ ( Φ ) for the matrices used in CS. Coherence bounds have been studied both for deterministic and randomized matrices. For example, there are known matrices Φ of size M × M 2 that achieve the coherence lower bound μ ( Φ ) = 1 / M , such as the Gabor frame generated from the Alltop sequence  [link] and more general equiangular tight frames  [link] . These constructions restrict the number of measurements needed to recover a K -sparse signal to be M = O ( K 2 log N ) . Furthermore, it can be shown that when the distribution used has zero mean and finite variance, then in the asymptotic regime (as M and N grow) the coherence converges to μ ( Φ ) = ( 2 log N ) / M   [link] , [link] , [link] . Such constructions would allow K to grow asymptotically as M = O ( K 2 log N ) , matching the known finite-dimensional bounds.

The measurement bounds dependent on coherence are handicapped by the squared dependence on the sparsity K , but it is possible to overcome this bottleneck by shifting the types of guarantees from worst-case/deterministic behavior, to average-case/probabilistic behavior  [link] , [link] . In this setting, we pose a probabilistic prior on the set of K -sparse signals x Σ K . It is then possible to show that if Φ has low coherence μ ( Φ ) and spectral norm Φ 2 , and if K = O ( μ - 2 ( Φ ) log N ) , then the signal x can be recovered from its CS measurements y = Φ x with high probability. Note that if we replace the Welch bound, then we obtain K = O ( M log N ) , which returns to the linear dependence of the measurement bound on the signal sparsity that appears in RIP-based results.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, An introduction to compressive sensing. OpenStax CNX. Apr 02, 2011 Download for free at http://legacy.cnx.org/content/col11133/1.5
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'An introduction to compressive sensing' conversation and receive update notifications?

Ask