<< Chapter < Page Chapter >> Page >

Supposing that f is piecewise smooth, however, with a finite number of discontinuities, then (as discussed in Sparse (Nonlinear) Models from Low-Dimensional Signal Models ) f will have a limited number of significant wavelet coefficients at fine scales. Because of theconcentration of these significant coefficients within each scale, the nonlinear approximation rate will remain σ K ( f ) 2 K - H as if there were no discontinuities present [link] .

Unfortunately, this resilience of wavelets to discontinuities does not extend to higher dimensions. Suppose, for example, that f is a C H smooth 2-D signal. Assuming the proper number of vanishing moments, a wavelet representation will achievethe optimal nonlinear approximation rate σ K ( f ) 2 K - H / 2 [link] , [link] . As in the 1-D case, this approximation rate is maintained when a finite number of point discontinuities are introduced into f . However, when f contains 1-D discontinuities (edges separating the smooth regions), the approximation rate will fall to σ K ( f ) 2 K - 1 / 2 [link] . The problem actually arises due to the isotropic, dyadic supports of the wavelets; instead of O ( 1 ) significant wavelets at each scale, there are now O ( 2 j ) wavelets overlapping the discontinuity. We revisit this important issue in Compression .

Despite the limited approximation capabilities for images with edges, nonlinear approximation in the wavelet domain typically offers a superior approximation to an image compared to linear approximation in the wavelet domain. As an example, [link] (b) shows the nonlinear approximation of the Cameraman test image obtained by keeping the largest scaling and wavelet coefficients. In this case, a number of high-frequency coefficients are selected, which gives an improved ability to represent features such as edges. Better concise transforms, which capture the image information in even fewer coefficients, would offer further improvements in terms of nonlinear approximation quality.

Finding approximations

As mentioned above, in the case where Ψ is an orthonormal basis and p = 2 , the solution to [link] is easily obtained by thresholding: one simply computes the coefficients α and keeps the K largest (setting the remaining coefficients to zero). Thresholding can also be shown to be optimal for arbitrary p norms in the special case where Ψ is the canonical basis. While the optimality of thresholding does notgeneralize to arbitrary norms and bases, thresholding can be shown to be a near-optimal approximation strategy for wavelet bases witharbitrary L p norms [link] .

In the case where Ψ is a redundant dictionary, however, the expansion coefficients α are not unique, and the optimization problem [link] can be much more difficult to solve. Indeed, supposing even that an exact K -term representation exists for f in the dictionary Ψ , finding that K -term approximation is NP-hard in general, requiring a combinatorial enumeration of the Z K possible sparse subspaces [link] . This search can be recast as the optimization problem

α ^ = arg min α 0 s.t. f = Ψ α .
While solving [link] is prohibitively complex, a variety of algorithms have been proposed as alternatives. One approachconvexifies the optimization problem by replacing the 0 fidelity criterion by an 1 criterion
α ^ = arg min α 1 s.t. f = Ψ α .
This problem, known as Basis Pursuit [link] , is significantly more approachable and can be solved with traditionallinear programming techniques whose computational complexities are polynomial in Z . The 1 criterion has the advantage of yielding a convex optimization problem while still encouraging sparse solutions due to the polytope geometry of the 1 unit ball (see for example [link] and [link] ). Iterative greedy algorithms such asMatching Pursuit (MP) and Orthogonal Matching Pursuit (OMP) [link] have also been suggested to find sparse representations α for a signal f . Both MP and OMP iteratively select the columns from Ψ that are most correlated with f , then subtract the contribution of each column, leaving a residual. OMP includes an additional step ateach iteration where the residual is orthogonalized against the previously selected columns.

Manifold approximation

We also consider the problem of finding the best manifold-based approximation to a signal (see [link] (c)). Suppose that F = { f θ : θ Θ } is a parametrized K -dimension manifold and that we are given a signal I that is believed to approximate f θ for an unknown θ Θ . From I we wish to recover an estimate of θ . Again, we may formulate this parameter estimation problem as an optimization, writing the objectivefunction (here we concentrate solely on the L 2 or 2 case)

D ( θ ) = f θ - I 2 2
and solving for
θ * = arg min θ Θ D ( θ ) .
We suppose that the minimum is uniquely defined.

Standard nonlinear parameter estimation  [link] tells us that, if D is differentiable, we can use Newton's method to iteratively refine a sequence of guesses θ ( 0 ) , θ ( 1 ) , θ ( 2 ) , to θ * and rapidly convergence to the true value. Supposing that F is a differentiable manifold, we would let

J = [ D / θ 0 D / θ 1 D / θ K - 1 ] T
be the gradient of D , and let H be the K × K Hessian, H i j = 2 D θ i θ j . Assuming D is differentiable, Newton's method specifies the following update step:
θ ( k + 1 ) θ ( k ) + [ H ( θ ( k ) ) ] - 1 J ( θ ( k ) ) .
To relate this method to the structure of the manifold, we can actually express the gradient and Hessian in terms of signals,writing
D ( θ ) = f θ - I 2 2 = ( f θ - I ) 2 d x = f θ 2 - 2 I f θ + I 2 d x .
Differentiating with respect to component θ i , we obtain
D θ i = J i = θ i f θ 2 - 2 I f θ + I 2 d x = θ i ( f θ 2 ) - 2 I θ i f θ d x = 2 f θ τ θ i - 2 I τ θ i d x = 2 f θ - I , τ θ i ,
where τ θ i = f θ θ i is a tangent signal. Continuing, we examine the Hessian,
2 D θ i θ j = H i j = θ j D θ i = θ j 2 f θ τ θ i - 2 I τ θ i d x = 2 τ θ i τ θ j + 2 f θ τ θ i j - 2 I τ θ i j d x = 2 τ θ i , τ θ j + 2 f θ - I , τ θ i j ,
where τ θ i j = 2 f θ θ i θ j denotes a second-derivative signal. Thus, we can interpret Newton's method geometrically as (essentially) asequence of successive projections onto tangent spaces on the manifold.

Again, the above discussion assumes the manifold to be differentiable. Many interesting parametric signal manifolds are in fact nowheredifferentiable — the tangent spaces demanded by Newton's method do not exist. However, in [link] we have identified a type of multiscale tangent structure to the manifold that permits a coarse-to-fine technique for parameter estimation.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Concise signal models. OpenStax CNX. Sep 14, 2009 Download for free at http://cnx.org/content/col10635/1.4
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Concise signal models' conversation and receive update notifications?

Ask