<< Chapter < Page Chapter >> Page >
r = 1 π ( r ) - π ( r + 1 ) 1 < .

The rest of proof is structured as follows. First, we show that the sequence of stable state distributions for the Markov chains (MC) used by the MCMC algorithm converges to a uniform distribution over the set of sequences that minimize the energy function as the iteration count t increases. Then, we show using  Theorem 11 and  Theorem 12 that the non-homogeneous MC used in the MCMC algorithm is strongly ergodic, which by the definition of strong ergodicity implies that MCMC always converges to the stable distribution found above. This implies that the outcome of the MCMC algorithm converges to a minimum-energy solution as t , completing the proof of Theorem  Theorem 10 .

We therefore begin by finding the stable state distribution for the non-homogeneous MC used by the MCMC algorithm. At each super-iteration r , the distribution defined as

π ( r ) ( w n ) exp ( - 1 t r ϵ ( w n ) ) z n α ˆ n exp ( - 1 t r ϵ ( z n ) ) = 1 z n α ˆ n exp ( - 1 t r ( ϵ ( z n ) - ϵ ( w n ) ) ) .

satisfies π ( r ) P ( r ) = π ( r ) . We can show that the distribution π ( r ) converges to a uniform distribution over the set of sequences that minimize the energy function, i.e.,

lim r π ( r ) ( w n ) = 0 w n H , 1 | H | w n H ,

where H = { w n α ˆ n s . t . ϵ ( w n ) = min z n α ˆ n ϵ ( z n ) } . To show [link] , we will show that π ( r ) ( w n ) is increasing for w n H and eventually decreasing for w n H C . Since for w n H and z n α ˆ n , ϵ ( z n ) - ϵ ( w n ) 0 , and so for r 1 < r 2 we have

z n α ˆ n exp ( - 1 t r 1 ( ϵ ( z n ) - ϵ ( w n ) ) ) w n α ˆ n exp ( - 1 t r 2 ( ϵ ( z n ) - ϵ ( w n ) ) ) ,

which together with [link] implies π ( r 1 ) ( w n ) π ( r 2 ) ( w n ) . On the other hand, if w n H , then

1 π ( r ) ( w n ) = z n : ϵ ( z n ) ϵ ( w n ) exp - 1 t r ( ϵ ( z n ) - ϵ ( w n ) ) + z n : ϵ ( z n ) < ϵ ( w n ) exp - 1 t r ( ϵ ( z n ) - ϵ ( y n ) ) .

For sufficiently small t , the right hand side (more precisely, the second and third lines) of [link] is dominated by the second term (third line), which increases when t decreases, and therefore π ( r ) ( w n ) decreases for w n H as t increases. Finally, since all sequences w n H have the same energy ϵ ( w n ) , it follows that the distribution is uniform over the symbols in H .

Having shown convergence of the non-homogenous Markov chain's stable state distributions, we now show that the non-homogeneous MC is strongly ergodic. The transition matrix P ( r ) of the MC at iteration t depends on the temperature t used within MCMC algorithm. We first show that the MC used in the MCMC algorithm is weakly ergodic via  Theorem 11 ; the proof of the following Lemma is given at the end of this section.

Lemma 2 The ergodic coefficient of P ( r ) for any r 0 is upper bounded by

δ P ( r ) 1 - exp - 1 t r n Δ k ,

where Δ k is defined in [link] .

Let w 1 n , w 2 n be two arbitrary sequences in α ˆ n . The probability of transitioning from a given state to a neighboring state in an iteration within iteration r ' of super iteration r of the MCMC algorithm is given by [link] , and can be rewritten as

P ( r , r ' ) ( w r ' = b | w n r ' ) = Pr ( w r ' = b | w n r ' ) = exp - 1 t r ϵ ( w r ' = b , w n r ' ) b ' α ˆ exp - 1 t r ϵ ( w r ' = b ' , w n r ' ) = exp - 1 t ϵ ( w 1 r ' - 1 b w r ' + 1 n ) - ϵ min , r ' ( w 1 r ' - 1 , w r ' + 1 n ) b ' α ˆ exp - 1 t ϵ ( w 1 r ' - 1 b ' w r ' + 1 n ) - ϵ min , r ' ( w 1 r ' - 1 , w r ' + 1 n ) exp ( - 1 t Δ k ) | α ˆ | ,

where ϵ min , r ' ( w 1 r ' - 1 b w r ' + 1 n ) = min b ' α ˆ ϵ ( w 1 r ' - 1 b ' w r ' + 1 n ) . Therefore the smallest probability of transition from w 1 n to w 2 n within super-iteration r is bounded by

min w 1 n , w 2 n α ˆ n P ( r ) ( w 2 n | w 1 n ) r ' = 1 n exp ( - 1 t r Δ k ) | α ˆ | = exp ( - 1 t r n Δ k ) | α ˆ | n .

Using the alternative definition of the ergodic coefficient given in [link] ,

δ P ( r ) = 1 - min w 1 n , w 2 n α ˆ n z n α ˆ n min ( P ( r ) ( z n | w 1 n ) , P ( r ) ( z n | w 2 n ) ) 1 - | α ˆ | n exp ( - 1 t r n Δ k ) | α ˆ | n = 1 - exp ( - 1 t r n Δ k ) .

Using Lemma 2 , we can evaluate the sum given in Theorem 11 as

r = 1 1 - δ P ( r ) r = 1 exp - 1 t r n Δ k = r = 1 1 r 1 / c = ,

and therefore the non-homogeneous MC defined by { P ( r ) } r = 1 is weakly ergodic. Now we can use Theorem 12 to show that the MC is strongly ergodic by proving that

r = 1 π ( r ) - π ( r + 1 ) 1 < .

Since we know from earlier in the proof that π ( r ) ( w n ) is increasing for w n H and eventually decreasing for w n H C , there exists a r 0 N such that for any r 1 > r 0 ,

r = r 0 r 1 π ( r ) - π ( r + 1 ) 1 = w n H r = r 0 r 1 π ( r + 1 ) ( w n ) - π ( r ) ( w n ) + w n H r = r 0 r 1 π ( r ) ( w n ) - π ( r + 1 ) ( w n ) = w n H π ( r 1 + 1 ) ( w n ) - π ( r 0 ) ( w n ) + w n H π ( r 0 ) ( w n ) - π ( r 1 + 1 ) ( w n ) = π ( r 1 + 1 ) - π ( r 0 ) 1 π ( r 1 + 1 ) 1 + π ( r 0 ) 1 = 2 .

Since the right hand side does not depend on r 1 , then we have that

r = 1 π ( r ) - π ( r + 1 ) 1 < .

This implies that the non-homogeneous MC used by MCMC algorithm is strongly ergodic, and thus completes the proof of  Theorem 10 .

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Universal algorithms in signal processing and communications. OpenStax CNX. May 16, 2013 Download for free at http://cnx.org/content/col11524/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Universal algorithms in signal processing and communications' conversation and receive update notifications?

Ask