<< Chapter < Page Chapter >> Page >

Consider a stationary ergodic source, where we no longer assume that it is parametric. We need to define notions of probability that fit the universal framework. For this,we study Kac's lemma  [link] .

We have a stationary source, { . . . , X - n , X - n + 1 , . . . , X 0 , X 1 , . . . } , z = X - + 1 0 = { X - + 1 , X - + 2 , . . . , X 0 } . Define N r as the number of shifts forward of a window of length until we see z again; this is called recurrence time. Define

Y k = 1 X k - + 1 k = z 0 else ,

e.g., Y 0 = 1 , then N r is the smallest positive number for which Y k = 1 . Note that { Y k } - + is a binary stationary ergodic source. Define Q k = Pr { Y k = 1 ; 1 j k - 1 , Y j = 0 | Y 0 = 1 } . Then the average recurrence time can be computed, μ = = 1 Q = E [ N r ] . We can now present Kac's lemma.

Lemma 3 [link] μ = 1 Pr ( Y = 1 ) and E N r = μ | X - + 1 0 = z = 1 Pr ( z ) .

Let A = { Y n = 1 for some - < n < + } . Because z just appeared, then its probability is positive. We will prove that μ = Pr ( A ) Pr ( Y 0 = 1 ) .

Define B + = { Y n = 1 , for some 0 n < } , and B - = { Y n = 1 , for some n < 0 } . Then A = B + B - = ( B + B - ) B + ( B - ) C ( B + ) C B - . We claim that Pr B + ( B - ) C = Pr ( B + ) C B - = 0 . This can be shown formally, but is easily seen by realizing that if z appears at any time n (say, positive) then it must appear at some negative time n with probability 1, because its probability is positive.

Therefore, we have

Pr ( A ) = Pr B + B - = j = 0 k = 1 Pr ( Y j = 1 , Y - k = 1 , Y n = 0 , - k < n < j ) = j = 0 k = 1 Pr ( Y - k = 1 ) Pr ( Y j = 1 , Y n = 0 , - k < n < j | Y - k = 1 ) = j = 0 k = 1 Pr ( Y - k = 1 ) Q j + k = j = 0 k = 1 Pr ( Y 0 = 1 ) Q j + k = Pr ( Y 0 = 1 ) i = 1 i Q i = Pr ( Y 0 = 1 ) μ .

Therefore, μ = Pr ( A ) Pr ( Y 0 = 1 ) . We conclude the proof by noting that Pr ( A ) = 0 .

Let us now develop a universal coding technique for stationary sources. Recall H = 1 E [ - log ( Pr ( X 1 ) ) ] . The asymptotic equipartition property (AEP) of information theory  [link] gives

Pr X 1 : - 1 log ( Pr ( X 1 ) ) - H > δ < ϵ ( δ , ) ,

where lim ϵ ( δ , ) = 0 . Define in this context a typical set T ( δ , ) that satisfies

2 - ( H + δ ) Pr ( X 1 ) 2 - ( H - δ ) .

For a typical sequence z , E [ N r | X 1 = z ] 2 ( H + δ ) . Then

Pr log ( N r ) H + 2 δ = Pr ( z T ( δ , ) ) Pr log ( N r ) H + 2 δ | z T ( δ , ) + Pr ( z T ( δ , ) ) Pr log ( N r ) H + 2 δ | z T ( δ , ) 2 - δ + ϵ ( δ , ) AEP 0 .

Consider our situation, we have a source with memory of length n and want to transmit X 1 .

  1. Choose n = 2 ( H + 2 δ ) .
  2. For z = X 1 , find the value of N r if it appears in memory.
  3. If it appears, then transmit a 0 flag bit followed by the value of N r .
  4. Else transmit a 1 followed by the uncompressed z .

Transmitting z via N r requires log ( N r ) bits, and so the expected coding length is

Pr log ( N r ) H + 2 δ 1 + 1 + log ( N r ) + 2 - δ ( 1 + log α ) + ϵ ( , δ ) ( 1 + 1 + log α ) ( 2 + ( H + δ ) ) + 2 - δ ( 1 + log α ) + ϵ ( , δ ) ( 2 + log α ) .

After we normalize by , the per symbol length converges to 2 + H + 2 δ .

Note that this analysis assumes that the entropy H for a block of symbols is known. If not, then we can have several sets that are each adjusted for some differententropy level.

Universal coding of the integers

Coding of an index also appears in other information theoretic problems such as indexing a coset in distributed source coding and a codeword in lossycoding. What are good ways to encode the index?

If the index n lies in the range n { 1 , . . . , N } and N is known, then we can encode the index using log ( N ) bits. However, sometimes the index can be unbounded, or there could be an(unknown) distribution that biases us strongly toward smaller indices. Therefore, we are interested in a universal encoding of the integerssuch that each n is encoded using roughly log ( n ) bits.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Universal algorithms in signal processing and communications. OpenStax CNX. May 16, 2013 Download for free at http://cnx.org/content/col11524/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Universal algorithms in signal processing and communications' conversation and receive update notifications?

Ask