<< Chapter < Page Chapter >> Page >

For i.i.d. sources, D ( P 1 ( x n ) | | P 2 ( x n ) ) = n D ( P 1 ( x i ) | | P 2 ( x i ) ) , which means that the divergence increases linearly with n . Not only does the divergence increase, but it does so by a constant per symbol.Therefore, based on typical sequence concepts that we have seen, for an x n generated by P 1 , its probability under P 2 vanishes. However, we can construct a distribution Q whose divergence with both P 1 abd P 2 is small,

Q ( x n ) = 1 2 P 1 ( x n ) + 1 2 P 2 ( x n ) .

We now have for P 1 ,

1 n D ( P 1 n | | Q ) = 1 n E log P 1 ( x n ) 1 2 P 1 ( x n ) + 1 2 P 2 ( x n ) 1 n log ( 2 ) = 1 n .

On the other hand, 1 n D ( P 1 ( x 1 n ) | | Q ( x 1 n ) ) 0 [link] , and so

1 n 1 n D ( P 1 ( x 1 n ) | | Q ( x 1 n ) ) 0 .

By symmetry, we see that Q is also close to P 2 in the divergence sense.

Intuitively, it might seem peculiar that Q is close to both P 1 and P 2 but they are far away from each other (in divergence terms). This intuition stems from the triangle inequality, which holds for all metrics. The contradiction is resolved by realizingthat the divergence is not a metric, and it does not satisfy the triangle inequality.

Note also that for two i.i.d. distributions P 1 and P 2 , the divergence

D ( P 1 ( x n ) | | P 2 ( x n ) ) = n D ( P 1 | | P 2 )

is linear in n . If Q were i.i.d., then D ( P 1 ( x n ) Q ( x 1 n ) ) must also be linear in n . But the divergence is not increasing linearly in n , it is upper bounded by 1. Therefore, we conclude that Q ( · ) is not an i.i.d. distribution. Instead, Q is a distribution that contains memory, and there is dependence in Q between collections of different symbols of x in the sense that they are either all drawn from P 1 or all drawn from P 2 . To take this one step further, consider K sources with

Q ( x n ) = i = 1 K 1 K P i ( x n ) ,

then in an analogous manner to before it can be shown that

D ( P i ( x 1 n ) | | Q ( x 1 n ) ) 1 n log ( K ) .

Sources with memory : Instead of the memoryless (i.i.d.) source,

P ( x n ) = i = 1 n P ( x i ) ,

let us now put forward a statistical model with memory,

P ( x n ) = i = 1 n P ( x i | x 1 i - 1 ) .

Stationary source : To understand the notion of a stationary source, consider an infinite stream of symbols, . . . , x - 1 , x 0 , x 1 , ... . A complete probabilistic description of a stationary distribution is given by the collection of allmarginal distribution of the following form for all t and n ,

P X t , X t + 1 , . . . , X t + n - 1 ( x t , x t + 1 , . . . , x t + n - 1 ) .

For a stationary source, this distribution is independent of t .

Entropy rate : We have defined the first order entropy of an i.i.d. random variable [link] , and let us discuss more advanced concepts for sources with memory.Such definitions appear in many standard textbooks, for example that by Gallager  [link] .

  1. The order- n entropy is defined,
    H n = 1 n H ( x 1 , . . . , x n ) = - 1 n E [ log ( P ( x 1 , . . . , x n ) ) ] .
  2. The entropy rate is the limit of order- n entropy, H ¯ = lim n H n . The existence of this limit will be shown soon.
  3. Conditional entropy is defined similarly to entropy as the expectation of the log of the conditional probability,
    H ( x n | x 1 , . . . , x n - 1 ) = - 1 n E [ log ( P ( x n | x 1 , . . . , x n - 1 ) ) ] ,
    where expectation is taken over the joint probability space, P ( x 1 , . . . , x n ) .

The entropy rate also satisfies H ¯ = lim n H ( x n | x 1 , . . . , x n ) .

Theorem 3 For a stationary source with bounded first order entropy, H 1 ( x ) < , the following hold.

  1. The conditional entropy H ( x n | x 1 , . . . , x n - 1 ) is monotone non-increasing in n.
  2. The order- n entropy is not smaller than the conditional entropy,
    H n ( x ) H ( x n | x 1 , . . . , x n - 1 ) .
  3. The order- n entropy H n ( x ) is monotone non-increasing.
  4. H ¯ ( x ) = lim n H n ( x ) = lim n H ( x n | x 1 , . . . , x n - 1 ) .

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Universal algorithms in signal processing and communications. OpenStax CNX. May 16, 2013 Download for free at http://cnx.org/content/col11524/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Universal algorithms in signal processing and communications' conversation and receive update notifications?

Ask