Loading [MathJax]/jax/output/HTML-CSS/fonts/TeX/fontdata.js

Exercise 4.5Z: Again Mutual Information

From LNTwww

Given joint PDF and
graph of differential entropies

The graph above shows the joint PDF  fXY(x,y)  to be considered in this task,  which is identical to the "green" constellation in  Exercise 4.5.

  • In this sketch  fXY(x,y)  is enlarged by a factor of  3  in   y–direction.
  • In the definition area highlighted in green, the joint PDF is constant equal to  C=1/F,  where  F  indicates the area of the parallelogram.


In Exercise 4.5 the following differential entropies were calculated:

h(X) = log(A),
h(Y)=log(Be),
h(XY)=log(F)=log(AB).

In this exercise, the parameter values  A=e2  and  B=e0.5  are now to be used.

According to the above diagram, the conditional differential entropies  h(Y|X)  and  h(X|Y)  should now also be determined and their relation to the mutual information  I(X;Y)  given.




Hints:

  • The exercise belongs to the chapter  AWGN channel capacity with continuous input.
  • If the results are to be given in "nat", this is achieved with "log"  ⇒  "ln".
  • If the results are to be given in "bit", this is achieved with "log"  ⇒  "log2".



Questions

1

State the following information theoretic quantities  "nat":

h(X) = 

 nat
h(Y) = 

 nat
h(XY) = 

 nat
I(X;Y) = 

 nat

2

What are the same quantities with the pseudo–unit  "bit"?

h(X) = 

 bit
h(Y) = 

 bit
h(XY) = 

 bit
I(X;Y) = 

 bit

3

Calculate the conditional differential entropy  h(Y|X).

h(Y|X) = 

 nat
h(Y|X) = 

 bit

4

Calculate the conditional differential entropy  h(X|Y).

h(X|Y) = 

 nat
h(X|Y) = 

 bit

5

Which of the following quantities is never negative?

Both  H(X)  and  H(Y)  in the discrete case.
The mutual information  I(X;Y)  in the discrete case.
The mutual information  I(X;Y)  in the continuous case.
Both  h(X)  and  h(Y)  in the continuous case.
Both  h(X|Y)  and  h(Y|X)  in the continuous case.
The joint entropy  h(XY)  in the continuous case.


Solution

(1)  Since the results are required in  "nat",  it is convenient to use the natural logarithm:

  • The random variable  X  is uniformly distributed between  0  and  1/e2=e2:
h(X)=ln(e2)=2nat_.
  • The random variable  Y  is triangularly distributed between  ±e0.5:
h(Y)=ln(ee)=ln(e)=+1nat_.
  • The area of the parallelogram is given by
F=AB=e2e0.5=e1.5.
  • Thus, the 2D-PDF in the area highlighted in green has constant height  C=1/F=e1.5  and we obtain for the joint entropy:
h(XY)=ln(F)=ln(e1.5)=1.5nat_.
  • From this we obtain for the mutual information:
I(X;Y)=h(X)+h(Y)h(XY)=2nat+1nat(1.5nat)=0.5nat_.


(2)  In general, the relation  log2(x)=ln(x)/ln(2) holds.  Thus, using the results of subtask  (1), we obtain:

h(X) = 2nat0.693nat/bit=2.886bit_,
h(Y) = +1nat0.693nat/bit=+1.443bit_,
h(XY) = 1.5nat0.693nat/bit=2.164bit_,
I(X;Y) = 0.5nat0.693nat/bit=0.721bit_.
  • Or also:
I(X;Y)=2.886bit+1.443bit+2.164bit=0.721bit.


(3)  The mutual information can also be written in the form  I(X;Y)=h(Y)h(YX) :

h(YX)=h(Y)I(X;Y)=1nat0.5nat=0.5nat=0.721bit_.


(4)  For the differential inference entropy, it holds correspondingly:

h(XY)=h(X)I(X;Y)=2nat0.5nat=2.5nat=3.607bit_.
Summary of all results of this exercise
  • All quantities calculated here are summarized in the graph. 
  • Arrows pointing up indicate a positive contribution, arrows pointing down indicate a negative contribution.


(5)  Correct are the  proposed solutions 1 to 3.

Again for clarification:

  • For the mutual information  I(X;Y)0 always holds.
  • In the discrete case there is no negative entropy, but in the continuous case there is.