Difference between revisions of "Aufgaben:Exercise 4.5: Mutual Information from 2D-PDF"
m (Text replacement - "value-continuous" to "continuous") |
|||
(3 intermediate revisions by one other user not shown) | |||
Line 12: | Line 12: | ||
respectively. Within each of the regions shown, let $f_{XY}(x, y) = C = \rm const.$ | respectively. Within each of the regions shown, let $f_{XY}(x, y) = C = \rm const.$ | ||
− | For example, the mutual information between the | + | For example, the mutual information between the continuous random variables $X$ and $Y$ can be calculated as follows: |
:$$I(X;Y) = h(X) + h(Y) - h(XY)\hspace{0.05cm}.$$ | :$$I(X;Y) = h(X) + h(Y) - h(XY)\hspace{0.05cm}.$$ | ||
Line 36: | Line 36: | ||
Hints: | Hints: | ||
− | *The exercise belongs to the chapter [[Information_Theory/AWGN_Channel_Capacity_for_Continuous_Input# | + | *The exercise belongs to the chapter [[Information_Theory/AWGN_Channel_Capacity_for_Continuous_Input#Mutual_information_between_continuous_random_variables|Mutual information with continuous input]]. |
*Let the following differential entropies also be given: | *Let the following differential entropies also be given: | ||
Line 78: | Line 78: | ||
===Solution=== | ===Solution=== | ||
{{ML-Kopf}} | {{ML-Kopf}} | ||
− | [[File:P_ID2887__Inf_A_4_5a.png|right|frame| | + | [[File:P_ID2887__Inf_A_4_5a.png|right|frame|"Red" probability density functions; <br>'''!''' Note: Ordinate of $f_{Y}(y)$ is directed to the left '''!''']] |
− | '''(1)''' For the rectangular | + | '''(1)''' For the rectangular two-dimensional PDF $f_{XY}(x, y)$ there are no statistical dependences between $X$ and $Y$ ⇒ $\underline{I(X;Y) = 0}$. |
Formally, this result can be proved with the following equation: | Formally, this result can be proved with the following equation: | ||
:$$I(X;Y) = h(X) \hspace{-0.05cm}+\hspace{-0.05cm} h(Y) \hspace{-0.05cm}- \hspace{-0.05cm}h(XY)\hspace{0.02cm}.$$ | :$$I(X;Y) = h(X) \hspace{-0.05cm}+\hspace{-0.05cm} h(Y) \hspace{-0.05cm}- \hspace{-0.05cm}h(XY)\hspace{0.02cm}.$$ | ||
− | *The red area | + | *The red area of the two-dimensional PDF $f_{XY}(x, y)$ is $F = 4$. |
− | *From this follows for the differential joint entropy in "bit": | + | *Since $f_{XY}(x, y)$ is constant in this area and the volume under $f_{XY}(x, y)$ must be equal to $1$, the height is $C = 1/F = 1/4$. |
+ | *From this follows for the differential joint entropy in "bit": | ||
:$$h(XY) \ = \ \hspace{0.1cm}-\hspace{0.2cm} \int \hspace{-0.9cm} \int\limits_{\hspace{-0.5cm}(x, y) \hspace{0.1cm}\in \hspace{0.1cm}{\rm supp} \hspace{0.03cm}(\hspace{-0.03cm}f_{XY}\hspace{-0.08cm})} | :$$h(XY) \ = \ \hspace{0.1cm}-\hspace{0.2cm} \int \hspace{-0.9cm} \int\limits_{\hspace{-0.5cm}(x, y) \hspace{0.1cm}\in \hspace{0.1cm}{\rm supp} \hspace{0.03cm}(\hspace{-0.03cm}f_{XY}\hspace{-0.08cm})} | ||
\hspace{-0.6cm} f_{XY}(x, y) \cdot {\rm log}_2 \hspace{0.1cm} [ f_{XY}(x, y) ] | \hspace{-0.6cm} f_{XY}(x, y) \cdot {\rm log}_2 \hspace{0.1cm} [ f_{XY}(x, y) ] | ||
Line 91: | Line 92: | ||
\hspace{-0.6cm} f_{XY}(x, y) | \hspace{-0.6cm} f_{XY}(x, y) | ||
\hspace{0.15cm}{\rm d}x\hspace{0.15cm}{\rm d}y = 2 \,{\rm bit}\hspace{0.05cm}.$$ | \hspace{0.15cm}{\rm d}x\hspace{0.15cm}{\rm d}y = 2 \,{\rm bit}\hspace{0.05cm}.$$ | ||
− | *It is considered that the double integral is equal to $1$ . The pseudo-unit "bit" corresponds to the | + | *It is considered that the double integral is equal to $1$ . The pseudo-unit "bit" corresponds to the "binary logarithm" ⇒ "log<sub>2</sub>". |
Line 98: | Line 99: | ||
* The marginal probability density functions $f_{X}(x)$ and $f_{Y}(y)$ are rectangular ⇒ uniform distribution between $0$ and $2$: | * The marginal probability density functions $f_{X}(x)$ and $f_{Y}(y)$ are rectangular ⇒ uniform distribution between $0$ and $2$: | ||
:$$h(X) = h(Y) = {\rm log}_2 \hspace{0.1cm} (2) = 1 \,{\rm bit}\hspace{0.05cm}.$$ | :$$h(X) = h(Y) = {\rm log}_2 \hspace{0.1cm} (2) = 1 \,{\rm bit}\hspace{0.05cm}.$$ | ||
+ | [[File:P_ID2888__Inf_A_4_5b_neu.png|right|frame|"Blue" | ||
+ | probability density functions]] | ||
+ | |||
* Substituting these results into the above equation, we obtain: | * Substituting these results into the above equation, we obtain: | ||
:$$I(X;Y) = h(X) + h(Y) - h(XY) = 1 \,{\rm bit} + 1 \,{\rm bit} - 2 \,{\rm bit} = 0 \,{\rm (bit)} | :$$I(X;Y) = h(X) + h(Y) - h(XY) = 1 \,{\rm bit} + 1 \,{\rm bit} - 2 \,{\rm bit} = 0 \,{\rm (bit)} | ||
Line 104: | Line 108: | ||
− | |||
'''(2)''' Also for this parallelogram we get $F = 4, \ C = 1/4$ as well as $h(XY) = 2$ bit. | '''(2)''' Also for this parallelogram we get $F = 4, \ C = 1/4$ as well as $h(XY) = 2$ bit. | ||
− | * Here, as in subtask '''(1)''' , the | + | * Here, as in subtask '''(1)''' , the variable $Y$ is uniformly distributed between $0$ and $2$ ⇒ $h(Y) = 1$ bit. |
*In contrast, $X$ is triangularly distributed between $0$ and $4$ $($with maximum at $2)$. | *In contrast, $X$ is triangularly distributed between $0$ and $4$ $($with maximum at $2)$. | ||
Line 115: | Line 118: | ||
\hspace{0.05cm}.$$ | \hspace{0.05cm}.$$ | ||
<br clear=all> | <br clear=all> | ||
− | [[File:P_ID2889__Inf_A_4_5c_neu.png|right|frame| | + | [[File:P_ID2889__Inf_A_4_5c_neu.png|right|frame|"„Green" probability density functions]] |
'''(3)''' The following properties are obtained for the green conditions: | '''(3)''' The following properties are obtained for the green conditions: | ||
:$$F = A \cdot B \hspace{0.3cm} \Rightarrow \hspace{0.3cm} C = \frac{1}{A \cdot B} | :$$F = A \cdot B \hspace{0.3cm} \Rightarrow \hspace{0.3cm} C = \frac{1}{A \cdot B} | ||
Line 121: | Line 124: | ||
\Rightarrow \hspace{0.3cm} h(XY) = {\rm log}_2 \hspace{0.1cm} (A \cdot B) | \Rightarrow \hspace{0.3cm} h(XY) = {\rm log}_2 \hspace{0.1cm} (A \cdot B) | ||
\hspace{0.05cm}.$$ | \hspace{0.05cm}.$$ | ||
− | *The random variable $Y$ is now uniformly distributed between $0$ and $A$ and the random variable $X$ is triangularly distributed between $0$ and $ | + | *The random variable $Y$ is now uniformly distributed between $0$ and $A$ and the random variable $X$ is triangularly distributed between $0$ and $2B$ $($with maximum at $B)$: |
:$$h(X) \ = \ {\rm log}_2 \hspace{0.1cm} (B \cdot \sqrt{ e}) | :$$h(X) \ = \ {\rm log}_2 \hspace{0.1cm} (B \cdot \sqrt{ e}) | ||
\hspace{0.05cm},$$ $$ | \hspace{0.05cm},$$ $$ | ||
Line 130: | Line 133: | ||
\hspace{0.05cm}.$$ | \hspace{0.05cm}.$$ | ||
[[File: P_ID2890__Inf_A_4_5d.png |right|frame|Other examples of 2D PDF $f_{XY}(x, y)$]] | [[File: P_ID2890__Inf_A_4_5d.png |right|frame|Other examples of 2D PDF $f_{XY}(x, y)$]] | ||
− | *$I(X;Y)$ thus independent of | + | *$I(X;Y)$ thus independent of th PDF parameters $A$ and $B$. |
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | '''(4)''' <u>All of the above conditions</u> are required. However, the requirements '''(2)''' and '''(3)''' are not satisfied for every parallelogram. | ||
+ | *The adjacent graph shows two constellations, where the random variable $X$ is equally distributed between $0$ and $1$ in each case. | ||
+ | *For the top graph, the plotted points lie at a height ⇒ $f_{Y}(y)$ is triangularly distributed ⇒ $I(X;Y) = 0.721$ bit. | ||
+ | *The lower composite PDF has a different mutual information, since the two plotted points are not at the same height <br>⇒ the PDF $f_{Y}(y)$ here has a trapezoidal shape. | ||
+ | *Feeling, I guess $I(X;Y) < 0.721$ bit, since the two-dimensional area is more approaching a rectangle. If you still feel like it, so check this statement. | ||
− | |||
− | |||
− | |||
− | |||
− | |||
{{ML-Fuß}} | {{ML-Fuß}} | ||
Latest revision as of 09:27, 11 October 2021
Given here are the three different two-dimensional regions $f_{XY}(x, y)$, which in the task are identified by their fill colors with
- red joint PDF,
- blue joint PDF,
- green joint PDF,
respectively. Within each of the regions shown, let $f_{XY}(x, y) = C = \rm const.$
For example, the mutual information between the continuous random variables $X$ and $Y$ can be calculated as follows:
- $$I(X;Y) = h(X) + h(Y) - h(XY)\hspace{0.05cm}.$$
For the differential entropies used here, the following equations apply:
- $$h(X) = -\hspace{-0.7cm} \int\limits_{x \hspace{0.05cm}\in \hspace{0.05cm}{\rm supp}(f_X)} \hspace{-0.55cm} f_X(x) \cdot {\rm log} \hspace{0.1cm} \big[f_X(x)\big] \hspace{0.1cm}{\rm d}x \hspace{0.05cm},$$
- $$h(Y) = -\hspace{-0.7cm} \int\limits_{y \hspace{0.05cm}\in \hspace{0.05cm}{\rm supp}(f_Y)} \hspace{-0.55cm} f_Y(y) \cdot {\rm log} \hspace{0.1cm} \big[f_Y(y)\big] \hspace{0.1cm}{\rm d}y \hspace{0.05cm},$$
- $$h(XY) = \hspace{0.1cm}-\hspace{0.2cm} \int \hspace{-0.9cm} \int\limits_{\hspace{-0.5cm}(x, y) \hspace{0.1cm}\in \hspace{0.1cm}{\rm supp} (f_{XY}\hspace{-0.08cm})} \hspace{-0.6cm} f_{XY}(x, y) \cdot {\rm log} \hspace{0.1cm} \big[ f_{XY}(x, y) \big] \hspace{0.15cm}{\rm d}x\hspace{0.15cm}{\rm d}y\hspace{0.05cm}.$$
- For the two marginal probability density functions, the following holds:
- $$f_X(x) = \hspace{-0.5cm} \int\limits_{\hspace{-0.2cm}y \hspace{0.1cm}\in \hspace{0.1cm}{\rm supp} (f_{Y}\hspace{-0.08cm})} \hspace{-0.4cm} f_{XY}(x, y) \hspace{0.15cm}{\rm d}y\hspace{0.05cm},$$
- $$f_Y(y) = \hspace{-0.5cm} \int\limits_{\hspace{-0.2cm}x \hspace{0.1cm}\in \hspace{0.1cm}{\rm supp} (f_{X}\hspace{-0.08cm})} \hspace{-0.4cm} f_{XY}(x, y) \hspace{0.15cm}{\rm d}x\hspace{0.05cm}.$$
Hints:
- The exercise belongs to the chapter Mutual information with continuous input.
- Let the following differential entropies also be given:
- If $X$ is triangularly distributed between $x_{\rm min}$ and $x_{\rm max}$, then:
- $$h(X) = {\rm log} \hspace{0.1cm} [\hspace{0.05cm}\sqrt{ e} \cdot (x_{\rm max} - x_{\rm min})/2\hspace{0.05cm}]\hspace{0.05cm}.$$
- If $Y$ is equally distributed between $y_{\rm min}$ and $y_{\rm max}$, then holds:
- $$h(Y) = {\rm log} \hspace{0.1cm} \big [\hspace{0.05cm}y_{\rm max} - y_{\rm min}\hspace{0.05cm}\big ]\hspace{0.05cm}.$$
- All results should be expressed in "bit". This is achieved with $\log$ ⇒ $\log_2$.
Questions
Solution
(1) For the rectangular two-dimensional PDF $f_{XY}(x, y)$ there are no statistical dependences between $X$ and $Y$ ⇒ $\underline{I(X;Y) = 0}$.
Formally, this result can be proved with the following equation:
- $$I(X;Y) = h(X) \hspace{-0.05cm}+\hspace{-0.05cm} h(Y) \hspace{-0.05cm}- \hspace{-0.05cm}h(XY)\hspace{0.02cm}.$$
- The red area of the two-dimensional PDF $f_{XY}(x, y)$ is $F = 4$.
- Since $f_{XY}(x, y)$ is constant in this area and the volume under $f_{XY}(x, y)$ must be equal to $1$, the height is $C = 1/F = 1/4$.
- From this follows for the differential joint entropy in "bit":
- $$h(XY) \ = \ \hspace{0.1cm}-\hspace{0.2cm} \int \hspace{-0.9cm} \int\limits_{\hspace{-0.5cm}(x, y) \hspace{0.1cm}\in \hspace{0.1cm}{\rm supp} \hspace{0.03cm}(\hspace{-0.03cm}f_{XY}\hspace{-0.08cm})} \hspace{-0.6cm} f_{XY}(x, y) \cdot {\rm log}_2 \hspace{0.1cm} [ f_{XY}(x, y) ] \hspace{0.15cm}{\rm d}x\hspace{0.15cm}{\rm d}y$$
- $$\Rightarrow \hspace{0.3cm} h(XY) \ = \ \ {\rm log}_2 \hspace{0.1cm} (4) \cdot \hspace{0.02cm} \int \hspace{-0.9cm} \int\limits_{\hspace{-0.5cm}(x, y) \hspace{0.1cm}\in \hspace{0.1cm}{\rm supp} \hspace{0.03cm}(\hspace{-0.03cm}f_{XY}\hspace{-0.08cm})} \hspace{-0.6cm} f_{XY}(x, y) \hspace{0.15cm}{\rm d}x\hspace{0.15cm}{\rm d}y = 2 \,{\rm bit}\hspace{0.05cm}.$$
- It is considered that the double integral is equal to $1$ . The pseudo-unit "bit" corresponds to the "binary logarithm" ⇒ "log2".
Furthermore:
- The marginal probability density functions $f_{X}(x)$ and $f_{Y}(y)$ are rectangular ⇒ uniform distribution between $0$ and $2$:
- $$h(X) = h(Y) = {\rm log}_2 \hspace{0.1cm} (2) = 1 \,{\rm bit}\hspace{0.05cm}.$$
- Substituting these results into the above equation, we obtain:
- $$I(X;Y) = h(X) + h(Y) - h(XY) = 1 \,{\rm bit} + 1 \,{\rm bit} - 2 \,{\rm bit} = 0 \,{\rm (bit)} \hspace{0.05cm}.$$
(2) Also for this parallelogram we get $F = 4, \ C = 1/4$ as well as $h(XY) = 2$ bit.
- Here, as in subtask (1) , the variable $Y$ is uniformly distributed between $0$ and $2$ ⇒ $h(Y) = 1$ bit.
- In contrast, $X$ is triangularly distributed between $0$ and $4$ $($with maximum at $2)$.
- This results in the same differential entropy $h(Y)$ as for a symmetric triangular distribution in the range between $±2$ (see specification sheet):
- $$h(X) = {\rm log}_2 \hspace{0.1cm} \big[\hspace{0.05cm}2 \cdot \sqrt{ e} \hspace{0.05cm}\big ] = 1.721 \,{\rm bit}$$
- $$\Rightarrow \hspace{0.3cm} I(X;Y) = 1.721 \,{\rm bit} + 1 \,{\rm bit} - 2 \,{\rm bit}\hspace{0.05cm}\underline{ = 0.721 \,{\rm (bit)}} \hspace{0.05cm}.$$
(3) The following properties are obtained for the green conditions:
- $$F = A \cdot B \hspace{0.3cm} \Rightarrow \hspace{0.3cm} C = \frac{1}{A \cdot B} \hspace{0.05cm}\hspace{0.3cm} \Rightarrow \hspace{0.3cm} h(XY) = {\rm log}_2 \hspace{0.1cm} (A \cdot B) \hspace{0.05cm}.$$
- The random variable $Y$ is now uniformly distributed between $0$ and $A$ and the random variable $X$ is triangularly distributed between $0$ and $2B$ $($with maximum at $B)$:
- $$h(X) \ = \ {\rm log}_2 \hspace{0.1cm} (B \cdot \sqrt{ e}) \hspace{0.05cm},$$ $$ h(Y) \ = \ {\rm log}_2 \hspace{0.1cm} (A)\hspace{0.05cm}.$$
- Thus, for the mutual information between $X$ and $Y$:
- $$I(X;Y) \ = {\rm log}_2 \hspace{0.1cm} (B \cdot \sqrt{ {\rm e}}) + {\rm log}_2 \hspace{0.1cm} (A) - {\rm log}_2 \hspace{0.1cm} (A \cdot B)$$
- $$\Rightarrow \hspace{0.3cm} I(X;Y) = \ {\rm log}_2 \hspace{0.1cm} \frac{B \cdot \sqrt{ {\rm e}} \cdot A}{A \cdot B} = {\rm log}_2 \hspace{0.1cm} (\sqrt{ {\rm e}})\hspace{0.15cm}\underline{= 0.721\,{\rm bit}} \hspace{0.05cm}.$$
- $I(X;Y)$ thus independent of th PDF parameters $A$ and $B$.
(4) All of the above conditions are required. However, the requirements (2) and (3) are not satisfied for every parallelogram.
- The adjacent graph shows two constellations, where the random variable $X$ is equally distributed between $0$ and $1$ in each case.
- For the top graph, the plotted points lie at a height ⇒ $f_{Y}(y)$ is triangularly distributed ⇒ $I(X;Y) = 0.721$ bit.
- The lower composite PDF has a different mutual information, since the two plotted points are not at the same height
⇒ the PDF $f_{Y}(y)$ here has a trapezoidal shape. - Feeling, I guess $I(X;Y) < 0.721$ bit, since the two-dimensional area is more approaching a rectangle. If you still feel like it, so check this statement.