Loading [MathJax]/jax/output/HTML-CSS/fonts/TeX/fontdata.js

Difference between revisions of "Aufgaben:Exercise 3.2: Expected Value Calculations"

From LNTwww
 
(17 intermediate revisions by 4 users not shown)
Line 1: Line 1:
  
{{quiz-Header|Buchseite=Informationstheorie/Einige Vorbemerkungen zu zweidimensionalen Zufallsgrößen
+
{{quiz-Header|Buchseite=Information_Theory/Some_Preliminary_Remarks_on_Two-Dimensional_Random_Variables
 
}}
 
}}
  
[[File:P_ID2751__Inf_A_3_2.png|right|2D–Wahrscheinlichkeitsfunktion]]
+
[[File:P_ID2751__Inf_A_3_2.png|right|frame|Two-dimensional <br>probability mass function]]
Wir betrachten folgende Wahrscheinlichkeitsfunktionen:
+
We consider the following probability mass functions&nbsp; (PMF):
  
: PX(X)=[1/2,1/8,0,3/8],
+
:$$P_X(X) = \big[1/2,\ 1/8,\ 0,\ 3/8 \big],$$
: PY(Y)=[1/2,1/4,1/4,0],
+
:$$P_Y(Y) = \big[1/2,\ 1/4,\ 1/4,\ 0 \big],$$
: PU(U)=[1/2,1/2],
+
:$$P_U(U) = \big[1/2,\ 1/2 \big],$$
: PV(V)=[3/4,1/4].
+
:$$P_V(V) = \big[3/4,\ 1/4\big].$$
  
Für die dazugehörigen Zufallsgrößen gelte:
+
For the associated random variables, let:
  
: X={0,1,2,3}, Y={0,1,2,3},
+
: $X= \{0,\ 1,\ 2,\ 3\}$, &nbsp; &nbsp;  $Y= \{0,\ 1,\ 2,\ 3\}$,&nbsp; &nbsp; $U = \{0,\ 1\}$, &nbsp; &nbsp; V={0,1}.
: U={0,1}, V={0,1}.
 
  
Oft muss man für solche diskreten Zufallsgrößen verschiedene Erwartungswerte der Form
+
Often, for such discrete random variables, one must have to calculate different expected values of the form
:$${\rm E} \left [ F(X)\right ] =\hspace{-0.3cm}  \sum_{x \hspace{0.05cm}\in \hspace{0.05cm}\hspace{-0.03cm}  {\rm supp} (P_X)}  \hspace{-0.1cm}  
+
:$${\rm E} \big [ F(X)\big ] =\hspace{-0.3cm}  \sum_{x \hspace{0.05cm}\in \hspace{0.05cm}\hspace{-0.03cm}  {\rm supp} (P_X)}  \hspace{-0.1cm}  
  P_{X}(x) \cdot F(x) $$
+
  P_{X}(x) \cdot F(x). $$
  
berechnen. Hierbei bedeuten:
+
Here, denote:
  
* PX(X) bezeichnet die <i>Wahrscheinlichkeitsfunktion</i> der diskreten Zufallsgröße X.
+
* PX(X)&nbsp; denotes the probability mass function of the discrete random variable &nbsp; X.
* Der <i>Support</i> von PX umfasst alle diejenigen Realisierungen x der Zufallsgröße X mit nicht verschwindender Wahrscheinlichkeit. <br>Formal kann hierfür geschrieben werden:
+
* The&nbsp; "support"&nbsp; of&nbsp; PX&nbsp; includes all those realisations&nbsp; x&nbsp; of the random variable&nbsp; X&nbsp; with non-vanishing probability.
:$${\rm supp} (P_X)  = \{ x: \hspace{0.25cm}x \in X \hspace{0.15cm}\underline{\rm und} \hspace{0.15cm} P_X(x) \ne 0 \} \hspace{0.05cm}.$$
+
*Formally, this can be written as
* F(X) ist eine (beliebige) reellwertige Funktion, die im gesamten Definitionsgebiet der Zufallsgröße angebbar ist.
+
:$${\rm supp} (P_X)  = \{ x: \hspace{0.25cm}x \in X \hspace{0.15cm}\underline{\rm and} \hspace{0.15cm} P_X(x) \ne 0 \} \hspace{0.05cm}.$$
 +
* F(X)&nbsp; is an (arbitrary) real-valued function that can be specified in the entire domain of definition of the random variable&nbsp; X&nbsp;.
  
  
In der Aufgabe sollen die Erwartungswerte für verschiedene Funktionen F(X) berechnet werden, unter anderem für
+
In the task, the expected values for various functions&nbsp; F(X)&nbsp; are to be calculated, among others for
  
: F(X)=1/PX(X),
+
#&nbsp; F(X)=1/PX(X),
: F(X)=PX(X),
+
#&nbsp; F(X)=PX(X),
: F(X)=log2 PX(X).
+
#&nbsp; F(X)=log2 PX(X).
  
  
''Hinweise:''
 
*Die Aufgabe gehört zum  Kapitel [[Informationstheorie/Einige_Vorbemerkungen_zu_zweidimensionalen_Zufallsgrößen|Einige Vorbemerkungen zu den 2D-Zufallsgrößen]].
 
* Die beiden &bdquo;eindimensionalen&rdquo; Wahrscheinlichkeitsfunktionen PX(X) und PY(Y) ergeben sich aus der dargestellten 2D&ndash;PMF PXY(X,Y), wie in [[Aufgaben:3.2Z_2D–Wahrscheinlichkeitsfunktion|Zusatzaufgabe 3.2Z]] gezeigt werden soll.
 
* Zu den binären Wahrscheinlichkeitsfunktionen PU(U) und PV(V) kommt man entsprechend den Modulo&ndash;Operationen U=Xmod2 sowie V=Ymod2.
 
*Sollte die Eingabe des Zahlenwertes &bdquo;0&rdquo; erforderlich sein, so geben Sie bitte &bdquo;0.&rdquo; ein.
 
  
  
===Fragebogen===
+
 
 +
 
 +
 
 +
 
 +
Hints:
 +
*The exercise belongs to the chapter&nbsp; [[Information_Theory/Einige_Vorbemerkungen_zu_zweidimensionalen_Zufallsgrößen|Some preliminary remarks on two-dimensional  random variables]].
 +
* The two one-dimensional  probability mass functions&nbsp; PX(X)&nbsp; and&nbsp; PY(Y)&nbsp; result from the presented 2D&ndash;PMF&nbsp; PXY(X, Y),&nbsp; as will be shown in&nbsp; [[Aufgaben:3.2Z_2D–Wahrscheinlichkeitsfunktion|Exercise 3.2Z]].
 +
* The binary probability mass functions&nbsp; PU(U)&nbsp; and&nbsp; PV(V)&nbsp; are obtained according to the modulo operations&nbsp; $U = X \hspace{0.1cm}\text{mod} \hspace{0.1cm}2&nbsp; and&nbsp;V = Y \hspace{0.1cm}\text{mod} \hspace{0.1cm} 2$.
 +
 +
 
 +
 
 +
===Questions===
  
 
<quiz display=simple>
 
<quiz display=simple>
{Welche Ergebnisse liefern die folgenden Erwartungswerte?
+
{What are the results of the following expected values?
 
|type="{}"}
 
|type="{}"}
E[1/PX(X)] =   { 3 3% }
+
${\rm E}\big[1/P_X(X)\big] \ = \ $  { 3 3% }
E[1/PY(Y)] = { 3 3% }
+
${\rm E}\big[1/P_{\hspace{0.04cm}Y}(\hspace{0.02cm}Y\hspace{0.02cm})\big] \ = \ ${ 3 3% }
  
  
{Geben Sie die folgenden Erwartungswerte an:
+
{Give the following expected values:
 
|type="{}"}
 
|type="{}"}
E[PX(X)] =  { 0.406 3% }
+
${\rm E}\big[P_X(X)\big] \ = \ $ { 0.406 3% }
E[PY(Y)] =  { 0.375 3% }
+
${\rm E}\big[P_Y(Y)\big] \ = \ $ { 0.375 3% }
  
  
{Berechnen Sie nun die folgenden Erwartungswerte:
+
{Now calculate the following expected values:
 
|type="{}"}
 
|type="{}"}
E[PY(X)] =  { 0.281 3% }
+
${\rm E}\big[P_Y(X)\big] \ = \ $ { 0.281 3% }
E[PX(Y)] =  { 0.281 3% }
+
${\rm E}\big[P_X(Y)\big] \ = \ $ { 0.281 3% }
  
  
{Welche der folgenden Aussagen sind zutreffend?
+
{Which of the following statements are true?
 
|type="[]"}
 
|type="[]"}
+ E[log2 PU(U)] ergibt die Entropie der Zufallsgröße U.
+
+ ${\rm E}\big[- \log_2 \ P_U(U)\big]$&nbsp; gives the entropy of the random variable&nbsp; U.
+ E[log2 PV(V)] ergibt die Entropie der Zufallsgröße V.
+
+ ${\rm E}\big[- \log_2 \ P_V(V)\big]$&nbsp; gives the entropy of the random variable&nbsp; V.
- E[log2 PV(U)] ergibt die Entropie der Zufallsgröße V.
+
- ${\rm E}\big[- \log_2 \ P_V(U)\big]$&nbsp; gives the entropy of the random variable&nbsp; V.
  
  
Line 73: Line 79:
 
</quiz>
 
</quiz>
  
===Musterlösung===
+
===Solution===
 
{{ML-Kopf}}
 
{{ML-Kopf}}
'''(1)'''&nbsp; Allgemein gilt für den Erwartungswert der Funktion <i>F</i>(<i>X</i>) hinsichtlich der Zufallsvariablen <i>X</i>:
+
'''(1)'''&nbsp; In general, the following applies to the expected value of the function&nbsp; $F(X)&nbsp; with respect to the random variable&nbsp;X$:
 
:$${\rm E} \left [ F(X)\right ] = \hspace{-0.4cm} \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} {\rm supp} (P_X)}  \hspace{-0.2cm}  
 
:$${\rm E} \left [ F(X)\right ] = \hspace{-0.4cm} \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} {\rm supp} (P_X)}  \hspace{-0.2cm}  
 
  P_{X}(x) \cdot F(x)  \hspace{0.05cm}.$$
 
  P_{X}(x) \cdot F(x)  \hspace{0.05cm}.$$
Im vorliegenden Beispiel gilt <i>X</i> = {0, 1, 2, 3} und <i>P<sub>X</sub></i>(<i>X</i>) = [1/2, 1/8, 0, 3/8]. Wegen <i>P<sub>X</sub></i>(<i>X</i> = 2) = 0 ergibt sich somit für die zu berücksichtigende Menge (&bdquo;Support&rdquo;) in obiger Summation:
+
In the present example,&nbsp; $X = \{0,\ 1,\ 2,\ 3\}&nbsp; and&nbsp;P_X(X) = \big [1/2, \ 1/8, \ 0, \ 3/8\big ]$.  
 +
*Because of&nbsp; $P_X(X = 2) = 0$&nbsp;, the quantity to be taken into account&nbsp; (the "support")&nbsp; in the above summation thus results in
 
:supp(PX)={0,1,3}.
 
:supp(PX)={0,1,3}.
Mit <i>F</i>(<i>X</i>) = 1/<i>P<sub>X</sub></i>(<i>X</i>) erhält man weiter:
+
*With&nbsp; $F(X) = 1/P_X(X)$&nbsp; one further obtains:
:$${\rm E} \left [ 1/P_X(X)\right ] = \hspace{-0.4cm} \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm}, 1\hspace{0.05cm},\hspace{0.05cm} 3 \}}  \hspace{-0.4cm} P_{X}(x) \cdot {1}/{P_X(x)}  
+
:$${\rm E} \big [ 1/P_X(X)\big ] = \hspace{-0.4cm} \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm},\hspace{0.05cm} 1\hspace{0.05cm},\hspace{0.05cm} 3 \}}  \hspace{-0.4cm} P_{X}(x) \cdot {1}/{P_X(x)}  
= \hspace{-0.4cm}  \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm}, 1\hspace{0.05cm},\hspace{0.05cm} 3 \}}  \hspace{-0.3cm} 1  
+
= \hspace{-0.4cm}  \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm},\hspace{0.05cm} 1\hspace{0.05cm},\hspace{0.05cm} 3 \}}  \hspace{-0.3cm} 1  
 
\hspace{0.15cm}\underline{ = 3} \hspace{0.05cm}.$$
 
\hspace{0.15cm}\underline{ = 3} \hspace{0.05cm}.$$
Der zweite Erwartungswert liefert mit supp(<i>P<sub>Y</sub></i>) = {0, 1, 2} das gleiche Ergebnis: E[1/<i>P<sub>Y</sub></i>(<i>Y</i>)] <u>= 3</u>.
+
*The second expected value gives the same result with&nbsp; ${\rm supp} (P_Y) = \{ 0\hspace{0.05cm}, 1\hspace{0.05cm}, 2 \} $&nbsp;:
 +
:$${\rm E} \left [ 1/P_Y(Y)\right ] \hspace{0.15cm}\underline{ = 3}.$$
  
  
'''(2)'''&nbsp; In beiden Fällen ist der Index der Wahrscheinlichkeitsfunktion mit der Zufallsvariablen (<i>X</i> bzw. <i>Y</i>) identisch und man erhält
+
 
:$${\rm E} \left [ P_X(X)\right ] =  \hspace{-0.3cm}  \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm}, 1\hspace{0.05cm},\hspace{0.05cm} 3 \}}  \hspace{-0.3cm}  P_{X}(x) \cdot {P_X(x)}  
+
'''(2)'''&nbsp; In both cases, the index of the probability mass function is identical with the random variable&nbsp; $(X&nbsp; or &nbsp;Y)$&nbsp; and we obtain
 +
:$${\rm E} \big [ P_X(X)\big ] =  \hspace{-0.3cm}  \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm},\hspace{0.05cm} 1\hspace{0.05cm},\hspace{0.05cm} 3 \}}  \hspace{-0.3cm}  P_{X}(x) \cdot {P_X(x)}  
 
= (1/2)^2 + (1/8)^2 + (3/8)^2 = 13/32
 
= (1/2)^2 + (1/8)^2 + (3/8)^2 = 13/32
 
\hspace{0.15cm}\underline{ \approx 0.406} \hspace{0.05cm},$$
 
\hspace{0.15cm}\underline{ \approx 0.406} \hspace{0.05cm},$$
:$${\rm E} \left [ P_Y(Y)\right ] = \hspace{-0.3cm}  \sum_{y \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm}, 1\hspace{0.05cm},\hspace{0.05cm} 2 \}}  \hspace{-0.3cm}  P_Y(y) \cdot P_Y(y) = (1/2)^2 + (1/4)^2 + (1/4)^2  
+
:$${\rm E} \big [ P_Y(Y)\big ] = \hspace{-0.3cm}  \sum_{y \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm},\hspace{0.05cm} 1\hspace{0.05cm},\hspace{0.05cm} 2 \}}  \hspace{-0.3cm}  P_Y(y) \cdot P_Y(y) = (1/2)^2 + (1/4)^2 + (1/4)^2  
 
\hspace{0.15cm}\underline{ = 0.375} \hspace{0.05cm}.$$
 
\hspace{0.15cm}\underline{ = 0.375} \hspace{0.05cm}.$$
  
'''(3)'''&nbsp; Hier gelten folgende Gleichungen:
+
 
:$${\rm E} \left [ P_Y(X)\right ] = \hspace{-0.3cm}  \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm}, 1\hspace{0.05cm},\hspace{0.05cm} 3 \}}  \hspace{-0.3cm}  P_{X}(x) \cdot {P_Y(x)} = \frac{1}{2} \cdot \frac{1}{2} + \frac{1}{8} \cdot \frac{1}{4} + \frac{3}{8} \cdot 0 = 9/32
+
 
 +
'''(3)'''&nbsp; The following equations apply here:
 +
:$${\rm E} \big [ P_Y(X)\big ] = \hspace{-0.3cm}  \sum_{x \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm},\hspace{0.05cm} 1\hspace{0.05cm},\hspace{0.05cm} 3 \}}  \hspace{-0.3cm}  P_{X}(x) \cdot {P_Y(x)} = \frac{1}{2} \cdot \frac{1}{2} + \frac{1}{8} \cdot \frac{1}{4} + \frac{3}{8} \cdot 0 = 9/32
 
\hspace{0.15cm}\underline{ \approx 0.281} \hspace{0.05cm},$$
 
\hspace{0.15cm}\underline{ \approx 0.281} \hspace{0.05cm},$$
Die Erwartungswertbildung bezieht sich hier auf <i>P<sub>X</sub></i>(&middot;), also auf die Zufallsgröße <i>X</i>. <i>P<sub>Y</sub></i>(&middot;) ist dabei die formale Funktion ohne (direkten) Bezug zur Zufallsgröße <i>Y</i>.
+
*The expected value formation here refers to&nbsp; $P_X(&middot;)$, i.e. to the random variable&nbsp; $X$.  
 +
*$P_Y(&middot;)&nbsp;is the formal function without (direct) reference to the random variable&nbsp;Y$.
 +
*The same numerical value is obtained for the second expected value&nbsp; (this does not have to be the case in general):
 +
:$${\rm E} \big [ P_X(Y)\big ] = \hspace{-0.3cm}  \sum_{y \hspace{0.05cm}\in \hspace{0.05cm} \{ 0\hspace{0.05cm},\hspace{0.05cm} 1\hspace{0.05cm},\hspace{0.05cm} 2 \}}  \hspace{-0.3cm}  P_{Y}(y) \cdot {P_X(y)} = \frac{1}{2} \cdot \frac{1}{2} + \frac{1}{4} \cdot \frac{1}{8} + \frac{1}{4} \cdot 0 = 9/32 \hspace{0.15cm}\underline{ \approx 0.281} \hspace{0.05cm}.$$
 +
 
  
Für den zweiten Erwartungswert erhält man im vorliegenden den gleichen Zahlenwert (das muss nicht so sein):
 
:E[PX(Y)]=y{0,1,2}PY(y)PX(y)=1212+1418+140=9/320.281_.
 
  
'''(4)'''&nbsp; Wir berechnen zunächst die drei Erwartungswerte:
+
'''(4)'''&nbsp; We first calculate the three expected values:
:$${\rm E} \left [-{\rm log}_2 \hspace{0.1cm} P_U(U)\right ]  
+
:$${\rm E} \big [-{\rm log}_2 \hspace{0.1cm} P_U(U)\big ]  
  = \frac{1}{2} \cdot {\rm log}_2 \hspace{0.1cm} \frac{2}{1} + \frac{1}{2} \cdot {\rm log}_2 \hspace{0.1cm} \frac{2}{1} \hspace{0.15cm}\underline{ = 1\,{\rm bit}} \hspace{0.05cm},$$
+
  = \frac{1}{2} \cdot {\rm log}_2 \hspace{0.1cm} \frac{2}{1} + \frac{1}{2} \cdot {\rm log}_2 \hspace{0.1cm} \frac{2}{1} \hspace{0.15cm}\underline{ = 1\ {\rm bit}} \hspace{0.05cm},$$
:$${\rm E} \left [-{\rm log}_2 \hspace{0.1cm} P_V(V)\right ]  
+
:$${\rm E} \big [-{\rm log}_2 \hspace{0.1cm} P_V(V)\big ]  
  = \frac{3}{4} \cdot {\rm log}_2 \hspace{0.1cm} \frac{4}{3} + \frac{1}{4} \cdot {\rm log}_2 \hspace{0.1cm} \frac{4}{1} \hspace{0.15cm}\underline{ = 0.811\,{\rm bit}} \hspace{0.05cm},$$
+
  = \frac{3}{4} \cdot {\rm log}_2 \hspace{0.1cm} \frac{4}{3} + \frac{1}{4} \cdot {\rm log}_2 \hspace{0.1cm} \frac{4}{1} \hspace{0.15cm}\underline{ = 0.811\ {\rm bit}} \hspace{0.05cm},$$
:$${\rm E} \left [-{\rm log}_2 \hspace{0.1cm} P_V(U)\right ]  
+
:$${\rm E} \big [-{\rm log}_2 \hspace{0.1cm} P_V(U)\big ]  
  = \frac{1}{2} \cdot {\rm log}_2 \hspace{0.1cm} \frac{4}{3} + \frac{1}{2} \cdot {\rm log}_2 \hspace{0.1cm} \frac{4}{1} \hspace{0.15cm}\underline{ = 1.208\,{\rm bit}} \hspace{0.05cm}.$$
+
  = \frac{1}{2} \cdot {\rm log}_2 \hspace{0.1cm} \frac{4}{3} + \frac{1}{2} \cdot {\rm log}_2 \hspace{0.1cm} \frac{4}{1} \hspace{0.15cm}\underline{ = 1.208\ {\rm bit}} \hspace{0.05cm}.$$
Richtig sind demnach die <u>beiden ersten Aussagen</u>:
+
Accordingly, the <u>first two statements</u> are correct:
* Die Entropie <i>H</i>(<i>U</i>) = 1 bit kann entsprechend der ersten Gleichung berechnet werden. Sie gilt für die binäre Zufallsgröße <i>U</i> mit gleichen Wahrscheinlichkeiten.
+
* The entropy&nbsp; $H(U) = 1$&nbsp; bit&nbsp; can be calculated according to the first equation.&nbsp; It applies to the binary random variable&nbsp; $U$&nbsp; with equal probabilities.
* Die Entropie <i>H</i>(<i>V</i>) = 0.811 bit berechnet sich entsprechend der zweiten Gleichung. Aufgrund der Wahrscheinlichkeiten 3/4 und 1/4 ist die Entropie (Unsicherheit) kleiner als für die Zufallsgröße <i>U</i>.
+
* The entropy&nbsp; $H(V) = 0.811$&nbsp; bit&nbsp; is calculated according to the second equation.&nbsp; Due to the probabilities&nbsp; $3/4&nbsp; and&nbsp;1/4$&nbsp;, the entropy (uncertainty) is smaller here than for the random variable&nbsp; $U$.
* Der dritte Erwartungswert kann schon allein vom Ergebnis her (1.208 bit) nicht die Entropie einer binären Zufallsgröße angeben, die stets auf 1 bit begrenzt ist.
+
* The third expected value cannot indicate the entropy of a binary random variable, which is always limited to&nbsp; 1&nbsp; (bit)&nbsp;, simply because of the result &nbsp; $(1.208$&nbsp; bit$)$&nbsp;.
 
{{ML-Fuß}}
 
{{ML-Fuß}}
  
  
  
[[Category:Aufgaben zu Informationstheorie|^3.1 Allgemeines zu 2D-Zufallsgrößen^]]
+
[[Category:Information Theory: Exercises|^3.1 General Information on 2D Random Variables^]]

Latest revision as of 10:11, 24 September 2021

Two-dimensional
probability mass function

We consider the following probability mass functions  (PMF):

PX(X)=[1/2, 1/8, 0, 3/8],
PY(Y)=[1/2, 1/4, 1/4, 0],
PU(U)=[1/2, 1/2],
PV(V)=[3/4, 1/4].

For the associated random variables, let:

X={0, 1, 2, 3},     Y={0, 1, 2, 3},    U={0, 1},     V={0,1}.

Often, for such discrete random variables, one must have to calculate different expected values of the form

E[F(X)]=xsupp(PX)PX(x)F(x).

Here, denote:

  • PX(X)  denotes the probability mass function of the discrete random variable   X.
  • The  "support"  of  PX  includes all those realisations  x  of the random variable  X  with non-vanishing probability.
  • Formally, this can be written as
supp(PX)={x:xXand_PX(x)0}.
  • F(X)  is an (arbitrary) real-valued function that can be specified in the entire domain of definition of the random variable  X .


In the task, the expected values for various functions  F(X)  are to be calculated, among others for

  1.   F(X)=1/PX(X),
  2.   F(X)=PX(X),
  3.   F(X)=log2 PX(X).





Hints:

  • The exercise belongs to the chapter  Some preliminary remarks on two-dimensional random variables.
  • The two one-dimensional probability mass functions  PX(X)  and  PY(Y)  result from the presented 2D–PMF  PXY(X, Y),  as will be shown in  Exercise 3.2Z.
  • The binary probability mass functions  PU(U)  and  PV(V)  are obtained according to the modulo operations  U=Xmod2  and  V=Ymod2.


Questions

1

What are the results of the following expected values?

E[1/PX(X)] = 

E[1/PY(Y)] = 

2

Give the following expected values:

E[PX(X)] = 

E[PY(Y)] = 

3

Now calculate the following expected values:

E[PY(X)] = 

E[PX(Y)] = 

4

Which of the following statements are true?

E[log2 PU(U)]  gives the entropy of the random variable  U.
E[log2 PV(V)]  gives the entropy of the random variable  V.
E[log2 PV(U)]  gives the entropy of the random variable  V.


Solution

(1)  In general, the following applies to the expected value of the function  F(X)  with respect to the random variable  X:

E[F(X)]=xsupp(PX)PX(x)F(x).

In the present example,  X={0, 1, 2, 3}  and  PX(X)=[1/2, 1/8, 0, 3/8].

  • Because of  PX(X=2)=0 , the quantity to be taken into account  (the "support")  in the above summation thus results in
supp(PX)={0,1,3}.
  • With  F(X)=1/PX(X)  one further obtains:
E[1/PX(X)]=x{0,1,3}PX(x)1/PX(x)=x{0,1,3}1=3_.
  • The second expected value gives the same result with  supp(PY)={0,1,2} :
E[1/PY(Y)]=3_.


(2)  In both cases, the index of the probability mass function is identical with the random variable  (X  or   Y)  and we obtain

E[PX(X)]=x{0,1,3}PX(x)PX(x)=(1/2)2+(1/8)2+(3/8)2=13/320.406_,
E[PY(Y)]=y{0,1,2}PY(y)PY(y)=(1/2)2+(1/4)2+(1/4)2=0.375_.


(3)  The following equations apply here:

E[PY(X)]=x{0,1,3}PX(x)PY(x)=1212+1814+380=9/320.281_,
  • The expected value formation here refers to  PX(·), i.e. to the random variable  X.
  • PY(·) is the formal function without (direct) reference to the random variable  Y.
  • The same numerical value is obtained for the second expected value  (this does not have to be the case in general):
E[PX(Y)]=y{0,1,2}PY(y)PX(y)=1212+1418+140=9/320.281_.


(4)  We first calculate the three expected values:

E[log2PU(U)]=12log221+12log221=1 bit_,
E[log2PV(V)]=34log243+14log241=0.811 bit_,
E[log2PV(U)]=12log243+12log241=1.208 bit_.

Accordingly, the first two statements are correct:

  • The entropy  H(U)=1  bit  can be calculated according to the first equation.  It applies to the binary random variable  U  with equal probabilities.
  • The entropy  H(V)=0.811  bit  is calculated according to the second equation.  Due to the probabilities  3/4  and  1/4 , the entropy (uncertainty) is smaller here than for the random variable  U.
  • The third expected value cannot indicate the entropy of a binary random variable, which is always limited to  1  (bit) , simply because of the result   (1.208  bit) .