Difference between revisions of "Aufgaben:Exercise 3.4: Entropy for Different PMF"
(32 intermediate revisions by 5 users not shown) | |||
Line 1: | Line 1: | ||
− | {{quiz-Header|Buchseite= | + | {{quiz-Header|Buchseite=Information_Theory/Some_Preliminary_Remarks_on_Two-Dimensional_Random_Variables |
}} | }} | ||
− | [[File: | + | [[File:EN_Inf_Z_3_3.png|right|frame|Probability functions, each with M=4 elements]] |
− | In | + | In the first row of the adjacent table, the probability mass function denoted by $\rm (a)$ is given in the following. |
− | $$H_{\rm a}(X) = {\rm E} \ | + | For this PMF PX(X)=[0.1, 0.2, 0.3, 0.4] the entropy is to be calculated in subtask '''(1)''' : |
+ | :$$H_{\rm a}(X) = {\rm E} \big [ {\rm log}_2 \hspace{0.1cm} \frac{1}{P_{X}(X)}\big ]= - {\rm E} \big [ {\rm log}_2 \hspace{0.1cm}{P_{X}(X)}\big ].$$ | ||
+ | Since the logarithm to the base 2 is used here, the pseudo-unit "bit" is to be added. | ||
− | + | In the further tasks, some probabilities are to be varied in each case in such a way that the greatest possible entropy results: | |
+ | |||
+ | * By suitably varying p3 and p4, one arrives at the maximum entropy Hb(X) under the condition p1=0.1 and $p_2 = 0.2$ ⇒ subtask '''(2)'''. | ||
+ | * By varying p2 and p3 appropriately, one arrives at the maximum entropy Hc(X) under the condition p1=0.1 and p4=0.4 ⇒ subtask '''(3)'''. | ||
+ | * In subtask '''(4)''' all four parameters are released for variation, which are to be determined according to the maximum entropy ⇒ Hmax(X) . | ||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
+ | Hints: | ||
+ | *The exercise belongs to the chapter [[Information_Theory/Einige_Vorbemerkungen_zu_zweidimensionalen_Zufallsgrößen|Some preliminary remarks on two-dimensional random variables]]. | ||
+ | *In particular, reference is made to the page [[Information_Theory/Einige_Vorbemerkungen_zu_zweidimensionalen_Zufallsgrößen#Probability_mass_function_and_entropy|Probability mass function and entropy]]. | ||
+ | |||
− | === | + | ===Questions=== |
<quiz display=simple> | <quiz display=simple> | ||
− | { | + | {To which entropy does the probability mass function PX(X)=[0.1, 0.2, 0.3, 0.4] lead? |
− | |type=" | + | |type="{}"} |
− | + | Ha(X) = { 1.846 0.5% } bit | |
− | |||
+ | {Let PX(X)=[0.1, 0.2, p3, p4] apply in general. What entropy is obtained if p3 and p4 are chosen as best as possible? | ||
+ | |type="{}"} | ||
+ | Hb(X) = { 1.857 0.5% } bit | ||
− | { | + | { Now let PX(X)=[0.1, p2, p3, 0.4]. What entropy is obtained if p2 and p3 are chosen as best as possible? |
|type="{}"} | |type="{}"} | ||
− | $\ | + | $H_{\rm c}(X) \ = \ $ { 1.861 0.5% } bit |
+ | { What entropy is obtained if all probabilities (p1, p2, p3, p4) can be chosen as best as possible? | ||
+ | |type="{}"} | ||
+ | Hmax(X) = { 2 1% } bit | ||
</quiz> | </quiz> | ||
− | === | + | ===Solution=== |
{{ML-Kopf}} | {{ML-Kopf}} | ||
− | '''1 | + | '''(1)''' With PX(X)=[0.1, 0.2, 0.3, 0.4] we get for the entropy: |
− | '''2 | + | :$$H_{\rm a}(X) = |
− | '''3 | + | 0.1 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.1} + |
− | ''' | + | 0.2 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.2} + |
− | ''' | + | 0.3 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.3} + |
− | ''' | + | 0.4 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.4} |
− | ''' | + | \hspace{0.15cm} \underline {= 1.846} \hspace{0.05cm}.$$ |
+ | Here (and in the other tasks) the pseudo-unit "bit" is to be added in each case. | ||
+ | |||
+ | |||
+ | |||
+ | '''(2)''' The entropy Hb(X) can be represented as the sum of two parts Hb1(X) and Hb2(X), with: | ||
+ | :$$H_{\rm b1}(X) = | ||
+ | 0.1 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.1} + | ||
+ | 0.2 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.2} = 0.797 \hspace{0.05cm},$$ | ||
+ | :$$H_{\rm b2}(X) = | ||
+ | p_3 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{p_3} + | ||
+ | (0.7-p_3) \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.7-p_3} \hspace{0.05cm}.$$ | ||
+ | |||
+ | *The second function is maximum for p3=p4=0.35. A similar relationship has been found for the binary entropy function. | ||
+ | *Thus one obtains: | ||
+ | |||
+ | :$$H_{\rm b2}(X) = 2 \cdot | ||
+ | p_3 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{p_3} = | ||
+ | 0.7 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.35} = 1.060 $$ | ||
+ | :⇒Hb(X)=Hb1(X)+Hb2(X)=0.797+1.060=1.857_. | ||
+ | |||
+ | |||
+ | |||
+ | '''(3)''' Analogous to subtask '''(2)''', p1=0.1 and p4=0.4 yield the maximum for p2=p3=0.25: | ||
+ | :$$H_{\rm c}(X) = | ||
+ | 0.1 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.1} + | ||
+ | 2 \cdot 0.25 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.25} + | ||
+ | 0.4 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.4} | ||
+ | \hspace{0.15cm} \underline {= 1.861} \hspace{0.05cm}.$$ | ||
+ | |||
+ | |||
+ | |||
+ | '''(4)''' The maximum entropy for the symbol range M=4 is obtained for equal probabilities, i.e. for p1=p2=p3=p4=0.25: | ||
+ | :$$H_{\rm max}(X) = | ||
+ | {\rm log}_2 \hspace{0.1cm} M | ||
+ | \hspace{0.15cm} \underline {= 2} \hspace{0.05cm}.$$ | ||
+ | |||
+ | *The difference of the entropies according to '''(4)''' and '''(3)''' gives ΔH(X)=0.139 bit. Here: | ||
+ | :$${\rm \Delta} H(X) = 1- | ||
+ | 0.1 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.1} - | ||
+ | 0.4 \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{0.4} | ||
+ | \hspace{0.05cm}.$$ | ||
+ | |||
+ | *With the binary entropy function | ||
+ | |||
+ | :$$H_{\rm bin}(p) = | ||
+ | p \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{p} + | ||
+ | (1-p) \cdot {\rm log}_2 \hspace{0.1cm} \frac{1}{1-p}$$ | ||
+ | |||
+ | :can also be written for this: | ||
+ | |||
+ | :$${\rm \Delta} H(X) = 0.5 \cdot \big [ 1- H_{\rm bin}(0.2) \big ] = | ||
+ | 0.5 \cdot \big [ 1- 0.722 \big ] = 0.139 | ||
+ | \hspace{0.05cm}.$$ | ||
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
{{ML-Fuß}} | {{ML-Fuß}} | ||
− | [[Category: | + | [[Category:Information Theory: Exercises|^3.1 General Information on 2D Random Variables^]] |
Latest revision as of 10:13, 24 September 2021
In the first row of the adjacent table, the probability mass function denoted by (a) is given in the following.
For this PMF PX(X)=[0.1, 0.2, 0.3, 0.4] the entropy is to be calculated in subtask (1) :
- Ha(X)=E[log21PX(X)]=−E[log2PX(X)].
Since the logarithm to the base 2 is used here, the pseudo-unit "bit" is to be added.
In the further tasks, some probabilities are to be varied in each case in such a way that the greatest possible entropy results:
- By suitably varying p3 and p4, one arrives at the maximum entropy Hb(X) under the condition p1=0.1 and p2=0.2 ⇒ subtask (2).
- By varying p2 and p3 appropriately, one arrives at the maximum entropy Hc(X) under the condition p1=0.1 and p4=0.4 ⇒ subtask (3).
- In subtask (4) all four parameters are released for variation, which are to be determined according to the maximum entropy ⇒ Hmax(X) .
Hints:
- The exercise belongs to the chapter Some preliminary remarks on two-dimensional random variables.
- In particular, reference is made to the page Probability mass function and entropy.
Questions
Solution
- Ha(X)=0.1⋅log210.1+0.2⋅log210.2+0.3⋅log210.3+0.4⋅log210.4=1.846_.
Here (and in the other tasks) the pseudo-unit "bit" is to be added in each case.
(2) The entropy Hb(X) can be represented as the sum of two parts Hb1(X) and Hb2(X), with:
- Hb1(X)=0.1⋅log210.1+0.2⋅log210.2=0.797,
- Hb2(X)=p3⋅log21p3+(0.7−p3)⋅log210.7−p3.
- The second function is maximum for p3=p4=0.35. A similar relationship has been found for the binary entropy function.
- Thus one obtains:
- Hb2(X)=2⋅p3⋅log21p3=0.7⋅log210.35=1.060
- ⇒Hb(X)=Hb1(X)+Hb2(X)=0.797+1.060=1.857_.
(3) Analogous to subtask (2), p1=0.1 and p4=0.4 yield the maximum for p2=p3=0.25:
- Hc(X)=0.1⋅log210.1+2⋅0.25⋅log210.25+0.4⋅log210.4=1.861_.
(4) The maximum entropy for the symbol range M=4 is obtained for equal probabilities, i.e. for p1=p2=p3=p4=0.25:
- Hmax(X)=log2M=2_.
- The difference of the entropies according to (4) and (3) gives ΔH(X)=0.139 bit. Here:
- ΔH(X)=1−0.1⋅log210.1−0.4⋅log210.4.
- With the binary entropy function
- Hbin(p)=p⋅log21p+(1−p)⋅log211−p
- can also be written for this:
- ΔH(X)=0.5⋅[1−Hbin(0.2)]=0.5⋅[1−0.722]=0.139.