Difference between revisions of "Aufgaben:Exercise 2.6: About the Huffman Coding"
(33 intermediate revisions by 4 users not shown) | |||
Line 1: | Line 1: | ||
− | {{quiz-Header|Buchseite= | + | {{quiz-Header|Buchseite=Information_Theory/Entropy_Coding_According_to_Huffman |
}} | }} | ||
− | [[File: | + | [[File:EN_Inf_A_2_6_v2.png|right|frame|Tree diagram of the Huffman method]] |
− | + | We consider here a source symbol sequence $\langle q_\nu \rangle$ with symbol set size $M = 8$: | |
− | :$$q_{\nu} | + | :$$q_{\nu} \in \{ \hspace{0.05cm}q_{\mu} \} = \{ \boldsymbol{\rm A} \hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm B}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm C}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm D}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm E}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm F}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm G}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm H}\hspace{0.05cm} |
\}\hspace{0.05cm}.$$ | \}\hspace{0.05cm}.$$ | ||
− | + | If the symbols are equally probable, i.e. $p_{\rm A} = p_{\rm B} =$ ... $ = p_{\rm H} = 1/M$ then source coding makes no sense. Already with the dual code $\rm A$ → <b>000</b>, $\rm B$ → <b>001</b>, ... , $\rm H$ → <b>111</b>, the average code word length $L_{\rm M}$ is its lower bound $H$ according to the source coding theorem $(H$ denotes the "source entropy" here$)$: | |
− | :$$L_{\rm M,\hspace{0.08cm}min} = H = 3 \hspace{0.15cm}{\rm bit/ | + | :$$L_{\rm M,\hspace{0.08cm}min} = H = 3 \hspace{0.15cm}{\rm bit/source \:symbol} \hspace{0.05cm}.$$ |
− | + | However, let the symbol probabilities be given as follows: | |
:$$p_{\rm A} \hspace{-0.05cm}= \hspace{-0.05cm} 0.04 \hspace{0.05cm},\hspace{0.1cm}p_{\rm B} \hspace{-0.05cm}= \hspace{-0.05cm} 0.08 \hspace{0.05cm},\hspace{0.1cm}p_{\rm C} \hspace{-0.05cm}= \hspace{-0.05cm} 0.14 \hspace{0.05cm},\hspace{0.1cm} | :$$p_{\rm A} \hspace{-0.05cm}= \hspace{-0.05cm} 0.04 \hspace{0.05cm},\hspace{0.1cm}p_{\rm B} \hspace{-0.05cm}= \hspace{-0.05cm} 0.08 \hspace{0.05cm},\hspace{0.1cm}p_{\rm C} \hspace{-0.05cm}= \hspace{-0.05cm} 0.14 \hspace{0.05cm},\hspace{0.1cm} | ||
− | p_{\rm D} \hspace{-0.05cm}= \hspace{-0.05cm} 0.25 \hspace{0.05cm}, | + | p_{\rm D} \hspace{-0.05cm}= \hspace{-0.05cm} 0.25 \hspace{0.05cm},\hspace{0.1cm} p_{\rm E} \hspace{-0.05cm}= \hspace{-0.05cm} 0.24 \hspace{0.05cm},\hspace{0.1cm}p_{\rm F} \hspace{-0.05cm}= \hspace{-0.05cm} 0.12 \hspace{0.05cm},\hspace{0.1cm}p_{\rm G} \hspace{-0.05cm}= \hspace{-0.05cm} 0.10 \hspace{0.05cm},\hspace{0.1cm} |
− | |||
p_{\rm H} \hspace{-0.05cm}= \hspace{-0.05cm} 0.03 \hspace{0.05cm}.$$ | p_{\rm H} \hspace{-0.05cm}= \hspace{-0.05cm} 0.03 \hspace{0.05cm}.$$ | ||
− | + | *So here we have a redundant source that can be compressed by Huffman coding. | |
+ | *This algorithm was published by [https://en.wikipedia.org/wiki/David_A._Huffman David Albert Huffman] – shortly after Shannon's groundbreaking work on information theory – and allows the construction of optimal prefix-free codes. | ||
− | |||
− | : | + | The algorithm is given here without derivation and proof, whereby we restrict ourselves to binary codes ⇒ the encoded sequence consists only of zeros and ones: |
− | : | + | :'''(1)''' Order the symbols according to decreasing probabilities of occurrence. |
− | : | + | :'''(2)''' Combine the two least likely symbols into a new symbol. |
− | : | + | :'''(3)''' Repeat steps '''(1)''' and '''(2)''', until only two (combined) symbols remain. |
− | : | + | :'''(4)''' Binary encode the more probable set of symbols with <b>1</b> , the other set with <b>0</b>. |
− | + | :'''(5)''' Step by step (from bottom to top) add <b>1</b> or <b>0</b> to the split partial codes. | |
− | |||
− | + | This algorithm is often illustrated by a tree diagram. The diagram above shows this for the case at hand. | |
− | : | + | You have the following tasks: |
− | + | :'''(a)''' Assign the symbols $\rm A$, ... , $\rm H$ to the inputs labelled '''[1]''', ... , '''[8]''' . | |
− | : | ||
− | |||
+ | :'''(b)''' Determination of the sum probabilities $U$, ... , $Z$ and $R$ ("root"). | ||
− | '' | + | :'''(c)''' Assignment of the symbols $\rm A$, ... , $\rm H$ to the corresponding Huffman binary sequences. A red connection corresponds to a <b>1</b>, a blue connection to a <b>0</b>. |
− | |||
− | |||
− | |||
+ | You will notice that the average code word length for Huffman coding, | ||
+ | :$$L_{\rm M} = \sum_{\mu = 1}^{M}\hspace{0.05cm} p_{\mu} \cdot L_{\mu}, $$ | ||
+ | is only slightly larger than the source entropy $H$. In this equation, the following values apply to the present case: | ||
+ | *$M = 8$, $p_1 = p_{\rm A}$, ... , $p_8 = p_{\rm H}$. | ||
+ | *The respective number of bits of the code symbols for $\rm A$, ... , $\rm H$ is denoted by $L_1$, ... , $L_8$ . | ||
− | === | + | |
+ | |||
+ | |||
+ | |||
+ | |||
+ | |||
+ | <u>Hints:</u> | ||
+ | *The exercise belongs to the chapter [[Information_Theory/Entropiecodierung_nach_Huffman|Entropy Coding according to Huffman]]. | ||
+ | *In particular, reference is made to the pages | ||
+ | ** [[Information_Theory/Entropiecodierung_nach_Huffman#The_Huffman_algorithm|The Huffman algorithm]], | ||
+ | **[[Information_Theory/Entropiecodierung_nach_Huffman#Representation_of_the_Huffman_code_as_a_tree_diagram|Representation of the Huffman code as a tree diagram]]. | ||
+ | *To check your results, please refer to the (German language) SWF module [[Applets:Huffman_Shannon_Fano|Coding according to Huffman and Shannon/Fano]]. | ||
+ | |||
+ | |||
+ | |||
+ | ===Questions=== | ||
<quiz display=simple> | <quiz display=simple> | ||
− | { | + | {Which inputs in the tree diagram stand for |
|type="{}"} | |type="{}"} | ||
− | + | Input number $ \ = \ $ { 7 } ⇒ symbol $\rm A$ | |
− | + | Input number $ \ = \ $ { 6 } ⇒ symbol $\rm B$ | |
− | + | Input number $ \ = \ $ { 3 } ⇒ symbol $\rm C$ | |
− | + | Input number $ \ = \ $ { 1 } ⇒ symbol $\rm D$ | |
− | { | + | {What numerical values (probabilities) should be at the nodes in the tree diagram? |
|type="{}"} | |type="{}"} | ||
− | + | Probability $ \ = \ $ { 0.07 1% } at node $\rm U$ | |
− | + | Probability $ \ = \ $ { 0.15 1% } at node $\rm V$ | |
− | + | Probability $ \ = \ $ { 0.22 1% } at node $\rm W$ | |
− | + | Probability $ \ = \ $ { 0.29 1% } at node $\rm X$ | |
− | + | Probability $ \ = \ $ { 0.46 1% } at node $\rm Y$ | |
+ | Probability $ \ = \ $ { 0.54 1% } at node $\rm Z$ | ||
+ | Probability $ \ = \ $ { 1 1% } at $\rm root$ | ||
− | { | + | {Which binary codes (represented with zeros and ones) result for |
|type="{}"} | |type="{}"} | ||
− | + | Binary code $ \ = \ $ { 11101 } ⇒ symbol $\rm A$ | |
− | + | Binary code $ \ = \ $ { 1111 } ⇒ symbol $\rm B$ | |
− | + | Binary code $ \ = \ $ { 110 } ⇒ symbol $\rm C$ | |
− | + | Binary code $ \ = \ $ { 10 } ⇒ symbol $\rm D$ | |
− | { | + | {What is the average code word length? |
|type="{}"} | |type="{}"} | ||
− | $ | + | $L_{\rm M} \ = \ $ { 2.73 3% } $\ \rm bit/source \hspace{0.15cm} symbol$ |
− | { | + | {What is the source entropy $H$? |
− | |type=" | + | |type="()"} |
− | + H = 2.71 bit/ | + | + $H = 2.71\ \rm bit/source \hspace{0.15cm}symbol.$ |
− | - H = 2.75 bit/ | + | - $H = 2.75\ \rm bit/source \hspace{0.15cm}symbol.$ |
− | - H = 3.00 bit/ | + | - $H = 3.00\ \rm bit/source \hspace{0.15cm}symbol.$ |
Line 92: | Line 108: | ||
</quiz> | </quiz> | ||
− | === | + | ===Solution=== |
{{ML-Kopf}} | {{ML-Kopf}} | ||
− | < | + | '''(1)''' Before the Huffman algorithm, the symbols must be sorted according to their probabilities of occurrence. |
+ | *Since the two least likely symbols are already combined in the first step, the probabilities of occurrence decrease from top to bottom <br>(from left to right in the diagram for this sample solution). | ||
+ | *By comparison with the data sheet, one obtains: | ||
+ | |||
+ | :Symbol $\rm A$: <u>Input 7</u>, Symbol $\rm B$: <u>Input 6</u>, Symbol $\rm C$: <u>Input 3</u>, Symbol $\rm D$: <u>Input 1</u>. | ||
+ | |||
+ | [[File:EN_Inf_Z_2_6b.png|right|right|frame|Tree diagram adapted to the exercise]] | ||
+ | |||
+ | |||
+ | '''(2)''' The node $\rm R$ is the tree ("root"). This is always assigned $\underline{R=1}$, regardless of the occurrence probabilities. | ||
− | : | + | The following applies to the other values: |
− | |||
− | + | :Step 1: $p_{\rm U} = p_{\rm A} + p_{\rm H} = 0.04 + 0.03 \hspace{0.15cm}\underline{ =0.07}$, | |
− | : | + | :Step 2: $p_{\rm V} = p_{\rm U} + p_{\rm B} = 0.07 + 0.08 \hspace{0.15cm}\underline{ =0.15}$, |
− | : | + | :Step 3: $p_{\rm W} = p_{\rm F} + p_{\rm G} = 0.12 + 0.10 \hspace{0.15cm}\underline{ =0.22}$, |
− | : | + | :Step 4: $p_{\rm X} = p_{\rm V} + p_{\rm C} = 0.15 + 0.14 \hspace{0.15cm}\underline{ =0.29}$, |
− | : | + | :Step 5: $p_{\rm Y} = p_{\rm W} + p_{\rm E} = 0.22 + 0.24 \hspace{0.15cm}\underline{ =0.46}$, |
− | : | + | :Step 6: $p_{\rm Z} = p_{\rm X} + p_{\rm D} = 0.29 + 0.25 \hspace{0.15cm}\underline{ =0.54}$. |
+ | <br clear=all> | ||
+ | '''(3)''' The Huffman code for symbol $\rm A$ is obtained by tracing the path from $\rm root$ (yellow dot) to symbol $\rm A$ and assigning a <b>1</b> to each red line and a <b>0</b> to each blue one. | ||
− | + | * Symbol $\rm A$: red–red–red–blue–red→ <b><u>11101</u></b>, | |
− | + | * Symbol $\rm B$: red–red–red–red → <b><u>1111</u></b>, | |
− | |||
− | + | * Symbol $\rm C$: red–red–blue → <b><u>110</u></b>, | |
− | + | * Symbol $\rm D$: red–blue– → <b><u>10</u></b>, | |
− | + | * Symbol $\rm E$: blue–red → <b>01</b>, | |
− | + | * Symbol $\rm F$: blue–blue–red → <b>001</b>, | |
− | + | * Symbol $\rm G$: blue–blue–blue → <b>000</b>, | |
− | + | * Symbol $\rm H$: red–red–red–blue–blue → <b>11100</b>. | |
− | |||
− | + | '''(4)''' The coding under point '''(3)''' has shown that | |
− | + | * the symbols $\rm D$ and $\rm E$ are each represented with two bits, | |
− | + | * the symbols $\rm C$, $\rm F$ and $\rm G$ with three bits, | |
− | + | * the symbols $\rm B$ with four bits, and | |
− | + | * the symbols $\rm A$ and $\rm H$ with five bits. | |
− | |||
− | : | + | Thus, for the average code word length (in "bit/source symbol"): |
+ | :$$L_{\rm M} \hspace{0.2cm} = \hspace{0.2cm} (p_{\rm D} + p_{\rm E}) \cdot 2 + (p_{\rm C} + p_{\rm F} + p_{\rm G}) \cdot 3 + p_{\rm B} \cdot 4 +(p_{\rm A} + p_{\rm H}) \cdot 5 = 0.49 \cdot 2 + 0.36 \cdot 3 +0.08 \cdot 4 +0.07 \cdot 5 | ||
+ | \hspace{0.15cm}\underline{= 2.73}\hspace{0.05cm}.$$ | ||
− | |||
− | |||
− | |||
− | |||
− | |||
− | + | '''(5)''' Only <u>answer 1</u> is correct: | |
+ | *The average code word length $L_{\rm M}$ cannot be smaller than the source entropy $H$. This eliminates answers 2 and 3. | ||
+ | *From the given probabilities of occurrence, the source entropy can actually be calculated to $H = 2.71$ bit/source symbol. | ||
+ | *It can be seen that this Huffman coding almost reaches the given limit ("Source Coding Theorem") $L_{\rm M, \ min } \ge H = 2.71$ bit/source symbol. | ||
{{ML-Fuß}} | {{ML-Fuß}} | ||
− | [[Category: | + | [[Category:Information Theory: Exercises|^2.3 Entropy Coding according to Huffman^]] |
Latest revision as of 15:54, 1 November 2022
We consider here a source symbol sequence $\langle q_\nu \rangle$ with symbol set size $M = 8$:
- $$q_{\nu} \in \{ \hspace{0.05cm}q_{\mu} \} = \{ \boldsymbol{\rm A} \hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm B}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm C}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm D}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm E}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm F}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm G}\hspace{0.05cm},\hspace{0.05cm} \boldsymbol{\rm H}\hspace{0.05cm} \}\hspace{0.05cm}.$$
If the symbols are equally probable, i.e. $p_{\rm A} = p_{\rm B} =$ ... $ = p_{\rm H} = 1/M$ then source coding makes no sense. Already with the dual code $\rm A$ → 000, $\rm B$ → 001, ... , $\rm H$ → 111, the average code word length $L_{\rm M}$ is its lower bound $H$ according to the source coding theorem $(H$ denotes the "source entropy" here$)$:
- $$L_{\rm M,\hspace{0.08cm}min} = H = 3 \hspace{0.15cm}{\rm bit/source \:symbol} \hspace{0.05cm}.$$
However, let the symbol probabilities be given as follows:
- $$p_{\rm A} \hspace{-0.05cm}= \hspace{-0.05cm} 0.04 \hspace{0.05cm},\hspace{0.1cm}p_{\rm B} \hspace{-0.05cm}= \hspace{-0.05cm} 0.08 \hspace{0.05cm},\hspace{0.1cm}p_{\rm C} \hspace{-0.05cm}= \hspace{-0.05cm} 0.14 \hspace{0.05cm},\hspace{0.1cm} p_{\rm D} \hspace{-0.05cm}= \hspace{-0.05cm} 0.25 \hspace{0.05cm},\hspace{0.1cm} p_{\rm E} \hspace{-0.05cm}= \hspace{-0.05cm} 0.24 \hspace{0.05cm},\hspace{0.1cm}p_{\rm F} \hspace{-0.05cm}= \hspace{-0.05cm} 0.12 \hspace{0.05cm},\hspace{0.1cm}p_{\rm G} \hspace{-0.05cm}= \hspace{-0.05cm} 0.10 \hspace{0.05cm},\hspace{0.1cm} p_{\rm H} \hspace{-0.05cm}= \hspace{-0.05cm} 0.03 \hspace{0.05cm}.$$
- So here we have a redundant source that can be compressed by Huffman coding.
- This algorithm was published by David Albert Huffman – shortly after Shannon's groundbreaking work on information theory – and allows the construction of optimal prefix-free codes.
The algorithm is given here without derivation and proof, whereby we restrict ourselves to binary codes ⇒ the encoded sequence consists only of zeros and ones:
- (1) Order the symbols according to decreasing probabilities of occurrence.
- (2) Combine the two least likely symbols into a new symbol.
- (3) Repeat steps (1) and (2), until only two (combined) symbols remain.
- (4) Binary encode the more probable set of symbols with 1 , the other set with 0.
- (5) Step by step (from bottom to top) add 1 or 0 to the split partial codes.
This algorithm is often illustrated by a tree diagram. The diagram above shows this for the case at hand.
You have the following tasks:
- (a) Assign the symbols $\rm A$, ... , $\rm H$ to the inputs labelled [1], ... , [8] .
- (b) Determination of the sum probabilities $U$, ... , $Z$ and $R$ ("root").
- (c) Assignment of the symbols $\rm A$, ... , $\rm H$ to the corresponding Huffman binary sequences. A red connection corresponds to a 1, a blue connection to a 0.
You will notice that the average code word length for Huffman coding,
- $$L_{\rm M} = \sum_{\mu = 1}^{M}\hspace{0.05cm} p_{\mu} \cdot L_{\mu}, $$
is only slightly larger than the source entropy $H$. In this equation, the following values apply to the present case:
- $M = 8$, $p_1 = p_{\rm A}$, ... , $p_8 = p_{\rm H}$.
- The respective number of bits of the code symbols for $\rm A$, ... , $\rm H$ is denoted by $L_1$, ... , $L_8$ .
Hints:
- The exercise belongs to the chapter Entropy Coding according to Huffman.
- In particular, reference is made to the pages
- To check your results, please refer to the (German language) SWF module Coding according to Huffman and Shannon/Fano.
Questions
Solution
- Since the two least likely symbols are already combined in the first step, the probabilities of occurrence decrease from top to bottom
(from left to right in the diagram for this sample solution). - By comparison with the data sheet, one obtains:
- Symbol $\rm A$: Input 7, Symbol $\rm B$: Input 6, Symbol $\rm C$: Input 3, Symbol $\rm D$: Input 1.
(2) The node $\rm R$ is the tree ("root"). This is always assigned $\underline{R=1}$, regardless of the occurrence probabilities.
The following applies to the other values:
- Step 1: $p_{\rm U} = p_{\rm A} + p_{\rm H} = 0.04 + 0.03 \hspace{0.15cm}\underline{ =0.07}$,
- Step 2: $p_{\rm V} = p_{\rm U} + p_{\rm B} = 0.07 + 0.08 \hspace{0.15cm}\underline{ =0.15}$,
- Step 3: $p_{\rm W} = p_{\rm F} + p_{\rm G} = 0.12 + 0.10 \hspace{0.15cm}\underline{ =0.22}$,
- Step 4: $p_{\rm X} = p_{\rm V} + p_{\rm C} = 0.15 + 0.14 \hspace{0.15cm}\underline{ =0.29}$,
- Step 5: $p_{\rm Y} = p_{\rm W} + p_{\rm E} = 0.22 + 0.24 \hspace{0.15cm}\underline{ =0.46}$,
- Step 6: $p_{\rm Z} = p_{\rm X} + p_{\rm D} = 0.29 + 0.25 \hspace{0.15cm}\underline{ =0.54}$.
(3) The Huffman code for symbol $\rm A$ is obtained by tracing the path from $\rm root$ (yellow dot) to symbol $\rm A$ and assigning a 1 to each red line and a 0 to each blue one.
- Symbol $\rm A$: red–red–red–blue–red→ 11101,
- Symbol $\rm B$: red–red–red–red → 1111,
- Symbol $\rm C$: red–red–blue → 110,
- Symbol $\rm D$: red–blue– → 10,
- Symbol $\rm E$: blue–red → 01,
- Symbol $\rm F$: blue–blue–red → 001,
- Symbol $\rm G$: blue–blue–blue → 000,
- Symbol $\rm H$: red–red–red–blue–blue → 11100.
(4) The coding under point (3) has shown that
- the symbols $\rm D$ and $\rm E$ are each represented with two bits,
- the symbols $\rm C$, $\rm F$ and $\rm G$ with three bits,
- the symbols $\rm B$ with four bits, and
- the symbols $\rm A$ and $\rm H$ with five bits.
Thus, for the average code word length (in "bit/source symbol"):
- $$L_{\rm M} \hspace{0.2cm} = \hspace{0.2cm} (p_{\rm D} + p_{\rm E}) \cdot 2 + (p_{\rm C} + p_{\rm F} + p_{\rm G}) \cdot 3 + p_{\rm B} \cdot 4 +(p_{\rm A} + p_{\rm H}) \cdot 5 = 0.49 \cdot 2 + 0.36 \cdot 3 +0.08 \cdot 4 +0.07 \cdot 5 \hspace{0.15cm}\underline{= 2.73}\hspace{0.05cm}.$$
(5) Only answer 1 is correct:
- The average code word length $L_{\rm M}$ cannot be smaller than the source entropy $H$. This eliminates answers 2 and 3.
- From the given probabilities of occurrence, the source entropy can actually be calculated to $H = 2.71$ bit/source symbol.
- It can be seen that this Huffman coding almost reaches the given limit ("Source Coding Theorem") $L_{\rm M, \ min } \ge H = 2.71$ bit/source symbol.