Difference between revisions of "Aufgaben:Exercise 3.14: Channel Coding Theorem"

From LNTwww
 
(4 intermediate revisions by 2 users not shown)
Line 1: Line 1:
  
{{quiz-Header|Buchseite=Informationstheorie/Anwendung auf die Digitalsignalübertragung
+
{{quiz-Header|Buchseite=Information_Theory/Application_to_Digital_Signal_Transmission
 
}}
 
}}
  
 
[[File:EN_Inf_A_3_14.png|right|frame|Information-theoretical quantities of the  $\rm BSC$   and  $\rm EUC models$]]
 
[[File:EN_Inf_A_3_14.png|right|frame|Information-theoretical quantities of the  $\rm BSC$   and  $\rm EUC models$]]
[https://en.wikipedia.org/wiki/Claude_Shannon Shannon's]  channel coding theorem states that an error-free transmission can be made over a ''discrete memoryless channel'' (DMC) with the code rate  $R$  as long as  $R$  is not greater than the channel capacity
+
[https://en.wikipedia.org/wiki/Claude_Shannon Shannon's]  channel coding theorem states that an error-free transmission can be made over a  "discrete memoryless channel"  $\rm(DMC)$  with the code rate  $R$  as long as  $R$  is not greater than the channel capacity
 
:$$C = \max_{P_X(X)} \hspace{0.15cm} I(X;Y) \hspace{0.05cm}.$$
 
:$$C = \max_{P_X(X)} \hspace{0.15cm} I(X;Y) \hspace{0.05cm}.$$
The channel coding theorem is to be evaluated numerically in this task, whereby two typical channel models are to be considered:
+
The  "Channel Coding Theorem"  is to be evaluated numerically in this task, whereby two typical channel models are to be considered:
* the  $\rm BSC$ model  (''Binary Symmetric Channel'')  with distortion probability $ε = 0.25$  and channel capacity  $C = 1 - H_{\rm bin}(ε),$
+
* The  $\rm BSC$  model  ("Binary Symmetric Channel")  with error probability $ε = 0.25$  and channel capacity  $C = 1 - H_{\rm bin}(ε),$
* the  $\rm EUC$ model  (from  ''Extremely Unsymmetric Channel'', this designation originates from us and is not common) according to)  [[Aufgaben:Aufgabe_3.11Z:_Extrem_unsymmetrischer_Kanal|Exercise 3.11Z]].
+
* the  $\rm EUC$  model  (from  "Extremely Unsymmetric Channel";  this designation originates from us and is not common) according to  [[Aufgaben:Aufgabe_3.11Z:_Extrem_unsymmetrischer_Kanal|Exercise 3.11Z]].
  
  
The graphs show the numerical values of the information-theoretical quantities for the two channels  $\rm BSC$  and  $\rm EUC$:
+
The graphs show the numerical values of the information-theoretical quantities for the two models  $\rm BSC$  and  $\rm EUC$:
* the source entropy  $H(X),$
+
* The source entropy  $H(X),$
 
* the equivocation  $H(X|Y),$
 
* the equivocation  $H(X|Y),$
 
* the mutual information  $I(X; Y),$
 
* the mutual information  $I(X; Y),$
* the irrelevance  $H(Y|X),$  and
+
* the irrelevance  $H(Y|X),$   
 
* the sink entropy  $H(Y).$
 
* the sink entropy  $H(Y).$
  
Line 29: Line 29:
 
Hints:
 
Hints:
 
*The exercise belongs to the chapter  [[Information_Theory/Anwendung_auf_die_Digitalsignalübertragung|Application to Digital Signal Transmission]].
 
*The exercise belongs to the chapter  [[Information_Theory/Anwendung_auf_die_Digitalsignalübertragung|Application to Digital Signal Transmission]].
*Reference is made in particular to the page     [[Information_Theory/Anwendung_auf_die_Digitalsignalübertragung#Definition_and_meaning_of_channel_capacity|Definition and meaning of channel capacity]].
+
*Reference is made in particular to the section     [[Information_Theory/Anwendung_auf_die_Digitalsignalübertragung#Definition_and_meaning_of_channel_capacity|Definition and meaning of channel capacity]].
 
   
 
   
  
Line 37: Line 37:
  
 
<quiz display=simple>
 
<quiz display=simple>
{Which statements are true for uncoded transmission &nbsp; &rArr; &nbsp;$\underline{R = 1}$ assuming&nbsp;$p_0 = p_1 = 0.5$&nbsp;?
+
{Which statements are true for uncoded transmission &nbsp; &rArr; &nbsp;$\underline{R = 1}$ assuming&nbsp; $p_0 = p_1 = 0.5$&nbsp;?
 
|type="()"}
 
|type="()"}
 
- BSC results in a smaller bit error probability.
 
- BSC results in a smaller bit error probability.
Line 43: Line 43:
 
+ Both models lead to the same bit error probability.
 
+ Both models lead to the same bit error probability.
 
   
 
   
{With &nbsp;$\underline{R = 1}$&nbsp; can the result be (formally) improved by other values of&nbsp; $p_0$ &nbsp;or&nbsp; $p_1&nbsp;?
+
{With &nbsp;$\underline{R = 1}$&nbsp; the result can be (formally) improved by other values of&nbsp; $p_0$ &nbsp;or&nbsp; $p_1&nbsp;?
 
|type="()"}
 
|type="()"}
- For both channels.
+
- With both channels.
 
- With the BSC model.
 
- With the BSC model.
 
+ With the EUC model.
 
+ With the EUC model.
Line 52: Line 52:
 
{Over which channel can error-free transmission be achieved with the&nbsp; $\underline{R = 0.16}$&nbsp;?
 
{Over which channel can error-free transmission be achieved with the&nbsp; $\underline{R = 0.16}$&nbsp;?
 
|type="()"}
 
|type="()"}
+ For both channels.
+
+ With both channels.
- For the BSC model.
+
- With the BSC model.
 
- With the EUC model.
 
- With the EUC model.
 
- With none of the models.
 
- With none of the models.
Line 60: Line 60:
 
|type="()"}
 
|type="()"}
 
- Over both channels.
 
- Over both channels.
- For the BSC model.
+
- With the BSC model.
 
+ With the EUC model.
 
+ With the EUC model.
- With no model.
+
- With none of the models.
  
 
{Over which channel can error-free transmission be achieved with the rate&nbsp; $\underline{R = 0.48}$&nbsp;?
 
{Over which channel can error-free transmission be achieved with the rate&nbsp; $\underline{R = 0.48}$&nbsp;?
 
|type="()"}
 
|type="()"}
- For both channels.
+
- With both channels.
- For the BSC model.
+
- With the BSC model.
 
- With the EUC model.
 
- With the EUC model.
 
+ With none of the models.
 
+ With none of the models.
Line 74: Line 74:
 
===Solution===
 
===Solution===
 
{{ML-Kopf}}
 
{{ML-Kopf}}
'''(1)''' &nbsp; <u>Proposed solution 3</u> is correct:
+
'''(1)''' &nbsp; <u>Proposed solution 3</u>&nbsp; is correct:
*The BSC error probability is&nbsp; &rArr; &nbsp; $R = 1$ with $p_0 = p_1 = 0.5$&nbsp; for uncoded transmission: &nbsp;  
+
*The BSC error probability is with $p_0 = p_1 = 0.5$&nbsp; for uncoded transmission&nbsp; &rArr; &nbsp; $R = 1$: &nbsp;  
 
:$$ p_{\rm B} = 0.5 \cdot 0.25 + 0.5 \cdot 0.25=0.25 \hspace{0.05cm}.$$
 
:$$ p_{\rm B} = 0.5 \cdot 0.25 + 0.5 \cdot 0.25=0.25 \hspace{0.05cm}.$$
 
*Correspondingly, with the same boundary conditions for the EUC model::
 
*Correspondingly, with the same boundary conditions for the EUC model::
Line 82: Line 82:
  
  
'''(2)''' &nbsp; <u>Proposed solution 3</u> is correct:
+
'''(2)''' &nbsp; <u>Proposed solution 3</u>&nbsp; is correct:
*In the BSC model with the distortion probability&nbsp; $ε = 0.25$&nbsp;, for uncoded transmission &nbsp; &rArr; &nbsp; $R = 1$ &nbsp;, the bit error probability is equal to&nbsp; $p_{\rm B} = 0.25$. regardless of&nbsp; $p_0$&nbsp; and&nbsp; $p_1$  
+
*In the BSC model with error probability&nbsp; $ε = 0.25$,&nbsp; for uncoded transmission &nbsp; &rArr; &nbsp; $R = 1$&nbsp; the bit error probability is&nbsp; $p_{\rm B} = 0.25$,&nbsp; regardless of&nbsp; $p_0$&nbsp; and&nbsp; $p_1$  
 
*In contrast, with the EUC model, for example, a smaller bit error probability is obtained with&nbsp; $p_0 = 0.6$&nbsp; and&nbsp; $p_1 = 0.4$&nbsp;:
 
*In contrast, with the EUC model, for example, a smaller bit error probability is obtained with&nbsp; $p_0 = 0.6$&nbsp; and&nbsp; $p_1 = 0.4$&nbsp;:
 
:$$p_{\rm B} = 0.6 \cdot 0 + 0.4 \cdot 0.5=0.2 \hspace{0.05cm}.$$
 
:$$p_{\rm B} = 0.6 \cdot 0 + 0.4 \cdot 0.5=0.2 \hspace{0.05cm}.$$
 
*Note, however, that now the source entropy is no longer&nbsp; $H(X) = 1\ \rm  (bit)$&nbsp; but only&nbsp; $H(X) = H_{bin} (0.6) = 0.971 \ \rm  (bit)$.  
 
*Note, however, that now the source entropy is no longer&nbsp; $H(X) = 1\ \rm  (bit)$&nbsp; but only&nbsp; $H(X) = H_{bin} (0.6) = 0.971 \ \rm  (bit)$.  
*In the limiting case&nbsp; $p_0 = 1$&nbsp;, only zeros are transmitted and&nbsp; $H(X) = 0$.&nbsp; For the bit error probability, however, the following then actually applies:
+
*In the limiting case&nbsp; $p_0 = 1$,&nbsp; only zeros are transmitted and&nbsp; $H(X) = 0$.&nbsp; For the bit error probability, however, the following then actually applies:
 
:$$ p_{\rm B} = 1 \cdot 0 + 0 \cdot 0.5=0 \hspace{0.05cm}.$$
 
:$$ p_{\rm B} = 1 \cdot 0 + 0 \cdot 0.5=0 \hspace{0.05cm}.$$
 
:So no information is transmitted, but it is transmitted with the bit error probability "zero".
 
:So no information is transmitted, but it is transmitted with the bit error probability "zero".
Line 93: Line 93:
  
  
'''(3)''' &nbsp; <u>Proposed solution 1</u> is correct:
+
'''(3)''' &nbsp; <u>Proposed solution 1</u>&nbsp; is correct:
*From the graph on the information page, it can be read for the capacities of the two channels:
+
*From the graph in the information section,&nbsp; it can be read for the capacities of the two channels:
 
:$$C_{\rm BSC} = 0.1887 \ \rm {bit/use}, \hspace{0.5cm}C_{rm EUC} = 0.3219 \ \rm {bit/use}.$$
 
:$$C_{\rm BSC} = 0.1887 \ \rm {bit/use}, \hspace{0.5cm}C_{rm EUC} = 0.3219 \ \rm {bit/use}.$$
*According to the channel coding theorem, a channel coding can be found at&nbsp; $R ≤ C$&nbsp; with which the error probability can be made zero.
+
*According to the channel coding theorem,&nbsp; a special channel coding can be found at&nbsp; $R ≤ C$&nbsp; with which the error probability can be made zero.
 
*For both channels this condition is true with rate&nbsp; $R = 0.16$&nbsp;.  
 
*For both channels this condition is true with rate&nbsp; $R = 0.16$&nbsp;.  
  
  
  
'''(4)''' &nbsp; <u>Proposed solution 3</u> is correct:
+
'''(4)''' &nbsp; <u>Proposed solution 3</u> &nbsp;is correct:
*With the EUC model, the necessary condition &nbsp; $R ≤ C$&nbsp; for an error-free transmission is fulfilled with &nbsp;$R = 0.32$ &nbsp;and&nbsp; $C = 0.3219$&nbsp;  
+
*With the EUC model, the necessary condition &nbsp; $R ≤ C$&nbsp; for an error-free transmission is fulfilled with &nbsp;$R = 0.32$ &nbsp;and&nbsp; $C = 0.3219$.&nbsp;  
 
*However, the prerequisite for this is the probability function &nbsp; $P_X(X) = (0.6,\ 0.4).$  
 
*However, the prerequisite for this is the probability function &nbsp; $P_X(X) = (0.6,\ 0.4).$  
*In contrast, for equally probable symbols &nbsp; &rArr; &nbsp; $P_X(X) = (0.5,\ 0.5)$&nbsp; the mutual information&nbsp; $I(X; Y) = 0.3113$ would result, <br>i.e. a smaller value than for the channel capacity&nbsp; $C$, and&nbsp; $I(X; Y) < R$ also applies.
+
*In contrast, for equally probable symbols &nbsp; &rArr; &nbsp; $P_X(X) = (0.5,\ 0.5)$&nbsp; the mutual information&nbsp; $I(X; Y) = 0.3113$&nbsp; would result, <br>i.e. a smaller value than for the channel capacity&nbsp; $C$, and&nbsp; $I(X; Y) < R$&nbsp; also applies.
*It can be seen that the EUC model offers more potential for the application of channel coding than the BSC model.&nbsp; Here, for example, it can be exploited in the code that a transmitted "0" is always transmitted without errors.
+
*It can be seen that the EUC model offers more potential for the application of channel coding than the BSC model.&nbsp; Here, for example, it can be exploited in the code that a transmitted&nbsp; "0"&nbsp; is always transmitted without errors.
  
  
'''(5)''' &nbsp; The comments on subtasks&nbsp; '''(3)'''&nbsp; and&nbsp; '''(4)'''&nbsp; show that the<u>proposed solution 4</u> applies.
+
'''(5)''' &nbsp; The comments on subtasks&nbsp; '''(3)'''&nbsp; and&nbsp; '''(4)'''&nbsp; show that the&nbsp; <u>proposed solution 4</u>&nbsp; applies.
  
 
{{ML-Fuß}}
 
{{ML-Fuß}}
Line 114: Line 114:
  
  
[[Category:Information Theory: Exercises|^3.3 Anwendung auf DSÜ-Kanäle^]]
+
[[Category:Information Theory: Exercises|^3.3 Application to Digital Signal Transmission^]]

Latest revision as of 14:50, 17 November 2022

Information-theoretical quantities of the  $\rm BSC$   and  $\rm EUC models$

Shannon's  channel coding theorem states that an error-free transmission can be made over a  "discrete memoryless channel"  $\rm(DMC)$  with the code rate  $R$  as long as  $R$  is not greater than the channel capacity

$$C = \max_{P_X(X)} \hspace{0.15cm} I(X;Y) \hspace{0.05cm}.$$

The  "Channel Coding Theorem"  is to be evaluated numerically in this task, whereby two typical channel models are to be considered:

  • The  $\rm BSC$  model  ("Binary Symmetric Channel")  with error probability $ε = 0.25$  and channel capacity  $C = 1 - H_{\rm bin}(ε),$
  • the  $\rm EUC$  model  (from  "Extremely Unsymmetric Channel";  this designation originates from us and is not common) according to  Exercise 3.11Z.


The graphs show the numerical values of the information-theoretical quantities for the two models  $\rm BSC$  and  $\rm EUC$:

  • The source entropy  $H(X),$
  • the equivocation  $H(X|Y),$
  • the mutual information  $I(X; Y),$
  • the irrelevance  $H(Y|X),$ 
  • the sink entropy  $H(Y).$


The parameter in these tables is  $p_0 = {\rm Pr}(X = 0)$  in the range between  $p_0 = 0.3$  to  $p_0 = 0.7$. 
For the second source symbol probability,   $p_1 = {\rm Pr}(X = 1) =1 - p_0$.




Hints:



Questions

1

Which statements are true for uncoded transmission   ⇒  $\underline{R = 1}$ assuming  $p_0 = p_1 = 0.5$ ?

BSC results in a smaller bit error probability.
EUC results in a smaller bit error probability.
Both models lead to the same bit error probability.

2

With  $\underline{R = 1}$  the result can be (formally) improved by other values of  $p_0$  or  $p_1 ?

With both channels.
With the BSC model.
With the EUC model.
Not for any model.

3

Over which channel can error-free transmission be achieved with the  $\underline{R = 0.16}$ ?

With both channels.
With the BSC model.
With the EUC model.
With none of the models.

4

Over which channel can error-free transmission be achieved with the  $\underline{R = 0.32}$ ?

Over both channels.
With the BSC model.
With the EUC model.
With none of the models.

5

Over which channel can error-free transmission be achieved with the rate  $\underline{R = 0.48}$ ?

With both channels.
With the BSC model.
With the EUC model.
With none of the models.


Solution

(1)   Proposed solution 3  is correct:

  • The BSC error probability is with $p_0 = p_1 = 0.5$  for uncoded transmission  ⇒   $R = 1$:  
$$ p_{\rm B} = 0.5 \cdot 0.25 + 0.5 \cdot 0.25=0.25 \hspace{0.05cm}.$$
  • Correspondingly, with the same boundary conditions for the EUC model::
$$ p_{\rm B} = 0.5 \cdot 0 + 0.5 \cdot 0.5=0.25 \hspace{0.05cm}.$$


(2)   Proposed solution 3  is correct:

  • In the BSC model with error probability  $ε = 0.25$,  for uncoded transmission   ⇒   $R = 1$  the bit error probability is  $p_{\rm B} = 0.25$,  regardless of  $p_0$  and  $p_1$
  • In contrast, with the EUC model, for example, a smaller bit error probability is obtained with  $p_0 = 0.6$  and  $p_1 = 0.4$ :
$$p_{\rm B} = 0.6 \cdot 0 + 0.4 \cdot 0.5=0.2 \hspace{0.05cm}.$$
  • Note, however, that now the source entropy is no longer  $H(X) = 1\ \rm (bit)$  but only  $H(X) = H_{bin} (0.6) = 0.971 \ \rm (bit)$.
  • In the limiting case  $p_0 = 1$,  only zeros are transmitted and  $H(X) = 0$.  For the bit error probability, however, the following then actually applies:
$$ p_{\rm B} = 1 \cdot 0 + 0 \cdot 0.5=0 \hspace{0.05cm}.$$
So no information is transmitted, but it is transmitted with the bit error probability "zero".


(3)   Proposed solution 1  is correct:

  • From the graph in the information section,  it can be read for the capacities of the two channels:
$$C_{\rm BSC} = 0.1887 \ \rm {bit/use}, \hspace{0.5cm}C_{rm EUC} = 0.3219 \ \rm {bit/use}.$$
  • According to the channel coding theorem,  a special channel coding can be found at  $R ≤ C$  with which the error probability can be made zero.
  • For both channels this condition is true with rate  $R = 0.16$ .


(4)   Proposed solution 3  is correct:

  • With the EUC model, the necessary condition   $R ≤ C$  for an error-free transmission is fulfilled with  $R = 0.32$  and  $C = 0.3219$. 
  • However, the prerequisite for this is the probability function   $P_X(X) = (0.6,\ 0.4).$
  • In contrast, for equally probable symbols   ⇒   $P_X(X) = (0.5,\ 0.5)$  the mutual information  $I(X; Y) = 0.3113$  would result,
    i.e. a smaller value than for the channel capacity  $C$, and  $I(X; Y) < R$  also applies.
  • It can be seen that the EUC model offers more potential for the application of channel coding than the BSC model.  Here, for example, it can be exploited in the code that a transmitted  "0"  is always transmitted without errors.


(5)   The comments on subtasks  (3)  and  (4)  show that the  proposed solution 4  applies.