Difference between revisions of "Aufgaben:Exercise 4.4: Extrinsic L-values at SPC"

From LNTwww
 
(16 intermediate revisions by 4 users not shown)
Line 1: Line 1:
{{quiz-Header|Buchseite=Kanalcodierung/Soft–in Soft–out Decoder}}
+
{{quiz-Header|Buchseite=Channel_Coding/Soft-in_Soft-Out_Decoder}}
  
[[File:P_ID2993__KC_A_4_4_v1.png|right|frame|Geeignete Hilfstabelle]]
+
[[File:EN_KC_A_4_4.png|right|frame|Suitable auxiliary table]]
Wir betrachten nochmals den  [[Kanalcodierung/Beispiele_bin%C3%A4rer_Blockcodes#Single_Parity.E2.80.93check_Codes|Single Parity–check Code]]. Bei einem solchen  ${\rm SPC} \ (n, \, n-1, \, 2)$  stammen von den  $n$  Bit eines Codewortes  $\underline{x}$  die ersten  $k = n -1$  Bit von der Quellenfolge  $\underline{u}$  und es wird nur ein einziges Prüfbit  $p$  hinzugefügt, und zwar derart, dass die Anzahl der Einsen im Codewort geradzahlig ist:
+
We consider again the  [[Channel_Coding/Examples_of_Binary_Block_Codes#Single_Parity-check_Codes|"single parity–check code"]].  In such a  ${\rm SPC} \ (n, \, n-1, \, 2)$  the  $n$  bits of a code word  $\underline{x}$  come from the  $k = n -1$  bits from the source sequence  $\underline{u}$  and only a single check bit  $p$  is added,  such that the number of  "ones"  in the code word  $\underline{x}$  is even:
 
:$$\underline{x} = \big ( \hspace{0.03cm}x_1, \hspace{0.03cm} x_2, \hspace{0.05cm} \text{...} \hspace{0.05cm} , x_{n-1}, \hspace{0.03cm} x_n \hspace{0.03cm} \big ) =  
 
:$$\underline{x} = \big ( \hspace{0.03cm}x_1, \hspace{0.03cm} x_2, \hspace{0.05cm} \text{...} \hspace{0.05cm} , x_{n-1}, \hspace{0.03cm} x_n \hspace{0.03cm} \big ) =  
 
\big ( \hspace{0.03cm}u_1, \hspace{0.03cm} u_2, \hspace{0.05cm} \text{...} \hspace{0.05cm} , u_{k}, \hspace{0.03cm} p \hspace{0.03cm} \big )\hspace{0.03cm}. $$
 
\big ( \hspace{0.03cm}u_1, \hspace{0.03cm} u_2, \hspace{0.05cm} \text{...} \hspace{0.05cm} , u_{k}, \hspace{0.03cm} p \hspace{0.03cm} \big )\hspace{0.03cm}. $$
  
Die extrinsische Information über das  $i$–te Codebit wird über alle anderen Symbole  $(j ≠ i)$  gebildet. Deshalb schreiben wir für das um ein Bit kürzere Codewort:
+
The extrinsic information about the  $i$th code bit is formed over all other bits  $(j ≠ i)$.  Therefore we write for the code word shorter by one bit:
 
:$$\underline{x}^{(-i)} = \big ( \hspace{0.03cm}x_1, \hspace{0.05cm} \text{...} \hspace{0.05cm} , \hspace{0.03cm} x_{i-1}, \hspace{0.43cm} x_{i+1},  \hspace{0.05cm} \text{...} \hspace{0.05cm} , x_{n} \hspace{0.03cm} \big )\hspace{0.03cm}. $$
 
:$$\underline{x}^{(-i)} = \big ( \hspace{0.03cm}x_1, \hspace{0.05cm} \text{...} \hspace{0.05cm} , \hspace{0.03cm} x_{i-1}, \hspace{0.43cm} x_{i+1},  \hspace{0.05cm} \text{...} \hspace{0.05cm} , x_{n} \hspace{0.03cm} \big )\hspace{0.03cm}. $$
  
Der extrinsische  $L$–Wert über das  $i$–te Codesymbol lautet mit dem  [[Kanalcodierung/Zielsetzung_der_Kanalcodierung#Einige_wichtige_Definitionen_zur_Blockcodierung|Hamming–Gewicht]]  $w_{\rm H}$  der verkürzten Folge  $\underline{x}^{(-i)}$:
+
The extrinsic  L–value over the  $i$th code symbol reads with the  [[Channel_Coding/Objective_of_Channel_Coding#Important_definitions_for_block_coding|"Hamming weight"]]  $w_{\rm H}$  of the truncated sequence  $\underline{x}^{(-i)}$:
:$$L_{\rm E}(i) = \frac{{\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm ist \hspace{0.15cm} gerade} \hspace{0.05cm} | \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}{{\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm ist \hspace{0.15cm} ungerade} \hspace{0.05cm} | \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}
+
:$$L_{\rm E}(i) = \frac{{\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm is \hspace{0.15cm} even} \hspace{0.05cm} | \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}{{\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm is \hspace{0.15cm} odd} \hspace{0.05cm} | \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}
 
  \hspace{0.05cm}.$$
 
  \hspace{0.05cm}.$$
  
*Ist die Wahrscheinlichkeit im Zähler größer als die im Nenner, so ist $L_{\rm E}(i) > 0$ und damit wird auch der Aposteriori–$L$–Wert $L_{\rm APP}(i) = L_{\rm A}(i) + L_{\rm E}(i)$ vergrößert, das heißt tendenziell in Richtung des Symbols $x_i = 0$ beeinflusst.  
+
*If the probability in the numerator is greater than that in the denominator,  then  $L_{\rm E}(i) > 0$  and thus the a-posteriori  L–value  $L_{\rm APP}(i) = L_{\rm A}(i) + L_{\rm E}(i)$  magnified,  that is tends to be affected in the direction of the symbol  $x_i = 0$.
*Bei $L_{\rm E}(i) < 0$ spricht aus Sicht der anderen Symbole $(j &ne; i)$ vieles dafür, dass $x_i = 1$ ist.
+
 +
*If&nbsp; $L_{\rm E}(i) < 0$&nbsp;  then there is much to be said for&nbsp; $x_i = 1$&nbsp; from the point of view of the other symbols&nbsp; $(j &ne; i)$.
  
  
Behandelt wird ausschließlich der SPC (4, 3, 4), wobei für die Wahrscheinlichkeiten $p_i = {\rm Pr}(x_i = 1)$ gilt:
+
Only the&nbsp; $\text{SPC (4, 3, 4)}$&nbsp; is treated,&nbsp; where for the probabilities&nbsp; $p_i = {\rm Pr}(x_i = 1)$&nbsp; holds:
 
:$$p_1 = 0.2 \hspace{0.05cm}, \hspace{0.3cm}
 
:$$p_1 = 0.2 \hspace{0.05cm}, \hspace{0.3cm}
 
p_2 = 0.9 \hspace{0.05cm}, \hspace{0.3cm}  
 
p_2 = 0.9 \hspace{0.05cm}, \hspace{0.3cm}  
Line 23: Line 24:
 
p_4 = 0.6  \hspace{0.05cm}.$$
 
p_4 = 0.6  \hspace{0.05cm}.$$
  
Daraus ergeben sich die Apriori&ndash;$L$&ndash;Werte zu:
+
From this the a-priori log likelihood ratios result to:
 
:$$L_{\rm A}(i) = {\rm ln} \hspace{0.1cm} \left [ \frac{{\rm Pr}(x_i = 0)}{{\rm Pr}(x_i = 1)}
 
:$$L_{\rm A}(i) = {\rm ln} \hspace{0.1cm} \left [ \frac{{\rm Pr}(x_i = 0)}{{\rm Pr}(x_i = 1)}
 
\right ] = {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_i}{p_i}
 
\right ] = {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_i}{p_i}
Line 33: Line 34:
  
  
''Hinweise:''
+
<u>Hints:</u>
* Die Aufgabe gehört zum Themengebiet des Kapitels [[Kanalcodierung/Soft%E2%80%93in_Soft%E2%80%93out_Decoder| Soft&ndash;in Soft&ndash;out Decoder]].
+
*This exercise belongs to the chapter&nbsp; [[Channel_Coding/Soft-in_Soft-Out_Decoder|"Soft&ndash;in Soft&ndash;out Decoder"]].
*Bezug genommen wird insbesondere auf die Seite [[Kanalcodierung/Soft–in_Soft–out_Decoder#Zur_Berechnung_der_extrinsischen_L.E2.80.93Werte|Zur Berechnung der extrinsischen L&ndash;Werte]].  
+
 
* In der oberen Tabelle sind für $p_i = 0$ bis $p_i = 1$ mit Schrittweite $0.1$ (Spalte 1) angegeben:
+
*Reference is made in particular to the section&nbsp; [[Channel_Coding/Soft-in_Soft-Out_Decoder#Calculation_of_extrinsic_log_likelihood_ratios|"Calculation of the extrinsic log likelihood ratios"]].
::In Spalte 2: &nbsp; die Wahrscheinlichkeit $q_i = {\rm Pr}(x_i = 0) = 1 - p_i$,
+
::in Spalte 3: &nbsp; die Werte für $1 - 2p_i$,
+
* In the table are given for&nbsp; $p_i = 0$&nbsp; to&nbsp; $p_i = 1$&nbsp; with step size&nbsp; $0.1$&nbsp; $($column 1$)$:
::in Spalte 4: &nbsp; die Apriori&ndash;$L$&ndash;Werte $L_i = \ln {[(1 - p_i)/p_i]} = L_{\rm A}(i)$.
+
::In column 2: &nbsp; the probability&nbsp; $q_i = {\rm Pr}(x_i = 0) = 1 - p_i$,
* Der <i>Tangens Hyperbolicus</i> ($\tanh$) von $L_i/2$ ist identisch mit $1-2p_i$ &nbsp;&#8658;&nbsp; Spalte 3.
+
::in column 3: &nbsp; the values for&nbsp; $1 - 2p_i$,
* In der [[Aufgaben:Aufgabe_4.4Z:_Ergänzung_zur_Aufgabe_4.4|Aufgabe 4.4Z]] wird gezeigt, dass für den extrinsischen $L$&ndash;Wert auch geschrieben werden kann:
+
::in column 4: &nbsp; the a-priori log likelihood ratios&nbsp; $L_i = \ln {\big [(1 - p_i)/p_ i \big ]} = L_{\rm A}(i)$.
 +
 
 +
* The&nbsp; "hyperbolic tangent"&nbsp; $(\tanh)$&nbsp; of $L_i/2$&nbsp; is identical to&nbsp; $1-2p_i$ &nbsp; &#8658; &nbsp; column 3.
 +
 
 +
* In&nbsp; [[Aufgaben:Exercise_4.4Z:_Supplement_to_Exercise_4.4|$\text{Exercise 4.4Z}$]]&nbsp; it is shown that for the extrinsic&nbsp; L&ndash;value can also be written:
 
:$$L_{\rm E}(i) = {\rm ln} \hspace{0.2cm}  \frac{1 + \pi}{1 - \pi}\hspace{0.05cm}, \hspace{0.3cm}
 
:$$L_{\rm E}(i) = {\rm ln} \hspace{0.2cm}  \frac{1 + \pi}{1 - \pi}\hspace{0.05cm}, \hspace{0.3cm}
 
{\rm mit} \hspace{0.3cm} \pi = \prod\limits_{j \ne i}^{n} \hspace{0.25cm}(1-2p_j)
 
{\rm mit} \hspace{0.3cm} \pi = \prod\limits_{j \ne i}^{n} \hspace{0.25cm}(1-2p_j)
Line 48: Line 53:
  
  
===Fragebogen===
+
===Questions===
 
<quiz display=simple>
 
<quiz display=simple>
{Es gelte $p_1 = 0.2, \ p_2 = 0.9, \ p_3 = 0.3, \ p_4 = 0.6$. Berechnen Sie daraus die Apriori&ndash;$L$&ndash;Werte des SPC (4, 3, 4) für Bit 1 und Bit 2.
+
{It holds&nbsp; $p_1 = 0.2, \ p_2 = 0.9, \ p_3 = 0.3, \ p_4 = 0.6$.&nbsp; From this,&nbsp; calculate the a-priori log likelihood ratios of the&nbsp; $\text{SPC (4, 3, 4)}$&nbsp; for bit 1 and bit 2.
 
|type="{}"}
 
|type="{}"}
 
$L_{\rm A}(i = 1) \ = \ ${ 1.386 3% }
 
$L_{\rm A}(i = 1) \ = \ ${ 1.386 3% }
 
$L_{\rm A}(i = 2) \ = \ ${ -2.26291--2.13109 }
 
$L_{\rm A}(i = 2) \ = \ ${ -2.26291--2.13109 }
  
{Wie lauten die extrinsischen $L$&ndash;Werte für Bit 1 und Bit 2.
+
{What are the extrinsic log likelihood ratios for bit 1 and bit 2.
 
|type="{}"}
 
|type="{}"}
 
$L_{\rm E}(i = 1) \ = \ ${ 0.128 3% }
 
$L_{\rm E}(i = 1) \ = \ ${ 0.128 3% }
 
$L_{\rm E}(i = 2) \ = \ ${ -0.09888--0.09312 }
 
$L_{\rm E}(i = 2) \ = \ ${ -0.09888--0.09312 }
  
{Welche Zusammenhänge bestehen zwischen $p_j$ und $L_j = L_{\rm A}(j)$?
+
{What are the relationships between&nbsp; $p_j$&nbsp; and&nbsp; $L_j = L_{\rm A}(j)$?
 
|type="[]"}
 
|type="[]"}
+ Es gilt $p_j = 1/(1 + {\rm e}^ {L_j})$.
+
+ It holds&nbsp; $p_j = 1/(1 + {\rm e}^ {L_j})$.
+ Es gilt $1-2p_([{\rm e}^ {L_j} - 1) \ / \ ({\rm e}^ {L_j} + 1)$.
+
+ It holds&nbsp; $1-2p_j = ({\rm e}^ {L_j} - 1) \ / \ ({\rm e}^ {L_j} + 1)$.
+ Es gilt $1-2p_j = \tanh {(L_j/2)}$.
+
+ It holds&nbsp; $1-2p_j = \tanh {(L_j/2)}$.
  
{Es gelte weiter $p_1 = 0.2, \ p_2 = 0.9, \ p_3$ und $p_4 = 0.6$. Berechnen Sie die extrinsischen $L$&ndash;Werte für Bit 3 und Bit 4, <br>Verwenden Sie hierzu verschiedene Gleichungen.
+
{It is further&nbsp; $p_1 = 0.2, \ p_2 = 0.9, \ p_3, \ p_4 = 0.6$.&nbsp; Calculate the extrinsic log likelihood ratios for bit 3 and bit 4.&nbsp; Use different equations for this purpose.
 
|type="{}"}
 
|type="{}"}
 
$L_{\rm E}(i = 3) \ = \ ${ 0.193 3% }
 
$L_{\rm E}(i = 3) \ = \ ${ 0.193 3% }
Line 72: Line 77:
 
</quiz>
 
</quiz>
  
===Musterlösung===
+
===Solution===
 
{{ML-Kopf}}
 
{{ML-Kopf}}
'''(1)'''&nbsp; Für die Apriori&ndash;$L$&ndash;Werte der beiden ersten Bits des Codewortes gilt:
+
'''(1)'''&nbsp; For the a-priori log likelihood ratios of the first two bits of the code word:
 
:$$L_{\rm A}(i = 1) \hspace{-0.15cm} \ = \ \hspace{-0.15cm} {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_1}{p_1} \right ] =  {\rm ln} \hspace{0.1cm} 4 \hspace{0.15cm}\underline{= +1.386}  
 
:$$L_{\rm A}(i = 1) \hspace{-0.15cm} \ = \ \hspace{-0.15cm} {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_1}{p_1} \right ] =  {\rm ln} \hspace{0.1cm} 4 \hspace{0.15cm}\underline{= +1.386}  
 
  \hspace{0.05cm},$$
 
  \hspace{0.05cm},$$
Line 80: Line 85:
 
  \hspace{0.05cm}.$$
 
  \hspace{0.05cm}.$$
  
Die Werte können aus der vierten Spalte der auf der Angabenseite beigefügten  Tabelle abgelesen werden.
+
*The values can be read from the fourth column of the table attached to the information page.
 +
 
  
  
'''(2)'''&nbsp; Zur Berechnung des extrinsischen $L$&ndash;Wertes über das $i$&ndash;te Bit dürfen nur die Informationen über die drei anderen Bits $(j &ne; i)$ herangezogen werden. Mit der angegebenen Gleichung gilt:
+
'''(2)'''&nbsp; To calculate the extrinsic L&ndash;value over the&nbsp; $i$th bit,&nbsp; only the information about the other three bits&nbsp; $(j &ne; i)$&nbsp; may be used.&nbsp; With the given equation holds:
 
:$$L_{\rm E}(i = 1) = {\rm ln} \hspace{0.2cm}  \frac{1 + \prod\limits_{j \ne 1} \hspace{0.25cm}(1-2p_j)}{1 - \prod\limits_{j \ne 1} \hspace{0.25cm}(1-2p_j)}
 
:$$L_{\rm E}(i = 1) = {\rm ln} \hspace{0.2cm}  \frac{1 + \prod\limits_{j \ne 1} \hspace{0.25cm}(1-2p_j)}{1 - \prod\limits_{j \ne 1} \hspace{0.25cm}(1-2p_j)}
 
  \hspace{0.05cm}.$$
 
  \hspace{0.05cm}.$$
  
Für das Produkt erhält man entsprechend der dritten Spalte der [[Aufgaben:4.4_Extrinsische_L%E2%80%93Werte_beim_SPC|Tabelle]]:
+
*For the product,&nbsp; we obtain according to the third column of the&nbsp; [[Aufgaben:Exercise_4.4:_Extrinsic_L-values_at_SPC|$\rm table$]]:
 
:$$\prod\limits_{j =2, \hspace{0.05cm}3,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) =  
 
:$$\prod\limits_{j =2, \hspace{0.05cm}3,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) =  
 
(-0.8) \cdot (+0.4) \cdot (-0.2) = 0.064
 
(-0.8) \cdot (+0.4) \cdot (-0.2) = 0.064
Line 94: Line 100:
 
  \hspace{0.05cm}.$$
 
  \hspace{0.05cm}.$$
  
Hinsichtlich Bit 2 erhält man entsprechend:
+
*In terms of bit 2,&nbsp; one obtains accordingly:
 
:$$\prod\limits_{j =1, \hspace{0.05cm}3,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) =  
 
:$$\prod\limits_{j =1, \hspace{0.05cm}3,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) =  
 
(+0.6) \cdot (+0.4) \cdot (-0.2) = -0.048
 
(+0.6) \cdot (+0.4) \cdot (-0.2) = -0.048
Line 102: Line 108:
  
  
'''(3)'''&nbsp; Für den Apriori&ndash;$L$&ndash;Wert gilt:
+
 
 +
'''(3)'''&nbsp; For the a-priori L&ndash; value holds:
 
:$$L_j = L_{\rm A}(j) = {\rm ln} \hspace{0.1cm} \left [ \frac{{\rm Pr}(x_j = 0)}{{\rm Pr}(x_j = 1)}
 
:$$L_j = L_{\rm A}(j) = {\rm ln} \hspace{0.1cm} \left [ \frac{{\rm Pr}(x_j = 0)}{{\rm Pr}(x_j = 1)}
 
\right ] = {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_j}{p_j}
 
\right ] = {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_j}{p_j}
Line 110: Line 117:
 
\hspace{0.05cm} .$$
 
\hspace{0.05cm} .$$
  
Damit gilt auch:
+
*Thus also applies:
 
:$$1- 2 \cdot p_j =  1 - \frac{2}{1+{\rm e}^{L_j} } = \frac{1+{\rm e}^{L_j}-2}{1+{\rm e}^{L_j} }
 
:$$1- 2 \cdot p_j =  1 - \frac{2}{1+{\rm e}^{L_j} } = \frac{1+{\rm e}^{L_j}-2}{1+{\rm e}^{L_j} }
 
= \frac{{\rm e}^{L_j}-1}{{\rm e}^{L_j} +1}\hspace{0.05cm} .$$
 
= \frac{{\rm e}^{L_j}-1}{{\rm e}^{L_j} +1}\hspace{0.05cm} .$$
  
Multipliziert man Zähler und Nenner noch mit ${\rm e}^{-L_j/2}$, so erhält man:
+
*Multiplying the numerator and denominator by&nbsp; ${\rm e}^{-L_j/2}$,&nbsp; we get:
 
:$$1- 2 \cdot p_j =  \frac{{\rm e}^{L_j/2}-{\rm e}^{-L_j/2}}{{\rm e}^{L_j/2}+{\rm e}^{-L_j/2}}={\rm tanh}  (L_j/2) \hspace{0.05cm} .$$
 
:$$1- 2 \cdot p_j =  \frac{{\rm e}^{L_j/2}-{\rm e}^{-L_j/2}}{{\rm e}^{L_j/2}+{\rm e}^{-L_j/2}}={\rm tanh}  (L_j/2) \hspace{0.05cm} .$$
  
Somit sind <u>alle Lösungsvorschläge</u> richtig. Die Funktion <i>Tangens Hyperbolicus</i> findet man zum Beispiel tabellarisch in Formelsammlungen oder in der letzten Spalte der vorne angegebenen Tabelle.
+
*Thus&nbsp; <u>all proposed solutions</u> are correct.
 +
 +
*The function&nbsp; "hyperbolic tangent"&nbsp; can be found,&nbsp; for example,&nbsp; in tabular form in formula collections or in the last column of the table given in front.
 +
 
  
  
'''(4)'''&nbsp; Wir berechnen $L_{\rm E}(i = 3)$ zunächst in gleicher Weise wie in der Teilaufgabe (2):
+
'''(4)'''&nbsp; We first calculate&nbsp; $L_{\rm E}(i = 3)$&nbsp; in the same way as in subtask&nbsp; '''(2)''':
 
:$$\prod\limits_{j =1, \hspace{0.05cm}2,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) =  
 
:$$\prod\limits_{j =1, \hspace{0.05cm}2,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) =  
 
(+0.6) \cdot (-0.8) \cdot (-0.2) = +0.096
 
(+0.6) \cdot (-0.8) \cdot (-0.2) = +0.096
Line 127: Line 137:
 
  \hspace{0.05cm}.$$
 
  \hspace{0.05cm}.$$
  
Den extrinsischen $L$&ndash;Wert hinsichtlich des letzten Bits berechnen wir nach der Gleichung
+
*We calculate the extrinsic L&ndash;value with respect to the last bit according to the equation
 
:$$L_{\rm E}(i = 4) = {\rm ln} \hspace{0.2cm}  \frac{1 + \pi}{1 - \pi}\hspace{0.05cm}, \hspace{0.3cm}
 
:$$L_{\rm E}(i = 4) = {\rm ln} \hspace{0.2cm}  \frac{1 + \pi}{1 - \pi}\hspace{0.05cm}, \hspace{0.3cm}
{\rm mit} \hspace{0.3cm} \pi = {\rm tanh}(L_1/2) \cdot {\rm tanh}(L_2/2) \cdot {\rm tanh}(L_3/2)
+
{\rm with} \hspace{0.3cm} \pi = {\rm tanh}(L_1/2) \cdot {\rm tanh}(L_2/2) \cdot {\rm tanh}(L_3/2)
 
  \hspace{0.05cm}.$$
 
  \hspace{0.05cm}.$$
  
Damit ergibt sich entsprechend der obigen [[Aufgaben:4.4_Extrinsische_L%E2%80%93Werte_beim_SPC|Tabelle]]:
+
*This results in accordance with the above&nbsp; [[Aufgaben:Exercise_4.4:_Extrinsic_L-values_at_SPC|$\rm table$]]:
 
:$$p_1 = 0.2 \hspace{0.2cm}\Rightarrow \hspace{0.2cm}
 
:$$p_1 = 0.2 \hspace{0.2cm}\Rightarrow \hspace{0.2cm}
 
L_1 = +1.386 \hspace{0.2cm}\Rightarrow \hspace{0.2cm}
 
L_1 = +1.386 \hspace{0.2cm}\Rightarrow \hspace{0.2cm}
Line 152: Line 162:
 
  = 0.4 \hspace{0.3cm}\Rightarrow \hspace{0.3cm}{\rm identisch \hspace{0.15cm}mit\hspace{0.15cm} }1-2\cdot p_3\hspace{0.05cm}.$$
 
  = 0.4 \hspace{0.3cm}\Rightarrow \hspace{0.3cm}{\rm identisch \hspace{0.15cm}mit\hspace{0.15cm} }1-2\cdot p_3\hspace{0.05cm}.$$
  
Das Endergebnis lautet somit:
+
*The final result is thus:
 
:$$\pi = (+0.6) \cdot (-0.8) \cdot (+0.4) = -0.192
 
:$$\pi = (+0.6) \cdot (-0.8) \cdot (+0.4) = -0.192
 
\hspace{0.3cm}\Rightarrow \hspace{0.3cm}
 
\hspace{0.3cm}\Rightarrow \hspace{0.3cm}
Line 161: Line 171:
  
  
[[Category:Aufgaben zu  Kanalcodierung|^4.1 Soft–in Soft–out Decoder^]]
+
[[Category:Channel Coding: Exercises|^4.1 Soft–in Soft–out Decoder^]]

Latest revision as of 16:25, 23 January 2023

Suitable auxiliary table

We consider again the  "single parity–check code".  In such a  ${\rm SPC} \ (n, \, n-1, \, 2)$  the  $n$  bits of a code word  $\underline{x}$  come from the  $k = n -1$  bits from the source sequence  $\underline{u}$  and only a single check bit  $p$  is added,  such that the number of  "ones"  in the code word  $\underline{x}$  is even:

$$\underline{x} = \big ( \hspace{0.03cm}x_1, \hspace{0.03cm} x_2, \hspace{0.05cm} \text{...} \hspace{0.05cm} , x_{n-1}, \hspace{0.03cm} x_n \hspace{0.03cm} \big ) = \big ( \hspace{0.03cm}u_1, \hspace{0.03cm} u_2, \hspace{0.05cm} \text{...} \hspace{0.05cm} , u_{k}, \hspace{0.03cm} p \hspace{0.03cm} \big )\hspace{0.03cm}. $$

The extrinsic information about the  $i$th code bit is formed over all other bits  $(j ≠ i)$.  Therefore we write for the code word shorter by one bit:

$$\underline{x}^{(-i)} = \big ( \hspace{0.03cm}x_1, \hspace{0.05cm} \text{...} \hspace{0.05cm} , \hspace{0.03cm} x_{i-1}, \hspace{0.43cm} x_{i+1}, \hspace{0.05cm} \text{...} \hspace{0.05cm} , x_{n} \hspace{0.03cm} \big )\hspace{0.03cm}. $$

The extrinsic  L–value over the  $i$th code symbol reads with the  "Hamming weight"  $w_{\rm H}$  of the truncated sequence  $\underline{x}^{(-i)}$:

$$L_{\rm E}(i) = \frac{{\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm is \hspace{0.15cm} even} \hspace{0.05cm} | \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}{{\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm is \hspace{0.15cm} odd} \hspace{0.05cm} | \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]} \hspace{0.05cm}.$$
  • If the probability in the numerator is greater than that in the denominator,  then  $L_{\rm E}(i) > 0$  and thus the a-posteriori L–value  $L_{\rm APP}(i) = L_{\rm A}(i) + L_{\rm E}(i)$  magnified,  that is tends to be affected in the direction of the symbol  $x_i = 0$.
  • If  $L_{\rm E}(i) < 0$  then there is much to be said for  $x_i = 1$  from the point of view of the other symbols  $(j ≠ i)$.


Only the  $\text{SPC (4, 3, 4)}$  is treated,  where for the probabilities  $p_i = {\rm Pr}(x_i = 1)$  holds:

$$p_1 = 0.2 \hspace{0.05cm}, \hspace{0.3cm} p_2 = 0.9 \hspace{0.05cm}, \hspace{0.3cm} p_3 = 0.3 \hspace{0.05cm}, \hspace{0.3cm} p_4 = 0.6 \hspace{0.05cm}.$$

From this the a-priori log likelihood ratios result to:

$$L_{\rm A}(i) = {\rm ln} \hspace{0.1cm} \left [ \frac{{\rm Pr}(x_i = 0)}{{\rm Pr}(x_i = 1)} \right ] = {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_i}{p_i} \right ] \hspace{0.05cm}.$$



Hints:

  • In the table are given for  $p_i = 0$  to  $p_i = 1$  with step size  $0.1$  $($column 1$)$:
In column 2:   the probability  $q_i = {\rm Pr}(x_i = 0) = 1 - p_i$,
in column 3:   the values for  $1 - 2p_i$,
in column 4:   the a-priori log likelihood ratios  $L_i = \ln {\big [(1 - p_i)/p_ i \big ]} = L_{\rm A}(i)$.
  • The  "hyperbolic tangent"  $(\tanh)$  of $L_i/2$  is identical to  $1-2p_i$   ⇒   column 3.
$$L_{\rm E}(i) = {\rm ln} \hspace{0.2cm} \frac{1 + \pi}{1 - \pi}\hspace{0.05cm}, \hspace{0.3cm} {\rm mit} \hspace{0.3cm} \pi = \prod\limits_{j \ne i}^{n} \hspace{0.25cm}(1-2p_j) \hspace{0.05cm}.$$


Questions

1

It holds  $p_1 = 0.2, \ p_2 = 0.9, \ p_3 = 0.3, \ p_4 = 0.6$.  From this,  calculate the a-priori log likelihood ratios of the  $\text{SPC (4, 3, 4)}$  for bit 1 and bit 2.

$L_{\rm A}(i = 1) \ = \ $

$L_{\rm A}(i = 2) \ = \ $

2

What are the extrinsic log likelihood ratios for bit 1 and bit 2.

$L_{\rm E}(i = 1) \ = \ $

$L_{\rm E}(i = 2) \ = \ $

3

What are the relationships between  $p_j$  and  $L_j = L_{\rm A}(j)$?

It holds  $p_j = 1/(1 + {\rm e}^ {L_j})$.
It holds  $1-2p_j = ({\rm e}^ {L_j} - 1) \ / \ ({\rm e}^ {L_j} + 1)$.
It holds  $1-2p_j = \tanh {(L_j/2)}$.

4

It is further  $p_1 = 0.2, \ p_2 = 0.9, \ p_3, \ p_4 = 0.6$.  Calculate the extrinsic log likelihood ratios for bit 3 and bit 4.  Use different equations for this purpose.

$L_{\rm E}(i = 3) \ = \ $

$L_{\rm E}(i = 4) \ = \ $


Solution

(1)  For the a-priori log likelihood ratios of the first two bits of the code word:

$$L_{\rm A}(i = 1) \hspace{-0.15cm} \ = \ \hspace{-0.15cm} {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_1}{p_1} \right ] = {\rm ln} \hspace{0.1cm} 4 \hspace{0.15cm}\underline{= +1.386} \hspace{0.05cm},$$
$$L_{\rm A}(i = 2) \hspace{-0.15cm} \ = \ \hspace{-0.15cm} {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_2}{p_2} \right ] = {\rm ln} \hspace{0.1cm} 1/9 \hspace{0.15cm}\underline{= -2.197} \hspace{0.05cm}.$$
  • The values can be read from the fourth column of the table attached to the information page.


(2)  To calculate the extrinsic L–value over the  $i$th bit,  only the information about the other three bits  $(j ≠ i)$  may be used.  With the given equation holds:

$$L_{\rm E}(i = 1) = {\rm ln} \hspace{0.2cm} \frac{1 + \prod\limits_{j \ne 1} \hspace{0.25cm}(1-2p_j)}{1 - \prod\limits_{j \ne 1} \hspace{0.25cm}(1-2p_j)} \hspace{0.05cm}.$$
  • For the product,  we obtain according to the third column of the  $\rm table$:
$$\prod\limits_{j =2, \hspace{0.05cm}3,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) = (-0.8) \cdot (+0.4) \cdot (-0.2) = 0.064 \hspace{0.05cm}\hspace{0.3cm} \Rightarrow \hspace{0.3cm}L_{\rm E}(i = 1) = {\rm ln} \hspace{0.2cm} \frac{1 + 0.064}{1 - 0.064} = {\rm ln} \hspace{0.1cm} (1.137)\hspace{0.15cm}\underline{= +0.128} \hspace{0.05cm}.$$
  • In terms of bit 2,  one obtains accordingly:
$$\prod\limits_{j =1, \hspace{0.05cm}3,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) = (+0.6) \cdot (+0.4) \cdot (-0.2) = -0.048 \hspace{0.05cm}\hspace{0.3cm} \Rightarrow \hspace{0.3cm}L_{\rm E}(i = 2) = {\rm ln} \hspace{0.2cm} \frac{1 -0.048}{1 +0.048} = {\rm ln} \hspace{0.1cm} (0.908)\hspace{0.15cm}\underline{= -0.096} \hspace{0.05cm}.$$


(3)  For the a-priori L– value holds:

$$L_j = L_{\rm A}(j) = {\rm ln} \hspace{0.1cm} \left [ \frac{{\rm Pr}(x_j = 0)}{{\rm Pr}(x_j = 1)} \right ] = {\rm ln} \hspace{0.1cm} \left [ \frac{1-p_j}{p_j} \right ]\hspace{0.3cm} \Rightarrow \hspace{0.3cm} 1-p_j = p_j \cdot {\rm e}^{L_j} \hspace{0.3cm} \Rightarrow \hspace{0.3cm} p_j = \frac{1}{1+{\rm e}^{L_j} } \hspace{0.05cm} .$$
  • Thus also applies:
$$1- 2 \cdot p_j = 1 - \frac{2}{1+{\rm e}^{L_j} } = \frac{1+{\rm e}^{L_j}-2}{1+{\rm e}^{L_j} } = \frac{{\rm e}^{L_j}-1}{{\rm e}^{L_j} +1}\hspace{0.05cm} .$$
  • Multiplying the numerator and denominator by  ${\rm e}^{-L_j/2}$,  we get:
$$1- 2 \cdot p_j = \frac{{\rm e}^{L_j/2}-{\rm e}^{-L_j/2}}{{\rm e}^{L_j/2}+{\rm e}^{-L_j/2}}={\rm tanh} (L_j/2) \hspace{0.05cm} .$$
  • Thus  all proposed solutions are correct.
  • The function  "hyperbolic tangent"  can be found,  for example,  in tabular form in formula collections or in the last column of the table given in front.


(4)  We first calculate  $L_{\rm E}(i = 3)$  in the same way as in subtask  (2):

$$\prod\limits_{j =1, \hspace{0.05cm}2,\hspace{0.05cm} 4} \hspace{0.05cm}(1-2p_j) = (+0.6) \cdot (-0.8) \cdot (-0.2) = +0.096 \hspace{0.05cm}\hspace{0.3cm} \Rightarrow \hspace{0.3cm}L_{\rm E}(i = 3) = {\rm ln} \hspace{0.2cm} \frac{1 +0.096}{1 -0.096} = {\rm ln} \hspace{0.1cm} (1.212)\hspace{0.15cm}\underline{= +0.193} \hspace{0.05cm}.$$
  • We calculate the extrinsic L–value with respect to the last bit according to the equation
$$L_{\rm E}(i = 4) = {\rm ln} \hspace{0.2cm} \frac{1 + \pi}{1 - \pi}\hspace{0.05cm}, \hspace{0.3cm} {\rm with} \hspace{0.3cm} \pi = {\rm tanh}(L_1/2) \cdot {\rm tanh}(L_2/2) \cdot {\rm tanh}(L_3/2) \hspace{0.05cm}.$$
  • This results in accordance with the above  $\rm table$:
$$p_1 = 0.2 \hspace{0.2cm}\Rightarrow \hspace{0.2cm} L_1 = +1.386 \hspace{0.2cm}\Rightarrow \hspace{0.2cm} L_1/2 = +0.693 \hspace{0.2cm} \Rightarrow \hspace{0.2cm} {\rm tanh}(L_1/2) = \frac{{\rm e}^{+0.693}-{\rm e}^{-0.693}}{{\rm e}^{+0.693}+{\rm e}^{-0.693}} = 0.6 \hspace{0.3cm}\Rightarrow \hspace{0.3cm}{\rm identisch \hspace{0.15cm}mit\hspace{0.15cm} }1-2\cdot p_1\hspace{0.05cm},$$
$$p_2 = 0.9 \hspace{0.2cm}\Rightarrow \hspace{0.2cm} L_2 = -2.197 \hspace{0.2cm}\Rightarrow \hspace{0.2cm} L_2/2 = -1.099\hspace{0.2cm} \Rightarrow \hspace{0.2cm} {\rm tanh}(L_2/2) = \frac{{\rm e}^{-1.099}-{\rm e}^{+1.099}}{{\rm e}^{-1.099}+{\rm e}^{+1.099}} = -0.8 \hspace{0.3cm}\Rightarrow \hspace{0.3cm}{\rm identisch \hspace{0.15cm}mit\hspace{0.15cm} }1-2\cdot p_2\hspace{0.05cm},$$
$$p_3 = 0.3 \hspace{0.2cm}\Rightarrow \hspace{0.2cm} L_3 = 0.847 \hspace{0.2cm}\Rightarrow \hspace{0.2cm} L_3/2 = +0.419 \hspace{0.2cm} \Rightarrow \hspace{0.2cm} {\rm tanh}(L_3/2) = \frac{{\rm e}^{+0.419}-{\rm e}^{-0.419}}{{\rm e}^{+0.419}+{\rm e}^{-0.419}} = 0.4 \hspace{0.3cm}\Rightarrow \hspace{0.3cm}{\rm identisch \hspace{0.15cm}mit\hspace{0.15cm} }1-2\cdot p_3\hspace{0.05cm}.$$
  • The final result is thus:
$$\pi = (+0.6) \cdot (-0.8) \cdot (+0.4) = -0.192 \hspace{0.3cm}\Rightarrow \hspace{0.3cm} L_{\rm E}(i = 4) = {\rm ln} \hspace{0.2cm} \frac{1 -0.192}{1 +0.192}\hspace{0.15cm}\underline{= -0.389} \hspace{0.05cm}.$$