Difference between revisions of "Channel Coding/Soft-in Soft-Out Decoder"

From LNTwww
 
(120 intermediate revisions by 7 users not shown)
Line 1: Line 1:
 
   
 
   
 
{{Header
 
{{Header
|Untermenü=Iterative Decodierverfahren
+
|Untermenü=Iterative Decoding Methods
|Vorherige Seite=Distanzeigenschaften und Fehlerwahrscheinlichkeitsschranken
+
|Vorherige Seite=Distance Characteristics and Error Probability Bounds
|Nächste Seite=Grundlegendes zu den Produktcodes
+
|Nächste Seite=The Basics of Product Codes
 
}}
 
}}
  
== Hard Decision vs. Soft Decision (1) ==
+
== # OVERVIEW OF THE FOURTH MAIN CHAPTER # ==
 
<br>
 
<br>
Zur Hinleitung auf die Thematik dieses vierten und letzten Kapitels und zur Motivation betrachten wir das folgende Nachrichtenübertragungssystem mit Codierung.<br>
 
  
[[File:P ID2971 KC T 4 1 S1a v8.png|center|frame|Betrachtetes Nachrichtenübertragungssystem mit Codierung|class=fit]]<br>
+
The last main chapter of the channel coding book describes&nbsp; &raquo;'''iterative decoding techniques'''&laquo;&nbsp; as used in most of today's&nbsp; (2017)&nbsp; communication systems.&nbsp; This is due to the following reasons:
  
Nachfolgend werden alle Symbole in bipolarer Darstellung angegeben: &bdquo;$0$&rdquo; &nbsp;&#8594;&nbsp; &bdquo;$+1$&rdquo; und &bdquo;$1$&rdquo; &nbsp;&#8594;&nbsp; &bdquo;$&ndash;1$&rdquo;.
+
*To approach the channel capacity,&nbsp; one needs very long codes.
*Die Symbolfolge $\underline{u} = (u_1, \ u_2)$ wird der Coderfolge $\underline{x} = (x_1, \ x_2, \ x_3) = (u_1, \ u_2, \ p)$ zugeordnet, wobei für das Paritybit $p = u_1 &oplus; u_2$ gilt &nbsp;&#8658;&nbsp; [[Kanalcodierung/Beispiele_bin%C3%A4rer_Blockcodes#Single_Parity.E2.80.93check_Codes_.281.29| Single Parity&ndash;check Code]] &nbsp;&#8658;&nbsp; ${\rm SPC} (3, 2, 2)$.<br>
 
  
*Der [[Kanalcodierung/Klassifizierung_von_Signalen#AWGN.E2.80.93Kanal_bei_bin.C3.A4rem_Eingang| AWGN&ndash;Kanal]] verändert die Binärsymbole $x_i &#8712; \{+1, \ &ndash;1\}$ zu reellwertigen Ausgangswerten $y_i$, z.B. im Block 4 der unteren Tabelle: &nbsp; $x_1 = +1$ &nbsp;&#8658;&nbsp; $y_1 = +0.9$ und $x_3 = \, &ndash;1$ &nbsp;&#8658;&nbsp; $y_3 = +0.1$.<br>
+
*But for long codes,&nbsp; "blockwise maximum likelihood decoding"&nbsp; is almost impossible.
  
*Die Decodierung geschieht gemäß dem Kriterium [[Kanalcodierung/Klassifizierung_von_Signalen#MAP.E2.80.93_und_ML.E2.80.93Kriterium_.282.29| Maximum Likelihood (block&ndash;wise ML)]], wobei zwischen <i>Hard Decision</i> (ML&ndash;HD) und <i>Soft Decision</i> (ML&ndash;SD) zu unterscheiden ist.<br>
 
  
*Das obige Blockschaltbild in seiner Gesamtheit entspricht ML&ndash;HD. Hier werden zur Detektion nur die Vorzeichen der AWGN&ndash;Ausgangswerte entsprechend $y_{\rm HD, \ \it i} = {\rm sign}[y_{\rm SD, \ \it i}]$ ausgewertet.<br>
+
The decoder complexity can be significantly reduced with almost the same quality if two&nbsp; $($or more$)$&nbsp; short channel codes are linked together and the newly acquired&nbsp; $($soft$)$&nbsp; information is exchanged between the decoders at the receiver in several steps &nbsp; &rArr; &nbsp; "iteratively".
  
*Bei <i>Soft Decision</i> (ML&ndash;SD) verzichtet man auf den schraffierten Block in obigem Blockschaltbild und wertet direkt die wertkontinuierlichen Eingangsgrößen $y_{\rm SD, \ \it i}$ aus.<br>
+
The breakthrough in the field came in the early 1990s with the invention of the&nbsp; "turbo codes"&nbsp; by&nbsp; [https://en.wikipedia.org/wiki/Claude_Berrou $\text{Claude Berrou}$]&nbsp; and shortly thereafter with the rediscovery of&nbsp; "low-density parity-check codes"&nbsp; by&nbsp; [https://en.wikipedia.org/wiki/David_J._C._MacKay $\text{David J. C. MacKay}$]&nbsp; and&nbsp; [https://en.wikipedia.org/wiki/Radford_M._Neal $\text{Radford M. Neal}$],&nbsp; after the LDPC codes developed as early as 1961 by&nbsp; [https://en.wikipedia.org/wiki/Robert_G._Gallager $\text{Robert G. Gallager}$]&nbsp; had been forgotten in the meantime.
  
[[File:P ID2972 KC T 4 1 S1b v4.png|center|frame|Gegenüberstellung von <i>Hard Decision</i> und <i>Soft Decision</i>|class=fit]]<br>
+
Specifically,&nbsp; the fourth main chapter discusses:
  
Aus der Beispieltabelle erkennt man:
+
*a comparison of&nbsp; &raquo;hard decision&laquo;&nbsp; and&nbsp; &raquo;soft decision&laquo;,
*<b>Hard Decision</b> &nbsp;&#8658; die Symbolfolge $\underline{\upsilon}_{\rm HD}$ ergibt sich aus den hart entschiedenen Kanalwerten $\underline{y}_{\rm HD}$ (blaue Hinterlegung): Es werden nur die beiden ersten Blöcke fehlerfrei decodiert.<br>
 
  
*<b>Soft Decision</b> &nbsp;&#8658; die Symbolfolge $\underline{\upsilon}_{\rm SD}$ ergibt sich aus den &bdquo;weichen&rdquo; Kanalausgangswerten $\underline{y}_{\rm SD}$ (grüne Hinterlegung): Nun wird in diesem Beispiel auch der dritte Block richtig entschieden.<br><br>
+
*the quantification of&nbsp; &raquo;reliability information&laquo;&nbsp; by&nbsp; &raquo;log likelihood ratios&laquo;,
  
Die Bildbeschreibung wird auf der nächsten Seite fortgesetzt.<br>
+
*the principle of symbol-wise&nbsp; &raquo;soft-in soft-out&nbsp; (SISO)&laquo;&nbsp; decoding,
  
== Hard Decision vs. Soft Decision (2) ==
+
*the definition of&nbsp; &raquo;a-priori&nbsp; L&ndash;value&laquo;,&nbsp; &raquo;a-posteriori&nbsp; L&ndash;value&laquo;&nbsp; and&nbsp; &raquo;extrinsic&nbsp; L&ndash;value&laquo;,
 +
 
 +
*the basic structure of&nbsp; &raquo;serially concatenated&laquo;&nbsp; resp.&nbsp; &raquo;parallel concatenated&laquo;&nbsp; coding systems,
 +
 
 +
*the characteristics of&nbsp; &raquo;product codes&laquo;&nbsp; and their&nbsp; &raquo;hard decision decoding&laquo;,
 +
 
 +
*the basic structure,&nbsp; decoding algorithm,&nbsp; and performance of&nbsp; &raquo;turbo codes&laquo;,
 +
 
 +
*basic information on the&nbsp; &raquo;low-density parity-check codes&laquo;&nbsp; and their applications.
 +
 
 +
 
 +
 
 +
 
 +
== Hard Decision vs. Soft Decision==
 
<br>
 
<br>
Für alle Spalten der folgenden Tabelle wird vorausgesetzt:
+
To introduce the topic discussed here, let us consider the following digital transmission system with coding.<br>
*der Nachrichtenblock $\underline{u} = (0, 1)$, bipolar darstellbar als $(+1, \, &ndash;1)$,<br>
 
  
*der ${\rm SPC} \ (3, 2)$&ndash;codierte Block $\underline{x} = (0, 1, 1)$ bzw. in Bipolardarstellung $(+1, \, &ndash;1, \, &ndash;1)$.<br><br>
+
*In the following, all symbols are given in bipolar representation:
 +
[[File:EN_KC_T_4_1_S1a.png|right|frame|Considered digital transmission system with coding|class=fit]]
  
Die vier Blöcke unterscheiden sich also nur durch unterschiedliche AWGN&ndash;Realisierungen.<br>
+
[[File:EN_KC_T_4_1_S1b.png|right|frame|Comparison of&nbsp; "hard decision" and&nbsp; "soft decision"; &nbsp; for all columns of this table it is assumed:<br>'''(1)''' &nbsp; the information block &nbsp; $\underline{u} = (0,\ 1)$,&nbsp; bipolar representable as&nbsp; $(+1, \, &ndash;1)$,<br>'''(2)''' &nbsp; the encoded block&nbsp; $\underline{x} = (0,\ 1,\ 1)$ &nbsp; &rArr; &nbsp; in bipolar representation:&nbsp; $(+1, \, -1, \, -1)$.<br>|class=fit]]
  
[[File:P ID2973 KC T 4 1 S1b v4.png|center|frame|Gegenüberstellung von <i>Hard Decision</i>  und <i>Soft Decision</i>|class=fit]]<br>
 
  
Die Tabelle ist wie folgt zu interpretieren:
+
*Bei idealem Kanal entsprechend Block 1 &nbsp;&#8658;&nbsp; $\underline{x} = \underline{y}_{\rm SD} = \underline{y}_{\rm HD}$ gibt es keinen Unterschied zwischen der (blauen) herkömmlichen ML&ndash;HD &ndash;Variante und der (grünen) ML&ndash;SD &ndash;Variante.<br>
+
::"$0$" &nbsp;&#8594;&nbsp; "$+1$",&nbsp; and&nbsp;  
 +
::"$1$" &nbsp;&#8594;&nbsp; "$-1$".
  
*Der Block 2 demonstriert geringe Signalverfälschungen. Wegen $\underline{y}_{\rm HD} = \underline{x}$ (das heißt, dass der Kanal die Vorzeichen nicht  verfälscht) liefert auch ML&ndash;HD das richtige Ergebnis $\underline{\upsilon}_{\rm HD} = \underline{u}$.<br>
+
*The symbol sequence&nbsp; $\underline{u} = (u_1, \ u_2)$&nbsp; of the digital source is assigned to the encoded sequence&nbsp;
 +
:$$\underline{x} = (x_1, \ x_2, \ x_3) = (u_1, \ u_2, \ p)$$
 +
:where for the parity bit &nbsp; $p = u_1 &oplus; u_2$ &nbsp; holds &nbsp; &#8658; &nbsp; [[Channel_Coding/Examples_of_Binary_Block_Codes#Single_Parity-check_Codes| $\text{ SPC (3, 2, 2)}$]].<br>
  
*Dagegen gilt im dritten Block $\underline{y}_{\rm HD} &ne; \underline{x}$ und es gibt auch keine ${\rm SPC} \ (3, 2)$&ndash;Zuordnung $\underline{u}$ &nbsp;&#8658;&nbsp; $\underline{y}_{\rm HD}$. Der ML&ndash;Decoder kann hier nur durch die Ausgabe $\underline{\upsilon}_{\rm HD} = \rm (E, \ E)$ vermelden, dass er bei der Decodierung dieses Blocks gescheitert ist. &bdquo;$\rm E$&rdquo; steht hierbei für [[Kanalcodierung/Klassifizierung_von_Signalen#Binary_Erasure_Channel_.E2.80.93_BEC| Erasure]] (deutsch: <i>Auslöschung</i>).<br>
+
*The&nbsp; [[Channel_Coding/Channel_Models_and_Decision_Structures#AWGN_channel_at_binary_input| $\text{AWGN channel}$]]&nbsp; changes the symbols&nbsp; $x_i &#8712; \{+1, \ &ndash;1\}$&nbsp; to real-valued output values&nbsp; $y_i$,&nbsp; for example according to&nbsp; $\text{channel 4}$&nbsp; in the table below: &nbsp;
 +
**$x_1 = +1$ &nbsp; &#8658; &nbsp; $y_1 = +0.9,$
 +
**$x_2 = -1$ &nbsp; &#8658; &nbsp; $y_1 = +0.1,$
 +
**$x_3 = -1$ &nbsp; &#8658; &nbsp; $y_1 = +0.1,$
  
*Auch der <i>Soft Decision</i> Decoder erkennt, dass eine Decodierung anhand der Vorzeichen nicht funktioniert. Anhand der $\underline{y}_{\rm SD}$&ndash;Werte erkennt er aber, dass mit großer Wahrscheinlichkeit das zweite Bit verfälscht wurde und entscheidet sich für die richtige Symbolfolge $\underline{\upsilon}_{\rm SD} = (+1, \, &ndash;1) = \underline{u}$.<br>
+
*Decoding is done according to the criterion&nbsp; [[Channel_Coding/Channel_Models_and_Decision_Structures#Criteria_.C2.BBMaximum-a-posteriori.C2.AB_and_.C2.BBMaximum-Likelihood.C2.AB|$\text{&raquo;block-wise maximum-likelihood&laquo;}$]]&nbsp; $\text{(ML)}$,&nbsp; distinguishing between&nbsp;
 +
**"hard decision" &nbsp; $\rm {(ML&ndash;HD)}$,&nbsp; and&nbsp;
 +
**"soft decision" &nbsp; $\rm {(ML&ndash;SD)}$&nbsp;.<br>
  
*Im vierten Block werden durch den AWGN&ndash;Kanal sowohl die Vorzeichen von Bit 2 als auch von Bit 3 verändert, was zum Ergebnis $\underline{\upsilon}_{\rm HD} = (+1, +1) &ne; \underline{u}(+1, \, &ndash;1)$ führt &nbsp;&#8658;&nbsp; ein Blockfehler und gleichzeitig ein Bitfehler. Auch der <i>Soft Decision</i> Decoder liefert hier das gleiche falsche Ergebnis.<br><br>
+
*The given block diagram corresponds to &nbsp;$\rm ML&ndash;HD$.&nbsp; Here,&nbsp; only the signs of the AWGN output values &nbsp; &rArr; &nbsp; $y_{\rm HD, \ \it i} = {\rm sign}\left [y_{\rm SD, \ \it i}\right ]$&nbsp; are evaluated for decoding.  
  
Die Decodiervariante ML&ndash;SD bietet gegenüber ML&ndash;HD zudem den Vorteil, dass man relativ einfach jedes Decodierergebnis mit einem Zuverlässigkeitswert versehen kann (in obiger Tabelle ist dieser allerdings nicht angegeben). Dieser Zuverlässigkeitswert
+
*With&nbsp; "soft decision"&nbsp; one omits the shaded block and directly evaluates the continuous value input variables&nbsp; $y_{\rm SD, \ \it i}$&nbsp;.<br>
*hätte bei Block 1 seinen Maximalwert,<br>
 
  
*wäre bei Block 2 deutlich kleiner,<br>
 
  
*läge bei Block 3 und 4 nahe bei $0$.<br>
+
Thus,&nbsp; the four columns in the table differ only by different AWGN realizations.<br>
  
== Zuverlässigkeitsinformation – Log Likelihood Ratio (1) ==
+
{{BlaueBox|TEXT=
<br>
+
$\text{Definitions:}$&nbsp; From the example table you can see:
Es sei $x &#8712; \{+1, \, &ndash;1\}$ eine binäre Zufallsvariable mit den Wahrscheinlichkeiten ${\rm Pr}(x = +1)$ und ${\rm Pr}(x = \, &ndash;1)$. Für die Codierungstheorie erweist es sich als zweckmäßig hinsichtlich der Rechenzeiten, wenn man anstelle der Wahrscheinlichkeiten ${\rm Pr}(x = \, &plusmn;1)$ den natürlichen Logarithmus des Quotienten heranzieht.<br>
+
*$\text{Hard Decision:}$ &nbsp; The sink symbol sequence&nbsp; $\underline{v}_{\rm HD}$&nbsp; results from the&nbsp; "hard decided channel values"&nbsp; $\underline{y}_{\rm HD}$&nbsp; $($blue background$)$. <br> &nbsp; &nbsp; In our example,&nbsp; only the constellations according to&nbsp; $\text{channel 1}$&nbsp; and&nbsp; $\text{channel 2}$&nbsp; are decoded without errors.<br>
 +
 
 +
*$\text{Soft Decision:}$ &nbsp; The sink symbol sequence&nbsp; $\underline{v}_{\rm SD}$&nbsp; results from the&nbsp; "soft channel output values"&nbsp; $\underline{y}_{\rm SD}$&nbsp; $($green background$)$. <br> &nbsp; &nbsp; Now,&nbsp;  in this example,&nbsp; it is also correctly decided at&nbsp; $\text{channel 3}$.&nbsp; }}<br>
 +
 
 +
The entries in the example table above are to be interpreted as follows:
 +
# For ideal channel &nbsp; &#8658; &nbsp; $\text{channel 1}$ &nbsp; &#8658; &nbsp; $\underline{x} = \underline{y}_{\rm SD} = \underline{y}_{\rm HD}$&nbsp; there is no difference between the (blue) conventional hard decision variant&nbsp; $\rm {(ML&ndash;HD)}$&nbsp; and the (green) soft decision variant&nbsp; $\rm {(ML&ndash;SD)}$.<br>
 +
#  The setting according to&nbsp; $\text{channel 2}$&nbsp; demonstrates low signal distortions.&nbsp; Because of&nbsp; $\underline{y}_{\rm HD} = \underline{x}$&nbsp; $($which means that the channel does not distort the signs$)$&nbsp; also&nbsp; $\rm ML&ndash;HD$&nbsp; gives the correct result&nbsp; $\underline{v}_{\rm HD} = \underline{u}$.<br>
 +
# At&nbsp; $\text{channel 3}$&nbsp; there is&nbsp; $\underline{y}_{\rm HD} &ne; \underline{x}$&nbsp; and there is also no&nbsp; ${\rm SPC} \ (3, 2)$&nbsp; assignment&nbsp;  $\underline{u}$ &nbsp; &#8658; &nbsp; $\underline{y}_{\rm HD}$.&nbsp; The maximum Likelihood decoder reports here by outputting&nbsp; $\underline{v}_{\rm HD} = \rm (E, \ E)$ that it failed in decoding this block.&nbsp; "$\rm E$" stands for an&nbsp; [[Channel_Coding/Channel_Models_and_Decision_Structures#Binary_Erasure_Channel_.E2.80.93_BEC|$\text{Erasure}$]]&nbsp;.<br>
 +
#  Also the&nbsp; soft decision decoder recognizes that decoding based on the signs does not work.&nbsp; Based on the&nbsp; $\underline{y}_{\rm SD}$&nbsp; values,&nbsp; however,&nbsp; it recognizes that with high probability the second bit has been falsified and decides to use the correct symbol sequence&nbsp; $\underline{v}_{\rm SD} = (+1, \, -1) = \underline{u}$.<br>
 +
#  In&nbsp; $\text{channel 4}$,&nbsp; both the signs of bit 2 and bit 3 are changed by the AWGN channel,&nbsp; leading to the result &nbsp; $\underline{v}_{\rm HD} = (+1, +1) &ne; \underline{u}(+1, \, -1)$ &nbsp; &#8658; &nbsp; a block error and a bit error at the same time.&nbsp; Also the soft decision decoder gives the same wrong result here.<br><br>
  
{{Definition}}''':''' Das <b>Log&ndash;Likelihood&ndash;Verhältnis</b> (kurz: der $L$&ndash;Wert, englisch: <i>Log&ndash;Likelihood Ratio</i>, LLR) der Zufallsgröße $x &#8712; \{+1, \, &ndash;1\}$ lautet:
+
The decoding variant&nbsp; $\rm ML&ndash;SD$&nbsp; also offers the advantage over&nbsp; $\rm ML&ndash;HD$&nbsp; that it is relatively easy to assign a reliability value to each decoding result&nbsp; $($but in the above table this is not specified$)$.&nbsp;  This reliability value would
 +
*have its maximum value for&nbsp; $\text{channel 1}$&nbsp;,<br>
  
:<math>L(x)={\rm ln} \hspace{0.15cm} \frac{{\rm Pr}(x = +1)}{{\rm Pr}(x = -1)}\hspace{0.05cm}.</math>
+
*be significantly smaller for&nbsp; $\text{channel 2}$&nbsp;,<br>
  
Bei unipolarer/symbolhafter Darstellung ($+1&nbsp;&#8594;&nbsp;0$ &nbsp;und&nbsp; $&ndash;1&nbsp;&#8594;&nbsp;1$) gilt entsprechend mit $\xi &#8712; \{0, 1\}$:
+
*be close to zero for&nbsp; $\text{channel 3}$&nbsp; and&nbsp; $\text{channel 4}$.<br>
  
:<math>L(\xi)={\rm ln} \hspace{0.15cm} \frac{{\rm Pr}(\xi = 0)}{{\rm Pr}(\xi = 1)}\hspace{0.05cm}.</math>{{end}}<br>
+
== Reliability information - Log likelihood ratio==
 +
<br>
 +
Let be&nbsp; $x &#8712; \{+1, \, -1\}$&nbsp; a binary random variable with probabilities&nbsp; ${\rm Pr}(x = +1)$&nbsp; and&nbsp; ${\rm Pr}(x = \, -1)$.&nbsp; For coding theory,&nbsp; it proves convenient in terms of computation times to use the natural logarithm of the quotient instead of the probabilities&nbsp; ${\rm Pr}(x = &plusmn;1)$&nbsp; .<br>
  
Nachfolgend ist der nichtlineare Zusammenhang zwischen ${\rm Pr}(x = &plusmn;1)$ und $L(x)$ angegeben. Ersetzt man ${\rm Pr}(x = +1)$ durch ${\rm Pr}(\xi = 0)$, so gibt die mittlere Zeile den $L$&ndash;Wert der unipolaren Zufallsgröße $\xi$ an.<br>
+
{{BlaueBox|TEXT= 
 +
$\text{Definition:}$&nbsp;  The&nbsp; <b>log likelihood ratio</b>&nbsp; $\rm (LLR)$&nbsp; of the random variable&nbsp; $x &#8712; \{+1, \, -1\}$&nbsp; is:
  
[[File:P ID2975 KC T 4 1 S2a v1.png|Wahrscheinlichkeit und <i>L</i>–Wert|class=fit]]<br>
+
::<math>L(x)={\rm ln} \hspace{0.15cm} \frac{ {\rm Pr}(x = +1)}{ {\rm Pr}(x = -1)}\hspace{0.05cm}.</math>
  
Man erkennt:
+
*For unipolar representation&nbsp; $(+1&nbsp; &#8594; &nbsp;0$ &nbsp; and &nbsp; $-1&nbsp; &#8594; &nbsp;1)$&nbsp; applies accordingly with&nbsp; $\xi &#8712; \{0, \, 1\}$:
*Der wahrscheinlichere Zufallswert von $x &#8712; \{+1, \, &ndash;1\}$ ist durch ${\rm sign} L(x)$ &nbsp;&#8658;&nbsp; <b>Vorzeichen</b> gegeben.<br>
 
  
*Dagegen gibt der <b>Betrag</b> $|L(x)|$ die Zuverlässigkeit für das Ergebnis ${\rm sign}(L(x))$ an.<br><br>
+
::<math>L(\xi)={\rm ln} \hspace{0.15cm} \frac{ {\rm Pr}(\xi = 0)}{ {\rm Pr}(\xi = 1)}\hspace{0.05cm}.</math>}}<br>
  
{{Beispiel}}''':'''
+
[[File:P ID2975 KC T 4 1 S2a v1.png|right|frame|Probability and log likelihood ratio|class=fit]]
[[File:P ID2976 KC T 4 1 S2b v2.png|right|frame|BSC–Modell]] Wir betrachten das  skizzierte [[Kanalcodierung/Klassifizierung_von_Signalen#Binary_Symmetric_Channel_.E2.80.93_BSC| BSC&ndash;Modell]] mit bipolarer Darstellung. Hier gilt mit  der Verfälschungswahrscheinlichkeit $\epsilon = 0.1$ sowie den beiden Zufallsgrößen $x &#8712; \{+1, \, &ndash;1\}$ und $y &#8712; \{+1, \, &ndash;1\}$ am Eingang und Ausgang des Kanals:
 
  
:<math>L(y\hspace{0.05cm}|\hspace{0.05cm}x) \hspace{-0.15cm}  = \hspace{-0.15cm}
+
The table gives the nonlinear relationship between&nbsp; ${\rm Pr}(x = &plusmn;1)$&nbsp; and&nbsp; $L(x)$.&nbsp; Replacing&nbsp; ${\rm Pr}(x = +1)$&nbsp; with&nbsp; ${\rm Pr}(\xi = 0)$,&nbsp; the middle row gives the&nbsp; $L$&ndash;value of the unipolar random variable&nbsp; $\xi$.<br>
{\rm ln} \hspace{0.15cm} \frac{{\rm Pr}(y\hspace{0.05cm}|\hspace{0.05cm}x = +1)}{{\rm Pr}(y\hspace{0.05cm}|\hspace{0.05cm}x = -1)}  = </math>
 
:<math>\hspace{-0.15cm}  =  \hspace{-0.15cm}
 
\left\{ \begin{array}{c} {\rm ln} \hspace{0.15cm} [(1 - \varepsilon)/\varepsilon]\\
 
{\rm ln} \hspace{0.15cm} [\varepsilon/(1 - \varepsilon)]  \end{array} \right.\quad
 
\begin{array}{*{1}c} {\rm f\ddot{u}r} \hspace{0.15cm} y = +1,
 
\\  {\rm f\ddot{u}r} \hspace{0.15cm} y = -1. \\ \end{array}</math>
 
  
Beispielsweise ergeben sich für $\epsilon = 0.1$ folgende Zahlenwerte (vergleiche obere Tabelle):
+
You can see:
 +
#The more likely random value of&nbsp; $x &#8712; \{+1, \, -1\}$&nbsp; is given by the&nbsp; <b>sign</b> &nbsp; &#8658; &nbsp; ${\rm sign} \ L(x)$.<br>
 +
#In contrast,&nbsp; the&nbsp; <b>absolute value</b>&nbsp; &nbsp; &#8658; &nbsp; $|L(x)|$&nbsp; indicates the reliability for the result&nbsp; "${\rm sign}(L(x))$".<br><br>
  
:<math>L(y = +1\hspace{0.05cm}|\hspace{0.05cm}x) =
+
{{GraueBox|TEXT=
{\rm ln} \hspace{0.15cm} \frac{0.9}{0.1} = +2.197\hspace{0.05cm}, \hspace{0.2cm}
+
[[File:P ID2976 KC T 4 1 S2b v2.png|right|frame|Given BSC model]] 
  L(y = -1\hspace{0.05cm}|\hspace{0.05cm}x) = -2.197\hspace{0.05cm}.</math>
+
$\text{Example 1:}$&nbsp;  We consider the outlined&nbsp; [[Channel_Coding/Channel_Models_and_Decision_Structures#Binary_Symmetric_Channel_.E2.80.93_BSC| $\text{BSC model}$]]&nbsp; with bipolar representation.&nbsp;
 +
*With the falsification probability&nbsp; $\varepsilon = 0.1$&nbsp; and the two random variables&nbsp; $x &#8712; \{+1, \, -1\}$&nbsp; and&nbsp; $y &#8712; \{+1, \, -1\}$&nbsp; at the input and output of the channel:
 +
::<math>L(y\hspace{0.05cm}\vert\hspace{0.05cm}x) =
 +
{\rm ln} \hspace{0.15cm} \frac{ {\rm Pr}(y\hspace{0.05cm}\vert\hspace{0.05cm}x = +1) }{ {\rm Pr}(y\hspace{0.05cm}\vert\hspace{0.05cm}x = -1)}  =
 +
\left\{ \begin{array}{c} {\rm ln} \hspace{0.15cm} \big[(1 - \varepsilon)/\varepsilon \big]\\
 +
  {\rm ln} \hspace{0.15cm}\big [\varepsilon/(1 - \varepsilon)\big]  \end{array} \right.\quad
 +
\begin{array}{*{1}c} {\rm for} \hspace{0.15cm} y = +1,
 +
\\  {\rm for} \hspace{0.15cm} y = -1. \\ \end{array}</math>
  
Dieses Beispiel zeigt, dass man die sog. $L$&ndash;Wert&ndash;Algebra auch auf bedingte Wahrscheinlichkeiten anwenden kann. In Aufgabe Z4.1 wird das BEC&ndash;Modell in ähnlicher Weise beschrieben.{{end}}<br>
+
*For example,&nbsp; $\varepsilon = 0.1$&nbsp; results in the following numerical values&nbsp; $($compare the upper table$)$:
 +
::<math>L(y = +1\hspace{0.05cm}\vert\hspace{0.05cm}x) =
 +
{\rm ln} \hspace{0.15cm} \frac{0.9}{0.1}  = +2.197\hspace{0.05cm}, \hspace{0.8cm}
 +
L(y = -1\hspace{0.05cm}\vert\hspace{0.05cm}x) = -2.197\hspace{0.05cm}.</math>
  
== Zuverlässigkeitsinformation – Log Likelihood Ratio (2) ==
+
*This example shows that the so-called&nbsp; &raquo;$\text{LLR algebra}$&laquo;  can also be applied to conditional probabilities. In &nbsp;[[Aufgaben:Exercise_4.1Z:_Log_Likelihood_Ratio_at_the_BEC_Model|"Exercise 4.1Z"]],&nbsp; the BEC&ndash;model is described in a similar way.}}<br>
<br>
+
 
Wir betrachten nun den AWGN&ndash;Kanal mit den beiden bedingten Wahrscheinlichkeitsdichtefunktionen
+
{{GraueBox|TEXT=  
 +
[[File:EN_KC_T_4_1_S2c.png|right|frame|Conditional AWGN probability density functions|class=fit]] 
 +
$\text{Example 2:}$&nbsp;
 +
In another example,&nbsp; we consider the&nbsp; [[Channel_Coding/Channel_Models_and_Decision_Structures#AWGN_channel_at_binary_input|$\rm AWGN$ ]]&nbsp; channel with the conditional probability density functions
  
:<math>f_{y \hspace{0.03cm}| \hspace{0.03cm}x=+1 } \hspace{0.05cm} (y \hspace{0.05cm}| \hspace{0.05cm}x=+1 )\hspace{-0.1cm} =  \hspace{-0.1cm}
+
::<math>f_{y \hspace{0.03cm}\vert \hspace{0.03cm}x=+1 } \hspace{0.05cm} (y \hspace{0.05cm}\vert\hspace{0.05cm}x=+1 )\hspace{-0.1cm} =  \hspace{-0.1cm}
 
\frac {1}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ -  {(y-1)^2}/(2\sigma^2) } \hspace{0.05cm},</math>  
 
\frac {1}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ -  {(y-1)^2}/(2\sigma^2) } \hspace{0.05cm},</math>  
:<math>f_{y \hspace{0.03cm}| \hspace{0.03cm}x=-1 } \hspace{0.05cm} (y \hspace{0.05cm}| \hspace{0.05cm}x=-1 )\hspace{-0.1cm} =  \hspace{-0.1cm}
+
::<math>f_{y \hspace{0.03cm}\vert \hspace{0.03cm}x=-1 } \hspace{0.05cm} (y \hspace{0.05cm}\vert\hspace{0.05cm}x=-1 )\hspace{-0.1cm} =  \hspace{-0.1cm}
 
\frac {1}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ -  {(y+1)^2}/(2\sigma^2) } \hspace{0.05cm}.</math>
 
\frac {1}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ -  {(y+1)^2}/(2\sigma^2) } \hspace{0.05cm}.</math>
  
In der Grafik sind zwei beispielhafte Gaußfunktionen als blaue bzw. rote Kurve dargestellt.<br>
+
&rArr; &nbsp; In the graph,&nbsp; two exemplary Gaussian functions are shown as blue and red curves, respectively.&nbsp;
  
[[File:P ID2977 KC T 4 1 S2c v3.png|Bedingte AWGN–Dichtefunktionen|class=fit]]<br>
+
*The total probability density function&nbsp; $\rm (PDF)$&nbsp; of the output signal&nbsp;$y$&nbsp; is obtained from the&nbsp; (equally)&nbsp; weighted sum:
  
Die gesamte WDF des Ausgangssignals <i>y</i> ergibt sich aus der (gleich) gewichteten Summe:
+
::<math>f_{y  } \hspace{0.05cm} (y  ) = 1/2 \cdot \big [ f_{y \hspace{0.03cm}\vert \hspace{0.03cm}x=+1 } \hspace{0.05cm} (y \hspace{0.05cm}\vert\hspace{0.05cm}x=+1 ) \hspace{0.1cm} +  \hspace{0.1cm}
 
+
f_{y \hspace{0.03cm}\vert \hspace{0.03cm}x=-1 } \hspace{0.05cm} (y \hspace{0.05cm}\vert \hspace{0.05cm}x=-1 )
:<math>f_{y  } \hspace{0.05cm} (y  ) = 1/2 \cdot \big [ f_{y \hspace{0.03cm}| \hspace{0.03cm}x=+1 } \hspace{0.05cm} (y \hspace{0.05cm}| \hspace{0.05cm}x=+1 ) \hspace{0.1cm} +  \hspace{0.1cm}
 
f_{y \hspace{0.03cm}| \hspace{0.03cm}x=-1 } \hspace{0.05cm} (y \hspace{0.05cm}| \hspace{0.05cm}x=-1 )
 
 
\big ]
 
\big ]
 
\hspace{0.05cm}.</math>
 
\hspace{0.05cm}.</math>
  
Berechnen wir nun die Wahrscheinlichkeit, dass der Empfangswert <i>y</i> in einem (sehr) schmalen Intervall der Breite <i>&Delta;</i> um <i>y</i><sub>0</sub> = 0.25 liegt, so erhält man näherungsweise
+
* We now calculate the probability that the received value&nbsp; $y$&nbsp; lies in a&nbsp; (very)&nbsp; narrow interval of width&nbsp; $\it \Delta$&nbsp; around&nbsp; $y_0 = 0.25$.&nbsp; One obtains approximately
  
:<math>{\rm Pr} (|y - y_0| \le{\it  \Delta}/2 \hspace{0.05cm}
+
::<math>{\rm Pr} (\vert y - y_0\vert  \le{\it  \Delta}/2 \hspace{0.05cm}
\Big \hspace{0.05cm}x=+1 )\hspace{-0.1cm} \approx  \hspace{-0.1cm}
+
\Big \vert  \hspace{0.05cm}x=+1 )\hspace{-0.1cm} \approx  \hspace{-0.1cm}
 
\frac {\it  \Delta}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ -  {(y_0-1)^2}/(2\sigma^2) } \hspace{0.05cm},</math>  
 
\frac {\it  \Delta}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ -  {(y_0-1)^2}/(2\sigma^2) } \hspace{0.05cm},</math>  
:<math>{\rm Pr} (|y - y_0| \le {\it  \Delta}/2 \hspace{0.05cm}
+
::<math>{\rm Pr} (\vert y - y_0\vert  \le {\it  \Delta}/2 \hspace{0.05cm}
\Big \hspace{0.05cm}x=-1 )\hspace{-0.1cm} \approx  \hspace{-0.1cm}
+
\Big \vert  \hspace{0.05cm}x=-1 )\hspace{-0.1cm} \approx  \hspace{-0.1cm}
 
\frac {\it  \Delta}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ -  {(y_0+1)^2}/(2\sigma^2) } \hspace{0.05cm}.</math>
 
\frac {\it  \Delta}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ -  {(y_0+1)^2}/(2\sigma^2) } \hspace{0.05cm}.</math>
  
Die etwas größeren senkrechten Striche bezeichnen die Bedingungen, die kleineren die Betragsbildung.<br>
+
:The slightly larger vertical lines denote the conditions,&nbsp; the smaller ones the absolute value.<br>
  
Der <i>L</i>&ndash;Wert der bedingten Wahrscheinlichkeit in Vorwärtsrichtung (das bedeutet: Ausgang <i>y</i> für einen gegebenen Eingang <i>x</i>) ergibt sich somit als der Logarithmus des Quotienten beider Ausdrücke:
+
*The&nbsp; log likelihood ratio&nbsp; $\rm (LLR)$&nbsp; of the conditional probability in the forward direction&nbsp; $($meaning: &nbsp; output &nbsp;$y$&nbsp; for a given input&nbsp; $x)$&nbsp; is thus obtained as the logarithm of the quotient of both expressions:
  
:<math>L(y = y_0\hspace{0.05cm}|\hspace{0.05cm}x) \hspace{-0.1cm} = \hspace{-0.1cm}
+
::<math>L(y = y_0\hspace{0.05cm}\vert\hspace{0.05cm}x) =
 
{\rm ln} \hspace{0.15cm} \left [ \frac{{\rm e} ^{ -  {(y_0-1)^2}/(2\sigma^2)}}{{\rm e} ^{ -  {(y_0+1)^2}/(2\sigma^2)}}  \right ] =  
 
{\rm ln} \hspace{0.15cm} \left [ \frac{{\rm e} ^{ -  {(y_0-1)^2}/(2\sigma^2)}}{{\rm e} ^{ -  {(y_0+1)^2}/(2\sigma^2)}}  \right ] =  
{\rm ln} \left [ {\rm e} ^{ - [ {(y_0-1)^2}+{(y_0+1)^2}]/(2\sigma^2)}  \right ]\hspace{0.15cm} = </math>
+
{\rm ln} \left [ {\rm e} ^{ - [ {(y_0-1)^2}+{(y_0+1)^2}]/(2\sigma^2)}  \right ] =  \frac{(y_0+1)^2-(y_0-1)^2}{2\cdot \sigma^2} = \frac{2 \cdot y_0}{\sigma^2}\hspace{0.05cm}. </math>
:<math>\hspace{2.4cm} \hspace{-0.1cm} \frac{(y_0+1)^2-(y_0-1)^2}{2\cdot \sigma^2} = \frac{2 \cdot y_0}{\sigma^2}\hspace{0.05cm}. </math>
 
  
Ersetzen wir nun die Hilfsgröße <i>y</i><sub>0</sub> durch die (allgemeine) Zufallsgröße <i>y</i> am AWGN&ndash;Ausgang, so lautet das Endergebnis:
+
*If we now replace the auxiliary&nbsp; $y_0$&nbsp; with the general random&nbsp; $y$&nbsp; at the AWGN output,&nbsp; the final result is:
  
:<math>L(y \hspace{0.05cm}|\hspace{0.05cm}x) = {2  \cdot y}/{\sigma^2} =K_{\rm L} \cdot y\hspace{0.05cm}. </math>
+
::<math>L(y \hspace{0.05cm}\vert\hspace{0.05cm}x) = {2  \cdot y}/{\sigma^2} =K_{\rm L} \cdot y\hspace{0.05cm}. </math>
  
Hierbei ist <i>K</i><sub>L</sub> = 2/<i>&sigma;</i><sup>2</sup> eine Konstante, die allein von der Streuung der Gaußschen Störung abhängt.<br>
+
:Here&nbsp; $K_{\rm L} = 2/\sigma^2$&nbsp; is a constant that depends solely on the standard deviation of the Gaussian noise component.}}<br>
  
== Symbolweise Soft–in Soft–out Decodierung ==
+
== Bit-wise soft-in soft-out decoding ==
 
<br>
 
<br>
Wir gehen  nun von einem (<i>n</i>, <i>k</i>)&ndash;Blockcode aus, wobei das Codewort <u><i>x</i></u> = (<i>x</i><sub>1</sub>, ... , <i>x<sub>n</sub></i>) durch den Kanal in das Empfangswort <u><i>y</i></u> = (<i>y</i><sub>1</sub>, ... , <i>y<sub>n</sub></i>) verfälscht wird. Bei langen Codes ist eine <i>Maximum&ndash;a&ndash;posteriori&ndash;Entscheidung auf Blockebene</i> &ndash; kurz: <b>[http://en.lntwww.de/Kanalcodierung/Klassifizierung_von_Signalen#MAP.E2.80.93_und_ML.E2.80.93Kriterium_.282.29 block&ndash;wise MAP]</b> &ndash; sehr aufwändig. Man müsste unter den 2<sup><i>k</i></sup> zulässigen Codeworten <i><u>x</u></i><sub><i>j</i></sub> &#8712; <i>C</i> dasjenige mit der größten Rückschlusswahrscheinlichkeit (englisch: <i>A Posteriori Probability</i>, APP) finden. Das auszugebende Codewort <i><u>z</u></i> wäre in diesem Fall
+
We now assume an&nbsp; $(n, \ k)$&nbsp;block code where the code word &nbsp; $\underline{x} = (x_1, \ \text{...} \ , \ x_n)$ &nbsp; is falsified by the channel into the received word&nbsp; $\underline{y} = (y_1, \ \text{...} \ , \ y_n)$.
 +
[[File:EN_KC_T_4_1_S4.png|right|frame|Model of bit-wise soft-in soft-out decoding|class=fit]]
 +
 +
*For long codes,&nbsp; a&nbsp; "maximum-a-posteriori block-level decision"&nbsp; $($for short: &nbsp;<b>[[Channel_Coding/Channel_Models_and_Decision_Structures#Criteria_.C2.BBMaximum-a-posteriori.C2.AB_and_.C2.BBMaximum-Likelihood.C2.AB|$\text{ block-wise MAP}$]]</b>&nbsp; &ndash; very elaborate.
 +
 +
*One would have to find among the&nbsp; $2^k&nbsp;$ allowable code words&nbsp; $\underline{x}_j &#8712; \mathcal{C}$&nbsp; the one with the greatest a-posteriori probability&nbsp; $\rm (APP)$.
 +
 +
*The code word to output&nbsp; $\underline{z}$&nbsp; in this case would be&nbsp;
 +
:$$\underline{z} = {\rm arg} \max_{\underline{x}_{\hspace{0.03cm}j} \hspace{0.03cm} \in \hspace{0.05cm} \mathcal{C}} \hspace{0.1cm} {\rm Pr}( \underline{x}_{\hspace{0.03cm}j} |\hspace{0.05cm} \underline{y} ) \hspace{0.05cm}.$$
 +
<br clear=all>
 +
A second possibility is decoding on bit level.&nbsp; The represented bit-wise&nbsp; $\text{soft-in soft out decoder}$&nbsp; has the task to decode all code word bits&nbsp; $x_i &#8712; \{0, \, 1\}$&nbsp; according to maximum a-posteriori probability&nbsp; ${\rm Pr}(x_i | \underline{y})$.&nbsp; With the control variable&nbsp; $i = 1, \text{...} , \ n$&nbsp; holds in this case:
  
:<math>\underline{z} = {\rm arg} \max_{\underline{x}_{\hspace{0.03cm}j} \hspace{0.03cm} \in \hspace{0.05cm} \mathcal{C}} \hspace{0.1cm} {\rm Pr}( \underline{x}_{\hspace{0.03cm}j} |\hspace{0.05cm} \underline{y} ) \hspace{0.05cm}.</math>
+
*The corresponding&nbsp; log likelihood ratio  for the&nbsp; $i$th&nbsp; code bit is:
 
 
[[File:P ID2981 KC T 4 1 S3a v6.png|Modell der symbolweisen Soft–in Soft–out Decodierung|class=fit]]<br>
 
 
 
Eine zweite Möglichkeit ist die Decodierung auf Bitebene. Der dargestellte symbolweise (oder bitweise) <span style="font-weight: bold;">Soft&ndash;in Soft&ndash;out Decoder</span>  hat die Aufgabe, alle Codewortbits <i>x<sub>i</sub></i> &#8712; {0, 1} entsprechend maximaler Rückschlusswahrscheinlichkeit Pr(<i>x<sub>i</sub></i> | <u><i>y</i></u>) zu decodieren. Mit der Laufvariablen <i>i</i> = 1, ... , <i>n</i> gilt dabei:
 
 
 
*Der entsprechende <i>L</i>&ndash;Wert (englisch: <i>Log Likelihood Ratio</i>, LLR) für das <i>i</i>&ndash;te Codebit lautet:
 
  
 
::<math>L_{\rm APP} (i) = L(x_i\hspace{0.05cm}|\hspace{0.05cm}\underline{y}) =
 
::<math>L_{\rm APP} (i) = L(x_i\hspace{0.05cm}|\hspace{0.05cm}\underline{y}) =
 
{\rm ln} \hspace{0.15cm} \frac{{\rm Pr}(x_i = 0\hspace{0.05cm}|\hspace{0.05cm}\underline{y})}{{\rm Pr}(x_i = 1\hspace{0.05cm}|\hspace{0.05cm}\underline{y})}\hspace{0.05cm} . </math>
 
{\rm ln} \hspace{0.15cm} \frac{{\rm Pr}(x_i = 0\hspace{0.05cm}|\hspace{0.05cm}\underline{y})}{{\rm Pr}(x_i = 1\hspace{0.05cm}|\hspace{0.05cm}\underline{y})}\hspace{0.05cm} . </math>
  
*Der Decoder arbeitet iterativ. Bei der Initialisierung (gekennzeichnet durch den Parameter <i>I</i>&nbsp;=&nbsp;0) ist <i>L</i><sub>APP</sub>(<i>i</i>) = <i>L</i><sub>K</sub>(<i>i</i>), wobei das Kanal&ndash;LLR <i>L</i><sub>K</sub>(<i>i</i>) durch den Empfangswert <i>y<sub>i</sub></i> gegeben ist.<br>
+
*The decoder works iteratively.&nbsp; At initialization&nbsp; $($indicated in the diagram by the parameter&nbsp; $I = 0)$&nbsp; is&nbsp; $L_{\rm APP}(i) = L_{\rm K}(i)$,&nbsp; where the channel&nbsp; $($German:&nbsp; "Kanal" &nbsp; &rArr; &nbsp; "K"$)$&nbsp; log likelihood ratio&nbsp; $L_{\rm K}(i)$&nbsp; is given by the received value&nbsp; $y_i$.<br>
  
*Berechnet wird zudem  der extrinsische <i>L</i>&ndash;Wert <i>L</i><sub>E</sub>(<i>i</i>), der die gesamte Information quantifiziert, die alle anderen Bits (<i>j</i> &ne; <i>i</i>) aufgrund der Code&ndash;Eigenschaften über das betrachtete <i>i</i>&ndash;te Bit liefern.<br>
+
*In addition,&nbsp; the extrinsic&nbsp; log likelihood ratio&nbsp; $L_{\rm E}(i)$&nbsp; is calculated,&nbsp; which quantifies the total information provided by all other bits &nbsp; $(j &ne; i)$ &nbsp; due to the code properties about the considered&nbsp; $i$th bit.<br>
  
*Bei der  nächsten Iteration (ab <i>I</i> = 1) wird <i>L</i><sub>E</sub>(<i>i</i>) bei der Berechnung von <i>L</i><sub>APP</sub>(<i>i</i>) als Apriori&ndash;Information  <i>L</i><sub>A</sub>(<i>i</i>) berücksichtigt. Für das neue Aposteriori&ndash;LLR in der Iteration <i>I</i>&nbsp;+&nbsp;1 gilt somit:
+
*At the next iteration&nbsp; $(I = 1)$&nbsp; $L_{\rm E}(i)$&nbsp; is taken into account in the computation of&nbsp; $L_{\rm APP}(i)$&nbsp; as a-priori information&nbsp; $L_{\rm A}(i)$.&nbsp; Thus,&nbsp; for the new a-posteriori log likelihood ratio in iteration&nbsp; $I + 1$&nbsp; holds:
  
 
::<math>L_{\hspace{0.1cm}\rm APP}^{(I+1)} (i) = L_{\hspace{0.1cm}\rm APP}^{(I)} (i) + L_{\hspace{0.1cm}\rm A}^{(I+1)} (i) = L_{\hspace{0.1cm}\rm APP}^{(I)} (i) + L_{\hspace{0.1cm}\rm E}^{(I)} (i)\hspace{0.05cm} .  </math>
 
::<math>L_{\hspace{0.1cm}\rm APP}^{(I+1)} (i) = L_{\hspace{0.1cm}\rm APP}^{(I)} (i) + L_{\hspace{0.1cm}\rm A}^{(I+1)} (i) = L_{\hspace{0.1cm}\rm APP}^{(I)} (i) + L_{\hspace{0.1cm}\rm E}^{(I)} (i)\hspace{0.05cm} .  </math>
  
*Die Iterationen werden fortgesetzt, bis alle |<i>L</i><sub>APP</sub>(<i>i</i>)| größer sind als ein vorzugebender Wert. Das wahrscheinlichste Codewort <i><u>z</u></i> ergibt sich dann aus den Vorzeichen aller <i>L</i><sub>APP</sub>(<i>i</i>), mit <i>i</i>&nbsp;=&nbsp;1,&nbsp;... ,&nbsp;<i>n</i>.<br>
+
*The iterations continue until all absolute values&nbsp; $|L_{\rm APP}(i)|$&nbsp; are greater than a given value.&nbsp; The most likely code word&nbsp; $\underline{z}$&nbsp; is then obtained from the signs of all&nbsp; $L_{\rm APP}(i)$,&nbsp; with&nbsp; $i = 1, \ \text{...} , \ n$.<br>
  
*Bei einem [http://en.lntwww.de/Kanalcodierung/Allgemeine_Beschreibung_linearer_Blockcodes#Systematische_Codes_.282.29 systematischen Code] geben die ersten <i>k</i> Bit von <u><i>z</i></u> das gesuchte Informationswort an, das mit großer Wahrscheinlichkeit mit der gesendeten  Nachricht <u><i>u</i></u> übereinstimmen wird.<br><br>
+
*For a&nbsp; [[Channel_Coding/General_Description_of_Linear_Block_Codes#Systematic_Codes| $\text{systematic code}$]]&nbsp; the first&nbsp; $k$&nbsp; bits of&nbsp; $\underline{z}$&nbsp; indicate the information word&nbsp; $\underline{\upsilon}$&nbsp; being searched for,&nbsp; which will most likely match the message&nbsp; $\underline{u}$&nbsp; being sent.<br><br>
  
Diese Beschreibung des SISO&ndash;Decodierers nach [Bos98]<ref>Bossert, M.: ''Kanalcodierung.'' Stuttgart: B. G. Teubner, 1998.</ref> soll in erster Linie die unterschiedlichen <i>L</i>&ndash;Werte verdeutlichen. Das große Potential der symbolweisen Decodierung erkennt man allerdings erst im Zusammenhang mit [http://en.lntwww.de/Kanalcodierung/Soft%E2%80%93in_Soft%E2%80%93out_Decoder#Grundstruktur_von_verketteten_Codiersystemen verketteten Codiersystemen.]<br>
+
This SISO decoder descriptione from&nbsp; [Bos99]<ref name='Bos99'>Bossert, M.:&nbsp; Channel Coding for Telecommunications.&nbsp; Chichester:&nbsp; Wiley,&nbsp; 1999.&nbsp; ISBN:&nbsp; 978-0-471-98277-7.</ref>&nbsp; is intended at this point primarily to clarify the different&nbsp; $L$&ndash;values.&nbsp; One recognizes the large potential of the bit-wise decoding only in connection with&nbsp; [[Channel_Coding/Soft-in_Soft-Out_Decoder#Basic_structure_of_concatenated_coding_systems|$\text{concatenated coding systems}$]].<br>
  
== Zur Berechnung der extrinsischen L–Werte (1) ==
+
== Calculation of extrinsic log likelihood ratios ==
 
<br>
 
<br>
Die Schwierigkeit bei der symbolweisen iterativen Decodierung ist im allgemeinen die Bereitstellung der extrinsischen <i>L</i>&ndash;Werte <i>L</i><sub>E</sub>(<i>i</i>). Bei einem Code der Länge <i>n</i> gilt hierbei für die Laufvariable: <i>i</i> = 1, ... , <i>n</i>.<br>
+
The difficulty in bit-wise&nbsp; $($or symbol-wise$)$&nbsp; iterative decoding is generally the provision of the extrinsic&nbsp; log likelihood ratios &nbsp; $L_{\rm E}(i)$.&nbsp; For a code of length&nbsp; $n$,&nbsp; the control variable is &nbsp; $i = 1, \ \text{...} , \ n$.<br>
  
{{Definition}}''':''' Der <b> extrinsische <i>L</i>&ndash;Wert</b> (englisch: <i>extrinsic LLR</i>) ist ein Maß für die Informationen, den die anderen Symbole (<i>j</i> &ne; <i>i</i>) des Codewortes über das <i>i</i>&ndash;te Codesymbol liefern, ausgedrückt als Log&ndash;Likelihood&ndash;Verhältnis. Wir benennen diesen Kennwert mit <i>L</i><sub>E</sub>(<i>i</i>).{{end}}<br>
+
{{BlaueBox|TEXT= 
 +
$\text{Definition:}$&nbsp;  The &nbsp; <b> extrinsic log likelihood ratio</b> &nbsp; is a measure of the information that the other symbols&nbsp; $(j &ne; i)$&nbsp; of the code word provide about the&nbsp; $i$&ndash;th encoded symbol,&nbsp; expressed as a log likelihood ratio.&nbsp; We denote this characteristic value by&nbsp; $L_{\rm E}(i)$.}}
  
Wir berechnen nun die extrinsischen <i>L</i>&ndash;Werte <i>L</i><sub>E</sub>(<i>i</i>) für zwei beispielhafte Codes.<br><br>
 
  
<b>Repetition Code &#8658; RC (<i>n</i>, 1, <i>n</i>)</b><br>
+
We now calculate the extrinsic log likelihood ratios&nbsp; $L_{\rm E}(i)$&nbsp; for two example codes.<br>
  
Ein Wiederholungscode zeichnet sich dadurch aus, dass alle <i>n</i> Codesymbole <i>x<sub>i</sub></i> &#8712; {0, 1} identisch sind. Der extrinsische <i>L</i>&ndash;Wert für das <i>i</i>&ndash;ten Symbol ist hier sehr einfach anzugeben und lautet:
+
$\text{(A) &nbsp;Repetition Code}$ &nbsp; &#8658; &nbsp; ${\rm RC} \ (n, 1, n)$</b><br>
  
:<math>L_{\rm E}(i) = \hspace{0.05cm}\sum_{j \ne i} \hspace{0.1cm} L_j  
+
A repetition code&nbsp; $\rm (RC)$&nbsp;is characterized by the fact that all&nbsp; $n$&nbsp; encoded symbols&nbsp; $x_i &#8712; \{0, \, 1\}$&nbsp; are identical.
\hspace{0.3cm}{\rm mit}\hspace{0.3cm}L_j = L_{\rm APP}(j)
+
 
 +
*Here,&nbsp; the extrinsic log likelihood ratio for the&nbsp; $i$&ndash;th symbol is very easy to specify:
 +
 
 +
::<math>L_{\rm E}(i) = \hspace{0.05cm}\sum_{j \ne i} \hspace{0.1cm} L_j  
 +
\hspace{0.3cm}{\rm with}\hspace{0.3cm}L_j = L_{\rm APP}(j)
 
\hspace{0.05cm}.</math>
 
\hspace{0.05cm}.</math>
 +
*If the sum over all&nbsp; $L_{j &ne; i}$&nbsp; is positive &nbsp; &rArr; &nbsp; this implies from the point of view of the other log likelihood ratio values a preference for the decision &nbsp; $x_i = 0$.
 +
 +
*On the other hand,&nbsp; if the sum is negative &nbsp; &rArr; &nbsp; $x_i = 1$&nbsp; is more likely.
 +
 +
*$L_{\rm E}(i) = 0$&nbsp; does not allow a prediction.<br>
 +
 +
 +
{{GraueBox|TEXT= 
 +
$\text{Example 5:}$&nbsp; We consider the decoding of the repetition code&nbsp; $\text{RC (4, 1,4)}$.&nbsp;  Thereby,&nbsp; we make three different assumptions for the log likelihood ratio&nbsp; $ \underline{L}_{\rm APP}^{(I=0)}=\underline{L}_{\rm K}.$
 +
 +
[[File:EN_KC_T_4_1_S3a.png|right|frame|Decoding example &nbsp;$\rm (5a)$&nbsp; for the&nbsp; ${\rm RC} \ (4, 1, 4)$|class=fit]]
 +
$\text{Decoding example (5a):}$
 +
:$$\underline{L}_{\rm APP}^{(I=0)} = (+1,  -1, +3, -1)\text{:}$$ 
 +
::$$L_{\rm E}(1) = -1+3-1 = +1\hspace{0.05cm},$$
 +
::$$L_{\rm E}(2) = +1+3-1 = +3\hspace{0.05cm},$$
 +
::$$L_{\rm E}(3) = +1-1 -1= -1\hspace{0.05cm},$$
 +
::$$L_{\rm E}(4) = +1-1 +3 = +3\hspace{0.05cm}$$
 +
:$$\hspace{0.3cm}\Rightarrow\hspace{0.3cm}\underline{L}_{\rm E}^{(I=0)}= (+1, +3, -1, +3)
 +
\hspace{0.3cm}\Rightarrow\hspace{0.3cm}\underline{L}_{\rm APP}^{(I=1)}=\underline{L}_{\rm APP}^{(I=0)}+ \underline{L}_{\rm E}^{(I=0)}= (+2, +2, +2, +2)$$
 +
*At the beginning&nbsp; $(I=0)$&nbsp; the positive&nbsp; $L_{\rm E}$&nbsp; values indicate &nbsp;$x_1 = x_2 = x_4 = 0$&nbsp; while &nbsp;$x_3 =1$&nbsp; is more likely &nbsp; $($because of negative&nbsp; $L_{\rm E})$ .
 +
 +
*Already after one iteration&nbsp; $(I=1)$&nbsp; all&nbsp; $L_{\rm APP}$&nbsp; values are positive &nbsp; &#8658; &nbsp; the information bit is decoded as&nbsp; $u = 0$.
 +
 +
*Further iterations yield nothing.<br>
 +
 +
 +
$\text{Decoding example (5b):}$
 +
[[File:EN_KC_T_4_1_S3ab.png|right|frame|Decoding example &nbsp;$\rm (5b)$&nbsp; for the&nbsp; ${\rm RC} \ (4, 1, 4)$]]
 +
 +
:$$\underline{L}_{\rm APP}^{(I=0)} = (+1,  +1, -4, +1)\text{:}$$
 +
:$$\underline{L}_{\rm E}^{(I=0)} = \ (-2,  -2, +3, -2)$$
 +
 +
*Although three signs were wrong at the beginning,&nbsp; after two iterations all&nbsp; $L_{\rm APP}$&nbsp; values are negative.
  
Ist die Summe über alle <i>L</i><sub><i>j</i>&ne;<i>i</i></sub> positiv, so bedeutet dies aus Sicht der anderen <i>L</i>&ndash;Werte eine Präferenz für &bdquo;<i>x<sub>i</sub></i> = 0&rdquo;. Bei negativer Summe ist &bdquo;<i>x<sub>i</sub></i> = 1&rdquo; wahrscheinlicher. <i>L</i><sub>E</sub>(<i>i</i>) = 0 erlaubt keine Vorhersage.<br>
+
*The information bit is decoded as&nbsp; $u = 1$.  
  
{{Beispiel}}''':''' Wir betrachten die Decodierung eines Wiederholungscodes (<i>n</i> = 4), wobei gelten soll:
 
*<u><i>L</i></u> = (+1, &ndash;1, +3, &ndash;1) &nbsp;&#8658;&nbsp;  <u><i>L</i></u><sub>E</sub> = (+1, +3, &ndash;1, +3): <i>L</i><sub>E</sub>(1) ist positiv, obwohl zwei der anderen
 
<i>L</i>&ndash;Werte (Mehrheit) negativ sind &nbsp;&#8658;&nbsp;  Präferenz für &bdquo;<i>x</i><sub>1</sub> = 0&rdquo;. Alle Aposteriori&ndash;<i>L</i>&ndash;Werte sind
 
nach einer Iteration positiv &nbsp;&#8658;&nbsp; Informationsbit ist <i>u</i> = 0. Weitere Iterationen bringen nichts.<br>
 
  
[[File:P ID3094 KC T 4 4 S3a neu v2.png|Decodierbeispiel 1 für den RC (3, 1)|class=fit]]<br>
 
  
*<u><i>L</i></u> = (+1, +1, &ndash;4, +1) &nbsp;&#8658;&nbsp;  <u><i>L</i></u><sub>E</sub> = (&ndash;2, &ndash;2, +3, &ndash;2): Alle Aposteriori&ndash;<i>L</i>&ndash;Werte sind
+
[[File:EN_KC_T_4_1_S3c.png|right|frame|Decoding example &nbsp;$\rm (5c)$&nbsp; for the&nbsp; ${\rm RC} \ (4, 1, 4)$]]<br>
nach zwei Iterationen negativ &nbsp;&#8658;&nbsp; Informationsbit ist <i>u</i> = 1 &nbsp;&#8658;&nbsp; zu Beginn waren drei Vorzeichen falsch.<br>
+
$\text{Decoding example (5c):}$
 +
:$$\underline{L}_{\rm APP}^{(I=0)} = (+1, +1, -3, +1)\text{:}$$
 +
:$$\underline{L}_{\rm E}^{(I=0)} = (-1, -1, +3, -1)$$
  
[[File:P ID3096 KC T 4 4 S3c neu v1.png|Decodierbeispiel 2 für den RC (3, 1)|class=fit]]<br>
+
*All&nbsp; $L_{\rm APP}$&nbsp; values are zero already after one iteration.
  
*<u><i>L</i></u> = (+1, +1, &ndash;3, +1) &nbsp;&#8658;&nbsp; <u><i>L</i></u><sub>E</sub> = (&ndash;1, &ndash;1, +3, &ndash;1): Alle Aposteriori&ndash;<i>L</i>&ndash;Werte sind
+
*The information bit cannot be decoded,&nbsp; although the initial situation was not much worse than with&nbsp; $\rm (5b)$.
nach einer Iteration 0 &nbsp;&#8658;&nbsp; Informationsbit kann nicht decodiert werden. Weitere Iterationen bringen nichts.<br>
+
 +
*Further iterations bring nothing here.<br>}}
  
[[File:P ID3095 KC T 4 4 S3b neu v1.png|Decodierbeispiel 3 für den RC (3, 1)|class=fit]]{{end}}<br>
 
  
== Zur Berechnung der extrinsischen L–Werte (2) ==
+
 
 +
$\text{(B) &nbsp; Single parity&ndash;check code}$ &nbsp; &#8658; &nbsp; ${\rm SPC} \ (n, \ n \, -1, \ 2)$
 
<br>
 
<br>
<b>Single Parity&ndash;check Code &#8658; SPC (<i>n</i>, <i>n</i> &ndash;1, 2)
 
</b><br>
 
  
Bei jedem <i>Single Parity&ndash;check Code</i> ist die Anzahl der Einsen in jedem Codewort geradzahlig. Oder anders ausgedrückt: Für jedes Codewort <u><i>x</i></u> ist das [http://en.lntwww.de/Kanalcodierung/Zielsetzung_der_Kanalcodierung#Einige_wichtige_Definitionen_zur_Blockcodierung Hamming&ndash;Gewicht] <i>w</i><sub>H</sub>(<u><i>x</i></u>) geradzahlig.<br>
+
For each single parity&ndash;check code,&nbsp; the number of&nbsp;  "ones"&nbsp;  in each code word is even.&nbsp;  Or,&nbsp;  put another way: &nbsp; For each code word&nbsp; $\underline{x}$&nbsp; the&nbsp; [[Channel_Coding/Objective_of_Channel_Coding#Important_definitions_for_block_coding|$\text{Hamming weight}$]]&nbsp; $w_{\rm H}(\underline{x})$&nbsp; is even.<br>
  
Das Codewort <u><i>x</i></u><sup>(&ndash;<i>i</i>)</sup> beinhalte alle Symbole mit Ausnahme von <i>x<sub>i</sub></i> &nbsp;&#8658;&nbsp; Vektor der Länge <i>n</i> &ndash; 1. Damit lautet der extrinsische <i>L</i>&ndash;Wert bezüglich dieses <i>i</i>&ndash;ten Symbols, wenn <u><i>x</i></u> empfangen wurde:
+
{{BlaueBox|TEXT= 
 +
$\text{Definition:}$&nbsp; Let the code word&nbsp; $\underline{x}^{(&ndash;i)}$&nbsp; contain all symbols except&nbsp; $x_i$ &nbsp; &#8658; &nbsp; vector of length&nbsp; $n&ndash;1$.&nbsp;  Thus,&nbsp;  the &nbsp; "extrinsic log likelihood ratio"&nbsp; will be with respect to the&nbsp; $i$&ndash;th symbol, when&nbsp; $\underline{x}$&nbsp; has been received:
  
:<math>L_{\rm E}(i) = {\rm ln} \hspace{0.15cm}\frac{{\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm ist \hspace{0.15cm} gerade} \hspace{0.05cm} | \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}{{\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm ist \hspace{0.15cm} ungerade} \hspace{0.05cm} | \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}
+
::<math>L_{\rm E}(i) = {\rm ln} \hspace{0.15cm}\frac{ {\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm is \hspace{0.15cm} even} \hspace{0.05cm} \vert\hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}{ {\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm is \hspace{0.15cm} odd} \hspace{0.05cm} \vert \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}
 
  \hspace{0.05cm}.</math>
 
  \hspace{0.05cm}.</math>
  
Wie in der Aufgabe A4.4 gezeigt werden soll, kann hierfür auch geschrieben werden:
+
*As will be shown in the&nbsp; [[Aufgaben:Aufgabe_4.4:_Extrinsische_L–Werte_beim_SPC|"Exercise 4.4"]],&nbsp; it is also possible to write for this:
 +
 
 +
::<math>L_{\rm E}(i) = 2 \cdot {\rm tanh}^{-1} \hspace{0.1cm} \left [  \prod\limits_{j \ne i}^{n} \hspace{0.15cm}{\rm tanh}(L_j/2) \right ]
 +
\hspace{0.3cm}{\rm with}\hspace{0.3cm}L_j = L_{\rm APP}(j)
 +
\hspace{0.05cm}.</math>}}
  
:<math>L_{\rm E}(i) = 2 \cdot {\rm tanh}^{-1} \hspace{0.1cm} \left [  \prod\limits_{j \ne i}^{n} \hspace{0.15cm}{\rm tanh}(L_j/2) \right ]
 
\hspace{0.3cm}{\rm mit}\hspace{0.3cm}L_j = L_{\rm APP}(j)
 
\hspace{0.05cm}.</math>
 
  
{{Beispiel}}''':''' Wir gehen vom <i>Single Parity&ndash;check Code</i> mit <i>n</i> = 3, <i>k</i> = 2 &nbsp;&#8658;&nbsp; kurz  <b>SPC (3, 2, 2)</b> aus. Die 2<sup><i>k</i></sup> = 4 gültigen Codeworte <u><i>x</i></u> = {<i>x</i><sub>1</sub>, <i>x</i><sub>2</sub>, <i>x</i><sub>3</sub>} lauten bei bipolarer Beschreibung &nbsp;&#8658;&nbsp; <i>x<sub>i</sub></i> &#8712; {&plusmn;1}:
+
{{GraueBox|TEXT= 
 +
$\text{Example 6:}$&nbsp; We assume a&nbsp; "Single parity&ndash;check code"&nbsp; with&nbsp; $n = 3, \ k = 2$ &nbsp; &#8658; &nbsp; briefly&nbsp; ${\rm SPC} \ (3, \ 2, \ 2)$.
 +
 +
*The &nbsp; $2^k = 4$ &nbsp; valid code words&nbsp; $\underline{x} = \{x_1,\ x_2,\ x_3\}$&nbsp; are for bipolar description &nbsp; &#8658; &nbsp; $x_i &#8712; \{&plusmn;1\}$:
 +
[[File:EN_KC_T_4_1_S3e.png|right|frame|Decoding example for the&nbsp; ${\rm SPC} \ (3, 2, 2)$|class=fit]]
  
:<math> \underline{x}_0 \hspace{-0.05cm}=\hspace{-0.05cm} (+1\hspace{-0.03cm},\hspace{-0.05cm}+1\hspace{-0.03cm},\hspace{-0.05cm}+1)\hspace{-0.05cm}, \hspace{0.3cm}
+
:$$ \underline{x}_0 \hspace{-0.05cm}=\hspace{-0.05cm} (+1\hspace{-0.03cm},\hspace{-0.05cm}+1\hspace{-0.03cm},\hspace{-0.05cm}+1)\hspace{-0.05cm}, $$
\underline{x}_1  \hspace{-0.05cm}=\hspace{-0.05cm} (+1\hspace{-0.03cm},\hspace{-0.05cm} -1\hspace{-0.03cm},\hspace{-0.05cm} -1)\hspace{-0.05cm}, \hspace{0.3cm}
+
:$$\underline{x}_1  \hspace{-0.05cm}=\hspace{-0.05cm} (+1\hspace{-0.03cm},\hspace{-0.05cm} -1\hspace{-0.03cm},\hspace{-0.05cm} -1)\hspace{-0.05cm}, $$
\underline{x}_2  \hspace{-0.05cm}=\hspace{-0.05cm} (-1\hspace{-0.03cm},\hspace{-0.05cm} +1\hspace{-0.03cm},\hspace{-0.05cm} -1)\hspace{-0.05cm}, \hspace{0.3cm}
+
:$$\underline{x}_2  \hspace{-0.05cm}=\hspace{-0.05cm} (-1\hspace{-0.03cm},\hspace{-0.05cm} +1\hspace{-0.03cm},\hspace{-0.05cm} -1)\hspace{-0.05cm}, $$
\underline{x}_3  \hspace{-0.05cm}=\hspace{-0.05cm} (-1\hspace{-0.03cm},\hspace{-0.05cm} -1\hspace{-0.03cm},\hspace{-0.05cm} +1)\hspace{-0.05cm}. </math>
+
:$$\underline{x}_3  \hspace{-0.05cm}=\hspace{-0.05cm} (-1\hspace{-0.03cm},\hspace{-0.05cm} -1\hspace{-0.03cm},\hspace{-0.05cm} +1)\hspace{-0.05cm}. $$
  
Bei diesem Code ist also das Produkt <i>x</i><sub>1</sub>&nbsp;&middot;&nbsp;<i>x</i><sub>2</sub>&nbsp;&middot;&nbsp;<i>x</i><sub>3</sub> stets positiv.<br>
+
:So,&nbsp;the product &nbsp; $x_1 \cdot x_2 \cdot x_3$ &nbsp; is always positive.<br>
  
[[File:P ID3098 KC T 4 1 S3e v1.png|Decodierbeispiel für den SPC (3, 2)|class=fit]]<br>
+
*The table shows the decoding process for&nbsp; $\underline{L}_{\rm APP} = (+2.0, +0.4, \, &ndash;1.6)$.
  
Die Tabelle zeigt den Decodiervorgang für <u><i>L</i></u><sub>APP</sub> = (+2.0, +0.4, &ndash;1.6). Die harte Entscheidung nach den Vorzeichen von <i>L</i><sub>APP</sub>(<i>i</i>) ergäbe hier (+1, +1, &ndash;1), also kein gültiges Codewort des SP(3, 2, 2).<br>
+
*The hard decision according to the signs of&nbsp; $L_{\rm APP}(i)$&nbsp; would yield here&nbsp; $(+1, +1, \, -1)$ &nbsp; &rArr; &nbsp; no valid code word of&nbsp; ${\rm SP}(3, \ 2, \ 2)$.<br>
  
Rechts sind in der Tabelle die dazugehörigen extrinsischen <i>L</i>&ndash;Werte eingetragen:<br>
+
*On the right side of the table, the corresponding extrinsic log likelihood ratios are entered:<br>
  
:<math>L_{\rm E}(1) \hspace{-0.15cm}  = \hspace{-0.15cm}  2 \cdot {\rm tanh}^{-1} \hspace{0.05cm}
+
::<math>L_{\rm E}(1) = 2 \cdot {\rm tanh}^{-1} \hspace{0.05cm}
 
  \left [  {\rm tanh} (0.2) \cdot {\rm tanh} (-0.8)\hspace{0.05cm}\right ] = -0.131\hspace{0.05cm}, </math>
 
  \left [  {\rm tanh} (0.2) \cdot {\rm tanh} (-0.8)\hspace{0.05cm}\right ] = -0.131\hspace{0.05cm}, </math>
:<math>L_{\rm E}(2) \hspace{-0.15cm}  = \hspace{-0.15cm}2 \cdot {\rm tanh}^{-1} \hspace{0.05cm}
+
::<math>L_{\rm E}(2) =2 \cdot {\rm tanh}^{-1} \hspace{0.05cm}
 
  \left [  {\rm tanh} (1.0) \cdot {\rm tanh} (-0.8)\hspace{0.05cm}\right ] = -0.518\hspace{0.05cm}, </math>
 
  \left [  {\rm tanh} (1.0) \cdot {\rm tanh} (-0.8)\hspace{0.05cm}\right ] = -0.518\hspace{0.05cm}, </math>
:<math>L_{\rm E}(3) \hspace{-0.15cm}  = \hspace{-0.15cm}2 \cdot {\rm tanh}^{-1} \hspace{0.05cm}
+
::<math>L_{\rm E}(3) =2 \cdot {\rm tanh}^{-1} \hspace{0.05cm}
 
  \left [  {\rm tanh} (1.0) \cdot {\rm tanh} (0.2)\hspace{0.05cm}\right ] = +0.151\hspace{0.05cm}.</math>
 
  \left [  {\rm tanh} (1.0) \cdot {\rm tanh} (0.2)\hspace{0.05cm}\right ] = +0.151\hspace{0.05cm}.</math>
  
Die zweite Gleichung lässt sich wie folgt interpretieren:
+
The second equation can be interpreted as follows:
*<i>L</i><sub>APP</sub>(1) = +2.0 und <i>L</i><sub>APP</sub>(3) = &ndash;1.6 sagen aus, dass Bit 1 eher &bdquo;+1&rdquo; als &bdquo;&ndash;1&rdquo; ist  und Bit 3 eher &bdquo;&ndash;1&rdquo; als &bdquo;+1&rdquo;. Die Zuverlässigkeit (Betrag) ist beim ersten Bit etwas größer als beim dritten.<br>
+
# $L_{\rm APP}(1) = +2.0$&nbsp; and&nbsp; $L_{\rm APP}(3) = \, -1.6$&nbsp; state that the first bit is more likely to be &nbsp;"$+1$"&nbsp; than &nbsp;"$-1$"&nbsp; and the third bit is more likely to be &nbsp;"$-1"$&nbsp; than &nbsp;"$+1$".
 +
#The reliability&nbsp; $($the amount$)$&nbsp; is slightly greater for the first bit than for the third.<br>
 +
#The extrinsic information &nbsp; $L_{\rm E}(2) = \, -0.518$&nbsp; considers only the information of bit 1 and bit 3 about bit 2.  
 +
#From their point of view,&nbsp; the second bit is &nbsp;"$-1$"&nbsp; with reliability&nbsp; $0.518$.<br>
 +
#The log likelihood ratio&nbsp;  $L_{\rm APP}^{I=0}(2) = +0.4$&nbsp; derived from the received value&nbsp; $y_2$&nbsp; has suggested &nbsp;"$+1$"&nbsp; for the second bit.
 +
#But the discrepancy is resolved here already in the iteration&nbsp; $I = 1$.&nbsp; The decision falls here for the code word&nbsp; $\underline{x}_1  \hspace{-0.05cm}=\hspace{-0.05cm} (+1\hspace{-0.03cm},\hspace{-0.05cm} -1\hspace{-0.03cm},\hspace{-0.05cm} -1)\hspace{-0.05cm},$.
 +
#For &nbsp; $0.518 < L_{\rm APP}(2) < 1.6$ &nbsp; the result&nbsp; $\underline{x}_1$&nbsp; would only be available after several iterations.&nbsp; For&nbsp; $L_{\rm APP}(2) > 1.6$&nbsp; on the other hand,&nbsp; the decoder returns&nbsp; $\underline{x}_0$.}}<br>
  
*Die extrinsische Information <i>L</i><sub>E</sub>(2) = &ndash;0.518 berücksichtigt nur die Informationen von Bit 1 und Bit 3 über Bit 2. Aus deren Sicht ist das zweite Bit eine &bdquo;&ndash;1&rdquo; mit der Zuverlässigkeit 0.518.<br>
+
== BCJR decoding: Forward-backward algorithm ==
 +
<br>
 +
An example of iterative decoding of convolutional codes is the&nbsp; &raquo;<b>BCJR algorithm</b>&laquo;,&nbsp; named after its inventors L. R. Bahl, J. Cocke, F. Jelinek, and J. Raviv &nbsp; &#8658; &nbsp; [BCJR74]<ref name='BCJR74'>Bahl, L.R.; Cocke, J.; Jelinek, F.; Raviv, J.:&nbsp; Optimal Decoding of Linear Codes for Minimizing Symbol Error Rate.&nbsp; In: IEEE Transactions on Information Theory, Vol. IT-20, pp. 284-287, 1974.</ref>.&nbsp; The algorithm has many parallels to the seven-year older Viterbi decoding,&nbsp; but also some significant differences:
 +
*While Viterbi estimates the total sequence &nbsp; &#8658; &nbsp; [[Channel_Coding/Channel_Models_and_Decision_Structures#Definitions_of_the_different_optimal_receivers|$\text{block&ndash;wise maximum likelihood}$]],&nbsp; the BCJR&ndash; Algorithm minimizes the bit error probability &nbsp; &#8658; &nbsp; [[Channel_Coding/Channel_Models_and_Decision_Structures#Definitions_of_the_different_optimal_receivers|$\text{bit&ndash;wise MAP}$]].<br>
  
*Der vom Empfangswert <i>y</i><sub>2</sub> abgeleitete <i>L</i>&ndash;Wert &nbsp;&#8658;&nbsp; <i>L</i><sub>APP</sub>(2) = +0.4 hat für das zweite Bit eine &bdquo;+1&rdquo; vermuten lassen. Die Diskrepanz wird hier bereits in der Iteration <i>I</i> = 1 aufgelöst.<br>
+
*The Viterbi algorithm&nbsp; $($in its original form$)$&nbsp; cannot process soft information.&nbsp; In contrast,&nbsp; the BCJR algorithm specifies a reliability value for each individual bit at each iteration,&nbsp; which is taken into account in subsequent iterations.<br>
  
*Entschieden wird hier für das Codewort <u><i>x</i></u><sub>1</sub>. Bei 0.518 < <i>L</i><sub>APP</sub>(2) < 1.6 würde das Ergebnis <u><i>x</i></u><sub>1</sub> erst nach mehreren Iterationen vorliegen. Für <i>L</i><sub>APP</sub>(2) > 1.6 liefert der Decoder dagegen <u><i>x</i></u><sub>0</sub>.{{end}}<br>
 
  
== BCJR–Decodierung: Vorwärts–Rückwärts–Algorithmus ==
+
The figure is intended  to illustrate&nbsp; $($in an almost inadmissibly simplified  way$)$&nbsp; the different approach of Viterbi and BCJR algorithm.&nbsp; This is based on a convolutional code with memory&nbsp; $m = 1$&nbsp; and length&nbsp; $L = 4$ &nbsp; &#8658; &nbsp; total length $($with termination$)$&nbsp; $L' = 5$.
<br>
+
[[File:EN_KC_T_4_1_S5.png|right|frame|Comparison of Viterbi and BCJR algorithm|class=fit]]
Ein Beispiel für die iterative Decodierung von Faltungscodes ist der <i>BCJR&ndash;Algorithmus</i>, benannt nach dessen Erfindern L. R. Bahl, J. Cocke, F. Jelinek und J. Raviv &nbsp;&#8658;&nbsp; [BCJR74]<ref>Bahl, L.R.; Cocke, J.; Jelinek, F.; Raviv, J.: ''Optimal Decoding of Linear Codes for Minimizing Symbol Error Rate.'' In: IEEE Transactions on Information Theory, Vol. IT-20, S. 284-287, 1974.</ref>. Der Algorithmus weist viele Parallelen zur sieben Jahren älteren Viterbi&ndash;Decodierung auf, doch auch signifikante Unterschiede:
 
*Während Viterbi die Gesamtsequenz schätzt &nbsp;&#8658;&nbsp; [http://en.lntwww.de/Kanalcodierung/Klassifizierung_von_Signalen#MAP.E2.80.93_und_ML.E2.80.93Kriterium_.282.29 block&ndash;wise Maximum Likelihood,] minimiert der BCJR&ndash;Algorithmus die Bitfehlerwahrscheinlichkeit &nbsp;&#8658;&nbsp; [http://en.lntwww.de/Kanalcodierung/Klassifizierung_von_Signalen#MAP.E2.80.93_und_ML.E2.80.93Kriterium_.282.29 bit&ndash;wise MAP.]<br>
 
 
 
*Der Viterbi&ndash;Algorithmus kann (in seiner ursprünglichen Form) keine Softinformation verarbeiten. Dagegen gibt der BCJR&ndash;Algorithmus bei jeder Iteration für jedes einzelne Symbol (Bit) einen Zuverlässigkeitswert an, der bei späteren Iterationen berücksichtigt wird.<br>
 
  
[[File:P ID3024 KC T 4 1 S5 v2.png|Gegenüberstellung von Viterbi– und BCJR–Algorithmus|class=fit]]<br>
+
*Viterbi searches and finds the most likely path from&nbsp; ${\it \Gamma}_0(S_0)$&nbsp; to&nbsp; ${\it \Gamma}_5(S_0)$, namely&nbsp; $S_0 &#8594; S_1 &#8594; S_0 &#8594; S_0 &#8594; S_1&#8594; S_0 $. We refer to the sample solution to&nbsp; [[Aufgaben:Exercise_3.09Z:_Viterbi_Algorithm_again|"Exercise 3.9Z"]].<br>
  
Die Abbildung soll &ndash; fast unzulässig vereinfacht &ndash; die unterschiedliche Vorgehensweise von Viterbi&ndash;Algorithmus (links) und BCJR&ndash;Algorithmus (rechts) verdeutlichen. Zugrunde liegt ein Faltungscode mit dem Gedächtnis <i>m</i> = 1 und der Länge <i>L</i> = 4 &nbsp;&#8658;&nbsp; Gesamtlänge <i>L</i>' = 5  (mit Terminierung).
+
*The sketch for the BCJR algorithm illustrates the extraction of the extrinsic&nbsp; L&ndash;value for the third symbol &nbsp; &#8658; &nbsp; $L_{\rm E}(3)$.&nbsp; The relevant area in the trellis is shaded:
*Der Viterbi&ndash;Algorithmus sucht und findet den wahrscheinlichsten Pfad von <i>&Gamma;</i><sub>0</sub>(<i>S</i><sub>0</sub>) nach <i>&Gamma;</i><sub>5</sub>(<i>S</i><sub>0</sub>), nämlich <i>S</i><sub>0</sub> &#8594; <i>S</i><sub>1</sub> &#8594; <i>S</i><sub>0</sub> &#8594; <i>S</i><sub>0</sub> &#8594; <i>S</i><sub>1</sub>. Wir verweisen auf die Musterlösung zur Aufgabe Z3.9.<br><br>
 
  
Die Skizze für den BCJR&ndash;Algorithmus verdeutlicht die Gewinnung des extrinsischen <i>L</i>&ndash;Wertes für das dritte Symbol &nbsp;&#8658;&nbsp; <i>L</i><sub>E</sub>(3). Der relevante Bereich im Trellis ist schraffiert:
+
* Working through the trellis diagram in the forward direction,&nbsp; one obtains in the same way as Viterbi the metrics&nbsp; $\alpha_1, \ \alpha_2, \ \text{...}\hspace{0.05cm} , \ \alpha_5$.&nbsp; To calculate&nbsp; $L_{\rm E}(3)$&nbsp; one needs from this&nbsp; $\alpha_2$.<br>
*Bei der Abarbeitung des Trellisdiagramms in Vorwärtsrichtung  gewinnt man &ndash; in gleicher Weise wie bei Viterbi &ndash; die Metriken <i>&alpha;</i><sub>1</sub>, <i>&alpha;</i><sub>2</sub>, ... <i>&alpha;</i><sub>5</sub>. Zur Berechnung von <i>L</i><sub>E</sub>(3) benötigt man hiervon <i>&alpha;</i><sub>2</sub>.<br>
 
  
*Anschließend durchläuft man das Trellisdiagramm rückwärts (also von rechts nach links) und erhält damit die Metriken <i>&beta;</i><sub>4</sub>, <i>&beta;</i><sub>3</sub>, .... , <i>&beta;</i><sub>0</sub> entsprechend der unteren Skizze.<br>
+
* Then we traverse the trellis diagram backwards&nbsp; $($i.e. from right to left$)$&nbsp; and thus obtain the metrics&nbsp; $\beta_4, \ \beta_3, \ \text{...}\hspace{0.05cm} , \ \beta_0$&nbsp; corresponding to the sketch below.<br>
  
*Der gesuchte extrinsische <i>L</i>&ndash;Wert <i>L</i><sub>E</sub>(3) ergibt sich aus den Metriken <i>&alpha;</i><sub>2</sub> (in Vorwärtsrichtung) und <i>&beta;</i><sub>3</sub> (in Rückwärtsrichtung) sowie der Apriori&ndash;Information <i>&gamma;</i><sub>3</sub> über das Symbol <i>i</i> = 3.<br>
+
* The extrinsic log likelihood ratio&nbsp; $L_{\rm E}(3)$&nbsp; is obtained from the forward direction metric&nbsp; $\alpha_2$&nbsp;  and the backward direction metric&nbsp; $\beta_3$&nbspand the a-priori information&nbsp; $\gamma_3$&nbsp; about the symbol&nbsp; $i = 3$.<br>
  
== Grundstruktur von verketteten Codiersystemen ==
+
== Basic structure of concatenated coding systems ==
 
<br>
 
<br>
Die wichtigsten Kommunikationssysteme der letzten Jahre verwenden zwei unterschiedliche Kanalcodes. Man spricht dann von <span style="font-weight: bold;">verketteten Codiersystemen</span> (englisch: <i>Concatenated Codes</i>). Auch bei relativ kurzen Komponentencodes <i>C</i><sub>1</sub> und <i>C</i><sub>2</sub>  ergibt sich für den verketteten Code <i>C</i> eine hinreichend große Codewortlänge <i>n</i>, die ja bekanntlich erforderlich ist, um sich der Kanalkapazität anzunähern.<br>
+
The most important communication systems of the last years use two different channel codes.&nbsp; One speaks then of&nbsp; &raquo;'''concatenated codes'''&laquo;.&nbsp; Even with relatively short component codes&nbsp; $\mathcal{C}_1$&nbsp; and&nbsp; $\mathcal{C}_2$&nbsp; the concatenated code&nbsp; $\mathcal{C}$&nbsp; results in a sufficiently large code word length&nbsp; $n$,&nbsp; which is known to be necessary to approach the channel capacity.<br>
  
Zunächst seien einige Beispiele aus dem Mobilfunk genannt:
+
To begin with,&nbsp; here are a few examples from mobile communications:
*Bei GSM (<i>Global System for Mobile Communications</i>, zweite Mobilfunkgeneration) wird zunächst die Datenbitrate von 9.6 kbit/s auf 12 kbit/s erhöht, um auch in leitungsvermittelten Netzen eine Fehlererkennung zu ermöglichen. Anschließend folgt ein punktierter Faltungscode mit der Ausgangsbitrate 22.8 kbit/s. Die Gesamtcoderate beträgt somit etwa 42.1%.
+
*In&nbsp; [[Examples_of_Communication_Systems/General_Description_of_GSM#.23_OVERVIEW_OF_THE_THIRD_MAIN_CHAPTER_.23|$\rm GSM$]]&nbsp; $($"Global System for Mobile Communications",&nbsp; the second generation of mobile communication systems$)$,&nbsp; the data bit rate is first increased from&nbsp; $9. 6 \ \rm kbit/s$&nbsp; to&nbsp; $12 \ \rm kbit/s$&nbsp; in order to enable error detection in circuit-switched networks as well.&nbsp; This is followed by a punctured convolutional code with the output bit rate&nbsp; $22.8 \ \rm kbit/s$.&nbsp; The total code rate is thus about $0.421$.
  
*Beim 3G&ndash;Mobilfunksystem UMTS (<i>Universal Mobile Telecommunications System</i>) verwendet man je nach den Randbedingungen (guter/schlechter Kanal, wenige/viele Teilnehmer in der Zelle) einen Faltungscode oder einen [http://en.lntwww.de/Kanalcodierung/Grundlegendes_zu_den_Turbocodes#Grundstruktur_eines_Turbocodes_.281.29 Turbocode] (darunter versteht man per se die Verkettung zweier Faltungscodierer). Beim 4G&ndash;Mobilfunksystem LTE (<i>Long Term Evolution</i>) verwendet man für kurze Kontrollsignale einen Faltungscode  und für die längeren Payload-Daten einen Turbocode.<br>
+
*In the 3G cellular system&nbsp; [[Examples_of_Communication_Systems/General_Description_of_UMTS#.23_OVERVIEW_OF_THE_FOURTH_MAIN_CHAPTER_.23|$\rm UMTS$]]&nbsp; $($"Universal Mobile Telecommunications System"),&nbsp; depending on the boundary conditions&nbsp; $($good/bad channel,&nbsp; few/many subscribers in the cell$)$&nbsp; one uses a&nbsp; [[Examples_of_Communication_Systems/General_Description_of_GSM#.23_OVERVIEW_OF_THE_THIRD_MAIN_CHAPTER_.23|$\text{convolutional code}$]]&nbsp; or a&nbsp; [[Channel_Coding/The_Basics_of_Turbo_Codes#Basic_structure_of_a_turbo_code|$\text{turbo code}$]]&nbsp; $($by this one understands per se the concatenation of two convolutional encoders$)$.&nbsp;  
  
[[File:P ID2998 KC T 4 1 S6 v1.png|Parallel verkettetes Codiersystem|class=fit]]<br>
+
*In 4G mobile communication systems&nbsp; [[Examples_of_Communication_Systems/General_Description_of_UMTS#.23_OVERVIEW_OF_THE_FOURTH_MAIN_CHAPTER_.23|$\rm LTE$]]&nbsp; $($"Long Term Evolution"$)$,&nbsp; a convolutional code is used for short control signals,&nbsp; turbo code for the longer payload data.<br>
  
Die Grafik zeigt die Grundstruktur eines parallel verketteten Codiersystems. Alle Vektoren bestehen aus <i>n</i> Elementen: <u><i>L</i></u> = (<i>L</i>(1), ... , <i>L</i>(<i>n</i>)). Die Berechnung aller <i>L</i>&ndash;Werte geschieht also auf Symbolebene. Nicht dargestellt ist hier der [http://en.lntwww.de/Kanalcodierung/Grundlegendes_zu_den_Turbocodes#Zweite_Voraussetzung_f.C3.BCr_Turbocodes:_Interleaving Interleaver,] der zum Beispiel bei den Turbocodes obligatorisch ist.
 
*Die Codesequenzen <u><i>x</i></u><sub>1</sub> und <u><i>x</i></u><sub>2</sub> werden zur gemeinsamen Übertragung über den Kanal durch einen Multiplexer zum Vektor <u><i>x</i></u> zusammengefügt. Am Empfänger wird die Sequenz <u><i>y</i></u> wieder in die Einzelteile <u><i>y</i></u><sub>1</sub> und <u><i>y</i></u><sub>2</sub> zerlegt. Daraus werden die Kanal&ndash;<i>L</i>&ndash;Werte <u><i>L</i></u><sub>K,1</sub> und <u><i>L</i></u><sub>K,2</sub> gebildet.<br>
 
  
*Der symbolweise Decoder ermittelt entsprechend der vorne beschriebenen [http://en.lntwww.de/Kanalcodierung/Soft%E2%80%93in_Soft%E2%80%93out_Decoder#Zur_Berechnung_der_extrinsischen_L.E2.80.93Werte_.281.29 Vorgehensweise] die extrinsischen <i>L</i>&ndash;Werte <u><i>L</i></u><sub>E,1</sub> und <u><i>L</i></u><sub>E,2</sub>, die gleichzeitig die Apriori&ndash;Informationen <u><i>L</i></u><sub>A,2</sub> und <u><i>L</i></u><sub>A,1</sub> für den jeweils anderen Decoder darstellen. <br>
 
  
*Nach ausreichend vielen Iterationen (also dann, wenn ein Abbruchkriterium erfüllt ist) liegt am Decoderausgang der Vektor der Aposteriori&ndash;Werte &nbsp;&#8658;&nbsp; <u><i>L</i></u><sub>APP</sub> an. Im Beispiel wird willkürlich der Wert im oberen Zweig genommen. Möglich wäre aber auch der untere  <i>L</i>&ndash;Wert.<br><br>
+
The graph shows the basic structure of a parallel concatenated coding system.&nbsp; All vectors consist of&nbsp; $n$&nbsp; elements:&nbsp;
 +
:$$\underline{L} = (L(1), \ \text{...}\hspace{0.05cm} , \ L(n)).$$
 +
[[File:EN_KC_T_4_1_S6.png|right|frame|Parallel concatenated codes|class=fit]]
 +
So the calculation of all&nbsp; L&ndash;values is done on symbol level.&nbsp; Not shown here is the&nbsp; [[Channel_Coding/The_Basics_of_Turbo_Codes#Second_requirement_for_turbo_codes:_Interleaving| $\rm interleaver$]],&nbsp; which is mandatory e.g. for turbo codes.
 +
# The sequences&nbsp; $\underline{x}_1$&nbsp; and&nbsp; $\underline{x}_2$&nbsp; are combined to form the vector&nbsp; $\underline{x}$&nbsp; for joint transmission over the channel by a multiplexer.
 +
# At the receiver, the sequence&nbsp; $\underline{y}$&nbsp; is again decomposed into the parts&nbsp; $\underline{y}_1$&nbsp; and&nbsp; $\underline{y}_2$.&nbsp;  From this the channel L&ndash;values&nbsp; $\underline{L}_{\rm K,\hspace{0.05cm}1}&nbsp;$ and&nbsp; $\underline{L}_{\rm K,\hspace{0.05cm}2}$&nbsp; are formed.<br>
 +
# The symbol-wise decoder determines the extrinsic log likelihood ratios &nbsp; $\underline{L}_{\rm E,\hspace{0.05cm} 1}$&nbsp; and&nbsp; $\underline{L}_{\rm E,\hspace{0.05cm} 2}$, which are also the a-priori information&nbsp; $\underline{L}_{\rm A,\hspace{0.05cm} 2}$&nbsp; and&nbsp; $\underline{L}_{\rm A,\hspace{0.05cm} 1}$&nbsp; for the other decoder. <br>
 +
# After a sufficient number of iterations&nbsp; $($i.e. when a termination criterion is met$)$,&nbsp; the vector&nbsp; $\underline{L}_{\rm APP}$&nbsp; of a-posteriori values is available at the decoder output.  
 +
# In the example,&nbsp;  the value in the upper branch is taken arbitrarily.&nbsp;
 +
# However,&nbsp; the lower log likelihood ratio would also be possible.<br><br>
  
Das obige Modell gilt insbesondere auch für die Decodierung der Turbo&ndash;Codes gemäß [http://en.lntwww.de/Kanalcodierung/Grundlegendes_zu_den_Turbocodes#Grundstruktur_eines_Turbocodes_.281.29 Kapitel 4.3.]<br>
+
The above model also applies in particular to the decoding of turbo codes according to the chapter&nbsp; [[Channel_Coding/The_Basics_of_Turbo_Codes#Basic_structure_of_a_turbo_code|"Basic structure of a turbo code"]].<br>
  
== Aufgaben ==
+
== Exercises for the chapter ==
 
<br>
 
<br>
[[Aufgaben:4.1 Zum „Log Likelihood Ratio”|A4.1 Zum „Log Likelihood Ratio”]]
+
[[Aufgaben:Exercise_4.1:_Log_Likelihood_Ratio|Exercise 4.1: Log Likelihood Ratio]]
 +
 
 +
[[Aufgaben:Exercise_4.1Z:_Log_Likelihood_Ratio_at_the_BEC_Model|Exercise 4.1Z: Log Likelihood Ratio at the BEC Model]]
 +
 
 +
[[Aufgaben:Exercise_4.2:_Channel_Log_Likelihood_Ratio_at_AWGN|Exercise 4.2: Channel Log Likelihood Ratio at AWGN]]
  
[[Zusatzaufgaben:4.1 L–Werte des BEC–Modells]]
+
[[Aufgaben:Exercise_4.3:_Iterative_Decoding_at_the_BSC|Exercise 4.3: Iterative Decoding at the BSC]]
  
[[Aufgaben:4.2 Kanal–LLR bei AWGN|A4.2 Kanal–LLR bei AWGN]]
+
[[Aufgaben:Exercise_4.3Z:_Conversions_of_L-value_and_S-value|Exercise 4.3Z: Conversions of L-value and S-value]]
  
[[Aufgaben:4.3 Iterative Decodierung beim BSC|A4.3 Iterative Decodierung beim BSC]]
+
[[Aufgaben:Exercise_4.4:_Extrinsic_L-values_at_SPC|Exercise 4.4: Extrinsic L-values at SPC]]
  
[[Zusatzaufgaben:4.3 Umrechnung von L–Wert und S–Wert]]
+
[[Aufgaben:Exercise_4.4Z:_Supplement_to_Exercise_4.4|Exercise 4.4Z: Supplement to Exercise 4.4]]
  
[[Aufgaben:4.4 Extrinsische L–Werte beim SPC|A4.4 Extrinsische L–Werte beim SPC]]
+
[[Aufgaben:Exercise_4.5:_On_the_Extrinsic_L-values_again|Exercise 4.5: On the Extrinsic L-values again]]
  
[[Zusatzaufgaben:4.4 Ergänzung zur Aufgabe A4.4]]
+
[[Aufgaben:Exercise_4.5Z:_Tangent_Hyperbolic_and_Inverse|Exercise 4.5Z: Tangent Hyperbolic and Inverse]]
  
==Quellenverzeichnis==
+
==References==
 
<references/>
 
<references/>
  
 
{{Display}}
 
{{Display}}

Latest revision as of 18:21, 12 January 2023

# OVERVIEW OF THE FOURTH MAIN CHAPTER #


The last main chapter of the channel coding book describes  »iterative decoding techniques«  as used in most of today's  (2017)  communication systems.  This is due to the following reasons:

  • To approach the channel capacity,  one needs very long codes.
  • But for long codes,  "blockwise maximum likelihood decoding"  is almost impossible.


The decoder complexity can be significantly reduced with almost the same quality if two  $($or more$)$  short channel codes are linked together and the newly acquired  $($soft$)$  information is exchanged between the decoders at the receiver in several steps   ⇒   "iteratively".

The breakthrough in the field came in the early 1990s with the invention of the  "turbo codes"  by  $\text{Claude Berrou}$  and shortly thereafter with the rediscovery of  "low-density parity-check codes"  by  $\text{David J. C. MacKay}$  and  $\text{Radford M. Neal}$,  after the LDPC codes developed as early as 1961 by  $\text{Robert G. Gallager}$  had been forgotten in the meantime.

Specifically,  the fourth main chapter discusses:

  • a comparison of  »hard decision«  and  »soft decision«,
  • the quantification of  »reliability information«  by  »log likelihood ratios«,
  • the principle of symbol-wise  »soft-in soft-out  (SISO)«  decoding,
  • the definition of  »a-priori  L–value«,  »a-posteriori  L–value«  and  »extrinsic  L–value«,
  • the basic structure of  »serially concatenated«  resp.  »parallel concatenated«  coding systems,
  • the characteristics of  »product codes«  and their  »hard decision decoding«,
  • the basic structure,  decoding algorithm,  and performance of  »turbo codes«,
  • basic information on the  »low-density parity-check codes«  and their applications.



Hard Decision vs. Soft Decision


To introduce the topic discussed here, let us consider the following digital transmission system with coding.

  • In the following, all symbols are given in bipolar representation:
Considered digital transmission system with coding
Comparison of  "hard decision" and  "soft decision";   for all columns of this table it is assumed:
(1)   the information block   $\underline{u} = (0,\ 1)$,  bipolar representable as  $(+1, \, –1)$,
(2)   the encoded block  $\underline{x} = (0,\ 1,\ 1)$   ⇒   in bipolar representation:  $(+1, \, -1, \, -1)$.


"$0$"  →  "$+1$",  and 
"$1$"  →  "$-1$".
  • The symbol sequence  $\underline{u} = (u_1, \ u_2)$  of the digital source is assigned to the encoded sequence 
$$\underline{x} = (x_1, \ x_2, \ x_3) = (u_1, \ u_2, \ p)$$
where for the parity bit   $p = u_1 ⊕ u_2$   holds   ⇒   $\text{ SPC (3, 2, 2)}$.
  • The  $\text{AWGN channel}$  changes the symbols  $x_i ∈ \{+1, \ –1\}$  to real-valued output values  $y_i$,  for example according to  $\text{channel 4}$  in the table below:  
    • $x_1 = +1$   ⇒   $y_1 = +0.9,$
    • $x_2 = -1$   ⇒   $y_1 = +0.1,$
    • $x_3 = -1$   ⇒   $y_1 = +0.1,$
  • Decoding is done according to the criterion  $\text{»block-wise maximum-likelihood«}$  $\text{(ML)}$,  distinguishing between 
    • "hard decision"   $\rm {(ML–HD)}$,  and 
    • "soft decision"   $\rm {(ML–SD)}$ .
  • The given block diagram corresponds to  $\rm ML–HD$.  Here,  only the signs of the AWGN output values   ⇒   $y_{\rm HD, \ \it i} = {\rm sign}\left [y_{\rm SD, \ \it i}\right ]$  are evaluated for decoding.
  • With  "soft decision"  one omits the shaded block and directly evaluates the continuous value input variables  $y_{\rm SD, \ \it i}$ .


Thus,  the four columns in the table differ only by different AWGN realizations.

$\text{Definitions:}$  From the example table you can see:

  • $\text{Hard Decision:}$   The sink symbol sequence  $\underline{v}_{\rm HD}$  results from the  "hard decided channel values"  $\underline{y}_{\rm HD}$  $($blue background$)$.
        In our example,  only the constellations according to  $\text{channel 1}$  and  $\text{channel 2}$  are decoded without errors.
  • $\text{Soft Decision:}$   The sink symbol sequence  $\underline{v}_{\rm SD}$  results from the  "soft channel output values"  $\underline{y}_{\rm SD}$  $($green background$)$.
        Now,  in this example,  it is also correctly decided at  $\text{channel 3}$. 


The entries in the example table above are to be interpreted as follows:

  1. For ideal channel   ⇒   $\text{channel 1}$   ⇒   $\underline{x} = \underline{y}_{\rm SD} = \underline{y}_{\rm HD}$  there is no difference between the (blue) conventional hard decision variant  $\rm {(ML–HD)}$  and the (green) soft decision variant  $\rm {(ML–SD)}$.
  2. The setting according to  $\text{channel 2}$  demonstrates low signal distortions.  Because of  $\underline{y}_{\rm HD} = \underline{x}$  $($which means that the channel does not distort the signs$)$  also  $\rm ML–HD$  gives the correct result  $\underline{v}_{\rm HD} = \underline{u}$.
  3. At  $\text{channel 3}$  there is  $\underline{y}_{\rm HD} ≠ \underline{x}$  and there is also no  ${\rm SPC} \ (3, 2)$  assignment  $\underline{u}$   ⇒   $\underline{y}_{\rm HD}$.  The maximum Likelihood decoder reports here by outputting  $\underline{v}_{\rm HD} = \rm (E, \ E)$ that it failed in decoding this block.  "$\rm E$" stands for an  $\text{Erasure}$ .
  4. Also the  soft decision decoder recognizes that decoding based on the signs does not work.  Based on the  $\underline{y}_{\rm SD}$  values,  however,  it recognizes that with high probability the second bit has been falsified and decides to use the correct symbol sequence  $\underline{v}_{\rm SD} = (+1, \, -1) = \underline{u}$.
  5. In  $\text{channel 4}$,  both the signs of bit 2 and bit 3 are changed by the AWGN channel,  leading to the result   $\underline{v}_{\rm HD} = (+1, +1) ≠ \underline{u}(+1, \, -1)$   ⇒   a block error and a bit error at the same time.  Also the soft decision decoder gives the same wrong result here.

The decoding variant  $\rm ML–SD$  also offers the advantage over  $\rm ML–HD$  that it is relatively easy to assign a reliability value to each decoding result  $($but in the above table this is not specified$)$.  This reliability value would

  • have its maximum value for  $\text{channel 1}$ ,
  • be significantly smaller for  $\text{channel 2}$ ,
  • be close to zero for  $\text{channel 3}$  and  $\text{channel 4}$.

Reliability information - Log likelihood ratio


Let be  $x ∈ \{+1, \, -1\}$  a binary random variable with probabilities  ${\rm Pr}(x = +1)$  and  ${\rm Pr}(x = \, -1)$.  For coding theory,  it proves convenient in terms of computation times to use the natural logarithm of the quotient instead of the probabilities  ${\rm Pr}(x = ±1)$  .

$\text{Definition:}$  The  log likelihood ratio  $\rm (LLR)$  of the random variable  $x ∈ \{+1, \, -1\}$  is:

\[L(x)={\rm ln} \hspace{0.15cm} \frac{ {\rm Pr}(x = +1)}{ {\rm Pr}(x = -1)}\hspace{0.05cm}.\]
  • For unipolar representation  $(+1  →  0$   and   $-1  →  1)$  applies accordingly with  $\xi ∈ \{0, \, 1\}$:
\[L(\xi)={\rm ln} \hspace{0.15cm} \frac{ {\rm Pr}(\xi = 0)}{ {\rm Pr}(\xi = 1)}\hspace{0.05cm}.\]


Probability and log likelihood ratio

The table gives the nonlinear relationship between  ${\rm Pr}(x = ±1)$  and  $L(x)$.  Replacing  ${\rm Pr}(x = +1)$  with  ${\rm Pr}(\xi = 0)$,  the middle row gives the  $L$–value of the unipolar random variable  $\xi$.

You can see:

  1. The more likely random value of  $x ∈ \{+1, \, -1\}$  is given by the  sign   ⇒   ${\rm sign} \ L(x)$.
  2. In contrast,  the  absolute value    ⇒   $|L(x)|$  indicates the reliability for the result  "${\rm sign}(L(x))$".

Given BSC model

$\text{Example 1:}$  We consider the outlined  $\text{BSC model}$  with bipolar representation. 

  • With the falsification probability  $\varepsilon = 0.1$  and the two random variables  $x ∈ \{+1, \, -1\}$  and  $y ∈ \{+1, \, -1\}$  at the input and output of the channel:
\[L(y\hspace{0.05cm}\vert\hspace{0.05cm}x) = {\rm ln} \hspace{0.15cm} \frac{ {\rm Pr}(y\hspace{0.05cm}\vert\hspace{0.05cm}x = +1) }{ {\rm Pr}(y\hspace{0.05cm}\vert\hspace{0.05cm}x = -1)} = \left\{ \begin{array}{c} {\rm ln} \hspace{0.15cm} \big[(1 - \varepsilon)/\varepsilon \big]\\ {\rm ln} \hspace{0.15cm}\big [\varepsilon/(1 - \varepsilon)\big] \end{array} \right.\quad \begin{array}{*{1}c} {\rm for} \hspace{0.15cm} y = +1, \\ {\rm for} \hspace{0.15cm} y = -1. \\ \end{array}\]
  • For example,  $\varepsilon = 0.1$  results in the following numerical values  $($compare the upper table$)$:
\[L(y = +1\hspace{0.05cm}\vert\hspace{0.05cm}x) = {\rm ln} \hspace{0.15cm} \frac{0.9}{0.1} = +2.197\hspace{0.05cm}, \hspace{0.8cm} L(y = -1\hspace{0.05cm}\vert\hspace{0.05cm}x) = -2.197\hspace{0.05cm}.\]
  • This example shows that the so-called  »$\text{LLR algebra}$« can also be applied to conditional probabilities. In  "Exercise 4.1Z",  the BEC–model is described in a similar way.


Conditional AWGN probability density functions

$\text{Example 2:}$  In another example,  we consider the  $\rm AWGN$   channel with the conditional probability density functions

\[f_{y \hspace{0.03cm}\vert \hspace{0.03cm}x=+1 } \hspace{0.05cm} (y \hspace{0.05cm}\vert\hspace{0.05cm}x=+1 )\hspace{-0.1cm} = \hspace{-0.1cm} \frac {1}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ - {(y-1)^2}/(2\sigma^2) } \hspace{0.05cm},\]
\[f_{y \hspace{0.03cm}\vert \hspace{0.03cm}x=-1 } \hspace{0.05cm} (y \hspace{0.05cm}\vert\hspace{0.05cm}x=-1 )\hspace{-0.1cm} = \hspace{-0.1cm} \frac {1}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ - {(y+1)^2}/(2\sigma^2) } \hspace{0.05cm}.\]

⇒   In the graph,  two exemplary Gaussian functions are shown as blue and red curves, respectively. 

  • The total probability density function  $\rm (PDF)$  of the output signal $y$  is obtained from the  (equally)  weighted sum:
\[f_{y } \hspace{0.05cm} (y ) = 1/2 \cdot \big [ f_{y \hspace{0.03cm}\vert \hspace{0.03cm}x=+1 } \hspace{0.05cm} (y \hspace{0.05cm}\vert\hspace{0.05cm}x=+1 ) \hspace{0.1cm} + \hspace{0.1cm} f_{y \hspace{0.03cm}\vert \hspace{0.03cm}x=-1 } \hspace{0.05cm} (y \hspace{0.05cm}\vert \hspace{0.05cm}x=-1 ) \big ] \hspace{0.05cm}.\]
  • We now calculate the probability that the received value  $y$  lies in a  (very)  narrow interval of width  $\it \Delta$  around  $y_0 = 0.25$.  One obtains approximately
\[{\rm Pr} (\vert y - y_0\vert \le{\it \Delta}/2 \hspace{0.05cm} \Big \vert \hspace{0.05cm}x=+1 )\hspace{-0.1cm} \approx \hspace{-0.1cm} \frac {\it \Delta}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ - {(y_0-1)^2}/(2\sigma^2) } \hspace{0.05cm},\]
\[{\rm Pr} (\vert y - y_0\vert \le {\it \Delta}/2 \hspace{0.05cm} \Big \vert \hspace{0.05cm}x=-1 )\hspace{-0.1cm} \approx \hspace{-0.1cm} \frac {\it \Delta}{\sqrt{2\pi} \cdot \sigma } \cdot {\rm e} ^{ - {(y_0+1)^2}/(2\sigma^2) } \hspace{0.05cm}.\]
The slightly larger vertical lines denote the conditions,  the smaller ones the absolute value.
  • The  log likelihood ratio  $\rm (LLR)$  of the conditional probability in the forward direction  $($meaning:   output  $y$  for a given input  $x)$  is thus obtained as the logarithm of the quotient of both expressions:
\[L(y = y_0\hspace{0.05cm}\vert\hspace{0.05cm}x) = {\rm ln} \hspace{0.15cm} \left [ \frac{{\rm e} ^{ - {(y_0-1)^2}/(2\sigma^2)}}{{\rm e} ^{ - {(y_0+1)^2}/(2\sigma^2)}} \right ] = {\rm ln} \left [ {\rm e} ^{ - [ {(y_0-1)^2}+{(y_0+1)^2}]/(2\sigma^2)} \right ] = \frac{(y_0+1)^2-(y_0-1)^2}{2\cdot \sigma^2} = \frac{2 \cdot y_0}{\sigma^2}\hspace{0.05cm}. \]
  • If we now replace the auxiliary  $y_0$  with the general random  $y$  at the AWGN output,  the final result is:
\[L(y \hspace{0.05cm}\vert\hspace{0.05cm}x) = {2 \cdot y}/{\sigma^2} =K_{\rm L} \cdot y\hspace{0.05cm}. \]
Here  $K_{\rm L} = 2/\sigma^2$  is a constant that depends solely on the standard deviation of the Gaussian noise component.


Bit-wise soft-in soft-out decoding


We now assume an  $(n, \ k)$ block code where the code word   $\underline{x} = (x_1, \ \text{...} \ , \ x_n)$   is falsified by the channel into the received word  $\underline{y} = (y_1, \ \text{...} \ , \ y_n)$.

Model of bit-wise soft-in soft-out decoding
  • For long codes,  a  "maximum-a-posteriori block-level decision"  $($for short:  $\text{ block-wise MAP}$  – very elaborate.
  • One would have to find among the  $2^k $ allowable code words  $\underline{x}_j ∈ \mathcal{C}$  the one with the greatest a-posteriori probability  $\rm (APP)$.
  • The code word to output  $\underline{z}$  in this case would be 
$$\underline{z} = {\rm arg} \max_{\underline{x}_{\hspace{0.03cm}j} \hspace{0.03cm} \in \hspace{0.05cm} \mathcal{C}} \hspace{0.1cm} {\rm Pr}( \underline{x}_{\hspace{0.03cm}j} |\hspace{0.05cm} \underline{y} ) \hspace{0.05cm}.$$


A second possibility is decoding on bit level.  The represented bit-wise  $\text{soft-in soft out decoder}$  has the task to decode all code word bits  $x_i ∈ \{0, \, 1\}$  according to maximum a-posteriori probability  ${\rm Pr}(x_i | \underline{y})$.  With the control variable  $i = 1, \text{...} , \ n$  holds in this case:

  • The corresponding  log likelihood ratio  for the  $i$th  code bit is:
\[L_{\rm APP} (i) = L(x_i\hspace{0.05cm}|\hspace{0.05cm}\underline{y}) = {\rm ln} \hspace{0.15cm} \frac{{\rm Pr}(x_i = 0\hspace{0.05cm}|\hspace{0.05cm}\underline{y})}{{\rm Pr}(x_i = 1\hspace{0.05cm}|\hspace{0.05cm}\underline{y})}\hspace{0.05cm} . \]
  • The decoder works iteratively.  At initialization  $($indicated in the diagram by the parameter  $I = 0)$  is  $L_{\rm APP}(i) = L_{\rm K}(i)$,  where the channel  $($German:  "Kanal"   ⇒   "K"$)$  log likelihood ratio  $L_{\rm K}(i)$  is given by the received value  $y_i$.
  • In addition,  the extrinsic  log likelihood ratio  $L_{\rm E}(i)$  is calculated,  which quantifies the total information provided by all other bits   $(j ≠ i)$   due to the code properties about the considered  $i$th bit.
  • At the next iteration  $(I = 1)$  $L_{\rm E}(i)$  is taken into account in the computation of  $L_{\rm APP}(i)$  as a-priori information  $L_{\rm A}(i)$.  Thus,  for the new a-posteriori log likelihood ratio in iteration  $I + 1$  holds:
\[L_{\hspace{0.1cm}\rm APP}^{(I+1)} (i) = L_{\hspace{0.1cm}\rm APP}^{(I)} (i) + L_{\hspace{0.1cm}\rm A}^{(I+1)} (i) = L_{\hspace{0.1cm}\rm APP}^{(I)} (i) + L_{\hspace{0.1cm}\rm E}^{(I)} (i)\hspace{0.05cm} . \]
  • The iterations continue until all absolute values  $|L_{\rm APP}(i)|$  are greater than a given value.  The most likely code word  $\underline{z}$  is then obtained from the signs of all  $L_{\rm APP}(i)$,  with  $i = 1, \ \text{...} , \ n$.
  • For a  $\text{systematic code}$  the first  $k$  bits of  $\underline{z}$  indicate the information word  $\underline{\upsilon}$  being searched for,  which will most likely match the message  $\underline{u}$  being sent.

This SISO decoder descriptione from  [Bos99][1]  is intended at this point primarily to clarify the different  $L$–values.  One recognizes the large potential of the bit-wise decoding only in connection with  $\text{concatenated coding systems}$.

Calculation of extrinsic log likelihood ratios


The difficulty in bit-wise  $($or symbol-wise$)$  iterative decoding is generally the provision of the extrinsic  log likelihood ratios   $L_{\rm E}(i)$.  For a code of length  $n$,  the control variable is   $i = 1, \ \text{...} , \ n$.

$\text{Definition:}$  The   extrinsic log likelihood ratio   is a measure of the information that the other symbols  $(j ≠ i)$  of the code word provide about the  $i$–th encoded symbol,  expressed as a log likelihood ratio.  We denote this characteristic value by  $L_{\rm E}(i)$.


We now calculate the extrinsic log likelihood ratios  $L_{\rm E}(i)$  for two example codes.

$\text{(A)  Repetition Code}$   ⇒   ${\rm RC} \ (n, 1, n)$

A repetition code  $\rm (RC)$ is characterized by the fact that all  $n$  encoded symbols  $x_i ∈ \{0, \, 1\}$  are identical.

  • Here,  the extrinsic log likelihood ratio for the  $i$–th symbol is very easy to specify:
\[L_{\rm E}(i) = \hspace{0.05cm}\sum_{j \ne i} \hspace{0.1cm} L_j \hspace{0.3cm}{\rm with}\hspace{0.3cm}L_j = L_{\rm APP}(j) \hspace{0.05cm}.\]
  • If the sum over all  $L_{j ≠ i}$  is positive   ⇒   this implies from the point of view of the other log likelihood ratio values a preference for the decision   $x_i = 0$.
  • On the other hand,  if the sum is negative   ⇒   $x_i = 1$  is more likely.
  • $L_{\rm E}(i) = 0$  does not allow a prediction.


$\text{Example 5:}$  We consider the decoding of the repetition code  $\text{RC (4, 1,4)}$.  Thereby,  we make three different assumptions for the log likelihood ratio  $ \underline{L}_{\rm APP}^{(I=0)}=\underline{L}_{\rm K}.$

Decoding example  $\rm (5a)$  for the  ${\rm RC} \ (4, 1, 4)$

$\text{Decoding example (5a):}$

$$\underline{L}_{\rm APP}^{(I=0)} = (+1, -1, +3, -1)\text{:}$$
$$L_{\rm E}(1) = -1+3-1 = +1\hspace{0.05cm},$$
$$L_{\rm E}(2) = +1+3-1 = +3\hspace{0.05cm},$$
$$L_{\rm E}(3) = +1-1 -1= -1\hspace{0.05cm},$$
$$L_{\rm E}(4) = +1-1 +3 = +3\hspace{0.05cm}$$
$$\hspace{0.3cm}\Rightarrow\hspace{0.3cm}\underline{L}_{\rm E}^{(I=0)}= (+1, +3, -1, +3) \hspace{0.3cm}\Rightarrow\hspace{0.3cm}\underline{L}_{\rm APP}^{(I=1)}=\underline{L}_{\rm APP}^{(I=0)}+ \underline{L}_{\rm E}^{(I=0)}= (+2, +2, +2, +2)$$
  • At the beginning  $(I=0)$  the positive  $L_{\rm E}$  values indicate  $x_1 = x_2 = x_4 = 0$  while  $x_3 =1$  is more likely   $($because of negative  $L_{\rm E})$ .
  • Already after one iteration  $(I=1)$  all  $L_{\rm APP}$  values are positive   ⇒   the information bit is decoded as  $u = 0$.
  • Further iterations yield nothing.


$\text{Decoding example (5b):}$

Decoding example  $\rm (5b)$  for the  ${\rm RC} \ (4, 1, 4)$
$$\underline{L}_{\rm APP}^{(I=0)} = (+1, +1, -4, +1)\text{:}$$
$$\underline{L}_{\rm E}^{(I=0)} = \ (-2, -2, +3, -2)$$
  • Although three signs were wrong at the beginning,  after two iterations all  $L_{\rm APP}$  values are negative.
  • The information bit is decoded as  $u = 1$.


Decoding example  $\rm (5c)$  for the  ${\rm RC} \ (4, 1, 4)$

$\text{Decoding example (5c):}$

$$\underline{L}_{\rm APP}^{(I=0)} = (+1, +1, -3, +1)\text{:}$$
$$\underline{L}_{\rm E}^{(I=0)} = (-1, -1, +3, -1)$$
  • All  $L_{\rm APP}$  values are zero already after one iteration.
  • The information bit cannot be decoded,  although the initial situation was not much worse than with  $\rm (5b)$.
  • Further iterations bring nothing here.


$\text{(B)   Single parity–check code}$   ⇒   ${\rm SPC} \ (n, \ n \, -1, \ 2)$

For each single parity–check code,  the number of  "ones"  in each code word is even.  Or,  put another way:   For each code word  $\underline{x}$  the  $\text{Hamming weight}$  $w_{\rm H}(\underline{x})$  is even.

$\text{Definition:}$  Let the code word  $\underline{x}^{(–i)}$  contain all symbols except  $x_i$   ⇒   vector of length  $n–1$.  Thus,  the   "extrinsic log likelihood ratio"  will be with respect to the  $i$–th symbol, when  $\underline{x}$  has been received:

\[L_{\rm E}(i) = {\rm ln} \hspace{0.15cm}\frac{ {\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm is \hspace{0.15cm} even} \hspace{0.05cm} \vert\hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]}{ {\rm Pr} \left [w_{\rm H}(\underline{x}^{(-i)})\hspace{0.15cm}{\rm is \hspace{0.15cm} odd} \hspace{0.05cm} \vert \hspace{0.05cm}\underline{y} \hspace{0.05cm}\right ]} \hspace{0.05cm}.\]
  • As will be shown in the  "Exercise 4.4",  it is also possible to write for this:
\[L_{\rm E}(i) = 2 \cdot {\rm tanh}^{-1} \hspace{0.1cm} \left [ \prod\limits_{j \ne i}^{n} \hspace{0.15cm}{\rm tanh}(L_j/2) \right ] \hspace{0.3cm}{\rm with}\hspace{0.3cm}L_j = L_{\rm APP}(j) \hspace{0.05cm}.\]


$\text{Example 6:}$  We assume a  "Single parity–check code"  with  $n = 3, \ k = 2$   ⇒   briefly  ${\rm SPC} \ (3, \ 2, \ 2)$.

  • The   $2^k = 4$   valid code words  $\underline{x} = \{x_1,\ x_2,\ x_3\}$  are for bipolar description   ⇒   $x_i ∈ \{±1\}$:
Decoding example for the  ${\rm SPC} \ (3, 2, 2)$
$$ \underline{x}_0 \hspace{-0.05cm}=\hspace{-0.05cm} (+1\hspace{-0.03cm},\hspace{-0.05cm}+1\hspace{-0.03cm},\hspace{-0.05cm}+1)\hspace{-0.05cm}, $$
$$\underline{x}_1 \hspace{-0.05cm}=\hspace{-0.05cm} (+1\hspace{-0.03cm},\hspace{-0.05cm} -1\hspace{-0.03cm},\hspace{-0.05cm} -1)\hspace{-0.05cm}, $$
$$\underline{x}_2 \hspace{-0.05cm}=\hspace{-0.05cm} (-1\hspace{-0.03cm},\hspace{-0.05cm} +1\hspace{-0.03cm},\hspace{-0.05cm} -1)\hspace{-0.05cm}, $$
$$\underline{x}_3 \hspace{-0.05cm}=\hspace{-0.05cm} (-1\hspace{-0.03cm},\hspace{-0.05cm} -1\hspace{-0.03cm},\hspace{-0.05cm} +1)\hspace{-0.05cm}. $$
So, the product   $x_1 \cdot x_2 \cdot x_3$   is always positive.
  • The table shows the decoding process for  $\underline{L}_{\rm APP} = (+2.0, +0.4, \, –1.6)$.
  • The hard decision according to the signs of  $L_{\rm APP}(i)$  would yield here  $(+1, +1, \, -1)$   ⇒   no valid code word of  ${\rm SP}(3, \ 2, \ 2)$.
  • On the right side of the table, the corresponding extrinsic log likelihood ratios are entered:
\[L_{\rm E}(1) = 2 \cdot {\rm tanh}^{-1} \hspace{0.05cm} \left [ {\rm tanh} (0.2) \cdot {\rm tanh} (-0.8)\hspace{0.05cm}\right ] = -0.131\hspace{0.05cm}, \]
\[L_{\rm E}(2) =2 \cdot {\rm tanh}^{-1} \hspace{0.05cm} \left [ {\rm tanh} (1.0) \cdot {\rm tanh} (-0.8)\hspace{0.05cm}\right ] = -0.518\hspace{0.05cm}, \]
\[L_{\rm E}(3) =2 \cdot {\rm tanh}^{-1} \hspace{0.05cm} \left [ {\rm tanh} (1.0) \cdot {\rm tanh} (0.2)\hspace{0.05cm}\right ] = +0.151\hspace{0.05cm}.\]

The second equation can be interpreted as follows:

  1. $L_{\rm APP}(1) = +2.0$  and  $L_{\rm APP}(3) = \, -1.6$  state that the first bit is more likely to be  "$+1$"  than  "$-1$"  and the third bit is more likely to be  "$-1"$  than  "$+1$".
  2. The reliability  $($the amount$)$  is slightly greater for the first bit than for the third.
  3. The extrinsic information   $L_{\rm E}(2) = \, -0.518$  considers only the information of bit 1 and bit 3 about bit 2.
  4. From their point of view,  the second bit is  "$-1$"  with reliability  $0.518$.
  5. The log likelihood ratio  $L_{\rm APP}^{I=0}(2) = +0.4$  derived from the received value  $y_2$  has suggested  "$+1$"  for the second bit.
  6. But the discrepancy is resolved here already in the iteration  $I = 1$.  The decision falls here for the code word  $\underline{x}_1 \hspace{-0.05cm}=\hspace{-0.05cm} (+1\hspace{-0.03cm},\hspace{-0.05cm} -1\hspace{-0.03cm},\hspace{-0.05cm} -1)\hspace{-0.05cm},$.
  7. For   $0.518 < L_{\rm APP}(2) < 1.6$   the result  $\underline{x}_1$  would only be available after several iterations.  For  $L_{\rm APP}(2) > 1.6$  on the other hand,  the decoder returns  $\underline{x}_0$.


BCJR decoding: Forward-backward algorithm


An example of iterative decoding of convolutional codes is the  »BCJR algorithm«,  named after its inventors L. R. Bahl, J. Cocke, F. Jelinek, and J. Raviv   ⇒   [BCJR74][2].  The algorithm has many parallels to the seven-year older Viterbi decoding,  but also some significant differences:

  • The Viterbi algorithm  $($in its original form$)$  cannot process soft information.  In contrast,  the BCJR algorithm specifies a reliability value for each individual bit at each iteration,  which is taken into account in subsequent iterations.


The figure is intended to illustrate  $($in an almost inadmissibly simplified way$)$  the different approach of Viterbi and BCJR algorithm.  This is based on a convolutional code with memory  $m = 1$  and length  $L = 4$   ⇒   total length $($with termination$)$  $L' = 5$.

Comparison of Viterbi and BCJR algorithm
  • Viterbi searches and finds the most likely path from  ${\it \Gamma}_0(S_0)$  to  ${\it \Gamma}_5(S_0)$, namely  $S_0 → S_1 → S_0 → S_0 → S_1→ S_0 $. We refer to the sample solution to  "Exercise 3.9Z".
  • The sketch for the BCJR algorithm illustrates the extraction of the extrinsic  L–value for the third symbol   ⇒   $L_{\rm E}(3)$.  The relevant area in the trellis is shaded:
  • Working through the trellis diagram in the forward direction,  one obtains in the same way as Viterbi the metrics  $\alpha_1, \ \alpha_2, \ \text{...}\hspace{0.05cm} , \ \alpha_5$.  To calculate  $L_{\rm E}(3)$  one needs from this  $\alpha_2$.
  • Then we traverse the trellis diagram backwards  $($i.e. from right to left$)$  and thus obtain the metrics  $\beta_4, \ \beta_3, \ \text{...}\hspace{0.05cm} , \ \beta_0$  corresponding to the sketch below.
  • The extrinsic log likelihood ratio  $L_{\rm E}(3)$  is obtained from the forward direction metric  $\alpha_2$  and the backward direction metric  $\beta_3$  and the a-priori information  $\gamma_3$  about the symbol  $i = 3$.

Basic structure of concatenated coding systems


The most important communication systems of the last years use two different channel codes.  One speaks then of  »concatenated codes«.  Even with relatively short component codes  $\mathcal{C}_1$  and  $\mathcal{C}_2$  the concatenated code  $\mathcal{C}$  results in a sufficiently large code word length  $n$,  which is known to be necessary to approach the channel capacity.

To begin with,  here are a few examples from mobile communications:

  • In  $\rm GSM$  $($"Global System for Mobile Communications",  the second generation of mobile communication systems$)$,  the data bit rate is first increased from  $9. 6 \ \rm kbit/s$  to  $12 \ \rm kbit/s$  in order to enable error detection in circuit-switched networks as well.  This is followed by a punctured convolutional code with the output bit rate  $22.8 \ \rm kbit/s$.  The total code rate is thus about $0.421$.
  • In the 3G cellular system  $\rm UMTS$  $($"Universal Mobile Telecommunications System"),  depending on the boundary conditions  $($good/bad channel,  few/many subscribers in the cell$)$  one uses a  $\text{convolutional code}$  or a  $\text{turbo code}$  $($by this one understands per se the concatenation of two convolutional encoders$)$. 
  • In 4G mobile communication systems  $\rm LTE$  $($"Long Term Evolution"$)$,  a convolutional code is used for short control signals,  turbo code for the longer payload data.


The graph shows the basic structure of a parallel concatenated coding system.  All vectors consist of  $n$  elements: 

$$\underline{L} = (L(1), \ \text{...}\hspace{0.05cm} , \ L(n)).$$
Parallel concatenated codes

So the calculation of all  L–values is done on symbol level.  Not shown here is the  $\rm interleaver$,  which is mandatory e.g. for turbo codes.

  1. The sequences  $\underline{x}_1$  and  $\underline{x}_2$  are combined to form the vector  $\underline{x}$  for joint transmission over the channel by a multiplexer.
  2. At the receiver, the sequence  $\underline{y}$  is again decomposed into the parts  $\underline{y}_1$  and  $\underline{y}_2$.  From this the channel L–values  $\underline{L}_{\rm K,\hspace{0.05cm}1} $ and  $\underline{L}_{\rm K,\hspace{0.05cm}2}$  are formed.
  3. The symbol-wise decoder determines the extrinsic log likelihood ratios   $\underline{L}_{\rm E,\hspace{0.05cm} 1}$  and  $\underline{L}_{\rm E,\hspace{0.05cm} 2}$, which are also the a-priori information  $\underline{L}_{\rm A,\hspace{0.05cm} 2}$  and  $\underline{L}_{\rm A,\hspace{0.05cm} 1}$  for the other decoder.
  4. After a sufficient number of iterations  $($i.e. when a termination criterion is met$)$,  the vector  $\underline{L}_{\rm APP}$  of a-posteriori values is available at the decoder output.
  5. In the example,  the value in the upper branch is taken arbitrarily. 
  6. However,  the lower log likelihood ratio would also be possible.

The above model also applies in particular to the decoding of turbo codes according to the chapter  "Basic structure of a turbo code".

Exercises for the chapter


Exercise 4.1: Log Likelihood Ratio

Exercise 4.1Z: Log Likelihood Ratio at the BEC Model

Exercise 4.2: Channel Log Likelihood Ratio at AWGN

Exercise 4.3: Iterative Decoding at the BSC

Exercise 4.3Z: Conversions of L-value and S-value

Exercise 4.4: Extrinsic L-values at SPC

Exercise 4.4Z: Supplement to Exercise 4.4

Exercise 4.5: On the Extrinsic L-values again

Exercise 4.5Z: Tangent Hyperbolic and Inverse

References

  1. Bossert, M.:  Channel Coding for Telecommunications.  Chichester:  Wiley,  1999.  ISBN:  978-0-471-98277-7.
  2. Bahl, L.R.; Cocke, J.; Jelinek, F.; Raviv, J.:  Optimal Decoding of Linear Codes for Minimizing Symbol Error Rate.  In: IEEE Transactions on Information Theory, Vol. IT-20, pp. 284-287, 1974.