Difference between revisions of "Digital Signal Transmission/Optimal Receiver Strategies"
(4 intermediate revisions by one other user not shown) | |||
Line 20: | Line 20: | ||
*In this model, any channel distortions are added to the transmitter and are thus already included in the basic transmission pulse gs(t) and the signal s(t). This measure is only for a simpler representation and is not a restriction.<br> | *In this model, any channel distortions are added to the transmitter and are thus already included in the basic transmission pulse gs(t) and the signal s(t). This measure is only for a simpler representation and is not a restriction.<br> | ||
− | *Knowing the currently applied received signal $ | + | *Knowing the currently applied received signal $r(t), the optimal receiver searches from the set \{Q_0,...,Q_{M-1}\} of the possible source symbol sequences, the receiver searches for the most likely transmitted sequence Q_j and outputs this as a sink symbol sequence V$. <br> |
*Before the actual decision algorithm, a numerical value Wi must be derived from the received signal r(t) for each possible sequence Qi by suitable signal preprocessing. The larger Wi is, the greater the inference probability that Qi was transmitted.<br> | *Before the actual decision algorithm, a numerical value Wi must be derived from the received signal r(t) for each possible sequence Qi by suitable signal preprocessing. The larger Wi is, the greater the inference probability that Qi was transmitted.<br> | ||
Line 65: | Line 65: | ||
Example 1: To illustrate the "ML" and the "MAP" decision rule, we now construct a very simple example with only two source symbols (M=2). | Example 1: To illustrate the "ML" and the "MAP" decision rule, we now construct a very simple example with only two source symbols (M=2). | ||
[[File:EN_Dig_T_3_7_S2.png|right|frame|For clarification of MAP and ML receiver|class=fit]] | [[File:EN_Dig_T_3_7_S2.png|right|frame|For clarification of MAP and ML receiver|class=fit]] | ||
− | <br><br>⇒ The two possible symbols Q0 and Q1 are represented by the transmitted signals s=0 and s=1. | + | <br><br>⇒ The two possible symbols Q0 and Q1 are represented by the transmitted signals s=0 and s=1. |
− | + | <br><br> | |
⇒ The received signal can – for whatever reason – take three different values, namely r=0, r=1 and additionally r=0.5. | ⇒ The received signal can – for whatever reason – take three different values, namely r=0, r=1 and additionally r=0.5. | ||
− | + | <br><br> | |
− | *The received values r=0 and r=1 will be assigned to the transmitter values s=0 (Q0) | + | <u>Note:</u> |
+ | *The received values r=0 and r=1 will be assigned to the transmitter values s=0 (Q0) resp. s=1 (Q1), by both, the ML and MAP decisions. | ||
*In contrast, the decisions will give a different result with respect to the received value r=0.5: | *In contrast, the decisions will give a different result with respect to the received value r=0.5: | ||
− | *The maximum likelihood decision rule leads to the source symbol Q0, because of | + | :*The maximum likelihood (ML) decision rule leads to the source symbol Q0, because of: |
− | :$${\rm Pr}\big [ r= 0.5\hspace{0.05cm}\vert\hspace{0.05cm} | + | ::$${\rm Pr}\big [ r= 0.5\hspace{0.05cm}\vert\hspace{0.05cm} |
Q_0\big ] = 0.4 > {\rm Pr}\big [ r= 0.5\hspace{0.05cm} \vert \hspace{0.05cm} | Q_0\big ] = 0.4 > {\rm Pr}\big [ r= 0.5\hspace{0.05cm} \vert \hspace{0.05cm} | ||
Q_1\big ] = 0.2 \hspace{0.05cm}.$$ | Q_1\big ] = 0.2 \hspace{0.05cm}.$$ | ||
− | *The MAP decision | + | :*The maximum–a–posteriori $\rm (MAP)$ decision rule leads to the source symbol Q1, since according to the incidental calculation in the graph: |
− | :$${\rm Pr}\big [Q_1 \hspace{0.05cm}\vert\hspace{0.05cm} | + | ::$${\rm Pr}\big [Q_1 \hspace{0.05cm}\vert\hspace{0.05cm} |
r= 0.5\big ] = 0.6 > {\rm Pr}\big [Q_0 \hspace{0.05cm}\vert\hspace{0.05cm} | r= 0.5\big ] = 0.6 > {\rm Pr}\big [Q_0 \hspace{0.05cm}\vert\hspace{0.05cm} | ||
r= 0.5\big ] = 0.4 \hspace{0.05cm}.$$}}<br> | r= 0.5\big ] = 0.4 \hspace{0.05cm}.$$}}<br> | ||
Line 85: | Line 86: | ||
== Maximum likelihood decision for Gaussian noise == | == Maximum likelihood decision for Gaussian noise == | ||
<br> | <br> | ||
− | We now assume that the received signal r(t) is additively composed of a useful | + | We now assume that the received signal r(t) is additively composed of a useful component s(t) and a noise component n(t), where the noise is assumed to be Gaussian distributed and white ⇒ [[Digital_Signal_Transmission/System_Components_of_a_Baseband_Transmission_System#Transmission_channel_and_interference|"AWGN noise"]]: |
:r(t)=s(t)+n(t). | :r(t)=s(t)+n(t). | ||
Any channel distortions are already applied to the signal s(t) for simplicity.<br> | Any channel distortions are already applied to the signal s(t) for simplicity.<br> | ||
− | The necessary noise power limitation is realized by an integrator; this corresponds to an averaging of the noise values in the time domain. If one limits the integration interval to the range t1 to t2, one can derive a quantity Wi for each source symbol sequence Qi, which is a measure for the conditional probability ${\rm Pr}\big [ r(t)\hspace{0.05cm} \vert \hspace{0.05cm} | + | The necessary noise power limitation is realized by an integrator; this corresponds to an averaging of the noise values in the time domain. If one limits the integration interval to the range t1 to t2, one can derive a quantity Wi for each source symbol sequence Qi, which is a measure for the conditional probability ${\rm Pr}\big [ r(t)\hspace{0.05cm} \vert \hspace{0.05cm} |
Q_i\big ] $: | Q_i\big ] $: | ||
:$$W_i = \int_{t_1}^{t_2} r(t) \cdot s_i(t) \,{\rm d} t - | :$$W_i = \int_{t_1}^{t_2} r(t) \cdot s_i(t) \,{\rm d} t - | ||
Line 96: | Line 97: | ||
I_i - {E_i}/{2} \hspace{0.05cm}.$$ | I_i - {E_i}/{2} \hspace{0.05cm}.$$ | ||
− | This decision variable Wi can be derived using the k–dimensionial [[Theory_of_Stochastic_Signals/Two-Dimensional_Random_Variables#Joint_probability_density_function|"joint probability density"]] of the noise (with k→∞) and some boundary crossings. The result can be interpreted as follows: | + | This decision variable Wi can be derived using the k–dimensionial [[Theory_of_Stochastic_Signals/Two-Dimensional_Random_Variables#Joint_probability_density_function|"joint probability density"]] of the noise (with k→∞) and some boundary crossings. The result can be interpreted as follows: |
− | *Integration is used for noise power reduction by averaging. If N binary symbols are decided simultaneously by the maximum likelihood detector, set t1=0 and t2=N⋅T for distortion-free channel. | + | *Integration is used for noise power reduction by averaging. If N binary symbols are decided simultaneously by the maximum likelihood detector, set t1=0 and t2=N⋅T for distortion-free channel. |
− | *The first term of the above decision variable Wi is equal to the [[Theory_of_Stochastic_Signals/Cross-Correlation_Function_and_Cross_Power-Spectral_Density#Definition_of_the_cross-correlation_function| "energy cross-correlation function"]] formed over the finite time interval NT between r(t) and si(t) at the point τ=0: | + | |
+ | *The first term of the above decision variable Wi is equal to the [[Theory_of_Stochastic_Signals/Cross-Correlation_Function_and_Cross_Power-Spectral_Density#Definition_of_the_cross-correlation_function| "energy cross-correlation function"]] formed over the finite time interval NT between r(t) and si(t) at the time point τ=0: | ||
:$$I_i = \varphi_{r, \hspace{0.08cm}s_i} (\tau = 0) = \int_{0}^{N \cdot T}r(t) \cdot s_i(t) \,{\rm d} t | :$$I_i = \varphi_{r, \hspace{0.08cm}s_i} (\tau = 0) = \int_{0}^{N \cdot T}r(t) \cdot s_i(t) \,{\rm d} t | ||
\hspace{0.05cm}.$$ | \hspace{0.05cm}.$$ | ||
− | *The second term gives the | + | *The second term gives half the energy of the considered useful signal si(t) to be subtracted. The energy is equal to the auto-correlation function $\rm (ACF)$ of si(t) at the time point τ=0: |
::<math>E_i = \varphi_{s_i} (\tau = 0) = \int_{0}^{N \cdot T} | ::<math>E_i = \varphi_{s_i} (\tau = 0) = \int_{0}^{N \cdot T} | ||
s_i^2(t) \,{\rm d} t \hspace{0.05cm}.</math> | s_i^2(t) \,{\rm d} t \hspace{0.05cm}.</math> | ||
− | *In case of distorting channel the impulse response hK(t) is not Dirac-shaped, but for example extended to the range −TK≤t≤+TK. In this case, t1=−TK and t2=N⋅T+TK must be used for the | + | *In the case of a distorting channel, the channel impulse response hK(t) is not Dirac-shaped, but for example extended to the range −TK≤t≤+TK. In this case, t1=−TK and t2=N⋅T+TK must be used for the integration limits.<br> |
== Matched filter receiver vs. correlation receiver == | == Matched filter receiver vs. correlation receiver == | ||
<br> | <br> | ||
− | There are various circuit implementations of the maximum likelihood receiver. | + | There are various circuit implementations of the maximum likelihood (ML) receiver. |
− | For example, the required integrals can be obtained by linear filtering and subsequent sampling. This realization form is called '''matched filter receiver''', because here the impulse responses of the M parallel filters have the same shape as the useful signals s0(t), ... , sM−1(t). <br> | + | ⇒ For example, the required integrals can be obtained by linear filtering and subsequent sampling. This realization form is called '''matched filter receiver''', because here the impulse responses of the M parallel filters have the same shape as the useful signals s0(t), ... , sM−1(t). <br> |
− | *The M decision variables Ii are then equal to the convolution products r(t)⋆si(t) at time t=0. | + | *The M decision variables Ii are then equal to the convolution products r(t)⋆si(t) at time t=0. |
− | *For example, the "optimal binary receiver" described in detail in the chapter [[Digital_Signal_Transmission/Optimization_of_Baseband_Transmission_Systems#Prerequisites_and_optimization_criterion|"Optimization of Baseband Transmission Systems"]] allows a maximum likelihood decision with | + | *For example, the "optimal binary receiver" described in detail in the chapter [[Digital_Signal_Transmission/Optimization_of_Baseband_Transmission_Systems#Prerequisites_and_optimization_criterion|"Optimization of Baseband Transmission Systems"]] allows a maximum likelihood (ML) decision with parameters M=2 and N=1.<br> |
− | A second form | + | ⇒ A second realization form is provided by the '''correlation receiver''' according to the following graph. One recognizes from this block diagram for the indicated parameters: |
+ | [[File:EN_Dig_T_3_7_S4.png|right|frame|Correlation receiver for N=3, t1=0, t2=3T and M=23=8 |class=fit]] | ||
− | |||
− | |||
− | |||
*The drawn correlation receiver forms a total of M=8 cross-correlation functions between the received signal r(t)=sk(t)+n(t) and the possible transmitted signals si(t), i=0, ... , M−1. The following description assumes that the useful signal sk(t) has been transmitted.<br> | *The drawn correlation receiver forms a total of M=8 cross-correlation functions between the received signal r(t)=sk(t)+n(t) and the possible transmitted signals si(t), i=0, ... , M−1. The following description assumes that the useful signal sk(t) has been transmitted.<br> | ||
− | * | + | *This receiver searches for the maximum value Wj of all correlation values and outputs the corresponding sequence Qj as sink symbol sequence V. Formally, the $\rm ML$ decision rule can be expressed as follows: |
− | :$$V = Q_j, \hspace{0.2cm}{\rm | + | :$$V = Q_j, \hspace{0.2cm}{\rm if}\hspace{0.2cm} W_i < W_j |
\hspace{0.2cm}{\rm for}\hspace{0.2cm} {\rm | \hspace{0.2cm}{\rm for}\hspace{0.2cm} {\rm | ||
all}\hspace{0.2cm} i \ne j \hspace{0.05cm}.$$ | all}\hspace{0.2cm} i \ne j \hspace{0.05cm}.$$ | ||
− | *If we further assume that all transmitted signals si(t) have | + | *If we further assume that all transmitted signals si(t) have same energy, we can dispense with the subtraction of Ei/2 in all branches. In this case, the following correlation values are compared (i=0, ... , M−1): |
::<math>I_i = \int_{0}^{NT} s_j(t) \cdot s_i(t) \,{\rm d} t + | ::<math>I_i = \int_{0}^{NT} s_j(t) \cdot s_i(t) \,{\rm d} t + | ||
\int_{0}^{NT} n(t) \cdot s_i(t) \,{\rm d} t | \int_{0}^{NT} n(t) \cdot s_i(t) \,{\rm d} t | ||
\hspace{0.05cm}.</math> | \hspace{0.05cm}.</math> | ||
− | *With high probability, Ij=Ik is larger than all other comparison values Ij≠k. However, if the noise n(t) is too large, the correlation receiver will | + | *With high probability, Ij=Ik is larger than all other comparison values Ij≠k ⇒ right decision. However, if the noise n(t) is too large, also the correlation receiver will make wrong decisions.<br> |
== Representation of the correlation receiver in the tree diagram== | == Representation of the correlation receiver in the tree diagram== | ||
<br> | <br> | ||
− | Let us illustrate | + | Let us illustrate the correlation receiver operation in the tree diagram, where the 23=8 possible source symbol sequences Qi of length N=3 are represented by bipolar rectangular transmitted signals si(t). |
− | [[File:P ID1458 Dig T 3 7 S5a version1.png| | + | [[File:P ID1458 Dig T 3 7 S5a version1.png|right|frame|All 23=8 possible bipolar transmitted signals si(t) for N=3|class=fit]] |
+ | The possible symbol sequences Q0=LLL, ... , Q7=HHH and the associated transmitted signals s0(t), ... , s7(t) are listed below. | ||
− | + | *Due to bipolar amplitude coefficients and the rectangular shape ⇒ all signal energies are equal: E0=...=E7=N⋅EB, where EB indicates the energy of a single pulse of duration T. | |
− | *Due to | + | |
− | *Therefore, subtraction of the Ei/2 term in all branches can be omitted ⇒ | + | *Therefore, the subtraction of the Ei/2 term in all branches can be omitted ⇒ the decision based on the correlation values Ii gives equally reliable results as maximizing the corrected values Wi. |
+ | <br clear=all> | ||
{{GraueBox|TEXT= | {{GraueBox|TEXT= | ||
− | Example 2: The graph shows the continuous integral values, assuming the actually transmitted signal s5(t) and the noise-free case. For this case, the time-dependent integral values and the integral end values | + | Example 2: The graph shows the continuous-valued integral values, assuming the actually transmitted signal s5(t) and the noise-free case. For this case, the time-dependent integral values and the integral end values: |
− | [[File:EN_Dig_T_3_7_S5b.png|right|frame| | + | [[File:EN_Dig_T_3_7_S5b.png|right|frame|Tree diagram of the correlation receiver in the noise-free case|class=fit]] |
:$$i_i(t) = \int_{0}^{t} r(\tau) \cdot s_i(\tau) \,{\rm d} | :$$i_i(t) = \int_{0}^{t} r(\tau) \cdot s_i(\tau) \,{\rm d} | ||
\tau = \int_{0}^{t} s_5(\tau) \cdot s_i(\tau) \,{\rm d} | \tau = \int_{0}^{t} s_5(\tau) \cdot s_i(\tau) \,{\rm d} | ||
Line 156: | Line 158: | ||
\Rightarrow \hspace{0.3cm}I_i = i_i(3T). $$ | \Rightarrow \hspace{0.3cm}I_i = i_i(3T). $$ | ||
The graph can be interpreted as follows: | The graph can be interpreted as follows: | ||
− | *Because of the rectangular shape of the signals si(t), all function curves ii(t) are rectilinear. The end values normalized to EB are +3, +1, −1 and −3.<br> | + | *Because of the rectangular shape of the signals si(t), all function curves ii(t) are rectilinear. The end values normalized to EB are +3, +1, −1 and −3.<br> |
− | *The maximum final value is I5=3⋅EB (red waveform), since signal s5(t) was actually sent. Without noise, the correlation receiver thus naturally always makes the correct decision.<br> | + | |
− | *The blue curve i1(t) leads to the final value I1=−EB+EB+EB=EB, since s1(t) differs from s5(t) only in the first bit. The comparison values I4 and I7 are also equal to EB.<br> | + | *The maximum final value is I5=3⋅EB (red waveform), since signal s5(t) was actually sent. Without noise, the correlation receiver thus naturally always makes the correct decision.<br> |
− | *Since s0(t), s3(t) and s6(t) differ from the transmitted s5(t) in two bits, I0=I3=I6=−EB. The green curve shows s6(t) initially increasing (first bit matches) and then decreasing over two bits. | + | |
− | *The purple curve leads to the final value I2=−3⋅EB. The corresponding signal s2(t) differs from s5(t) in all three symbols and s2(t)=−s5(t) holds.}}<br><br> | + | *The blue curve i1(t) leads to the final value I1=−EB+EB+EB=EB, since s1(t) differs from s5(t) only in the first bit. The comparison values I4 and I7 are also equal to EB.<br> |
+ | |||
+ | *Since s0(t), s3(t) and s6(t) differ from the transmitted s5(t) in two bits, I0=I3=I6=−EB. The green curve shows s6(t) initially increasing (first bit matches) and then decreasing over two bits. | ||
+ | |||
+ | *The purple curve leads to the final value I2=−3⋅EB. The corresponding signal s2(t) differs from s5(t) in all three symbols and s2(t)=−s5(t) holds.}}<br><br> | ||
{{GraueBox|TEXT= | {{GraueBox|TEXT= | ||
− | Example 3: The graph | + | Example 3: The graph describes the same situation as Example 2, but now the received signal r(t)=s5(t)+n(t) is assumed. The variance of the AWGN noise n(t) here is σ2n=4⋅EB/T. |
− | [[File:EN_Dig_T_3_7_S5c_neu.png|right|frame| | + | [[File:EN_Dig_T_3_7_S5c_neu.png|right|frame|Tree diagram of the correlation receiver with noise (σ2n=4⋅EB/T) |class=fit]] |
− | One can see from this graph compared to the noise-free case: | + | <br><br><br>One can see from this graph compared to the noise-free case: |
− | *The | + | *The curves are now no longer straight due to the noise component n(t) and there are also slightly different final values than without noise. |
− | *In the considered example, | + | |
− | * | + | *In the considered example, the correlation receiver decides correctly with high probability, since the difference between I5 and the next value I7 is relatively large: 1.65⋅EB. <br> |
− | + | ||
+ | *The error probability in this example is not better than that of the matched filter receiver with symbol-wise decision. In accordance with the chapter [[Digital_Signal_Transmission/Optimization_of_Baseband_Transmission_Systems#Prerequisites_and_optimization_criterion|"Optimization of Baseband Transmission Systems"]], the following also applies here: | ||
:$$p_{\rm S} = {\rm Q} \left( \sqrt{ {2 \cdot E_{\rm B} }/{N_0} }\right) | :$$p_{\rm S} = {\rm Q} \left( \sqrt{ {2 \cdot E_{\rm B} }/{N_0} }\right) | ||
= {1}/{2} \cdot {\rm erfc} \left( \sqrt{ { E_{\rm B} }/{N_0} }\right) \hspace{0.05cm}.$$}} | = {1}/{2} \cdot {\rm erfc} \left( \sqrt{ { E_{\rm B} }/{N_0} }\right) \hspace{0.05cm}.$$}} | ||
Line 175: | Line 182: | ||
{{BlaueBox|TEXT= | {{BlaueBox|TEXT= | ||
− | $\text{ | + | $\text{Conclusions:}$ |
− | + | #If the input signal does not have statistical bindings $\text{(Example 2)}$, there is no improvement by joint decision of N symbols over symbol-wise decision <br>⇒ pS=Q(√2⋅EB/N0). | |
− | + | #In the presence of statistical bindings $\text{(Example 3)}$, the joint decision of N symbols noticeably reduces the error probability, since the maximum likelihood receiver takes the bindings into account. | |
− | + | #Such bindings can be either deliberately created by transmission-side coding $($see the LNTwww book [[Channel_Coding|"Channel Coding"]]) or unintentionally caused by (linear) channel distortions.<br> | |
− | + | #In the presence of such "intersymbol interferences", the calculation of the error probability is much more difficult. However, comparable approximations as for the Viterbi receiver can be used, which are given at the [[Digital_Signal_Transmission/Viterbi_Receiver#Bit_error_probability_with_maximum_likelihood_decision|end of the next chapter]]. }}<br> | |
== Correlation receiver with unipolar signaling == | == Correlation receiver with unipolar signaling == | ||
<br> | <br> | ||
− | So far, we have always assumed binary ''bipolar'' signaling when describing the correlation receiver: | + | So far, we have always assumed binary '''bipolar''' signaling when describing the correlation receiver: |
:$$a_\nu = \left\{ \begin{array}{c} +1 \\ | :$$a_\nu = \left\{ \begin{array}{c} +1 \\ | ||
-1 \\ \end{array} \right.\quad | -1 \\ \end{array} \right.\quad | ||
Line 191: | Line 198: | ||
q_\nu = \mathbf{L} \hspace{0.05cm}. \\ | q_\nu = \mathbf{L} \hspace{0.05cm}. \\ | ||
\end{array}$$ | \end{array}$$ | ||
− | Now we consider the case of binary ''unipolar'' digital signaling holds: | + | Now we consider the case of binary '''unipolar''' digital signaling holds: |
:$$a_\nu = \left\{ \begin{array}{c} 1 \\ | :$$a_\nu = \left\{ \begin{array}{c} 1 \\ | ||
0 \\ \end{array} \right.\quad | 0 \\ \end{array} \right.\quad | ||
Line 199: | Line 206: | ||
q_\nu = \mathbf{L} \hspace{0.05cm}. \\ | q_\nu = \mathbf{L} \hspace{0.05cm}. \\ | ||
\end{array}$$ | \end{array}$$ | ||
+ | [[File:P ID1462 Dig T 3 7 S5c version1.png|right|frame|Possible unipolar transmitted signals for N=3|class=fit]] | ||
+ | The 23=8 possible source symbol sequences Qi of length N=3 are now represented by unipolar rectangular transmitted signals si(t). | ||
− | + | Listed on the right are the eight symbol sequences and the transmitted signals | |
− | + | :$$Q_0 = \rm LLL, \text{ ... },\ Q_7 = \rm HHH,$$ | |
− | + | :$$s_0(t), \text{ ... },\ s_7(t).$$ | |
− | By comparing with the [[Digital_Signal_Transmission/Optimal_Receiver_Strategies#Representation_of_the_correlation_receiver_in_the_tree_diagram|"corresponding table"]] for bipolar signaling, one can see: | + | By comparing with the [[Digital_Signal_Transmission/Optimal_Receiver_Strategies#Representation_of_the_correlation_receiver_in_the_tree_diagram|"corresponding table"]] for bipolar signaling, one can see: |
− | *Due to the unipolar amplitude coefficients, the signal energies Ei are now different, | + | *Due to the unipolar amplitude coefficients, the signal energies Ei are now different, e.g. E0=0 and E7=3⋅EB. |
− | *Here the decision based on the integral | + | |
− | + | *Here the decision based on the integral values Ii does not lead to the correct result. Instead, the corrected comparison values Wi=Ii−Ei/2 must now be used.<br> | |
{{GraueBox|TEXT= | {{GraueBox|TEXT= | ||
− | Example 4: The graph shows the | + | Example 4: The graph shows the integral values Ii, again assuming the actual transmitted signal s5(t) and the noise-free case. The corresponding bipolar equivalent was considered in [[Digital_Signal_Transmission/Optimal_Receiver_Strategies#Representation_of_the_correlation_receiver_in_the_tree_diagram|Example 2]]. |
− | [[File:EN_Dig_T_3_7_S5d.png|right|frame|Tree diagram of the correlation receiver (unipolar)|class=fit]] | + | [[File:EN_Dig_T_3_7_S5d.png|right|frame|Tree diagram of the correlation receiver (unipolar signaling)|class=fit]] |
− | For this example, the following comparison values result, each normalized to EB: | + | For this example, the following comparison values result, each normalized to EB: |
:$$I_5 = I_7 = 2, \hspace{0.2cm}I_1 = I_3 = I_4= I_6 = 1 \hspace{0.2cm}, | :$$I_5 = I_7 = 2, \hspace{0.2cm}I_1 = I_3 = I_4= I_6 = 1 \hspace{0.2cm}, | ||
\hspace{0.2cm}I_0 = I_2 = 0 | \hspace{0.2cm}I_0 = I_2 = 0 | ||
Line 223: | Line 232: | ||
This means: | This means: | ||
− | *When compared in terms of maximum Ii values, the source symbol sequences Q5 and Q7 would be equivalent. | + | *When compared in terms of maximum Ii values, the source symbol sequences Q5 and Q7 would be equivalent. |
− | *On the other hand, if the different energies (E5=2, E7=3) are taken into account, the decision is clearly in favor of the sequence Q5 because of W5>W7. | + | |
− | *The correlation receiver according to Wi=Ii−Ei/2 therefore decides correctly on s(t)=s5(t) even with unipolar signaling. }}<br> | + | *On the other hand, if the different energies (E5=2, E7=3) are taken into account, the decision is clearly in favor of the sequence Q5 because of W5>W7. |
+ | |||
+ | *The correlation receiver according to Wi=Ii−Ei/2 therefore decides correctly on s(t)=s5(t) even with unipolar signaling. }}<br> | ||
== Exercises for the chapter== | == Exercises for the chapter== |
Latest revision as of 14:16, 11 July 2022
Contents
- 1 Considered scenario and prerequisites
- 2 Maximum-a-posteriori and maximum–likelihood decision rule
- 3 Maximum likelihood decision for Gaussian noise
- 4 Matched filter receiver vs. correlation receiver
- 5 Representation of the correlation receiver in the tree diagram
- 6 Correlation receiver with unipolar signaling
- 7 Exercises for the chapter
Considered scenario and prerequisites
All digital receivers described so far always make symbol-wise decisions. If, on the other hand, several symbols are decided simultaneously, statistical bindings between the received signal samples can be taken into account during detection, which results in a lower error probability – but at the cost of an additional delay time.
In this (partly also in the next chapter) the following transmission model is assumed. Compared to the last two chapters, the following differences arise:
- Q∈{Qi} with i=0, ... , M−1 denotes a time-constrained source symbol sequence ⟨qν⟩ whose symbols are to be jointly decided by the receiver.
- If the source Q describes a sequence of N redundancy-free binary symbols, set M=2N. On the other hand, if the decision is symbol-wise, M specifies the level number of the digital source.
- In this model, any channel distortions are added to the transmitter and are thus already included in the basic transmission pulse gs(t) and the signal s(t). This measure is only for a simpler representation and is not a restriction.
- Knowing the currently applied received signal r(t), the optimal receiver searches from the set {Q0, ... , QM−1} of the possible source symbol sequences, the receiver searches for the most likely transmitted sequence Qj and outputs this as a sink symbol sequence V.
- Before the actual decision algorithm, a numerical value Wi must be derived from the received signal r(t) for each possible sequence Qi by suitable signal preprocessing. The larger Wi is, the greater the inference probability that Qi was transmitted.
- Signal preprocessing must provide for the necessary noise power limitation and – in the case of strong channel distortions – for sufficient pre-equalization of the resulting intersymbol interferences. In addition, preprocessing also includes sampling for time discretization.
Maximum-a-posteriori and maximum–likelihood decision rule
The (unconstrained) optimal receiver is called the "MAP receiver", where "MAP" stands for "maximum–a–posteriori".
Definition: The maximum–a–posteriori receiver (abbreviated MAP) determines the M inference probabilities Pr[Qi|r(t)], and sets the output sequence V according to the decision rule, where the index is i=0, ... , M−1 as well as i≠j:
- Pr[Qj|r(t)]>Pr[Qi|r(t)].
- The "inference probability" Pr[Qi|r(t)] indicates the probability with which the sequence Qi was sent when the received signal r(t) is present at the decision. Using "Bayes' theorem", this probability can be calculated as follows:
- Pr[Qi|r(t)]=Pr[r(t)|Qi]⋅Pr[Qi]Pr[r(t)].
- The MAP decision rule can thus be reformulated or simplified as follows: Let the sink symbol sequence V=Qj, if for all i≠j holds:
- Pr[r(t)|Qj]⋅Pr[Qj)Pr[r(t)]>Pr[r(t)|Qi]⋅Pr[Qi]Pr[r(t)]⇒Pr[r(t)|Qj]⋅Pr[Qj]>Pr[r(t)|Qi]⋅Pr[Qi].
A further simplification of this MAP decision rule leads to the "ML receiver", where "ML" stands for "maximum likelihood".
Definition: The maximum likelihood receiver (abbreviated ML) decides according to the conditional forward probabilities Pr[r(t)|Qi], and sets the output sequence V=Qj, if for all i≠j holds:
- Pr[r(t)|Qj]>Pr[r(t)|Qi].
A comparison of these two definitions shows:
- For equally probable source symbols, the "ML receiver" and the "MAP receiver" use the same decision rules. Thus, they are equivalent.
- For symbols that are not equally probable, the "ML receiver" is inferior to the "MAP receiver" because it does not use all the available information for detection.
Example 1: To illustrate the "ML" and the "MAP" decision rule, we now construct a very simple example with only two source symbols (M=2).
⇒ The two possible symbols Q0 and Q1 are represented by the transmitted signals s=0 and s=1.
⇒ The received signal can – for whatever reason – take three different values, namely r=0, r=1 and additionally r=0.5.
Note:
- The received values r=0 and r=1 will be assigned to the transmitter values s=0 (Q0) resp. s=1 (Q1), by both, the ML and MAP decisions.
- In contrast, the decisions will give a different result with respect to the received value r=0.5:
- The maximum likelihood (ML) decision rule leads to the source symbol Q0, because of:
- Pr[r=0.5|Q0]=0.4>Pr[r=0.5|Q1]=0.2.
- The maximum–a–posteriori (MAP) decision rule leads to the source symbol Q1, since according to the incidental calculation in the graph:
- Pr[Q1|r=0.5]=0.6>Pr[Q0|r=0.5]=0.4.
Maximum likelihood decision for Gaussian noise
We now assume that the received signal r(t) is additively composed of a useful component s(t) and a noise component n(t), where the noise is assumed to be Gaussian distributed and white ⇒ "AWGN noise":
- r(t)=s(t)+n(t).
Any channel distortions are already applied to the signal s(t) for simplicity.
The necessary noise power limitation is realized by an integrator; this corresponds to an averaging of the noise values in the time domain. If one limits the integration interval to the range t1 to t2, one can derive a quantity Wi for each source symbol sequence Qi, which is a measure for the conditional probability Pr[r(t)|Qi]:
- Wi=∫t2t1r(t)⋅si(t)dt−1/2⋅∫t2t1s2i(t)dt=Ii−Ei/2.
This decision variable Wi can be derived using the k–dimensionial "joint probability density" of the noise (with k→∞) and some boundary crossings. The result can be interpreted as follows:
- Integration is used for noise power reduction by averaging. If N binary symbols are decided simultaneously by the maximum likelihood detector, set t1=0 and t2=N⋅T for distortion-free channel.
- The first term of the above decision variable Wi is equal to the "energy cross-correlation function" formed over the finite time interval NT between r(t) and si(t) at the time point τ=0:
- Ii=φr,si(τ=0)=∫N⋅T0r(t)⋅si(t)dt.
- The second term gives half the energy of the considered useful signal si(t) to be subtracted. The energy is equal to the auto-correlation function (ACF) of si(t) at the time point τ=0:
- Ei=φsi(τ=0)=∫N⋅T0s2i(t)dt.
- In the case of a distorting channel, the channel impulse response hK(t) is not Dirac-shaped, but for example extended to the range −TK≤t≤+TK. In this case, t1=−TK and t2=N⋅T+TK must be used for the integration limits.
Matched filter receiver vs. correlation receiver
There are various circuit implementations of the maximum likelihood (ML) receiver.
⇒ For example, the required integrals can be obtained by linear filtering and subsequent sampling. This realization form is called matched filter receiver, because here the impulse responses of the M parallel filters have the same shape as the useful signals s0(t), ... , sM−1(t).
- The M decision variables Ii are then equal to the convolution products r(t)⋆si(t) at time t=0.
- For example, the "optimal binary receiver" described in detail in the chapter "Optimization of Baseband Transmission Systems" allows a maximum likelihood (ML) decision with parameters M=2 and N=1.
⇒ A second realization form is provided by the correlation receiver according to the following graph. One recognizes from this block diagram for the indicated parameters:
- The drawn correlation receiver forms a total of M=8 cross-correlation functions between the received signal r(t)=sk(t)+n(t) and the possible transmitted signals si(t), i=0, ... , M−1. The following description assumes that the useful signal sk(t) has been transmitted.
- This receiver searches for the maximum value Wj of all correlation values and outputs the corresponding sequence Qj as sink symbol sequence V. Formally, the ML decision rule can be expressed as follows:
- V=Qj,ifWi<Wjforalli≠j.
- If we further assume that all transmitted signals si(t) have same energy, we can dispense with the subtraction of Ei/2 in all branches. In this case, the following correlation values are compared (i=0, ... , M−1):
- Ii=∫NT0sj(t)⋅si(t)dt+∫NT0n(t)⋅si(t)dt.
- With high probability, Ij=Ik is larger than all other comparison values Ij≠k ⇒ right decision. However, if the noise n(t) is too large, also the correlation receiver will make wrong decisions.
Representation of the correlation receiver in the tree diagram
Let us illustrate the correlation receiver operation in the tree diagram, where the 23=8 possible source symbol sequences Qi of length N=3 are represented by bipolar rectangular transmitted signals si(t).
The possible symbol sequences Q0=LLL, ... , Q7=HHH and the associated transmitted signals s0(t), ... , s7(t) are listed below.
- Due to bipolar amplitude coefficients and the rectangular shape ⇒ all signal energies are equal: E0=...=E7=N⋅EB, where EB indicates the energy of a single pulse of duration T.
- Therefore, the subtraction of the Ei/2 term in all branches can be omitted ⇒ the decision based on the correlation values Ii gives equally reliable results as maximizing the corrected values Wi.
Example 2: The graph shows the continuous-valued integral values, assuming the actually transmitted signal s5(t) and the noise-free case. For this case, the time-dependent integral values and the integral end values:
- ii(t)=∫t0r(τ)⋅si(τ)dτ=∫t0s5(τ)⋅si(τ)dτ⇒Ii=ii(3T).
The graph can be interpreted as follows:
- Because of the rectangular shape of the signals si(t), all function curves ii(t) are rectilinear. The end values normalized to EB are +3, +1, −1 and −3.
- The maximum final value is I5=3⋅EB (red waveform), since signal s5(t) was actually sent. Without noise, the correlation receiver thus naturally always makes the correct decision.
- The blue curve i1(t) leads to the final value I1=−EB+EB+EB=EB, since s1(t) differs from s5(t) only in the first bit. The comparison values I4 and I7 are also equal to EB.
- Since s0(t), s3(t) and s6(t) differ from the transmitted s5(t) in two bits, I0=I3=I6=−EB. The green curve shows s6(t) initially increasing (first bit matches) and then decreasing over two bits.
- The purple curve leads to the final value I2=−3⋅EB. The corresponding signal s2(t) differs from s5(t) in all three symbols and s2(t)=−s5(t) holds.
Example 3: The graph describes the same situation as Example 2, but now the received signal r(t)=s5(t)+n(t) is assumed. The variance of the AWGN noise n(t) here is σ2n=4⋅EB/T.
One can see from this graph compared to the noise-free case:
- The curves are now no longer straight due to the noise component n(t) and there are also slightly different final values than without noise.
- In the considered example, the correlation receiver decides correctly with high probability, since the difference between I5 and the next value I7 is relatively large: 1.65⋅EB.
- The error probability in this example is not better than that of the matched filter receiver with symbol-wise decision. In accordance with the chapter "Optimization of Baseband Transmission Systems", the following also applies here:
- pS=Q(√2⋅EB/N0)=1/2⋅erfc(√EB/N0).
Conclusions:
- If the input signal does not have statistical bindings (Example 2), there is no improvement by joint decision of N symbols over symbol-wise decision
⇒ pS=Q(√2⋅EB/N0). - In the presence of statistical bindings (Example 3), the joint decision of N symbols noticeably reduces the error probability, since the maximum likelihood receiver takes the bindings into account.
- Such bindings can be either deliberately created by transmission-side coding (see the LNTwww book "Channel Coding") or unintentionally caused by (linear) channel distortions.
- In the presence of such "intersymbol interferences", the calculation of the error probability is much more difficult. However, comparable approximations as for the Viterbi receiver can be used, which are given at the end of the next chapter.
Correlation receiver with unipolar signaling
So far, we have always assumed binary bipolar signaling when describing the correlation receiver:
- aν={+1−1f¨urforqν=H,qν=L.
Now we consider the case of binary unipolar digital signaling holds:
- aν={10forforqν=H,qν=L.
The 23=8 possible source symbol sequences Qi of length N=3 are now represented by unipolar rectangular transmitted signals si(t).
Listed on the right are the eight symbol sequences and the transmitted signals
- Q0=LLL, ... , Q7=HHH,
- s0(t), ... , s7(t).
By comparing with the "corresponding table" for bipolar signaling, one can see:
- Due to the unipolar amplitude coefficients, the signal energies Ei are now different, e.g. E0=0 and E7=3⋅EB.
- Here the decision based on the integral values Ii does not lead to the correct result. Instead, the corrected comparison values Wi=Ii−Ei/2 must now be used.
Example 4: The graph shows the integral values Ii, again assuming the actual transmitted signal s5(t) and the noise-free case. The corresponding bipolar equivalent was considered in Example 2.
For this example, the following comparison values result, each normalized to EB:
- I5=I7=2,I1=I3=I4=I6=1,I0=I2=0,
- W5=1,W1=W4=W7=0.5,W0=W3=W6=0,W2=−0.5.
This means:
- When compared in terms of maximum Ii values, the source symbol sequences Q5 and Q7 would be equivalent.
- On the other hand, if the different energies (E5=2, E7=3) are taken into account, the decision is clearly in favor of the sequence Q5 because of W5>W7.
- The correlation receiver according to Wi=Ii−Ei/2 therefore decides correctly on s(t)=s5(t) even with unipolar signaling.
Exercises for the chapter
Exercise 3.09: Correlation Receiver for Unipolar Signaling
Exercise 3.10: Maximum Likelihood Tree Diagram