Contents
Considered scenario and prerequisites
The "correlation receiver" is optimal in the sense of the "maximum likelihood decision rule", that is, it leads to the minimum error probability for equally likely source symbols. Disadvantage:
- The realization cost increases exponentially with the length N of the symbol sequence to be detected.
- Since the sequence is decided jointly, long delays occur when N is large.
In the 1970s, Andrew J. Viterbi proposed a maximum likelihood receiver that allows detection of parts of the received message and for which the implementation cost is limited even for infinitely long sequences.
It should be noted with regard to the individual components of the block diagram:
- Adjusted to the basic reception pulse gr(t) and the noise, the "matched filter" HMF(f) serves for "noise power limitation". The matched filter output signal m(t) or the sequence ⟨mν⟩ of equidistant signal values after sampling has the best possible signal-to-noise power ratio (SNR).
- The task of the decorrelation filter HDF(f) is to extract from the sequence ⟨mν⟩ the detection samples dν=dSν+dNν whose noise components dNν are uncorrelated. This filter is therefore also called a "whitening filter".
- The Viterbi decision, which is the focus of the following considerations, obtains the binary output sequence ⟨vν⟩ from the continuous-valued input sequence ⟨dν⟩ according to the maximum likelihood rule with the smallest possible error probability Pr(vν≠qν).
In order to describe the Viterbi algorithm as simply as possible, some simplifying assumptions are made here:
- Let the amplitude coefficients be unipolar ⇒ aν∈{0,1}. Note that using bipolar coefficients ⇒ aν∈{−1,+1} requires only a few modifications. These difficulties concern the description rather than the implementation.
- The basic pulse gd(t) after the decorrelation filter HDF(f) consists only of the main value g0=gd(t=TD) and a single precursor g−1=gd(t=TD−T). There are no trailers ("postcursors") in our model: gν=gd(t=TD+ν⋅T)≡0. Also, let other precursors be excluded for now: g−2=g−3= ...=0.
- This gives for the samples of the continuous decorrelation filter output signal d(t) at time ν⋅T: dν=aν⋅g0+aν+1⋅g−1+dNν, where the noise component dNν is assumed to be Gaussian distributed with standard deviation σd.
Note: The algorithm is not more complex with bipolar signaling. ∙ In contrast, the computational cost increases when the basic detection pulse becomes broader and has more than one precursor g−1. ∙ The neglect of precursors in the description is not a fundamental limitation because any pulse can satisfy this condition by appropriate choice of the detection time TD. ∙ In the following all signal values are normalized to 1.
Example 1: In the graph, the samples dSν of the noiseless detection signal are entered as (blue) crosses, where the corresponding amplitude coefficients a1=1, a2=1, a3=0, ... can be read from the source signal q(t) plotted in green.
- The basic pulse values are assumed to be g0=0.7 and g−1=0.3. It can be further seen that dSν can take only four different values, namely 0, g−1=0.3, g0=0.7, g0+g−1=1.
- The noisy samples present at the Viterbi decision (red dots) are d0=0.2, d1=0.7, d2=0.5, d3=0, ... , where the differences dNν=dν−dSν originate from an AWGN noise source.
- A threshold decision (with the threshold at E=0.5) would make at least one wrong decision at these ten bits shown (at time t=4T), and possibly another at t=2T if the sample d2=0.5 is nevertheless slightly smaller than the threshold E=0.5.
- In contrast, the Viterbi receiver will correctly decide this sequence of length 10, as shown in the next sections.
Metric and accumulated metric
As in the chapter "Optimal Receiver Strategies", Q∈{Qi} specifies the time-limited source symbol sequence consisting of N binary symbols.
- Thus, the number of possible symbol sequences Qi is 2N.
- V denotes the sink symbol sequence of length N set equal to the most probable sequence Qj by the Viterbi decision.
Definitions:
- The metric εν(i) denotes the squared deviation between the actual noisy sample dν and the noiseless sample dSν belonging to the sequence Qi:
- εν(i)=|dν−dSν(i)|2.
- In the literature, this is sometimes referred to as error quantity.
- The accumulated metric γν(i) characterizes the sum of all metrics up to time ν:
- γν(i)=ν∑k=0εk(i).
- In this context, one also speaks of the total error quantity.
The graph illustrates the quantities defined above in a tree structure, from which it can be seen that the accumulated metric can be calculated iteratively:

∙ i, i′, i″ are different index variables.
∙ Definition is valid for a basic pulse with main value g0 and precursor g−1.
∙ With v precursors the above sum would have to start at k=1−v.
∙ The parameter i∈{0,...,2N+1−1} is usually represented in binary.
∙ It describes the amplitude coefficients a1, ... , aν+1 (each 0 or 1).
- γν(i)=γν−1(i′)+εν(i″).
It should be noted about this graph:
- The nodes of the tree diagram represent the "accumulated metric" γν(i). Their number is doubled with each iteration step. At time ν, there are 2ν+1 such nodes. For example, for ν=3, exactly 24=16 nodes can be seen.
- The "amplitude coefficients" associated with the accumulated metric γν(i) are obtained by following the path from the initial node to the node under consideration. It is agreed that an upward branch is assigned the coefficient aν=1 and a downward branch is assigned the coefficient aν=0.
- The green highlighted node γ3(1100) denotes the accumulated metric under the hypothetical assumption that the symbols a1=1, a2=1, a3=0, a4=0 were sent. This assignment can also be read from the directions of the arrows in the tree diagram: First twice upwards, then twice downwards.
- Due to the precursor, the coefficient a4 must already be taken into account at time ν=3. All nodes γν(i) computed under the assumption aν+1=1 are represented by rectangles, while the hypothesis aν+1=0 is symbolized by a rounded rectangle in each case, for example γ2(110) or γ3(1100).
- The branches in the tree diagram are assigned to the "metrics" εν(i). With the assumed basic pulse (only g0 and g−1), there are exactly four different metrics at each time point except for the initial state (ν=0):
- εν(00)=|dν|2,εν(01)=|dν−g−1|2,
- εν(10)=|dν−g0|2,εν(11)=|dν−g0−g−1|2.
- The accumulated metric γν(i) is the sum of the preceding node γν−1(i′) and the intervening branch εν(i″). For example, for the highlighted nodes:
- γ1(11)=γ0(1)+ε1(11),
- γ2(110)=γ1(11)+ε2(10),
- γ3(1100)=γ2(110)+ε3(00).
- For the first nodes γ0(0) and γ0(1), it is considered that, by convention, the symbol a0=0 is always transmitted before the actual transmission (a1, a2, ...). It follows that:
- γ0(0)=ε0(00)=|d0|2,γ0(1)=ε0(01)=|d0−g−1|2.
The following two examples will hopefully clarify these somewhat tedious statements.
Example 2: As in Example 1, we consider the unipolar source symbol sequence of length N=3 with the following parameter values:
- g0=0.7,g−1=0.3,d0=0.2,d1=0.7,d2=0.5,d3=0.
Then, for the metrics εν(i) at time points ν=0 to ν=3:
- ν=0:ε0(00)=[0.2−(0⋅0.7+0⋅0.3)]2=0.04,ε0(01)=[0.2−(0⋅0.7+1⋅0.3)]2=0.01;
- ν=1:ε1(00)=[0.7−(0⋅0.7+0⋅0.3)]2=0.49,ε1(01)=[0.7−(0⋅0.7+1⋅0.3)]2=0.16,
- ε1(10)=[0.7−(1⋅0.7+0⋅0.3)]2=0.00ε1(11)=[0.7−(1⋅0.7+1⋅0.3)]2=0.09;
- ν=2:ε2(00)=[0.5−(0⋅0.7+0⋅0.3)]2=0.25,ε2(01)=[0.5−(0⋅0.7+1⋅0.3)]2=0.04,
- ε2(10)=[0.5−(1⋅0.7+0⋅0.3)]2=0.04,ε2(11)=[0.5−(1⋅0.7+1⋅0.3)]2=0.25;
- ν=3:ε3(00)=[0.0−(0⋅0.7+0⋅0.3)]2=0.00,ε3(01)=[0.0−(0⋅0.7+1⋅0.3)]2=0.09,
- ε3(10)=[0.0−(1⋅0.7+0⋅0.3)]2=0.49,ε3(11)=[0.0−(1⋅0.7+1⋅0.3)]2=1.00.
Example 3: With the metrics εν(i) determined in Example 2, the accumulated metrics γν(i) can now also be calculated. Below is the tree diagram with the accumulated metrics γν(i) as nodes for the time points ν=0 to ν=3.
- The minimum accumulated metric at time ν=3 is γ3(1100)=0.14. From this, the coefficients of the most likely transmitted (unipolar) sequence according to the present signal values d0=0.2, d1=0.7, d2=0.5 and d3=0 are a1=1, a2=1 and a3=0 (green path).
- If the sequence length is N=3 (that is: only three symbols are jointly decided by the Viterbi receiver), the decision a4=0 is also certainly the correct one, since all coefficients aν>3 were assumed to be zero.
- However, for longer sequences (N>3), it cannot necessarily be concluded from the minimum value γ3(1100) that a1=1, a2=1, a3=0 is actually part of the most probable sequence.
- Considering further sample values (d4, d5, ...) this preliminary result could well change.
Minimum accumulated metric
We continue from the numerical values of the last examples:
- d0=0.2,d1=0.7,d2=0.5,d3=0,g0=0.7,g−1=0.3.
Thus, nothing changes in the tree diagram compared to Example 3. Some colored markings in this diagram only indicate in which way the tree diagram can be simplified according to Viterbi's suggestions. Important properties of the Viterbi decision can be seen in the graph, e.g. at time ν=2:
- At time ν=2, the minimum accumulated metric is γ2(101)=0.05 (highlighted in brown).
- This means: A decision at ν=2 – based on d0, d1 and d2 – would have been in favor of sequence "101" instead of the transmitted sequence "110".
- It follows: A too early final determination should be avoided at all costs. However, at any time ν one can already exclude several partial symbol sequences which need not be considered at later times.
- To γ2(111)=0.35 the same metrics ε3(11) and ε3(10) are added as to γ2(101)=0.05. Because of γ2(111)>γ2(101) it is therefore already certain at ν=2 that "111" cannot be part of the most probable sequence.
- The same is true for "001" and "011". Therefore, these nodes and all their successors need not be considered further (brown coverages).
- Also, the rounded nodes γ2(000)=0.78, γ2(010)=0.24, γ2(100)=0.26 are not part of the most probable sequence, since they are larger than γ2(110)=0.14, which is marked in green. Also, these and their successors need not be considered from time ν=3 (green covering).
Conclusion:
- Of the eight nodes at ν=2, only two need to be followed up, namely the rectangle γ2(101)=0.05 and the rounded rectangle γ2(110)=0.14.
- These describe the minimum accumulated metrics assuming that a3=0 and a3=1, respectively.
Representation with the trellis diagram
We now generalize the result of this example. Under the still valid assumption that the basic pulse gd(t) has only one precursor (g−1) besides the main value (g0), the two minimum accumulated metrics at time ν are formally given by
- Γν(0)=Min[Γν−1(0)+εν(00),Γν−1(1)+εν(10)],
- Γν(1)=Min[Γν−1(0)+εν(01),Γν−1(1)+εν(11)].
The procedure "minimization of the accumulated metrics" can be clearly illustrated in the trellis diagram.
A comparison with the figures in the last sections shows:
- The lower branch represents the minimum accumulated metric Γν(0) calculated at each time ν under the hypothesis aν+1=0 (blue rounded squares).
- The upper branch describes the minimum accumulated metric Γν(1) under the assumption aν+1=1 (red squares).
- Besides Γν(0) and Γν(1), the maximum likelihood decision must also store the associated symbol sequences ("paths"). These branches are highlighted in red and blue, resp.
- If Γν arises from the node Γν−1(0) – that is, if the lower of the two incoming branches is highlighted – the associated symbol is "0", otherwise the symbol is "1".
- For ν=3: Both, Γ3(0)=0.14 and Γ3(1)=0.23, result from the precursor Γ2(0), so that both selected paths point to the symbol "0" (blue branches).
Simplified trellis diagram
The advantage of the trellis diagram is that the number of nodes and branches does not double at each iteration step. By selecting the minimum accumulated metrics, only those symbol sequences are considered further which could be part of the most probable sequence.
The trellis diagram can be further simplified by drawing in only the selected branches. This is illustrated in the lower part of the diagram with our numerical example:
- The actual transmitted amplitude coefficients are a1=1, a2=1 and a3=0.
- The trellis diagram drawn above was calculated assuming that due to the pulse values g−1=0.3 and g0=0.7 as well as AWGN noise.
- The noisy input values d0=0.2, d1=0.7, d2=0.5 and d3=0 are applied to the maximum likelihood decision.
This simplified trellis diagram allows the following statements:
- The decision a1=1 can be made immediately at time ν=1, since under both hypotheses – the following symbol is a2=0 or this symbol is a2=1 – the same result is obtained.
- At time ν=2 the final decision about a2 cannot yet be made. Assuming that a3=0, one would have to decide for a2=1, while a3=1 would lead to the fixed choice of a2=0.
- At ν=3 the decisions for a1=1, a2=1, a3=0 are final, since both continuous paths (blue and red) suggest this (in this case correct) sequence.
- If one were to postpone the decision to time ν=4, one would not have more usable information about a1, a2 and a3.
Extension to two precursors
If the basic pulse gd(t) is described by the samples g0≠0, g−1≠0 and g−2≠0, then four metrics Γν(00), Γν(01), Γν(10) and Γν(11) must be determined in the trellis diagram at each time ν.
For example, Γν(01) describes the minimum accumulated metric for the detection of the symbol aν under the hypothesis "aν−1=0 and aν−2=1". It holds:
- Γν(01)=Min[Γν−1(00)+εν(001),Γν−1(10)+εν(101)].
Exercise 3.12 discusses in detail the metrics calculation and the minimization of the accumulated metrics for the case of two precursors.
Example 4: We consider here an exemplary trellis diagram, which reflects the (error-free) detection of the following symbol sequence:
- a1=0,a2=0,a3=1,a4=1,a5=1,a6=0,...
This simplified trellis diagram allows the following statements:
- All branches departing from Γν(00) or Γν(01) are assigned to symbol "0" and drawn in blue. The red branches departing from the two upper states denote symbol "1".
- If one follows the continuous paths, one recognizes the indicated sequence. Since at time ν=6 only blue branches arrive, the first six bits of the sequence are finally fixed here.
- However, partial sequences could also be output already at times ν=1, ν=3 and ν=4, since at these times the same partial paths result for all four states.
- In contrast, at ν=2 and ν=5, decisions may not be made immediately. For example, at ν=5 it is only certain that either "a5=0, a6=1" will be, or "a5=1, a6=0".
Bit error probability with maximum likelihood decision
The exact bit error probability calculation for binary maximum likelihood decision (e.g. with the correlation or the Viterbi receiver) is very complex. Thereby
- determine the difference energies between all possible symbol sequences Qi and Qj≠i, where the error probability is essentially determined by the minimum difference energy;
- also the influences of matched filter HMF(f) and decorrelation filter HDF(f) have to be taken into account and additionally the rms value σd of the detection noise signal is determined.
Without proof: A simple approximation for the (average) bit error probability in maximum likelihood decision is:
- pML=Q(gmax/σd)withgmax=Max|gν|.
- The approximation is valid only for binary bipolar signaling. For unipolar signaling, the argument of the Q–function must be halved.
For the following interpretation and example, it is assumed that v+1 basic pulse values (including the main value) are different from zero. Then holds:
- The Viterbi decision must consider all these basic pulse values. This means that a trellis diagram with 2v states has to be processed.
- The precondition for the validity of the above equation is the uncorrelatedness of the noise at the decision, which is achieved by the decorrelation filter.
- For the comparison with threshold decision (pThD) or decision feedback (pDFE), the noise rms value σd is assumed to be constant.
- Optimization of the maximum likelihood system results in very narrowband filters, since all pulse tails can be factored out by the ML algorithm.
- Thus, with constant noise power density N0 (at receiver input), the noise rms value σd (at decision) is for the ML system smaller than for other variants.
- This means: The signal-to-noise ratio gain by the ML decision may be even larger than the following example (with σd=const.) expresses.
Example 5: We consider the basic pulse values g0=0.6 and g−1=g+1=0.2 (=g1). In addition, we assume the constant noise rms value σd=0.1. For simplicity, these quantities are normalized.
- For a binary receiver with simple threshold decision (ThD) according to the chapter "Error Probability with Intersymbol Interference", the (average) bit error probability for bipolar signaling is:
- pThD=1/4⋅Q(g0+2⋅g1σd)+1/2⋅Q(g0σd)+1/4⋅Q(g0−2⋅g1σd)≈1/4⋅Q(g0−2⋅g1σd)=Q(2)4=0.57%.
- For the receiver with decision feedback equalization (DFE), considering g+1=0 (full compensation of the "postcursor"), we obtain:
- pDFE=1/2⋅Q(g0+g−1σd)+1/2⋅Q(g0−g−1σd)≈1/2⋅Q(g0−g−1σd)=Q(4)4=0.16⋅10−4.
- In contrast, the application of the maximum likelihood decision with "correlation receiver" or "Viterbi receiver" (ML) leads to the much smaller error probability
- pML=Q(g0σd)≈Q(6)=10−9.
- This corresponds to a SNR gain of about 3 dB (versus DFE) and 7.5 dB (versus ThD) over the other two systems. The result of this simple approximation was essentially confirmed by simulations.
Note:
- To apply the described Viterbi algorithm directly, the (normalized) basic pulse values g0=0.2, g−1=0.6 and g−2=0.2 must be set in the equations.
- Indeed, a time shift by multiples of the symbol duration T with respect to the coordinate system does not change the performance of the Viterbi decision.
- The maximum likelihood error probability according to the above equation depends solely on the largest basic pulse value.
- It is quite possible that a "precursor" (here: g−1=0.6) is larger than the main value g0=0.2. In case of threshold decision this would lead to a "closed eye".
Conclusion:
- Maximum likelihood decision is only advantageous in the presence of intersymbol interference.
- In the case of Nyquist equalization (that is: only the basic pulse value g0 is non-zero) the maximum likelihood receiver also operates symbolically and with the same bit error probability Q(g0/σd) as the other receivers.
Exercises for the chapter
Exercise 3.11: Viterbi Receiver and Trellis Diagram
Exercise 3.11Z: Metric and Accumutated Metric
Exercise 3.12: Trellis Diagram for Two Precursors
Exercise 3.13: Threshold Decision vs. DFE vs. Maximum Liekelihood