The present invention relates to mechanisms for maximum a posteriori (MAP) decoding of signals encoded using error correction codes, and, more particularly, to MAP decoding of signals encoded using binary Hamming codes and related error correction codes.
Error correction codes are often employed in communication systems to provide increased resilience to channel noise. Generally, error correction encoding techniques typically employ convolutional, block or concatenated coding of the data before the signal is modulated so that errors introduced by noise and interference on the channel may be corrected. One popular family of linear low rate error correction codes having large minimum distances is first order Reed-Muller codes. Each family of error correction codes has a dual code. The dual code for binary Reed-Muller codes are the binary Hamming codes.
The received signal, which is typically corrupted by the noise and interference on the channel, must be decoded at the receiver. Typically, maximum a posteriori decoding techniques are applied to decode signals encoded using error correction codes to make maximum probability decisions about each transmitted bit. For a detailed discussion of Hamming codes and the decoding thereof, see, for example, T. Kasami et al., “On the Optimum Bit Orders With Respect to the State Complexity of Trellis Diagrams for Binary Linear Codes,” IEEE Trans. Inf. Theory, vol. IT-39, 242–45 (January, 1993) or T. Kasami et al, “On Complexity of Trellis Structure of Linear Block Codes,” IEEE Trans. Inf. Theory, vol. IT-39, 1057–64 (January, 1993), incorporated by reference herein.
One particular implementation of the MAP decoding algorithm, commonly referred to as the BCJR algorithm, was proposed in L. R. Bahl, J. Cocke, F. Jelinek and J. Raviv, “Optimal Decoding of Linear Codes for Minimizing Symbol Error Rate”, IEEE Trans. Inform. Theory, V. IT-20, 284–287 (March, 1974), incorporated by reference herein. The BCJR algorithm is a symbol-by-symbol decoder based on a trellis representation of a code. The BCJR algorithm for binary Hamming codes has a complexity that is proportional to n2, where n is the code length. A need exists for a method and apparatus for decoding Hamming and extended Hamming codes having reduced complexity.
Generally, a method and apparatus are disclosed for MAP decoding of signals encoded using error correction codes to make maximum probability decisions about each transmitted bit. The disclosed MAP decoding algorithm exploits properties of Hamming error correction codes to provide a decoding algorithm with significantly reduced complexity. While conventional decoding algorithms for binary Hamming codes, such as the Hartman-Rudolph algorithm, have a complexity that is proportional to n3, where n is the code length and the BCJR algorithm has complexity proportional to n2, the disclosed decoding algorithm has a complexity that is proportional to n log n for Hamming and extended Hamming codes.
The disclosed MAP decoding algorithm determines the probability that a given symbol is each possible value and selects a value for each symbol based on the value having the maximum probability. The disclosed MAP decoding algorithm extends the work of Hartman and Rudolph. A first portion of the Hartman-Rudolph decoding algorithm computes a difference, ρ, in the probability of receiving a binary zero and binary one based on characteristics of the channel and received signal and then, for each codeword of the dual code, determines the product of the ρl values corresponding to non-zero positions of codewords of the dual code, in a known manner. As used herein, the symbol “ρ” indicates the set of ρl values for each of the l positions of a codeword. According to one feature of the invention, a more efficient technique for computing the product of the ρ values corresponding to non-zero positions of codewords of the dual code is provided. Generally, the present invention recognizes that in the case of Hamming codes, a real vector and a specially designed 2[2]-vector fast Walsh-Hadamard transform can be used to more efficiently compute the product of the ρl values corresponding to non-zero positions of codewords.
A second portion of the conventional Hartman-Rudolph decoding algorithm processes all positions to determine a sum of modified products of ρ values for each position whose sign indicates the most probable value of a transmitted symbol for a given position. Generally, the second portion of the Hartman-Rudolph decoding algorithm goes through each position of the code and modifies the products of the ρl values (computed by the first aspect) and then sums the modified products of ρl values. For a given position of each dual codeword, the initial product of ρl values (computed by the first aspect) is (i) divided by the p, value for the current position if the dual codeword has a binary one (1) in the current position, and (ii) multiplied by the ρl value for the current position if the dual codeword has a binary zero (0) in the current position. According to another feature of the invention, a more efficient technique for computing the sum of modified products of the ρl values for each position is provided. Generally, the present invention recognizes that in the case of Hamming codes, the real vector fast Walsh-Hadamard transforms can be used to more efficiently compute the sum of modified products of ρl values for each position.
A more complete understanding of the present invention, as well as further features and advantages of the present invention, will be obtained by reference to the following detailed description and drawings.
The present invention recognizes that for extended Hamming codes and Hamming codes, the complexity of conventional MAP decoding algorithms can be significantly reduced. As discussed more fully below, the disclosed MAP decoding algorithm 200 for Hamming codes has a complexity that is proportional to nlog2 n. While the present invention is illustrated herein using exemplary Hamming codes, the present invention may be applied to any error correction code for which the dual of the code may be represented as a product of sparse matrices, such as Reed-Muller codes. For a detailed discussion of a system and method for decoding Reed-Muller codes, see, for example, U.S. patent application Ser. No. 10/112,574, filed Mar. 29, 2002, entitled “Method and Apparatus for MAP Decoding of First-Order Reed-Muller Codes and Related Error Correction Codes,” incorporated by reference herein.
Let q be the Galois field of size q (i.e., a finite field of q elements) and let C be an [n,k] linear code over q. It is assumed that codewords of C are transmitted through a memoryless channel 120. Let Q be an output sampling space, either finite or infinite, of the channel 120. In the case of finite Q, it is assumed that conditional probabilities Pr(y|α) of receiving symbol yεQ under the condition that symbol αεq has been transmitted are known. If Q is infinite, then y is a continuous random variable whose density function, say ƒα(y), is defined by a transmitted symbol αεq. For instance, if a binary code with binary modulation is used in a channel 120 with white additive Gaussian noise then we have two density functions:
where
and is the signal to noise ratio. It is assumed that functions ƒα(y) are known for all αεFq. Since a memoryless channel is assumed, the conditional probability of receiving a vector yεQn under the condition that a vector cεFqn has been transmitted, can be written as the following product:
It is also assumed that all codewords have the same probability to be transmitted and therefore this probability equals
It is assumed that a vector c=(c0,c1, . . . , Cn−1)εC is transmitted and a vector yεQn is received. The goal is to determine what was the j-th transmitted symbol. To do this, the conditional probabilities Pr(cj=α|y) are estimated for all αεq and the maximum one is selected.
Let Cj(α)={c=(c0,c1, . . . , cn−1)εC:cj=α}, i.e., Cj)α) is the set of all codewords whose j-th symbol is α. In the case of finite Q, using Bayes rule and equation (1), the conditional probabilities can be expressed as follows:
Using similar arguments, in the case of infinite Q, the conditional probabilities can be expressed as follows:
The denominator of equations (2) and (3) is the same for αεq. Therefore, for MAP decoding, it is enough to estimate only the sum of products of Pr(yi|bj) or ƒb
Hereinafter, it is assumed that Q is finite and the conditional probabilities Pr(yi|a) will be employed. In the case of infinite Q, one has to replace the conditional probabilities with density functions ƒα(yj).
In what follows we will denote by ⊕ summation by mod 2. Let C be an [n, k] linear code over 2, and let Cbe the dual of C. It is assumed that code words of Care enumerated and denoted by c′, i.e., cij′ is the j-th element of the i-th codeword of C. In what follows we will call code words of Cdual code words.
C. Hartman and L. Rudolph developed an algorithm for the MAP decoding of a linear code using its dual code. See, R. P. Hartman and L. D. Rudolph, “An Optimum Symbol-by-Symbol Decoding Rule for Linear Codes,” IEEE Trans. Inf. Theory, vol. IT-22, 514–17 (September 1976), incorporated by reference herein. In particular, Hartman and Rudolph proved the following theorem:
Theorem 1:
Generally, Hartman and Rudolph provided an algorithm that uses the dual code to decode a given code.
Let:
It was noted by Hartman and Rudolph that dividing equation (5) by
and further by
and using identity, as follows:
the following expression is obtained:
Similarly, to evaluate the sum of probabilities, which is proportional to the sum of Aj(0) and Aj(1), the following is obtained:
From equations (6) and (7), it follows that
It is noted that the sign of the difference in probabilities (8) can be evaluated to determine if the received data is a binary zero (0) or binary one (1), and the magnitude of the difference in probabilities can be evaluated for an indication of the reliability. The present invention provides techniques for computing equations 6 and 7 efficiently for Hamming codes. The numerator and denominator in Equation 8 can be considered as a sum of products of the ρl values.
Initially, a first portion of the Hartman-Rudolph decoding algorithm determines the product of the ρl values corresponding to non-zero positions of codewords of the dual code. According to one aspect of the present invention, a more efficient technique for computing the product of the ρl values corresponding to non-zero positions of codewords of the dual code is provided. Generally, the present invention recognizes that in the case of Hamming codes, the real vector and specially designed 2[2]-vector fast Walsh-Hadamard transforms can be used to more efficiently compute the product of the ρl values corresponding to non-zero positions of codewords.
Thereafter, a second portion of the Hartman-Rudolph decoding algorithm processes all positions to determine a sum of modified products of ρ values for each position whose sign indicates the most probable value of a transmitted symbol on a given position. Generally, the second portion of the Hartman-Rudolph decoding algorithm goes through each position of the code and modifies the products of the ρl values (computed by the first aspect) and then sums the modified products of ρl values. For a given position of each dual codeword, the initial product of ρl values (computed by the first aspect) is (i) divided by the ρl value for the current position if the dual codeword has a binary one (1) in the current position, and (ii) multiplied by the ρl value for the current position if the dual codeword has a binary zero (0) in the current position.
According to another aspect of the present invention, a more efficient technique for computing the sum of modified products of ρl values for each position is provided. Generally, the present invention recognizes that in the case of Hamming codes, the real vector fast Walsh-Hadamard transform can be used to more efficiently compute the sum of modified products of ρl values for each position.
The Hamming codes, m, and extended Hamming codes, m, are [2m-1,2m-1-m,3] and [2m, 2m-m-1,4], respectively. They are optimal, i.e., they have the largest possible minimum distance among all codes with the same length and size. The Hamming code m can be defined as the code whose parity check matrix consists of all nonzero binary m-tuples ordered lexicographically. The extended Hamming code m is the code whose parity check matrix is constructed from the parity check matrix of m by adding to the all-zero m-tuple and then adding the all-one row. For instance, if m equals 3, then parity check matrices of m and m are as follows:
respectively. The dual of the extended Hamming code is the first order Reed-Muller code (l, m) and the dual of the Hamming code is so called the simplex code.
Consider the meaning of expression (8) on the example of codes 3 and 3. The set of code words of the code 3 and 3, that are dual codes of 3 and 3, are
respectively. It is convenient to enumerate positions from 0 to 7 for 3 code and from 1 to 7 for 3.
Consider the first positions for 3 and 3. Then, according to expression (8) for 3, the following expression is obtained:
For 3, the following expression is obtained:
Let (δ0(0),δ1(0), . . . ,δn−1(0);δ0(1),δ1(1), . . . ,δn−1(1)) and (β0(0),β1(0), . . . ,βn−1(0);β0(1),β1(1), . . . , βn−1(1)) be binary vectors. The 2[2]-vector fast Walsh-Hadamard type transform, discussed below, may be expressed as follows:
As discussed hereinafter, a first portion of the MAP decoding algorithm 200 (steps 210 through 270) determines the product of the ρl values corresponding to non-zero positions of codewords of the dual code. As shown in
Thereafter, the natural log of each ρ is computed during step 220 to permit the efficient computation of the sum of logarithms of probabilities, as opposed to an equivalent product of probabilities, as follows:
σj=In|ρj|, j=0, . . . ,n−1;
Since the polarity (+/−) of each of the probability difference values, ρ, was masked by the absolute value employed during step 220, the following β values are computed during step 230 to maintain the polarity information:
The real vector Walsh-Hadamard transform is applied to the τ (magnitude) values during step 240 to find the sum of logs of ρ values, as follows:
(s0,S1, . . . ,sn−1)=(σ0,τ1, . . . ,σn−1).
The results of the real vector Walsh-Hadamard transform are then processed during step 250 to compute:
(ti(0) is the sum over all τj values, corresponding to non-zero positions of the n+i-th dual codeword and ti(1) is the sum over all τj values, corresponding to non-zero positions of the i-th dual codeword).
Since the exemplary MAP decoding algorithm 200 employs sums of logarithms of probabilities, as opposed to an equivalent product of probabilities, exponentiation is performed during step 260, to obtain the product of p corresponding to non-zero positions for all dual codewords, as follows:
hi(0)=t
Each h value is a product of the absolute value of ρ values corresponding non-zero entries for all codewords of first order Reed-Muller code. At this point, the MAP decoding algorithm 200 has products of absolute values. To prepare to reintroduce the polarity information and obtain the δj values, the specially designed 2[2]-vector Walsh-Hadamard transform is applied during step 265 to the β(polarity) values, as follows:
The polarity of the probability difference values is then reintroduced during step 270, as follows:
A second portion of the MAP decoding algorithm 200 (steps 275 through 290) processes all positions of all dual codewords to determine a sum for each position that indicates the reliability that a received bit is a given value for a given position. Generally, the second portion of the MAP decoding algorithm 200 goes through each position of the codes and modifies the product of the ρl values (computed by the first aspect) and then sums the modified product of ρl values. For a given position of each dual codeword, the initial product of ρl values (computed by the first aspect) is (i) divided by the ρl value for the current position if the dual codeword has a binary one (1) in the current position, and (ii) multiplied by the ρl value for the current position if the dual codeword has a binary zero (0) in the current position. For example, if the dual codeword is “1100,” the product of ρl values computed by the first portion of the algorithm is ρlρ2 and the modified product of ρl values for the codeword for the first position would be ρlρ2/ρ1. Similarly, if the codeword is “0110,” the product of ρl values computed by the first portion of the algorithm is ρlρ3 and the modified product of ρl values for the codeword for the first position would be ρlρ3×ρl.
In preparation for another real vector Walsh-Hadamard transform, the difference between h values is obtained during step 275, as follows:
xi=hi(0)−hi(1),i=0, . . . ,n−1.
and q is computed during step 280 to find the sum of all h values, as follows:
The real vector Walsh-Hadamard transform is applied during step 285, as follows:
(r0,r1, . . . , rn−1)=(x0, x1, . . . , xn−1).
The dual codewords are conceptually divided into two groups based on the binary value of the current position of each dual codeword, and the following values are computed during step 290:
where wj(0) and wj(1) are the sum of products of ρl values corresponding to non-zero positions of dual codewords with a binary zero (0) and a binary one (1) on the j-th position, respectively. Finally, a decision is made about the received bit during step 295, as follows:
It is noted that the first term in the numerator multiplies the initial product of ρl values (computed by the first aspect) by the ρl value for the current position (if the codeword has a binary zero (0) in the current position) and the second term in the numerator divides the initial product of ρl values (computed by the first aspect) by the ρl value for the current position (if the codeword has a binary one (1) in the current position). Thus, wj(0)ρj and wj(1)/ρj are sums of modified products of ρ values. This allows avoiding computations of modified products for all dual cadewords on step 295, thus contributing to the efficiency gains of the present invention.
The following discussion establishes that the MAP decoding algorithm 200 actually computes equation (8). Let i0 and i(1) be the sets of positions where the i-th codeword of (l, m) has zeros and ones respectively.
From the definition of one can easily see that Step 240 provides:
Noting that
the following is obtained:
From the definition of FWH
Thus, δj(0) and δj(1) provide signs of the products
and
respectively. Hence,
It can be seen that:
Taking into account the following equality:
the following is obtained:
Finally,
To examine the complexity of the algorithm, it is noted that complexities of all steps except steps 240, 270, and 285 are linear, the steps 240 and 270 require n log2n summations of real numbers, and step 270 requires 2n log2n binary summations. Therefore, the overall complexity of the algorithm is proportional to n log2n.
In the case of the Hamming code, m, a vector (y1, y2, . . . , yn) is received from the channel. Using arguments similar to those discussed above, one can easily see that the following algorithm provides a MAP decoding of the Hamming code and its complexity is proportional to n log2n.
As shown in
Thereafter, the natural log of each ρ is computed during step 320 to permit the efficient computation of the sum of logarithms of probabilities, as opposed to an equivalent product of probabilities, as follows:
τj=In|ρj|, j=0, . . . , n−1;
Since the polarity (+/−) of each of the probability difference values, ρ, was masked by the absolute value employed during step 320, the following β values are computed during step 330 to maintain the polarity information:
The real vector Walsh-Hadamard transform is applied to the τ (magnitude) values during step 340 to find the sum of logs of ρ values, as follows:
(s0,s1, . . . ,sn−1)=(τ0,τ1, . . . , τn−1).
The results of the real vector Walsh-Hadamard transform are then processed during step 350 to compute:
It is noted that the t(0) term used for the extended Hamming code has been dropped because the Hamming code has only half the number of codewords of the extended Hamming code.
The hi(0) term that was used for the extended Hamming code is set to zero during step 355 (again, because the Hamming code has only half the number of codewords of the extended Hamming code). Since the exemplary MAP decoding algorithm 300 employs sums of logarithms of probabilities, as opposed to an equivalent product of probabilities, exponentiation is performed during step 360, to obtain the product of ρ values for non-zero positions for all codewords, as follows:
hi(1)=et
Each h value is a product of the absolute value of ρ values with non-zero entries for all codewords of first order Reed-Muller code. At this point, the MAP decoding algorithm 300 has products of absolute values. To prepare to reintroduce the polarity information and obtain the δj values, a specially designed 2[2]-vector Walsh-Hadamard transform is again applied during step 365 to the p (polarity) values, as follows:
The polarity of the probability difference values is then reintroduced during step 370, as follows:
The second portion of the MAP decoding algorithm 300 (steps 375 through 390) proceeds in the same manner as the corresponding steps of the MAP decoding algorithm 200, discussed above in conjunction with
It is to be understood that the embodiments and variations shown and described herein are merely illustrative of the principles of this invention and that various modifications may be implemented by those skilled in the art without departing from the scope and spirit of the invention.
Number | Name | Date | Kind |
---|---|---|---|
5870414 | Chaib et al. | Feb 1999 | A |
5926488 | Khayrallah | Jul 1999 | A |
6487258 | Jedwab et al. | Nov 2002 | B1 |
Number | Date | Country | |
---|---|---|---|
20040088645 A1 | May 2004 | US |