Reed-Solomon (RS) codes are prevalent in many communication systems and data storage devices. They are powerful codes that can be efficiently decoded using traditional algebraic hard-decision decoding (HDD) methods such as the Berlekamp-Massey algorithm (denoted by BM-HDDA) and Euclid's algorithm. However, these traditional methods fail to exploit soft reliability information readily available at the output of the detector. With moderate complexity, but also with little improvement in performance for practical values of signal-to-noise ratio (SNR), generalized minimum distance (GMD) decoding, Chase decoding, and their variations were devised to make use of soft reliability information. Reliability-based decoding can also be applied to the binary images of RS codes. Based on the list decoding algorithm of Guruswami and Sudan, Koetter and Vardy (KV) proposed an algebraic soft-decision (ASD) decoding method for RS codes that incorporates reliability to improve performance over previous algebraic decoding algorithms. Analysis, extensions, modifications, refinements, and variations of KV algorithm are numerous in the art.
To improve performance beyond the ASD algorithms, non-algebraic methods have been proposed by Vardy and Be'ery to achieve the maximum-likelihood decoding (MLD) performance based on binary representation of an RS code as a union of cosets. However, the running time of these decoding methods grows exponentially with the length of the code. Indeed, it has been shown by Guruswami and Vardy that MLD of RS codes is NP-hard. Except for codes of lengths up to 15, decoding based on the trellis representation of RS codes is not feasible due to the enormous number of states in the trellis. Liu and Lin devised a turbo algorithm for soft-decision decoding (SDD) of RS codes based on decomposing the binary image of an RS code into binary component codes with relatively small trellis complexities, which are concatenated in a way that permits turbo decoding of RS codes of lengths up to 127.
Iterative decoding methods based on belief propagation (BP), such as the sum-product algorithm (SPA) and the min-sum algorithm (MSA), were proposed to decode low-density parity-check (LDPC) codes resulting in excellent performance with reasonable complexity that allows these codes to be used in practical applications. These decoding algorithms work on the parity-check matrices of LDPC codes which are sparse and their associated Tanner graphs have small number of short cycles, if any. However, due to the high-density of nonzero elements in their parity-check matrices and large number of short cycles, especially cycles of length four, in their associated Tanner graphs, direct application of these decoding methods is not suitable for RS codes. To overcome these difficulties, Jiang and Narayanan (JN) proposed an iterative BP-algorithm for decoding RS codes that uses a binary parity-check matrix for the code that is adapted using Gaussian elimination before each decoding iteration, in order to make the variable nodes (VNs) of lowest reliability connected with single edges to check nodes (CNs). Since this algorithm adapts parity-check matrices before running a decoding iteration based on BP, it is commonly referred to as an adaptive-BP-algorithm (or simply ABPA). Although the JN-ABPA gives good performance, its complexity may not be tolerable in practical applications. To reduce the complexity, it was proposed to use a redundant fixed set of parity-check equations corresponding to low-weight codewords in the binary image of the dual code, and select a subset of these equations based on the least reliable bits for decoding. For further improvement in performance, El-Khamy and McEliece (KM) presented an algorithm that combines the KV-ASD and the JN-ABPA. However, this improved ABPA, called KM-ABP-ASD algorithm (or simply KM-ABP-ASDA), still requires a very large computational complexity for practical applications. There is a need in the art for systems and methods for efficiently encoding and/or decoding information bits with reduced complexity.
Certain embodiments disclose a method for joint encoding. The method includes, in part, obtaining a sequence of encoded symbols. The sequence of encoded symbols may be generated through GFT. The method further includes, in part, jointly decoding the sequence of encoded symbols using an iterative soft decision decoding algorithm to generate a decoded sequence, transforming the decoded sequence to generate a plurality of cyclic codewords, and decoding the plurality of cyclic codewords to generate a plurality of decoded information symbols.
In one embodiment, jointly encoding the plurality of cyclic codewords includes, in part, permuting each of the plurality of cyclic codewords to generate a plurality of permuted codewords, combining the plurality of permuted codewords to generate a cascaded codeword, interleaving the cascaded codeword using an interleaving matrix to obtain an interleaved codeword, and performing a Galois Fourier Transform on the interleaved codeword to generate the at least one combined codeword.
In one embodiment, the plurality of permuted codewords are generated using Hadamard permutation algorithm. Any other permutation method may also be used without departing from the teachings of the present disclosure. In one embodiment, the plurality of cyclic codewords correspond to a first finite field and the at least one combined codeword corresponds to a second finite field, larger than the first finite field.
In one embodiment, the combined codeword includes a quasi-cyclic low density parity check codeword. In one embodiment, the quasi-cyclic low density parity check codeword corresponds to a binary parity check matrix.
In one embodiment, the plurality of cyclic codewords are generated based on Reed-Solomon code. In another embodiment, the plurality of cyclic codewords are generated based on Bose-Chaudhuri-Hocquenghem (BCH) codes. In yet another embodiment, the plurality of cyclic codewords are generated based on Quadratic Residue (QR) codes. In another embodiment, the plurality of cyclic codewords are generated based on Reed-Muller codes. It should be noted that in general, the plurality of cyclic codewords may be generated using any cyclic code generation algorithm known in the art, or a combination thereof, without departing from the teachings of the present disclosure.
In one embodiment, generating the plurality of cyclic codewords comprises encoding the sequence of information symbols using a single parity check cyclic code algorithm. In one embodiment, the method further includes, in part, generating at least one shortened cyclic codeword of prime length by performing a shortening procedure on at least one of the plurality of cyclic codewords when a length of the plurality of cyclic codewords is a composite number, and jointly encoding the at least one shortened cyclic codes to generate the combined codewords.
In one embodiment, the method further includes, in part, encoding the at least one combined codeword using an error-correcting coding algorithm to generate at least one post-GFT encoded codeword, and providing the at least one post-GFT encoded codeword to the device.
Certain embodiments disclose a method for joint decoding. The method includes, in part, obtaining a sequence of encoded symbols. In one embodiment, the sequence of encoded symbols is generated through GFT. The method further includes jointly decoding the sequence of encoded symbols using an iterative soft decision decoding algorithm to generate a decoded sequence, transforming the decoded sequence to generate a plurality of cyclic codewords, and decoding the plurality of cyclic codewords to generate a plurality of decoded information symbols.
In one embodiment, transforming the decoded sequence to generate a plurality of cyclic codewords includes performing an inverse Galois Fourier Transform on the decoded sequence to generate a plurality of transformed codewords, de-interleaving the plurality of transformed codewords to generate a plurality of de-interleaved codewords, and performing inverse permutation on the plurality of de-interleaved codewords to generate a plurality of cyclic codewords.
In one embodiment, performing inverse permutation on the plurality of de-interleaved codewords further includes, in part, de-combining the plurality of de-interleaved codewords to generate a plurality of de-combined codewords, and permuting the plurality of de-combined codewords to generate the plurality of cyclic codewords.
In one embodiment, the sequence of encoded symbols corresponds to a first finite field, and the plurality of transformed codewords correspond to a second finite field. The first finite field is larger than the second finite field.
In one embodiment, the method further includes, in part, decoding the sequence of encoded symbols using hard decision decoding to generate a sequence of corrected symbols, and jointly decoding the sequence of corrected symbols using the iterative soft decision decoding algorithm to generate the decoded sequence.
In one embodiment, the iterative soft decision decoding algorithm is Min-Sum algorithm. In one embodiment, a length of at least one of the plurality of cyclic codewords is a prime number. In one embodiment, the plurality of cyclic codewords correspond to a finite field of characteristic two. In one embodiment, the plurality of cyclic codewords includes Reed-Solomon codes, BCH codes, QR codes, Reed Muller codes, and the like.
Certain embodiments disclose an apparatus for joint encoding. The apparatus includes, in part, a memory, and at least one processor coupled to the memory. The at least one processor is configured to obtain a sequence of information symbols, generate a plurality of cyclic codewords, each cyclic codeword corresponding to a portion of the sequence of information symbols, and jointly encode the plurality of cyclic codewords to generate at least one combined codeword. In one embodiment, the at least one combined codeword is generated through Galois Fourier Transform. The at least one processor is further configured to provide the at least one combined codeword to a device.
Certain embodiments disclose an apparatus for joint decoding. The apparatus includes, in part, a memory, and at least one processor coupled to the memory. The at least one processor is configured to obtain a sequence of encoded symbols, wherein the sequence of encoded symbols is generated through Galois Fourier Transform, jointly decode the sequence of encoded symbols using an iterative soft decision decoding algorithm to generate a decoded sequence, transform the decoded sequence to generate a plurality of cyclic codewords, and decode the plurality of cyclic codewords to generate a plurality of decoded information symbols.
Certain embodiments disclose a non-transitory processor-readable medium for joint encoding. The non-transitory processor-readable medium includes, in part, processor-readable instructions configured to cause one or more processors to obtain a sequence of information symbols, generate a plurality of cyclic codewords, each cyclic codeword corresponding to a portion of the sequence of information symbols, and jointly encode the plurality of cyclic codewords to generate at least one combined codeword. The at least one combined codeword is generated through Galois Fourier Transform. The processor-readable instructions are further configured to provide the at least one combined codeword to a device.
Certain embodiments disclose non-transitory processor-readable medium for joint decoding. The non-transitory processor-readable medium includes processor-readable instructions configured to cause one or more processors to obtain a sequence of encoded symbols, wherein the sequence of encoded symbols is generated through Galois Fourier Transform, jointly decode the sequence of encoded symbols using an iterative soft decision decoding algorithm to generate a decoded sequence, transform the decoded sequence to generate a plurality of cyclic codewords, and decode the plurality of cyclic codewords to generate a plurality of decoded information symbols.
An understanding of the nature and advantages of various embodiments may be realized by reference to the following figures. In the appended figures, similar components or features may have the same reference label. Further, various components of the same type may be distinguished by following the reference label by a dash and a second label that distinguishes among the similar components. If only the first reference label is used in the specification, the description is applicable to any one of the similar components having the same first reference label irrespective of the second reference label.
A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention encompasses numerous alternatives, modifications and equivalents. Numerous specific details are set forth in the following description in order to provide a thorough understanding of the invention. These details are provided for the purpose of example and the invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the invention is not unnecessarily obscured.
The term “block code” is used herein to refer to any member of family of error-correcting codes that encode data in blocks of predetermined lengths. In addition, “cyclic code” is used herein to refer to a block code in which circular shifts of each codeword results in another codeword that belongs to the same code. The term “cyclic encoding algorithm” is used herein to refer to an algorithm that is used to encode a message using a cyclic code. The term codeword is used herein to refer to a sequence of encoded symbols that are generated by applying an encoding algorithm to a sequence of information bits.
The terms “communication channel” or “channel” are used interchangeably to refer to a communication medium through which a signal is transmitted. In a wireless communication, the communication channel may be air. In wired transmission the communication channel may be the metal in the wire or fiber optic.
The term “storage channel” is used herein to refer to a storage device, such as a hard drive or any other type of memory that can be used to store information for a period of time. The term “finite field” or “Galois field” is used herein to refer to a “field” that contains a finite number of elements (as is well known in mathematics). As with any field, a finite field is a set on which the operations of multiplication, addition, subtraction and division are defined and satisfy certain basic rules. The number of elements of a finite field is called its order. A finite field of order q exists if and only if the order q is a prime power pk (where p is a prime number and k is a positive integer). In a field of order pk, adding p copies of any element always results in zero. Therefore, the “characteristic” of the field is considered to be equal to p.
Certain embodiments disclose a novel coding scheme for encoding and iterative soft decision decoding (SDD) of a class of cyclic codes. One embodiment operates on a class of cyclic codes over finite fields of prime lengths which include all RS and BCH codes of prime lengths over fields of characteristic two, as well as all quadratic residue (QR) codes, and the like. The key to this coding scheme is to map a sequence of codewords in a cyclic code C and its permuted equivalents into a codeword in a powerful quasi-cyclic (QC) LDPC code CLDPC whose parity-check matrix HLDPC is an array of binary circulant permutation matrices (CPMs), the Tanner graph of which is free of cycles of length four. At the decoder, the received sequence is decoded iteratively based on a binary parity-check matrix of the QC-LDPC code CLDPC with a binary iterative soft-decision BP-algorithm. The decoded sequence is then transformed into a sequence of decoded codewords in C through inverse mapping and inverse permutations. Errors not corrected by the LDPC decoder can be decoded using an HDD decoder based on the code C. In one embodiment, the mapping involves symbol permutation within each codeword, combining codewords into a codeword in a code over an extension field, interleaving the permuted codewords, and Galois Fourier transform (GFT). In one embodiment, the inverse mapping involves the inverse GFT, de-interleaving, de-combining codewords, and inverse symbol permutation of each decoded codeword. In general, any other transformation may also be performed on the codewords without departing from the teachings of the present disclosure.
In one embodiment, the proposed coding scheme decodes a cascade of a cyclic code C and a number of its permuted equivalents as a single LDPC code in the GFT domain. In this example, although the LDPC code is non-binary, its parity-check matrix is binary and most importantly the decoding is carried out in binary using a binary iterative soft-decision BP-algorithm. This binary decoding can be performed efficiently and reduces the decoding complexity significantly. For decoding RS and BCH codes, the proposed decoding scheme not only requires much lower decoding complexity than other SDD schemes or algorithms, but also yields superior performance. Computing GFTs and their inverses can be implemented efficiently with fast algorithms. Since the LDPC code CLDPC is QC, the QC-structure simplifies its decoder implementation in terms of wire routing and allows partial parallel decoding. Furthermore, since RS and BCH codes are cyclic, their encoding can be easily implemented with simple feedback shift-registers. It should be emphasized that the coding scheme described herein operates at a coding rate which is at least equal to if not more than the coding rate of the cyclic code C.
One of the most important features of the proposed decoding scheme is that the decoding is performed on a collection of received codewords jointly. During the decoding process, the reliability information of each decoded codeword is shared by the others to enhance the overall reliability of all the decoded codewords. This joint-decoding and information sharing may result in an error performance per decoded codeword better than that of a received codeword decoded individually using maximum likelihood decoding (MLD). This will be demonstrated through examples.
Embodiments of the present disclosure may be used in various systems, including but not limited to high-speed optical communications, mobile and/or cellular communication systems, high-density storage systems, such as 3-dimensional flash memory, and/or any other type of system that needs to correct errors that are caused by noise and/or interference from internal and/or external components. It may also find applications in a combination of multiple-access and broadcast communications (downlinks) for error control. The communication systems may include optical communications, satellite communications, cellular communications, peer to peer communications, or any other type of communications systems.
The term “channel encoding” is used herein to refer to a process of encoding a message using an error control coding. In general, channel encoding is performed by adding extra bits (such as parity or checksum bits) to a codeword that can be used at the receiver to correct any possible errors in the received codeword. The term “channel decoding” is used herein to refer to the process of decoding a received message using a channel decoder to retrieve the original un-coded information bits and/or symbols. The channel decoder utilizes the redundant bits in the received message to correct any possible errors.
In this example, the storage unit may be considered as a storage channel. The servers may perform channel encoding before storing the information in the storage unit, and perform channel decoding after retrieving the information from the storage unit. In general, the servers may perform any other type of processing on the information before storing and/or after retrieving the information from the storage unit without departing from the teachings of the present disclosure.
The channel encoder 202 receives information bit sequence, encodes the information bits by adding redundancy to generate encoded codewords. The codewords may then be transmitted through the channel 204 and/or saved in a storage system. Noise, interference and/or any other disturbances may cause errors in the transmitted bits. At the receiver side, the channel decoder 206 may receive a signal that may include the encoded bits and noise. The channel decoder 206 decodes the received signal to correct the errors and generate the decoded bits.
After the collective soft decoding, the n embedded codewords are retrieved through inverse GFTs 404, de-interleaving 406 and inverse Hadamard-permutations 408. If any of the decoded codewords is detected in error, the cyclic code decoder 410, may perform hard decision decoding (HDD) based on the cyclic code Ccyc to correct residual errors. The joint-decoding and information sharing may result in an error performance per decoded codeword better than the error performance of a received codeword decoded individually using MLD as demonstrated in several examples. In one embodiment, the joint decoder scheme decodes a Q-ary received sequence in binary based on a binary LDPC matrix. The binary iterative decoding can be performed efficiently and reduces the decoding complexity significantly.
The rest of this disclosure is organized as follows. First, a class of cyclic codes of prime lengths is defined over finite fields of characteristic two in terms of parity-check matrices. This class of cyclic codes contains a subclass of RS codes and a subclass of BCH codes, namely all RS and BCH codes of prime lengths over such fields. In addition, Hadamard-powers of parity-check matrices are defined and their null spaces are characterized. In one embodiment, the GFT is used to transform a collection of codewords in the cyclic code into a codeword in an LDPC code over a larger field that has a binary parity-check matrix. We show that this matrix has good structural properties which allows for excellent performance when used for decoding by iterative SDD. In one embodiment, a coding scheme is developed for encoding and decoding in the GFT domain. In another embodiment, a technique is disclosed for shortening a cyclic code of prime length to a desired length and its associated LDPC code and a method is presented to reduce the code rate to various degrees but maintain the code length. In yet another embodiment, the proposed iterative SDD scheme is applied to RS codes. It is shown that the proposed scheme not only requires much lower decoding complexity than other SDD schemes or algorithms, but also yields superior performance. In later sections two example embodiments are presented in which the proposed decoding scheme is applied to binary BCH codes and QR codes. It is shown that even binary cyclic Hamming codes can perform extremely well.
A Class of Cyclic Codes of Prime Lengths Over Finite Fields
In this section, cyclic codes of prime lengths over finite fields of characteristic two are defined. Next, the Hadamard-equivalents of the cyclic codes of prime length are explained. It should be noted that although cyclic codes of prime length are used in the following embodiments, in general, the codes may have any length (e.g., prime or composite), without departing from the teachings of the present disclosure.
Cyclic Codes of Prime Lengths
Let GF(q) be a finite field of characteristic two, say q=2κ for some positive integer K, and C be an (n, n−m) cyclic code over GF(q) of odd prime length n and dimension n m. Then, C has a generator polynomial g(X)=Πi=0m-1(X−βl
The matrix B given by (1) is an m×n matrix over GF(qτ). A vector v=(v0, v1, . . . , vn-1) of length n over GF(q) is a codeword in C if and only if it is orthogonal to every row in B, i.e., vBT=0, where T denotes transpose. In polynomial form, the vector v is represented by a polynomial v(X)=v0+v1X+ . . . +vn-1Xn-1 over GF(q) of degree n−1 or less. Then, v(X) is a code polynomial if and only if v(X) is divisible by the generator polynomial g(X). This is the case if and only if v(βl
The matrix B in (1) satisfies the following property which is relevant for iterative decoding as shown later.
Lemma 1
Any 2×2 submatrix of B given by (1) is nonsingular.
Proof:
For 0≤i≠i′<m, 0≤j≠j′<n, the 2×2 submatrix of B composed of the elements in rows i and i′ and columns j and j′ has determinant βjl
There are two special cases that we are particularly interested in and they are important for practical applications. However, the scope of the present disclosure is not limited to these cases and, in general, the proposed coding/decoding technique can be applied to any codes without departing from the teachings of the present disclosure.
In one embodiment, τ=1 and l0=1, l1=2, . . . , lm-1=m. In this case, the cyclic code C given by the null space of the parity-check matrix B in the form of (1) is an (n, n−m) RS code over GF(q), denoted by CRS, of length n and dimension n−m. The length n of the code is a prime factor of q−1. The generator polynomial gRS(X) of this RS code has m consecutive powers of β, namely, β, β2, . . . , βm, as its roots, and the minimum distance of the code is exactly equal to m+1. If q−1 is a prime, then n=q−1 and β is a primitive element of GF(q). In this case, the code CRS is a primitive RS code of length q−1. If q−1 is not a prime, then n is a prime factor of q−1 and CRS is a nonprimitive RS code. As an example, let q=27, i.e., κ=7. Then, q−1=127 is a prime and β is a primitive element of the field GF(27). In one example, m=8 and β, β2, β3, . . . , β8 can be used to form an 8×127 parity-check matrix, denoted by BRS, over GF(27) in the form of (1). Then, the null space over GF(27) of BRS gives a 128-ary (127,119) RS code of length 127. In this example, the code has minimum distance 9 and rate 0.937. The code in this example is a 4-symbol-error-correcting RS code.
In another embodiment, q=2, i.e., κ=1, and β is an element of order n in GF(2τ) which is a prime factor of 2τ−1. Consider a cyclic code of length n over GF(2) whose generator polynomial gBCH(X) has m roots which include 2t consecutive powers of β, namely, β, β2, . . . , β2t, and their conjugates. It is clear that 2t≤m. Then, the null space over GF(2) of B (denoted by BBCH) in the form of (1) gives a t-error-correcting BCH code, denoted by CBCH, of length n and dimension n−m. If 2τ−1 is a prime, then β is a primitive element of GF(2τ) and n=2τ−1. In this case, CBCH is a primitive BCH code. If 2τ−1 is not a prime and n is a prime factor of 2τ−1, then CBCH is a nonprimitive BCH code. The t-error-correcting binary BCH code of length n is a subfield subcode of the t-symbol-error-correcting RS code of length n over GF(2τ) and dimension n−2t. As an example, let κ=1, τ=7, q=2, and t=4. Then, there exists a 4-error-correcting (127,99) primitive BCH code whose generator polynomial has β, β2, . . . , β8 and their conjugates as roots (a total of 28 roots). In this case, the parity-check matrix BBCH in the form of (1) is a 28×127 matrix over GF(27). This code is a subfield subcode of the 128-ary (127,119) RS code given above.
In yet another embodiment, a class of binary QR codes is considered as an example. This class of binary QR codes has good minimum distance properties but is difficult to be decoded algebraically. The most well-known QR code is the (23,12) Golay code with minimum distance 7, which is a perfect code.
Hadamard Equivalents of a Cyclic Code
Let t be a nonnegative integer. The t-th Hadamard-power of B is defined as a matrix, denoted by B°t=[βtjl
For 1≤t<n, let C(t) be the code given by the null space over GF(q) of B°t. Then, C(t) is an (n, n−m) code over GF(q) which is equivalent to the (n, n−m) cyclic code C given by the null space over GF(q) of B defined by (1). The code C(t) can be obtained from C by applying the permutation πt to its codewords. In this disclosure, the term C(t) is used to refer to the t-th Hadamard-equivalent of C. Hereafter, the notation πt is used to denote both the t-th Hadamard-permutation of the columns of the matrix B, and the t-th Hadamard-permutation of the codewords of the code C. Applying the permutation πt to a codeword in C results in a codeword in C(t). Note that (v0, v1, . . . , vn-1) is in the null space of B°t if and only if the polynomial Σi=0n-1viXi has βtl
Theorem 1.
For 1≤t<n, the t-th Hadamard-power B°t of B is a column-permutation of B specified by πt. The code C(t) given by the null space over GF(q) of B°t is a cyclic code equivalent to the code C given by the null space over GF(q) of B. For t=0, B°0 is an all-one matrix of size m×n whose null space gives an (n, n−1) SPC code over GF(q).
Transforming a Cyclic Code to an LDPC Code
In one embodiment, r copies of the cyclic code C(t) of length n over GF(q) for each t, 0≤t<n, are combined to form a code Cτ(t) of length n over GF(qτ). Then, the codes Cτ(t), are cascaded for 0≤t<n, to obtain a code Ccasc of length n2 over GF(qτ). The subscript “casc” stands for “cascade.” A permutation, denoted by wt, is performed on Ccasc to obtain a code Ccascπ and then the GFT is finally applied to Ccascπ that results in a code Ccascπ,F. This final code turns out to be an LDPC code over GF(qτ) as it has an LDPC matrix. As an LDPC code, we denote it by CLDPC. Starting with the parity-check matrix B of the cyclic code, in one embodiment, we track the above transformations to determine a low-density parity-check matrix for the final LDPC code.
The Code Cτ(t) and its Parity-Check Matrix B°t
In one embodiment, let C be an (n, n−m) cyclic code over GF(q) with parity-check matrix B given by (1). Then, for 0≤t<n, C(t) is the cyclic code with parity-check matrix B°t=[βtjl
It should be noted that the code Cτ(t) has the same parity-check matrix B°t as C(t). Indeed, since every codeword in Cτ(t) is a weighted sum of codewords in C(t), it follows that Cτ(t) is in the null space of B°t. As B°t is of rank 1 if t=0 and rank m if 1≤t<n, and Cτ(t) is of dimension n−1 if t=0 and dimension n−m if 1≤t<n, it follows that the null space of B°t over GF(qτ) is precisely C(t). Note that from the codeword ct in Cτ(t), it is possible to retrieve the codewords c0,t, c1,t, . . . , cτ-1,t in C(t) as every element in GF(qτ) can be written in a unique way as a polynomial in a of degree less than T.
In one embodiment, C is an RS code over GF(q), τ=1 and, for each 0≤t<n, the code Cτ(t) is the same as C(t).
The Code Ccasc and its Parity-Check Matrix HCasc
In one embodiment, let Ccasc be the code over GF(qτ) of length n2 obtained by cascading codewords in Cτ(0), Cτ(1), . . . , Cτ(n-1) in this order. A codeword ccasc in Ccasc is in the form of ccasc=(c0, c1, . . . , cn-1) with ct in Cτ(t) for 0≤t<n. The code Ccasc has length n2, dimension (n−1)+(n−m)(n−1)=(n−m+1)(n−1), and a parity-check matrix of the form:
which is an n×n diagonal array of m×n matrices B°0, B°1, . . . , B°(n-1). The array, Hcasc, is an mn×n2 matrix over GF(qτ). The rank of Hcasc is 1+m(n−1).
The Code Ccascπ and its Parity-Check Matrix Hcascπ
In one embodiment, coordinates of the codewords in Ccasc and the rows and columns of its parity-check matrix Hcasc are permuted. First, the following index sequences are defined:
πrow(0)=[0,m,2m, . . . ,(n−1)m], (3)
πrow=[πrow(0),πrow(0)+1, . . . ,πrow(0)+m−1], (4)
πcol(0)=[0,n,2n, . . . ,(n−1)n], (5)
πcol=[πcol(0),πcol(0)+1, . . . ,πcol(0)+n−1], (6)
In one embodiment, πrow and πcol define a permutation on {0, 1, . . . , mn−1} and a permutation on {0, 1, . . . , n2−1}, respectively. In particular, πrow(mi′+i″)=ni″+i′ for 0≤i′<n, 0≤i″<m, and πcol(nj′+j″)=nj″+j′ for 0≤j′,j″<n. Let ccasc=(c0, c1, . . . , cn-1) be a codeword in Ccasc. Then, ccasc is a vector of length n2 over GF(qτ). By labeling the coordinates of ccasc from 0 to n2−1, and applying the permutation πcol, we obtain a vector ccascπ=(c0π, c1π, . . . , cn-1π), where ctπ, 0≤t<n, is a vector of length n in which its j-th component, 0≤j<n, is the t-th component in cj. Note that the vector ccascπ is obtained by interleaving the n codewords c0, c1, . . . , cn-1 from Cτ(0), Cτ(1), . . . , Cτ(n-1), respectively. In one embodiment, the interleaving can be performed by arranging the n codewords c0, c1, . . . , cn-1 as rows of an n×n array Λ. Then, the n columns of this array Λ are c0π, c1π, . . . , cn-1π. Let Ccascπ be the collection of all vectors, ccascπ, corresponding to all codewords, ccasc, in Ccasc. Then, Ccascπ is a linear code over GF(qτ) of length n2 and dimension (n−m+1)(n−1) which are equal to those of Ccasc.
In one embodiment, a parity-check matrix for Ccascπ can be obtained by applying the permutations πcol and πrow to the columns and rows of the parity-check matrix Hcasc of Ccasc, respectively. First, the rows of Hcasc are labeled from 0 to mn−1 and the columns from 0 to n2−1. If the rows of Hcasc are first permuted based on πrow and then the columns of Hcasc are permuted based on πcol, an mn×n2 matrix, Hcascπ, is obtained over GF(qτ), where π denotes πrow and πcol, collectively, i.e., π=(πrow, πcol). The matrix Hcascπ is a parity-check matrix for the code Ccascπ. The (mi′+i″, nj′+j″)-entry in Hcasc is the (ni″+i′, nj″+j′)-entry in Hcascπ, where 0≤i″<m, 0≤i′,j′,j″<n. Based on the matrix Hcasc given in (2), it can be seen that the (mi′+i″, nj′+j″)-entry in Hcasc, which is the (ni″+i′, nj″+j′)-entry in Hcascπ, is 0 unless i′=j′ in which case it equals βi″j″t, where t=i′=j′. Hence, Hcascπ is an m×n array of n×n diagonal matrices as follows:
Hcascπ=[De,f]0≤e<m,0≤f<n, (7)
where, for 0≤e<m and 0≤f<n, De,f is an n×n diagonal matrix over GF(qτ) of the following form:
De,f=diag(1,βfl
The Code Ccascπ,F and its Parity-Check Matrix Hcascπ,F
Let a=(a0, a1, . . . , an-1) be a vector over GF(qτ) and β, as before, be an element of order n in GF(qτ)). According to one embodiment, the Galois Fourier transform (GFT), denoted by F(a), of a is a vector b=(b0, b1, . . . , bn-1) whose t-th component, for 0≤t<n, is given by
bt=Σk=0n-1akβkt=a0+a1βt+a2β2t+ . . . +an-1β(n-1)t, (9)
which is simply the inner product of (a0, a1, . . . , an-1) and (1, βt, β2t, . . . , β(n-1)t). The vector a is called the inverse GFT of the vector b, denoted by F−1(b), and can be retrieved using the relation:
at=Σk=0n-1bkβ−kt=b0+b1β−t+b2β−2t+ . . . +bn-1β−(n-1)t, (10)
for 0≤t<n. Let V=[βij]0≤i,j<n. Then, V is an n×n Vandermonde matrix which is nonsingular with inverse V−1=[β−ij]0≤i,j<n. It is easy to check that F(a)=aV and F−1(b)=bV−1.
Let ccascπ=(c0π, c1π, cn-1π) be a codeword in Ccascπ where each of c0π, c1π, . . . , cn-1π is a sequence over GF(qτ) of length n. For 0≤i<n, let ciπ,F=F(ciπ)=ciπV and
The vector ccascπ,F is referred to as the GFT of ccascπ. The inverse GFT of ccascπ,F gives back the vector ccascccascπ, i.e.,
Let Ccascπ,F be the collection of all vectors, ccascπ,F, corresponding to all codewords, ccascccascπ, in Ccascπ. Then, Ccascπ,F is a linear code over GF(qτ) of length n2 and dimension (n−m+1)(n−1), which are equal to those of Ccascπ. In one embodiment, from (11) and the fact that Hcascπ is a parity-check matrix for Ccascπ, it follows that the matrix Hcascπ,F
is a parity-check matrix of Ccascπ,F. It is an mn×n2 matrix over GF(qτ) of rank 1+(n−1)m.
The Code CLDPC and its Parity-Check Matrix HLDPC
In this subsection, structure of the parity-check matrix Hcascπ,F
Hcascπ,F
It is straightforward to check from (8) that for 0≤e<m and 0≤f<n, VDe,fV−1 gives an n×n matrix whose (i,j)-entry, 0≤i, j<n, equals
Σk=0n-1β(i-j+fl
The sum given by (15) equals to zero unless i−j assumes the unique value modulo n satisfying βi-j+fl
Hcascπ,F
is an m×n array of binary CPMs of size n×n. Each row in Hcascπ,F
According to one embodiment, it follows from the above developments that Hcascπ,F
According to one embodiment, for large n, Hcascπ,F
Structural Properties of HLDPC
The parity-check matrix HLDPC of the qτ-ary QC-LDPC code CLDPC has several properties which are relevant to its use in iterative decoding based on BP. First, since, from Lemma 1, the base matrix B satisfies the 2×2 SNS-constraint, HLDPC satisfies the RC-constraint, i.e., no two rows (or two columns) in HLDPC have more than one place where they both have 1-components. This RC-constraint ensures that the Tanner graph associated with HLDPC has girth at least six, which is typically required for iterative decoding algorithms, such as the SPA or the MSA, to achieve good performance.
It follows from the RC-constrained structure of HLDPC and the orthogonal principle of one-step majority-logic decoding that the QC-LDPC code CLDPC has minimum distance at least m+1. Note that, while the cascaded code Ccasc of C(0), C(1), . . . , C(n-1) with parity-check matrix Hcasc has minimum distance 2, as C(0) has minimum distance equal to two, the LDPC code CLDPC which is the null space of HLDPC, in one embodiment, has minimum distance at least m+1, which is the Singleton upper bound on the minimum distance of the codes C=C(1), C(2), . . . , C(n-1).
It should be noted that since the LDPC matrix HLDPC is RC-constrained with constant column weight m, it has no absorbing set of size less than [m/2]+1 or a small trapping set of size less than m−3. This implies that the qτ-ary QC-LDPC code CLDPC given by the null space of HLDPC decoded with SPA or the MSA does not suffer from high error-floors if m is reasonably large. Furthermore, note that HLDPC is an array of CPMs which simplifies wire routing in the decoder and allows for using the reduced-complexity decoding schemes to reduce the hardware implementation complexity of an iterative decoder.
According to one embodiment, HLDPC is a binary matrix. This binary property considerably simplifies the decoding of CLDPC, a code over GF(qτ), where q=2κ. A vector r=(r0, r1, . . . , rn
To summarize the disclosure so far, one embodiment starts with the mn×n2 matrix Hcasc given by (2) over GF(qτ) of rank 1+m(n−1), which is an n×n diagonal array of m×n submatrices. Row and column permutations denoted by π are performed to obtain an mn×n2 matrix Hcascπ, which is an m×n array of n×n diagonal submatrices. This is followed by applying the GFT as implemented using (13) to obtain an RC-constrained mn×n2 matrix over GF(2), Hcascπ,F
What is gained the most in terms of decoding complexity is an LDPC matrix which is binary. In one embodiment, this binary structure can be used effectively for decoding a cascaded cyclic code Ccasc over GF(qτ) in the GFT domain as a powerful nonbinary QC-LDPC code. The decoding of the nonbinary QC-LDPC code can be carried out in binary. This is the most important advantage in decoding a cascaded code as an LDPC code in the GFT domain, in terms of decoding complexity.
The code CLDPC is referred to herein as the LDPC code associated with the cyclic code C. The following theorem summarizes this section.
Theorem 2
Let C be an (n, n−m) cyclic code over GF(q) of prime length n with generator polynomial g(X)=Πi=0m-1(X−βl
The developments given in this section show that for every cyclic code C of prime length n over a field GF(q) of characteristic two, there is a QC-LDPC code CLDPC over GF(q), an extension field of GF(q), associated with it.
In one embodiment, consider the (5,3) cyclic code C over GF(24) generated by the polynomial g(X)=(X−β)(X−β2), where β is an element of order 5 in GF(24). Then, q=24, κ=4, and τ=1. The code C is a 16-ary nonprimitive RS code with a parity-check matrix of the form (1) shown as below:
which satisfies the 2×2 SNS-constraint. The Hadamard-powers of B are
and B°1=B. The matrix Hcasc in (2) is then
Applying the permutations πrow and πcol to the rows and columns of Hcasc, respectively, the following 2×5 array of 5×5 diagonal matrices over GF(24) is obtained:
The ten 5×5 diagonal submatrices in Hcascπ are De,f, 0≤e<2, 0≤f<5 (see (7) and (8)). Let V=[βij]0≤i,j<5. For each e and f, VDe,fV−1 is a 5×5 binary CPM, denoted by De,fF
which satisfies the RC-constraint. Note that the matrix Hcasπ,F
Given a cyclic code C of prime length n over GF(q), where q=2κ, with parity-check matrix B as in (1), in one embodiment, a coding scheme is disclosed for encoding and decoding the cascaded code Ccasc of length n2 over GF(qτ). A codeword in Ccasc is composed of n component words, c0, c1, . . . , cn-1, each of length n, where ct, 0≤t<n, is a codeword in the code Cτ(t) over GF(qτ) with parity-check matrix B°t. The proposed coding scheme encodes a codeword in Ccasc into a codeword in the LDPC code CLDPC of length n2 over GF(qτ) using the GFT. The codeword is transmitted over the channel (or stored in a storage system). The received vector is decoded using iterative SDD, and converted back, using the inverse GFT, to a sequence of estimates of the component words of the codeword in Ccasc. Since Cτ(0) is an (n, n−1) code over GF(qτ), it encodes a message composed of n−1 symbols in GF(qτ). For 1≤t<n, C(t) is an (n, n−m) code over GF(qτ), and hence encodes a message of length n−m over GF(qτ). Thus, in one embodiment, the proposed coding scheme encodes in total (n−1)+(n−1)(n−m)=(n−m+1)(n−1) symbols in GF(q) into a codeword in CLDPC composed of n2 symbols in GF(qτ), resulting in a code rate of (n−1)(n−m+1)/n2.
Note that, in one embodiment, both the encoding and decoding are performed collectively on nτ codewords over GF(q). This scheme should be compared with the classical scheme in which codewords from the (n, n−m) cyclic code C over GF(q) are transmitted over the channel and decoded individually giving a code rate of (n−m)/n. The coding scheme proposed herein has a higher code rate for m>1. The difference, which is in the favor of our scheme, can be used to transmit more information or be sacrificed to enhance the error-correcting capability as shown in the rest of this document.
Encoding of a Cascaded Code in the GFT Domain
At 604, the device combines the plurality of permuted codewords to generate a cascaded codeword. As an example, the device combines the z codewords in C(t) to form a codeword ct of length n over GF(qτ) in the code Cτ(t) for each t, 0≤t<n, and form the cascaded codeword Ccasc=(c0, c1, . . . , cn-1) in the cascaded code Ccasc.
At 606, the device interleaves the cascaded codeword using an interleaving matrix to obtain an interleaved codeword. As an example, the device apply the permutation πcol, defined by Eqn. (6), to ccasc, i.e., interleave ccasc, to obtain the interleaved cascaded codeword ccascπ=(c0π, c1π, . . . , cn-1) in Ccascπ.
At 608, the device performs a Galois Fourier Transform on the interleaved codeword to generate the at least one combined codeword. In one embodiment, the plurality of cyclic codewords correspond to a first finite field and the at least one combined codeword corresponds to a second finite field which is larger than the first finite field. In one embodiment, the combined codeword comprises a quasi-cyclic low density parity check codeword. In one embodiment, the device takes the GFT of each component word ctπ, 0≤t<n, in ccascπ as to obtain ctπ,F and form the vector ccascπ,F=(c0π,F, c1π,F, . . . , cn-1π,F) of length n2 over GF(qτ) for transmission. The vector c0π,F is a codeword in the LDPC code CLDPC which has HLDPC given by (14) as a parity-check matrix.
In the following, we elaborate on each step of the proposed encoding scheme. Let Mi,t, 0≤i<τ, 0≤t<n, be nτ messages with symbols from GF(q). For t=0, each message Mi,0 is composed of n−1 information symbols. For 1≤t<n, each message Mi,t is composed of n−m information symbols. The message Mi,0, for 0≤i<τ, is encoded into a codeword ci,0=vi,0 in C(0). For 1≤t<n, the message Mi,t composed of n−m symbols is encoded into a codeword vi,t in C. (e.g., Step 504). In one embodiment, the Hadamard permutation πt is applied in Step 602 to vi,t resulting in a codeword ci,t in C(t). This generates the τ codewords c0,t, c1,t, . . . , cτ-1,t in C(t) for each t, 1≤t<n. In Step 604, for each t, 0≤t<n, the vector ct=Σi=0τ-1ci,tαi is formed, where α is a primitive element in GF(qτ). The resulting vector ct is a codeword in the code Cτ(t). Step 606 can be performed by arranging the sequences c0, c1, . . . , cn-1 as rows in an n×n array, denoted by Λ. Then, the columns of this array are c0π, c1π, . . . , cn-1π from which the interleaved cascaded codeword ccascπ in Ccasc is formed. Finally, in Step 608, the GFT is applied to each ctπ, 0≤t<n, to obtain the vector ctπ,F over GF(qτ). The vector ccascπ,F=(c0π,F, c1π,F, . . . , cn-1π,F) is formed, which is a codeword in the LDPC code CLDPC with parity-check matrix HLDPC. The vector ccascπ,F is transmitted over the channel or stored in a memory (e.g., Step 508).
The implementation of above encoding process is relatively simple. In one embodiment, two encoders may be used, one for the SPC code C(0) (also a cyclic code with generator polynomial g(X)=X−1) and one for the base cyclic code C. Since C(0) is an SPC code, its encoder is very simple. Since C is a cyclic code, in one embodiment, its encoder can be implemented with a simple feedback shift-register with feedback connections based on the coefficients of its generator polynomial g(X). Combining the τ codewords in C(t) to form a codeword ct in Cτ(t) is based on representing a sequence of τ symbols in GF(q) as a symbol in GF(qτ). In one embodiment, this can be easily implemented by a look-up-table, or any other method. Note that for most practical applications, τ is small, e.g., equal to 8. This step is only performed when τ>1 since in case τ=1, which includes RS codes, ct=c0,t and Cτ(t)=C(t). In one embodiment, the Hadamard-permutation and interleaving operations can be performed together using a memory unit to store the code symbols of c0, c1, . . . , cn-1 in an n×n array Λ. Computing the GFTs of the columns of the array Λ can be implemented efficiently with a fast algorithm. The GFT is performed on each column of the array Λ as it is read out from the memory unit for transmission.
Decoding of a Cascaded Code in the GFT Domain
At 706, the device transforms the decoded sequence to generate a plurality of cyclic codewords. At 708, the device decodes the plurality of cyclic codewords to generate a plurality of decoded information symbols.
As illustrated, at 802, the device performs an inverse Galois Fourier Transform on the decoded sequence to generate a plurality of transformed codewords. In one embodiment, the device takes the inverse GFT of the decoded sequence to obtain an estimate ccascπ of the interleaved cascaded codeword ccascπ formed by the encoder at Step 606.
At 804, the device de-interleaves the plurality of transformed codewords to generate a plurality of de-interleaved codewords. In one embodiment, the device applies the inverse permutation πcol−1 to πcascπ, i.e., de-interleave ccascπ, to obtain a sequence ccasc=(c0, c1, . . . , cn-1) of n vectors of length n which is an estimate of ccasc formed by the encoder at Step 604.
At 806, the device de-combines the plurality of de-interleaved codewords to generate a plurality of de-combined codewords. In one embodiment, the device de-combines ct, for each t, 0≤t<n, into τ sequences c0,t, c1,t, . . . , cτ-1,t, which are the estimates of the codewords c0,t, c1,t, . . . , cτ-1,t in C(t) formed by the encoder at Step 602.
At 808, the device permutes the plurality of de-combined codewords to generate the plurality of cyclic codewords. In one embodiment, the device permutes ci,t, for 0≤i<τ and 1≤t<n, using πt−1 to obtain an estimate vi,t for vi,t formed by the encoder at Step 504.
In one embodiment, the decoder may optionally apply HDD to decode vi,t, for 0≤i<τ, 1≤t<n, based on the code C(t) to correct errors if detected. For example, the device decodes the sequence of encoded symbols using HDD to generate a sequence of corrected symbols, and jointly decodes the sequence of corrected symbols using the iterative soft decision decoding algorithm to generate the decoded sequence.
In the following, each decoding step is explained in more detail. In one embodiment, in Step 704, the received vector r is decoded based on the LDPC matrix HLDPC using an iterative soft-decision BP-algorithm. Recall that HLDPC is a binary matrix. Based on the properties described earlier in this disclosure, the received vector τ=(r0, r1, . . . , rn
The κτ binary decodings result in κτ decoded binary vectors. We combine these decoded binary vectors into a vector, denoted by ccascπ,F, of length n2 over GF(qτ) by grouping each set of κτ corresponding bits in the κτ decoded binary vectors into an element in GF(qτ). Then, ccascπ,F is an estimate of the transmitted vector ccascπ,F.
Let ccascπ,F=(c0π,F, c1π,F, . . . , cn-1π,F), where each ctπ,F is of length n over GF(qτ). In Step 802, the inverse GFT is applied to each ctπ,F, 0≤t<n, to yield ctπ and the vector ccascπ=(c0π, c1π, . . . , cn-1π) is formed as an estimate of ccascπ=(c0π, c1π, . . . , cn-1π) in the encoding process.
In one embodiment, Step 804, in which the inverse permutation πcol−1 is applied to ccascπ can be implemented by storing c0π, c1π, . . . , cn-1π as columns in an n×n array
The complexity of decoding the LDPC code CLDPC over GF(qτ) based on the binary parity-check matrix HLDPC is generally in the order of κτ times that of decoding the LDPC code over GF(2) given by the null space of HLDPC. Using the reduced-complexity iterative decoding scheme, the hardware implementation complexity of the decoder based on HLDPC can be reduced by a factor as large as n, the size of the constituent CPMs in HLDPC, i.e., the number of codewords in a cascaded codeword in Ccasc.
With the coding scheme proposed herein, a collection of nτ codewords in a base code C and C(0) is combined into a single coded sequence which is transmitted in the GFT domain over the channel. At the received end, these nτ codewords are decoded collectively in the GFT domain. This collection of nτ codewords is referred to herein as a coded frame.
Note that the decoder has the option of performing HDD for C(t), 1≤t<n, as the final decoding step. Since C(0) is an SPC code with minimum distance of 2, it does not have any error-correction capability. In one embodiment, the scheme can be easily modified to allow for error-correction by restricting Mi,0, for 0≤i<τ, to be a message composed of n−m q-ary symbols rather than n−1 q-ary symbols as follows. For example, in Step 504, we encode each message Mi,0 of n−m q-ary symbols into a codeword vi,0 in the code C over GF(q) for 0≤i<τ. Next, we combine the τ codewords, v0,0, v1,0, . . . , vτ-1,0 to form a codeword v0 of length n over GF(qτ) as in Step 604. Let v0=(v0,0, v1,0, v2,0, . . . vn-1,0), where vj,0 is an element in GF(qτ). Next, let c0=(v0,0, βl
In one embodiment, at the receiver, the decoder goes through Steps 704, 802 and 804 as explained in above which results in an estimate c0=({tilde over (c)}0,0, {tilde over (c)}1,0, {tilde over (c)}2,0, . . . {tilde over (c)}n-1,0) of c0. However, before implementing Step 806, the decoder forms v0=({tilde over (c)}0,0, β−l
In some scenarios, whether or not HDD is performed on C at the final stage of decoding may have a small effect on performance (e.g., in one embodiment, the effect may not exceed one-tenth of a dB in SNR). This can be attributed to the fact that the iterative decoding based on BP of the LDPC code CLDPC corrects most of the errors, and in case it fails then the residual errors exceed the error correcting capability of the cyclic base code C.
Belief Propagation (BP) Decoding Modes
For certain embodiments, the κτ binary constituent vectors of a received vector r can be decoded in three simple modes: (1) decode them in parallel by using κτ identical binary BP-decoders; (2) decode them in serial, one at a time, using a single BP-decoder; and (3) decode them in partial-parallel using κ (or τ) BP-decoders. Other partial-parallel decoding modes can also be performed without departing from the teachings of the present disclosure. Different decoding modes provide various trade-offs between decoding speed and implementation complexity. In one embodiment, fully-parallel decoding of all κτ binary constituent vectors may be performed. Grouping operation is performed when and only when all the binary constituent received vectors are successfully decoded with a chosen BP-decoding algorithm. Before all the binary constituent received vectors are successfully decoded, those that have been successfully decoded are saved in the memory, and the rest are continuously processed until all of them are successfully decoded and stored in the memory as a κτ×n2 array. At this point, grouping the κτ decoded binary vectors is completed. If not all of κτ binary received constituent vectors are successfully decoded by the time when a preset maximum number of BP-iterations for all the BP-decoders is reached, a decoding failure may be declared.
Let Imax denote the preset maximum number of BP-iterations to be performed for decoding each of the κτ binary received constituent vectors rb,i, 0≤i<κτ. For 0≤l<Imax and 0≤i<κτ, let rb,i(l) denote the updated hard-decision of rb,i at the end of the l-th iteration. For l=0, rb,i(0)=rb,i, 0≤i<κτ. In one embodiment, the iterative soft decision decoding (e.g., belief propagation (BP) decoding) of a received vector r is carried out as in the following steps:
BP-1. Initialization: Set l=0 and log-likelihood ratios (LLR's) for the received code bits of all the κτ constituent received vectors based on the channel output detector.
BP-2. Perform the l-th BP-iteration to update the LLR's of the bits of rb,i, 0≤i<κτ for which HLDPC(rb,i(l−1))T≠0. Compute rb,i(l).
BP-3. Check HLDPC bb,i(l))T, 0≤i<κτ. If HLDPC(rb,i(l))T=0 for all 0≤i<κτ, go to Step BP-4. If not, save rb,i(l)'s for which HLDPC(rb,i(l))T=0. If l=lmax, declare a decoding failure; otherwise, set l←l+1 and go to Step BP-2.
BP-4. Stop BP-decoding and group the κτ decoded binary vectors into an estimate of the transmitted word ccascπ,F.
The decoding part of the proposed coding scheme for cyclic codes of prime lengths is an iterative soft-decision decoding (ISDD) scheme in the GFT domain. For simplicity, we refer to this decoding algorithm as GFT-ISDD scheme in the rest of this disclosure.
One of the most important features of the proposed coding scheme is that the decoding is performed on a collection of received codewords jointly. During the decoding process, the reliability information of each decoded codeword is shared by the others to enhance the overall reliability of all the decoded codewords. This joint-decoding and information sharing may result in an error performance per decoded codeword better than the error performance of a received codeword decoded individually using MLD. This will be demonstrated in several example embodiments described in this document. This gain over MLD is referred to as joint-decoding gain. For a long code, the joint-decoding gain can be large.
However, for a long code, computing its MLD performance is practically impossible. In this case, the union bound (UB) on MLD performance of the large codes, denoted by UB-MLD, are used for comparison. For large SNR, the UB is very tight.
Measures of the Error Performance of the BP-Decoding
For certain embodiments, in measuring the error performance of a cascaded code Ccasc of a base code C at the output of the LDPC decoder, we use frame error probability (i.e., the probability that a frame of nτ codewords is decoded incorrectly) in terms of frame error rate (FER), denoted by PFER. After de-interleaving of the decoded frame and performing the inverse of the Hadamard-permutation on the n symbols of each decoded codeword, the block error probability (i.e., the probability that a codeword in a frame is decoded incorrectly) in terms of block error rate (BLER), denoted by PBLER, is computed. The block error probability PBLER of a codeword in a decoded frame can be computed from the frame error probability PFER. Let λ be the average number of codewords in a decoded frame which are incorrectly decoded. Then, the error probability PBLER of a decoded codeword is PBLER=λ/(nτ)PFER. It is clear that the frame error probability PFER is an upper bound on the block error probability PBLER of a decoded codeword. In simulations, we find that the gap between these two error probabilities is very small corresponding to a difference of one-tenth of a dB in SNR. This reflects the fact that if a frame is not decoded correctly, then most of the nτ codewords are not decoded correctly.
Suppose we transmit nτ codewords in the base code C independently and decode them individually using a certain decoding algorithm. Let P*BLER be the block error probability of a single decoded codeword. Then, the total error probability, denoted by Ptotal, of nτ individually decoded codewords (i.e., the probability that at least one of the nτ decoded codewords is incorrect) is at least P*BLER.
In the later sections, when we compare the error performance of a cascaded code constructed from a base code C using the above proposed decoding scheme with the error performance of the base code C with other decoding algorithms or schemes in which each received vector is decoded individually, we compare PBLER with P*BLER. In comparison, we also include the frame error probability PFER of a cascaded code to show the gap between PFER and PBLER.
Computational Complexity of the GFT-ISSD Scheme in Conjunction with the MSA
It should be noted that, in general, any iterative soft-decision BP-decoding algorithm can be used in conjunction with the proposed GFT-ISDD scheme to decode the qτ-ary LDPC code CLDPC with q=2κ based on its binary mn×n2 parity-check matrix HLDPC. The two commonly used iterative BP-algorithms for decoding LDPC codes are the SPA and the MSA. Both algorithms require real number computations including multiplications, additions, and comparisons. The MSA is a simplified (or approximated) version of the SPA which requires mainly real number additions and comparisons, but much less multiplications which are required when scaling is performed. Since the MSA is an approximated version of the SPA, decoding an LDPC code with the MSA results in some performance loss. However, if optimal scaling is used in updating LLR's of the received symbols, the MSA may perform just as well as (or very close to) the SPA. For this reason, the MSA or its variations are most commonly used in practice.
For certain embodiments, the MSA is used in conjunction with the GFT-ISDD scheme in decoding the qτ-ary LDPC code CLDPC based on the binary parity-check matrix HLDPC. We denote such a combination as the GFT-ISDD MSA. However, any other soft decoding scheme may also be used instead of MSA, without departing from the teachings of the present disclosure. In computing the complexity of the GFT-ISDD/MSA, we only count the number of real number comparisons and additions. For a scaled MSA, each check node (CN) requires only 2 real number multiplications, which is a very small part of the total computations. Thus, in one embodiment, we ignore the real number multiplications when estimating the computational complexity of the GFT-ISDD/MSA. In the remaining of this disclosure, “a real number computation” refers to either a real number addition or comparison.
Recall that using the GFT-ISDD scheme to decode a qτ-ary received vector r of length n2 with q=2κ, we decode its κτ binary constituent vectors rb,i=(r0,i, r1,i, . . . rn
Since every iteration updates the reliabilities of κτ binary received vectors, in one embodiment, the total number of real number computations required for updating a qτ-ary received vector per iteration of the GFT-ISSDD/MSA is
NMSA=κτmn(3n+┌log2 n┘−2)≈3κτmn2. (23)
Let lavg be the average number of iterations required to decode a qτ-ary received vector r with the GFT-ISDD/MSA. Then, the average total number of real number computations (excluding the computations required for GFTs over GF(qτ)) required for decoding r is
NGFT-ISDD/MSA=NMSAIavg=κτmn(3n+┌log2n┘−2)Iavg≈3κτmn2Iavg. (24)
Since the GFT-ISDD/MSA decoding is performed on a collection of (n−1)τ codewords in the base code C and τ codewords in the SPC code C(0) (can be made as a codeword in C as described earlier), then the average number of real number computations required to decode a single codeword in C using the GFT-ISDD/MSA is
Navg=κm(3n+┌log2 n┘−2)lavg≈3κmnIavg. (25)
If we set the maximum number of iterations to be performed for each MSA decoder to Imax, then the maximum number of real computations required to decode a qτ-ary received vector with the GFT-ISDD/MSA is
Nmax=κm(3n+┌log2 n┘−2)Imax≈3κmnImax. (26)
Note that mn is the total number of nonzero entries in the parity-check matrix B of the base cyclic code C of prime length, see (1).
In the remaining of this document, we show that decoding the LDPC code CLDPC associated with a cyclic base code C using the GFT-ISDD/MSA converges very fast. For example, setting Imax to 5 gives an error performance only 0.2 dB away from the error performance obtained by setting Imax to 50, especially for decoding RS and BCH codes.
Code Shortening and Rate Reduction
Shortening the Codes
In many applications, it is desired to use a code of composite length, i.e., not of prime length. In one embodiment, a code of prime length can be shortened to a desired length. Note that for any (n, k) cyclic code over GF(q), the last (rightmost) k symbols in each codeword can be chosen to be information symbols, i.e., they can be chosen arbitrarily, and the first n−k symbols (parity-check symbols) can then be calculated to produce a codeword in the code. A code in this form is said to be systematic. By considering all possible choices of the k information symbols, we obtain the qk codewords in the code. To shorten the code to length n′, where 1≤n−n′<k, we consider all codewords in the code ending with n−n′0's. There are qk-(n-n′) such codewords. By deleting the last n−n′ symbols from each such codeword, we obtain the shortened (n′, k−(n−n′)) code. Note that, in general, this shortened code is not cyclic. However, it can be implemented using the same encoder and decoder as the (n, k) cyclic code, and in this sense it is called a shortened cyclic code. A parity-check matrix for the shortened (n′, k−(n−n′)) code can be obtained by deleting the last n−n′ columns of a parity-check matrix of the (n, k) code.
In one embodiment, the proposed coding scheme is applied to the shortened cyclic code of length n′ obtained by shortening the (n, n−m) cyclic code C of prime length n to the desired length n′, where n′>m. Let M′i,t, for 0≤i<τ, 1≤t<n, be a message composed of n′−m q-symbols and M′i,0, for 0≤i<τ, be a message composed of n′−1 q-ary symbols. For 0≤i<τ and 0≤t<n, we append each message M′i,t with n−n′0's and encode it into a codeword ci,t in the code C(t) of length n. Then, ci,t ends with n−n′0's. This completes Step 602 in the encoding process, as shown in
At the receiving end, the decoder first decodes the received sequence of length nn′ based on the LDPC matrix H′LDPC using an iterative soft-decision BP-algorithm and appends the decoded sequence with n(n−n′) 0's. Then, it follows Steps 802-806 as illustrated in
In one embodiment, encoding steps 504 and 602 as illustrated in
Reduction of Code Rate
Using the shortening technique described above, we not only reduce the length of the final QC-LDPC code for information transmission but also reduce the code rate. However, in some applications, we may want to reduce the code rate but maintain the code length. One approach to accomplish this goal is presented in the following.
One embodiment starts with an (n, n−m) cyclic code C over GF(q) of prime length n as the base code for the construction of a cascaded code Ccasc,*π,F of length n2 in the GFT domain using the encoding scheme presented in
Different choices of the set {λ0, λ1, . . . , λn-1} of parameters result in different codes, Ccasc,*π,F, of the same length n2 but with different rates. The choice λ0=λ1= . . . =λn-1=τ gives the code Ccascπ,F. All these codes can be decoded with the same decoder as the code Ccascπ,F.
Decoding RS Codes of Prime Lengths with the GFT-ISDD Scheme
One embodiment applies the GFT-ISDD scheme proposed herein to cyclic RS codes of prime lengths and show that the proposed GFT-ISDD scheme not only requires much lower decoding complexity than other SDD schemes or algorithms for RS codes, but also yields superior performance demonstrated through examples.
The GFT-ISDD Scheme for Decoding RS Codes
In one embodiment, let β be an element of order n in a field GF(q) of characteristic two, where n is a prime factor of q−1 and q=2κ. Consider the cyclic RS code CRS of length n and dimension n−m over GF(q) whose generator polynomial gRS(X) has β, β2, . . . , βm as roots with 1≤m<n. Then, the generator polynomial of this RS code is gRS(X)=(X−β)(X−β2) . . . (X−βm), which is a polynomial of degree m over GF(q) and divides Xn−1. The parity-check matrix of the RS code CRS in terms of the roots of its generator polynomial is given by the following m×n matrix over GF(q):
Typically, the decoding of an RS code is carried out based on a parity-check matrix in the form of (27) using either an HDD or an SDD algorithm. Since the RS code is over GF(q) and β is an element in GF(q), we have τ=1. Hence, the code Cτ(t) is the same as C(t) for 0≤t<n and the code Ccasc is obtained by concatenating codewords in C(0), C(1), C(2), . . . , C(n-1), where C(0) is the (n, n−1) SPC code, C(1)=CRS is the RS code, and C(2), . . . , C(n-1) are the Hadamard-equivalents of CRS. We call the cascaded code constructed using CRS as the base code the cascaded RS code which is denoted by CRS,casc.
In one embodiment, encoding follows with Steps 504 and 602-608 as described corresponding to
In one embodiment, we use the proposed GFT-ISDD scheme to decode a received vector r over GF(q) following decoding steps illustrated in
In the following, we use a number of example embodiments to demonstrate the effectiveness of the proposed GFT-ISDD scheme for decoding RS codes of prime lengths. Hereafter, in every example, we use the binary MSA for decoding every binary constituent received vector, i.e., we use the GFT-ISDD/MSA for decoding. In performance simulations, we assume transmission over the binary-input additive white Gaussian noise (AWGN) channel using BPSK signaling. It should be noted that the list of example embodiments are non-limiting and any other code, with any other code parameter may also be used without departing from the teachings of the present disclosure. In addition, although GFT-ISDD/MSA is considered as an example decoding algorithm, in general, any other decoding procedure may be used to decode the received sequence without departing from the teachings of the present disclosure.
In the given examples, we may compare the performance of a cascaded RS code CRS,casc decoded with the GFT-ISDD/MSA with the MLD performance (or the UB-MLD) of its base RS code CRS.
In one embodiment, decoding of an RS code CRS over GF(25) using the GFT-ISDD/MSA is considered. In this example, q=25 and 25−1=31 is a prime, the only prime factor of q−1 is n=31. Let β be a primitive element of GF(25). Suppose we construct an RS code CRS whose generator polynomial has β, β2, β3, β4, β5, β6 as roots. This RS code is a (31,25) primitive RS code over GF(25) with rate 0.806 and minimum distance 7 (a triple-error-correcting code). Its parity-check matrix BRS in terms of the roots of its generator polynomial is a 6×31 matrix over GF(25) in the form given by (27). From this matrix, we construct a 6×31 RS-LDPC array HRS,LDPC of binary CPMs of size 31×31, which is a 186×961 matrix with column and row weights 6 and 31, respectively. The Tanner graph associated with HRS,LDPC has girth at least six.
For 1≤t<31, the code CRS(t) given by the null space of B°RSt (the t-th Hadamard-power of BRS) is also a (31,25) code over GF(25) which is equivalent to CRS. For t=0, the code CRS(0) given by the null space of the 0-th Hadamard-power B°RS0 of BRS is a (31,30) SPC code over GF(25). By cascading CRS(0) and the 30 Hadamard-equivalents of CRS, we obtain a cascaded RS code CRS,casc of length 961 over GF(25). Interleaving the 31 component codewords in each codeword of CRS,casc using the permutation πcol defined by (6) followed by the GFT, we obtain a transmitted vector ccascπ,F of 961 symbols over GF(25). This is a codeword in the 32-ary (961,780) RS-LDPC code CRS,LDPC given by the null space over GF(25) of the binary 186×961 RS-LDPC matrix HRS,LDPC.
Next, we decode the cascaded RS code CRS,casc based on the binary 186×961 RS-LDPC matrix HRS,LDPC constructed above using the proposed GFT-ISDD/MSA. We first decompose a received vector r into 5 binary constituent received vectors rb,0, rb,1, rb,2, rb,3, rb,4. Each of these 5 binary constituent received vectors is decoded using the MSA with a scaling factor 0.625. The MSA decoding results in 5 decoded binary vectors. We combine these 5 decoded binary vectors into a vector over GF(25), which is an estimate of the transmitted codeword ccascπ,F in the GFT domain. Applying the inverse GFT, followed by de-interleaving and then inverse Hadamard permutations, we obtain the estimates of the 31 transmitted codewords, one from C(0) and the 30 codewords from the (31,25) RS code CRS. Suppose we decode the RS-LDPC code CRS,LDPC using 50 iterations of the MSA without performing HDD on these estimates.
Also included in
From
It follows from Eqn. (23) that the number of real number computations in each iteration of the GFT-ISDD/MSA is 89,280. Since each iteration is performed on 31 codewords from the (31,25) RS base code CRS. The average number of real computations required for updating each codeword in CRS is 2880. At the SNR (Eb/No) of 4.9 dB, the code achieves a BLER of 10−6 and the decoding takes an average of about 2 iterations to converge. Therefore, from Eqn. (25), at the SNR=4.9 dB, the average number of real number computations required to decode a codeword in CRS is 5760. If the decoding is performed up to a maximum of 5 iterations, then from Eqn. (26), the number of real number computations required is at most 14,400. If the MSA is used in the KM-ABP-ASD(20,10), the number of real number computations required to perform 20 inner iterations and 10 outer iterations is at least 1,161,000, as explained later in this section.
In the next example, we consider a longer RS code with a high rate over a larger field than that used in Example 2. It should be noted that the embodiments presented in this disclosure are mere examples. One of ordinary skill in the art would readily understand that these code parameters are not limiting. In general, the coding scheme presented herein can be used with a variety of code types corresponding to a variety of parameters, and be decoded with any decoding procedure without departing from the teachings of the present disclosure.
In one embodiment, decoding of an RS code CRS over GF(27) using the GFT-ISDD/MSA is considered. In this example, q=27 and 27−1=127 is a prime, the only prime factor of q−1 is n=127. Let P be a primitive element of GF(27). Suppose we construct an RS code CRS whose generator polynomial has β, β2, β3, β4, β5, β6, β7, β8 as roots. The RS code CRS is a (127,119) primitive RS code over GF(27) with rate 0.937 and symbol-error-correcting capability of four. Its parity-check matrix BRS in terms of the roots of its generator polynomial is an 8×127 matrix over GF(27) in the form given by Eqn. (27). Using this matrix and its 127 Hadamard-powers B°RSt, 0≤t<127, we construct an 8×127 RS-LDPC array HRS,LDPC of binary CPMs of size 127×127, which is a 1016×16129 matrix with column and row weights 8 and 127, respectively. The Tanner graph associated with HRS,LDPC has girth at least 6.
For 1≤t<127, the code CRS(t) given by the null space of B°RSt (the t-th Hadamard-power of BRS) is also a (127,119) code over GF(27) which is equivalent to CRS. For t=0, the code CRS(0) given by the null space of the 0-th Hadamard-power B°RS0 of BRS is a (127,126) SPC code over GF(27). By cascading CRS(0) and the 126 Hadamard-equivalents of CRS, we obtain a cascaded RS code CRS,casc of length 16129 over GF(27). A codeword in CRS,casc is composed of 127 words: one word in the (127,126) SPC code and 126 words in CRS and its Hadamard-equivalents. Interleaving the 127 words in each codeword of CRS,casc using the permutation πcol defined by Eqn. (6) followed by the GFT performed over GF(27), we obtain a transmitted vector ccasc of 16129 symbols over GF(27). This is a codeword in the 128-ary (16129,15120) QC-RS-LDPC code CRS,LDPC given by the null space over GF(27) of the binary 1016×16129 RS-LDPC matrix HRS,LDPC.
To decode the cascaded RS code CRS,casc based on HRS,LDPC using the GFT-ISDD scheme, we first decompose a received vector r into 7 binary constituent received vectors. In decoding of these 7 binary constituent received vectors, we use the binary MSA with a scaling factor of 0.5. The MSA decoding results in 7 decoded binary vectors. We combine these 7 decoded binary vectors into a vector over GF(27), which is an estimate of the transmitted codeword ccascπ,F in the GFT domain. Applying the inverse GFT, followed by de-interleaving and then inverse Hadamard permutations, we obtain the estimates of the 127 transmitted codewords, one from C(0) and 126 codewords from the (127,119) RS code CRS. Suppose we do not perform error corrections on these estimates.
Also included in
In the last two examples, we gave the error performances of two primitive RS codes decoded using the proposed GFT-ISDD scheme. In the next example, we consider a nonprimitive RS code.
In one embodiment, decoding of a nonprimitive RS code CRS over GF(211) using the GFT-ISDD/MSA is considered. In this example, q−1=2048−1=2047, which can be factored as the product of two prime numbers, 23 and 89. Based on these two prime factors, we can construct two nonprimitive RS codes over GF(211).
Let α be a primitive element of GF(211) and β=α23. The order of β is n=89. Let CRS,0 be the RS code of length 89 over GF(211) whose generator polynomial has β, β2, β3, β4 as roots. Then, the code CRS,0 is an (89,85) RS code over GF(211) with rate 0.955. We decode the cascaded code of this RS code and its Hadamard-equivalents using the GFT-ISDD/MSA. Each received vector r consists of 7921 received symbols which are elements in GF(211). In decoding, we decode the 11 binary constituent received vectors of r, each with 50 iterations of the MSA.
In one embodiment, a shortened cascaded RS code in the GFT domain is considered. As an example, consider the (127,121) RS code CRS over GF(27) whose generator polynomial has β, β2, β3, β4, β5, β6 as roots where β is a primitive element of GF(27). CRS is a triple-error-correcting RS code with rate 0.9527. Suppose we shorten this RS code by a degree of 63. Shortening results in a (64,58) shortened RS code. Using this shortened RS code as the base code and following the encoding process described above, we obtain a shortened (8128,7371) QC-RS-LDPC code over GF(27) with rate 0.9068 in the GFT domain. Each transmitted codeword in the cascaded shortened (64,58) RS code consists of 126 codewords in the shortened (64,58) RS code and one codeword in the (64,63) SPC code in the GFT domain.
Computational Complexity and Comparison
In the following, we give a comparison between the computational complexity of the GFT-ISDD/MSA and that of an ABPA (either the JN-ABP-HDDA or the KM-ABP-ASDA) for decoding an RS code.
In one embodiment, the MSA is used in conjunction with the proposed GFT-ISDD scheme for decoding the cascaded RS code CRS,casc in the GFT domain. It follows from Eqn. (25) that the average number of real computations required to decode a single codeword in the base RS code CRS using the GFT-ISDD/MSA in the iterative process of decoding its associated RS-LDPC code CRS,LDPC is
NRS,avg=κm(3n+┌log2 n┘−2)IRS,LDPC≈3κmnIRS,avg, (28)
where IRS,avg is the average number of iterations required to decode a q-ary received vector using the GFT-ISDD/MSA.
For decoding of an (n, n−m) RS code CRS over GF(2κ) with the ABPA, the decoding of a codeword consists of N1 inner ABP iterations and N2 outer iterations. Each ABP iteration consists of the four steps. The first step is to sort the components of the LLR vector associated with the binary received vector of length nκ, which is the binary expansion (or binary image) of the received vector of length n over GF(2κ), to find mκ least reliable bits. This step requires real number comparisons. The second step is to adapt a new parity-check matrix in which the part (submatrix) corresponding to the mκ least reliable bits is sparse. This step requires binary operations to perform Gaussian elimination. The third step is to perform a BP algorithm to update the reliabilities of the received symbols based on the newly adapted parity-check matrix. This step requires real number computations (including real number multiplications, additions, and comparisons). After the completion of each ABP iteration, if the hard-decision vector formed based on the updated LLR vector does not converge to a codeword, HDD is performed on the hard-decision vector using an HDDA. For the JN-ABP-HDDA, HDD is performed with the BM-HDDA. For the KM-ABP-ASDA, the KV-ASDA is used. Any other algorithm may also be used without departing from the teachings of the present disclosure. The KV-ASDA requires much more computation complexity than the BM-HDDA.
To show that the computational complexity of the proposed GFT-ISDD/MSA is much lower than that of the ABPA (either the JN-ABP-HDDA or the KM-ABP-ASDA), we only compare the number of real number computations required by the GFT-ISDD/MSA and that of the ABP iteration step (which is common for both the JN-ABP-HDDA and the KM-ABP-ASDA). Suppose the MSA is applied to update LLRs during the ABP iteration step. In this example, we only compute a lower bound on the number of real computations (including real number additions and comparisons) at the ABP/MSA iteration step. The adapted parity-check matrix used for each inner ABP/MSA iteration consists of two submatrices, one sparse and one dense. The dense submatrix is a κm×κ(n−m) matrix with average row weight κ(n−m)/2 and average column weight κm/2. The sparse submatrix is a κm×κm matrix with column weight 1 (can be made weight 2 which will increase the computational complexity). Then, the average number of real number computations in each inner ABP/MSA iteration is lower bounded by
κm(3κ(n−m)/2+┌log2(1+κ(n−m)/2)┘)≈3/2κ2m(n−m). (29)
Therefore, in this example, a lower bound on the number of real number computations required at the ABP/MSA decoding step with total number N1N2 of inner and outer iterations is
NABP=N1N2κm(3κ(n−m)/2+┌log2(1+κ(n−m)/2)┘)≈3/2N1N2κ2m(n−m). (30)
Comparing the number of real number computations required for one GFT-ISDD/MSA iteration with that of one ABP/MSA iteration, we can easily check that for an RS code of rate greater than 2/κ, the GFT-ISDD/MSA requires lower computational complexity than the ABP/MSA for each LLR updating iteration. For example, consider the (31,25) RS code given in Example 2. Each GFT-ISDD/MSA iteration requires 2880 real number computations, while each ABP/MSA iteration of the JN-(or the KM-) ABPA requires 5805 real number computations. For the (127,119) RS code over GF(27) given in Example 3, each GFT-ISDD/MSA iteration requires 21,616 real number computations to update the LLRs of a codeword in the cascaded (127,199) RS code, while each ABP/MSA iteration of the JN-(or the KM-) ABPA requires at least 70,476 real number computations (ignoring all the additional computations required in other steps).
For the JN-ABPA (or the KM-ABPA), to achieve good error performance, N1 and N2 must be reasonably large (even using the SPA for reliability updating). Large N1 and N2 increase the computational complexity drastically and make the implementation of the ABP decoder for an RS code impractical. Again, consider the (31,25) RS code over GF(25). For this code to achieve MLD performance, the KM-ABP-ASDA requires N1=20 and N2=10; then, the total number of real computations required to decode a codeword is at least 1,161,000. However, decoding the code with the GFT-ISDD/MSA achieves MLD performance at the SNR of 4.8 dB as shown in
Another significant advantage of the GFT-ISDD scheme over the ABPA is that the decoding matrix is fixed without new adaptation after each decoding iteration. Furthermore, since the parity-check matrix HRS,LDPC of the cascaded RS-LDPC code CRS,LDPC is an array of binary CMPs of size n×n, the reduced-complexity revolving iterative decoding (RID) scheme can be used in conjunction with the GFT-ISDD/MSA to reduce the hardware decoder complexity by a factor as large as n (the size of a CPM).
Decoding of BCH Codes of Prime Lengths with the Proposed GFT-ISDD Scheme
In one embodiment, the GFT-ISDD scheme is used to decode binary BCH codes of prime lengths. Similar to RS codes, BCH codes form an important class of error-correcting codes, discovered in 1959 and 1960. They have been widely used for error control in various communications and storage systems since the 1960s. BCH codes are generally decoded with the BM-HDDA. Several SDD algorithms or schemes (mostly reliability-based list decoding) have also been devised over the years. These decoding algorithms or schemes either achieve relatively small coding gains over the BM-HDDA or require large decoding complexity. In the following, we apply the GFT-ISDD scheme to decode BCH codes of prime lengths and show that superior error performance can be achieved with practically implementable complexity.
In one embodiment, for this subclass of codes, we set κ=1, q=2, and τ≥3. Let n be a prime factor of 2τ−1 and β be an element of order n in GF(2τ). Let t and m be two positive integers such that 1≤2t≤m<n. Let S be a set of m elements in the cyclic subgroup {1, β, β2, . . . , βn-1} of the field GF(2T) which consists of 2t consecutive powers of β, say β, β2, β3, . . . , β2t, and their conjugates, i.e., if η is an element of S, then η(2
From the m×n parity-check matrix BBCH over GF(2τ) for CBCH in the form given by Eqn. (1), we can construct the m×n array HBCH,LDPC of binary CPMs of size n×n, which is an mn×n2 binary matrix with column and row weights m and n, respectively. Note that the 2τ-ary BCH-LDPC code CBCH,LDPC given by the null space over GF(2τ) of the BCH-LDPC parity-check matrix HBCH,LDPC has minimum distance at least m+1, which is in general greater than the designed distance 2t+1 of the t-error-correcting base BCH code CBCH.
Encoding follows steps as shown in
In the following, we use several examples to demonstrate the effectiveness of the GFT-ISDD scheme for decoding BCH codes of prime lengths. This is even true for cyclic binary Hamming codes.
In one embodiment, we consider the triple-error-correcting (31,16) primitive BCH code CBCH whose generator polynomial has β, β2, β3, β4, β5, β6 and their conjugates as roots, a total of 15 roots, where β is a primitive element in GF(25). This code has minimum distance seven. Using the 15 roots of the generator polynomial of this BCH code, we form a 15×31 parity-check matrix BBCH over GF(25) in the form of Eqn. (1). Based on the parity-check matrix BBCH, we form the BCH-LDPC matrix HBCH,LDPC, a 15×31 array of 31×31 binary CPMs. The cascaded code CBCH,casc of the (31,16) BCH code CBCH is of length 961. The GFT of a codeword in the interleaved code CBCH,cascπ of the cascaded BCH code CBCH,casc is a codeword in the 32-ary BCH-LDPC code CBCH,LDPC given by the null space over GF(25) of the BCH-LDPC matrix HBCH,LDPC. Since HBCH,LDPC satisfies the RC-constraint and has column weight 15, the BCH-LDPC code CBCH,LDPC associated with the base BCH code CBCH has minimum distance at least 16 which is more than twice the minimum distance of the base code CBCH.
We decode the cascaded BCH code CBCH,casc using the GFT-ISDD scheme based on the parity-check array HBCH,LDPC. At the BCH-LDPC decoding stage, the received vector over GF(25) is decomposed into 5 binary constituent received vectors and each of these 5 binary vectors is decoded with 50 iterations of the MSA.
In one embodiment, a binary double-error-correcting primitive BCH code CBCH constructed from the field GF(27) is considered, which is a (127,113) cyclic code of prime length with rate 0.89. Let β be a primitive element of GF(27). The generator polynomial of this code has β, β2, β3, β4 and their conjugates as roots, a total of 14 roots. This BCH code is a subfield subcode of the (127,123) RS code over GF(27). Based on the 14 roots of its generator polynomial, we construct a 14×127 BCH-LDPC array HBCH,LDPC of binary CPMs of size 127×127. It is a 1778×16129 binary matrix with column weight 14 and row weight 127. Even though the (127,113) BCH base code has minimum distance 5, its associated 128-ary BCH-LDPC code CBCH,LDPC given by the null space over GF(27) of the BCH-LDPC matrix HBCH,LDPC has minimum distance at least 15.
In the next example, we show that even a cyclic binary Hamming code of prime length decoded with the proposed GFT-ISDD/MSA can achieve a very good error performance.
In one embodiment, a (127,120) cyclic binary Hamming code CHam is considered. The CHam code has a rate 0.945 and minimum distance 3, and is generated by a primitive polynomial of degree 7 over GF(2), g(X)=X7+X3+1. The subscript “Ham” stands for “Hamming”. This primitive polynomial has a primitive element β of GF(27) and its 6 conjugates as roots which are β2, β4, β8, β16, β32, and β64. Using these 7 roots, we form a 7×127 array HHam,LDPC of 127×127 binary CPMs, which is an 889×16129 binary matrix with column and row weights 7 and 127, respectively. We decode the (127,120) Hamming code in cascaded form in the GFT domain using the GFT-ISDD/MSA based on the Hamming-LDPC matrix HHam,LDPC. Suppose the decoding of the 128-ary Hamming-LDPC code CHam,LDPC associated with the base Hamming code CHam is performed with 5, 10, and 50 iterations of the MSA.
Decoding of QR Codes of Prime Lengths with the GFT-ISDD Scheme
Another interesting class of cyclic codes of prime lengths is the class of binary quadratic residue (QR) codes. QR codes were first introduced by Gleason and later were investigated by many mathematicians and coding theorists. QR codes are good in terms of their minimum distances, i.e., for a given length n and dimension k, say an (n, k) QR code may have the largest possible minimum distance compared to other linear block codes of the same length and dimension. Many short binary QR codes with good (even the best) minimum distances have been found. Among them, the most well-known QR code is the (23,12) Golay code with minimum distance seven. QR codes are good but in general, they are difficult to decode algebraically up to their error-correcting capabilities guaranteed by their minimum distances. Since QR codes are cyclic and of prime lengths, in one embodiment, the proposed GFT-ISDD scheme can be applied to decode them. In this section, we apply the GFT-ISDD scheme to decode two short binary QR codes to demonstrate that they can achieve very good error performances.
Before applying the GFT-ISDD scheme to QR codes, we give a brief introduction to these codes. Let n be a prime. Then, there exists a prime field, denoted by GF(n), with integer elements, 0,1,2, . . . , n−1, under modulo-n addition and multiplication. There are exactly half ((n−1)/2) of the nonzero elements that have square roots in GF(n), i.e., they are even powers of a primitive element in GF(n). These elements are referred to as quadratic residues. Let a be a primitive element of GF(n) and m=(n−1)/2. Then, m even powers of α, namely, (α0)n, (α2)n, (α4)n, . . . , (α2(m-1))n, give the (n−1)/2 quadratic residues in GF(n). Note that after the power (α2(m-1))n, the set of quadratic residues will repeat. As an example, consider the prime field GF(7)={0,1,2,3,4,5,6} under modulo-7 addition and multiplication. The integer 5 is a primitive element of GF(7) and m=(7−1)/2=3. If we take the powers, (50)7, (52)7, and (52×2)7, we obtain the quadratic residues, 1, 4, and 2, in GF(7). Note that (52×3)7=1.
From here on, the elements in the set {1,4,2} of quadratic residues will repeat. Another way to find the (n−1)/2 quadratic residues in GF(n) is to take the powers of two of the nonzero elements, 1, 2, 3, . . . , m, i.e., (12)n, (22)n, (32)n, . . . , (m2)n. Consider the prime field GF(7) given above. The powers of two of the elements, 1, 2, and 3, (i.e., (12)7, (22)7, and (32)7) give the 3 quadratic residues 1, 4, and 2 in GF(7) which are the same as computed above.
Let n be an odd prime and fln be the set of m=(n−1)/2 quadratic residues in the prime field GF(n). Let β be an element of order n in the field GF(2τ), where r be the smallest positive integer such that n is a factor of 2τ−1. Define the following polynomial:
gQR(X)=ΣlϵΩ
The polynomial gQR(X) defined by Eqn. (31) is a polynomial of degree m over GF(2), which divides Xn−1. The cyclic code with gQR(X) as the generator polynomial is a binary (n, n−m) QR code, denoted by CQR, of rate (n+1)/2n. Many short binary QR codes of this type have been constructed and they have very good (or even the best) minimum distances.
In one embodiment, the QR codes may be encoded following the steps described in
In one embodiment, consider the QR code of length 17. The set of quadratic residues in this case is Ω17={1,2,4,8,9,13,15,16}. The smallest positive integer τ for which 2τ−1 is divisible by 17 is τ=8. Let a be a primitive element of GF(28) which is a root of the primitive polynomial X8+X4+X3+X2+1, and β=α15. Then, β is an element of order 17 in GF(28). In this case, the generator polynomial gQR(X) given in (31) is X6+X4+X2+X+1, which is an irreducible polynomial with roots β, β2, β4, β8, β9, β13, β15, and β16 in GF(28). The cyclic code generated by gQR(X) is a binary (17,9) QR code CQR with minimum distance 5, which is the best linear code in terms of minimum distance for length 17 and dimension 9.
In one embodiment, suppose we decode this code in cascaded form using the GFT-ISDD/MSA. Based on the 8 roots of gQR(X), we form an 8×17 binary QR-LDPC array HQR,LDPC of CPMs of size 17×17, which is a 136×289 matrix with column and row weights 8 and 17, respectively. The QR-LDPC code associated with CQR is a 256-ary QC-QR-LDPC code of length 289 with minimum distance at least 9. We decode the cascaded code CQR,casc with CQR as the base code using the GFT-ISDD scheme with 50 iterations of the MSA scaled by a factor 0.5.
One embodiment considers the most well-known QR code (e.g., the (23,12) Golay code CGolay with minimum distance of 7), which is a perfect code. The set of 11 quadratic residues in the prime field GF(23) is Ω23={1,2,3,4,6,8,9,12,13,16,18}. The smallest positive integer τ for which 23 divides 2τ−1 is τ=11. Let a be a primitive element of GF(211) which is a root of the primitive polynomial X11+X2+1, and β=α89. Then the order of β is 23. The generator polynomial of the (23,12) Golay code CGolay is
gGolay(X)=X11+X9+X7+X6+X5+X+1,
which has β, β2, β3, β4, β6, β8, β9, β12, β13, β16, and β18 as roots.
Note that 89 is also a prime factor of 211−1. In one embodiment, using the quadratic residues of the prime field GF(89), an (89,45) QR code with minimum distance 15 can be constructed.
As described above, in one embodiment, the novel and effective coding scheme disclosed herein can be used for encoding and iterative soft-decision decoding of a class of cyclic codes of prime lengths, n, over finite fields of characteristic two. This class of cyclic codes includes all RS and BCH of prime lengths and all QR codes over such fields. Two of the keys to the disclosed coding scheme are joint-encoding and joint-decoding of a collection of codewords from a code. In one embodiment, the coding scheme includes applying Hadamard-permutations (or their inverses) to symbols of the encoded or decoded codewords in the collection; combining (or de-combining) encoded or decoded codewords into a codeword over a larger field; interleaving (or de-interleaving) encoded or decoded codewords; and taking GFTs (or their inverses) to the sequence of interleaved encoded or decoded codewords.
The joint-decoding and information sharing according to one embodiment results in an error performance per decoded codeword better than the error performance of a received codeword decoded individually using MLD as demonstrated in several embodiments. Another important feature of the proposed GFT-ISDD scheme, according to one embodiment, is that the decoding of a non-binary received sequence is carried out in binary based on a binary LDPC matrix. The binary iterative decoding can be performed efficiently and reduces the decoding complexity significantly. For decoding RS, BCH, and QR codes of prime lengths, the proposed coding scheme not only requires much lower decoding complexity than other existing soft-decision decoding algorithms for these codes, but also yields superior performance as demonstrated by the examples. In one embodiment, it is shown that with the proposed GFT-ISDD scheme, even a cyclic Hamming code can achieve an excellent error performance. It should be noted that although RS, BCH and QR codes are mentioned as examples, in general, the proposed coding scheme (e.g., collective encoding and/or collective decoding) can be applied to any type of code without departing from the teachings of the present disclosure.
It should be noted that one of the important structures of the QC-LDPC code CLDPC is that its parity check matrix HLDPC is binary, an array of CPMs over GF(2), whose Tanner graph is free of cycles of length four. The binary and QC-structures of the LDPC matrix HLDPC significantly facilitate and simplify the proposed decoding at the receiver side. Since both the base code C and the (n, n−1) SPC code are cyclic, the encoding of a collection of these codes can, in one embodiment, be accomplished with two simple feedback shift-registers and a buffer memory for permuting and interleaving the n generated codewords. The GFT of the interleaved sequence of n permuted codewords can be done with a fast Fourier transform algorithm.
In this disclosure, we also showed that the proposed coding scheme can be applied to shortened cyclic codes of prime lengths and presented a technique to reduce the rate of a cascaded code in the GFT-domain to various degrees while maintaining the code length. Besides RS, BCH, and QR codes of prime lengths, there are other well-known cyclic codes of prime lengths, such as cyclic Reed-Muller (RM) codes, finite geometry codes, difference set codes, generalized RM codes, and polynomial codes. All these codes are majority-logic decodable, which makes HDD much easier. However, if the proposed decoding scheme is applied to any of these codes, it yields superior performance.
The computer system 1900 is shown comprising hardware elements that can be electrically coupled via a bus 1905 (or may otherwise be in communication, as appropriate). The hardware elements may include one or more processors 1910, including without limitation one or more general-purpose processors and/or one or more special-purpose processors (such as digital signal processing chips, graphics acceleration processors, video decoders, and/or the like); one or more input devices 1915, which can include without limitation a mouse, a keyboard, remote control, and/or the like; and one or more output devices 1920, which can include without limitation a display device, a printer, and/or the like.
The computer system 1900 may further include (and/or be in communication with) one or more non-transitory storage devices 1925, which can comprise, without limitation, local and/or network accessible storage, and/or can include, without limitation, a disk drive, a drive array, an optical storage device, a solid-state storage device, such as a random access memory (“RAM”), and/or a read-only memory (“ROM”), which can be programmable, flash-updateable and/or the like. Such storage devices may be configured to implement any appropriate data stores, including without limitation, various file systems, database structures, and/or the like.
The computer system 1900 might also include a communications subsystem 1930, which can include without limitation a modem, a network card (wireless or wired), an infrared communication device, a wireless communication device, and/or a chipset (such as a Bluetooth™ device, an 802.11 device, a Wi-Fi device, a WiMax device, cellular communication device, GSM, CDMA, WCDMA, LTE, LTE-A, LTE-U, etc.), and/or the like. The communications subsystem 1930 may permit data to be exchanged with a network (such as the network described below, to name one example), other computer systems, and/or any other devices described herein. In many embodiments, the computer system 1900 will further comprise a working memory 1935, which can include a RAM or ROM device, as described above.
The computer system 1900 also can comprise software elements, shown as being currently located within the working memory 1935, including an operating system 1940, device drivers, executable libraries, and/or other code, such as one or more application programs 1945, which may comprise computer programs provided by various embodiments, and/or may be designed to implement methods, and/or configure systems, provided by other embodiments, as described herein. Merely by way of example, one or more procedures described with respect to the method(s) discussed above might be implemented as code and/or instructions executable by a computer (and/or a processor within a computer); in an aspect, then, such code and/or instructions can be used to configure and/or adapt a general purpose computer (or other device) to perform one or more operations in accordance with the described methods.
A set of these instructions and/or code might be stored on a non-transitory computer-readable storage medium, such as the non-transitory storage device(s) 1925 described above. In some cases, the storage medium might be incorporated within a computer system, such as computer system 1900. In other embodiments, the storage medium might be separate from a computer system (e.g., a removable medium, such as a compact disc), and/or provided in an installation package, such that the storage medium can be used to program, configure, and/or adapt a general purpose computer with the instructions/code stored thereon. These instructions might take the form of executable code, which is executable by the computer system 1900 and/or might take the form of source and/or installable code, which, upon compilation and/or installation on the computer system 1900 (e.g., using any of a variety of generally available compilers, installation programs, compression/decompression utilities, etc.), then takes the form of executable code.
It will be apparent to those skilled in the art that substantial variations may be made in accordance with specific requirements. For example, customized hardware might also be used, and/or particular elements might be implemented in hardware, software (including portable software, such as applets, etc.), or both. Further, connection to other computing devices such as network input/output devices may be employed.
As mentioned above, in one aspect, some embodiments may employ a computer system (such as the computer system 1900) to perform methods in accordance with various embodiments of the invention. According to a set of embodiments, some or all of the procedures of such methods are performed by the computer system 1900 in response to processor 1910 executing one or more sequences of one or more instructions (which might be incorporated into the operating system 1940 and/or other code, such as an application program 1945) contained in the working memory 1935. Such instructions may be read into the working memory 1935 from another computer-readable medium, such as one or more of the non-transitory storage device(s) 1925. Merely by way of example, execution of the sequences of instructions contained in the working memory 1935 might cause the processor(s) 1910 to perform one or more procedures of the methods described herein.
The terms “machine-readable medium,” “computer-readable storage medium” and “computer-readable medium,” as used herein, refer to any medium that participates in providing data that causes a machine to operate in a specific fashion. These mediums may be non-transitory. In an embodiment implemented using the computer system 1900, various computer-readable media might be involved in providing instructions/code to processor(s) 1910 for execution and/or might be used to store and/or carry such instructions/code. In many implementations, a computer-readable medium is a physical and/or tangible storage medium. Such a medium may take the form of a non-volatile media or volatile media. Non-volatile media include, for example, optical and/or magnetic disks, such as the non-transitory storage device(s) 1925. Volatile media include, without limitation, dynamic memory, such as the working memory 1935.
Common forms of physical and/or tangible computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, or any other magnetic medium, a CD-ROM, any other optical medium, any other physical medium with patterns of marks, a RAM, a PROM, EPROM, a FLASH-EPROM, any other memory chip or cartridge, or any other medium from which a computer can read instructions and/or code.
Various forms of computer-readable media may be involved in carrying one or more sequences of one or more instructions to the processor(s) 1910 for execution. Merely by way of example, the instructions may initially be carried on a magnetic disk and/or optical disc of a remote computer. A remote computer might load the instructions into its dynamic memory and send the instructions as signals over a transmission medium to be received and/or executed by the computer system 1900.
The communications subsystem 1930 (and/or components thereof) generally will receive signals, and the bus 1905 then might carry the signals (and/or the data, instructions, etc. carried by the signals) to the working memory 1935, from which the processor(s) 1910 retrieves and executes the instructions. The instructions received by the working memory 1935 may optionally be stored on a non-transitory storage device 1925 either before or after execution by the processor(s) 1910.
It should further be understood that the components of computer system 1900 can be distributed across a network. For example, some processing may be performed in one location using a first processor while other processing may be performed by another processor remote from the first processor. Other components of computer system 1900 may be similarly distributed. As such, computer system 1900 may be interpreted as a distributed computing system that performs processing in multiple locations. In some instances, computer system 1900 may be interpreted as a single computing device, such as a distinct laptop, desktop computer, or the like, depending on the context.
The methods, systems, and devices discussed above are examples. Various configurations may omit, substitute, or add various procedures or components as appropriate. For instance, in alternative configurations, the methods may be performed in an order different from that described, and/or various stages may be added, omitted, and/or combined. Also, features described with respect to certain configurations may be combined in various other configurations. Different aspects and elements of the configurations may be combined in a similar manner. Also, technology evolves and, thus, many of the elements are examples and do not limit the scope of the disclosure or claims.
Specific details are given in the description to provide a thorough understanding of example configurations (including implementations). However, configurations may be practiced without these specific details. For example, well-known circuits, processes, algorithms, structures, and techniques have been shown without unnecessary detail in order to avoid obscuring the configurations. This description provides example configurations only, and does not limit the scope, applicability, or configurations of the claims. Rather, the preceding description of the configurations will provide those skilled in the art with an enabling description for implementing described techniques. Various changes may be made in the function and arrangement of elements without departing from the spirit or scope of the disclosure.
Also, configurations may be described as a process which is depicted as a flow diagram or block diagram. Although each may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be rearranged. A process may have additional steps not included in the figure. Furthermore, examples of the methods may be implemented by hardware, software, firmware, middleware, microcode, hardware description languages, or any combination thereof. When implemented in software, firmware, middleware, or microcode, the program code or code segments to perform the necessary tasks may be stored in a non-transitory computer-readable medium such as a storage medium. Processors may perform the described tasks.
Having described several example configurations, various modifications, alternative constructions, and equivalents may be used without departing from the spirit of the disclosure. For example, the above elements may be components of a larger system, wherein other rules may take precedence over or otherwise modify the application of the invention. Also, a number of steps may be undertaken before, during, or after the above elements are considered.
The present application claims priority to Provisional Application No. 62/257,587 entitled “SYSTEMS, APPARATUSES AND METHODS FOR ERROR CORRECTION USING CYCLIC CODE-BASED LDPC CODES,” filed Nov. 19, 2015, which is assigned to the assignee hereof and expressly incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
7810015 | Paumier | Oct 2010 | B2 |
20100077275 | Yu | Mar 2010 | A1 |
20120307928 | Jia | Dec 2012 | A1 |
20150381204 | Zhang | Dec 2015 | A1 |
20160248442 | Myung | Aug 2016 | A1 |
Entry |
---|
Barnault et al., Fast Decoding Algorithm for LDPC Over GF(2/sup q/), IEEE Information Theory Workshop, Paris, France, Apr. 4, 2003, pp. 70-73. |
Bellorado et al., Low-Complexity Soft-Decoding Algorithms for Reed-Solomon Codes part I: An Algebraic Soft-In hard-Out Chase Decoder, IEEE Trans. Inf. Theory, vol. 56, No. 3, Mar. 11, 2010, 13 pages. |
Bellorado et al., Low-Complexity Soft-Decoding Algorithms for Reed-Solomon Codes- Part II: Soft-Input Soft-Out Iterative Decoding, IEEE Trans. Inf. Theory, vol. 56, No. 3, Mar. 2010, pp. 960-967. |
BERLEKAMP, Algebraic Coding Theory, McGraw-Hill, World Scientific Publishing Co. Pte. Ltd., 2015, pp. 119-199. |
BLAHUT, Theory and Practice of Error Control Codes, Addison-Wesley Publishing Company, Jan. 1, 1983, pp. 93-322. |
Blake et al., The Mathematical Theory of Coding, Academic Press, 1975, pp. 139-143. |
Bose et al., On a Class of Error Correcting Binary Group Codes, Information and Control, vol. 3, No. 1, Mar. 1960, pp. 68-79. |
BRIGHAM, Fast Fourier Transform and Its Applications, Upper Saddle River, NJ: Prentice Hall, 1988, pp. 131-166. |
Cassuto et al., On the Average Complexity of Reed-Solomon List Decoders, California Institute of Technology, IEEE Transactions on Information Theory, vol. 59, No. 4, Apr. 2013, pp. 2336-2351. |
CHASE, Class of Algorithms for Decoding Block Codes with Channel Measurement Information, IEEE Trans. Information Theory, vol. 18, No. 1, Jan. 1972, pp. 170-182. |
Chen et al., Near Optimum Universal Belief Propagation Based Decoding of Low-Density Parity-Check Codes, IEEE Trans. Inf. Theory, vol. 50, No. 3, Mar. 2002, pp. 406-414. |
Chen et al., Overlapped Message Passing for Quasi-Cyclic Low-Density Parity Check Codes, IEEE Trans. Circuits and Systems I, vol. 51, No. 6, Jun. 2004, pp. 1106-1113. |
Chen et al., Performance and Analysis of Quadratic Residue Codes of Lengths Less than 100, Available online at: https://arxiv.org/abs/1408.5674, Aug. 25, 2014, 31 pages. |
Chen et al., Progressive Algebraic Soft-Decision Decoding of Reed-Solomon Codes, IEEE Trans. Commun., vol. 61, No. 2, Feb. 2013, pp. 433-442. |
Davey et al., Low Density Parity Check Codes Over GF(q), IEEE Commun. Lett., vol. 2, No. 6,, Jun. 1998, pp. 165-167. |
Declercq et al., Decoding Algorithms for Nonbinary LDPC Codes Over GF(q), IEEE Transactions on Communication, vol. 55, No. 4, Apr. 16, 2007, pp. 633-643. |
Diao et al., A Matrix-Theoretic Approach for Analyzing Quasi-Cyclic Low-Density Parity-Check Codes, IEEE Trans. Inf. Theory, vol. 58, No. 6, Jun. 2012, pp. 4030-4048. |
Diao et al., A Transform Approach for Analyzing and Constructing Quasi-Cyclic Low-Density Parity-Check codes, Proc. Inf. Theory Applic. Workshop, La Jolla, CA., Feb. 6-11, 2011, pp. 1-8. |
Diao et al., LDPC Codes on Partial Geometries: Construction, Trapping Set Structure, and Puncturing, IEEE Trans. Inf. Theory, vol. 59, No. 12, Dec. 2013, pp. 7898-7914. |
FORNEY, Generalized Minimum Distance Decoding, IEEE Trans. Inf. Theory, vol. 12, No. 2, Apr. 1966, pp. 125-131. |
FOSSORIER, Reduced Complexity Iterative Decoding of Low-Density Parity Check Codes Based on Belief Propagation, IEEE Transactions on Communication, vol. 47, No. 5, May 1999, pp. 673-680. |
FOSSORIER, Reliability-Based Decoding of Reed-Solomon Codes Using their Binary Image, IEEE Commun. Lett., vol. 8, No. 7, Jul. 2004, pp. 452-454. |
GALLAGER, Low Density Parity-Check Codes, IRE Trans. Inf. Theory, vol. IT-8, No. 1, Jan. 1962, pp. 21-28. |
GLEASON, Weight polynomials of self-dual codes and the MacWilliams identities, Actes Congr. Int. Math., vol. 3, Gauthier-Viliars, Paris, 1970, pp. 211-215. |
GOLAY, Notes on digital coding, Proc. IRE, vol. 37. Jun. 1949, 1 page. |
Gross et al., Applications of algebraic soft-decision decoding of Reed-Solomon codes, IEEE Trans. Commun., vol. 54, No. 7, Jul. 2006, pp. 1224-1234. |
Guruswami et al., Improved Decoding of Reed-Solomon and Algebraic-Geometry Codes, IEEE Trans. Inf. Theory, vol. 45, No. 6, Sep. 1999, pp. 1757-1767. |
Guruswami et al., Maximum-Likelihood Decoding of Reed-Solomon codes is NP-hard, IEEE Trans Inf, Theory, vol. 51, No. 7, Jul. 2005, pp. 2249-2256. |
HAMMING, Error Detecting and Error Correcting Codes, Bell Syst. Tech. J., vol. 29, No. 2, Apr. 1950, 14 pages. |
HOCQUENGHEM, Codes Correcteurs d'erreurs, Chiffres, vol. 2, http://www.citeulike.org/user/alexzeh /articie/2211978https://en.wikipedia.org/wiki/Alexis_Hocquenghem, Sep. 1959, 3 pages. |
Horn et al., Matrix Analysis, Cambridge University Press. Cambridge Books Online, 1985, pp. 425-505. 18 pages. |
Hu et al., An Efficient Hybrid Decoding Algorithm for Reed-Solomon Codes Based on Bit Reliability, IEEE Trans. Communication, vol. 51, No. 7, Jul. 28, 2003, pp. 1073-1081. |
Jiang et al,, Algebraic Soft-Decision Decoding of Reed-Solomon Codes Using Bit-Level Soft Information, IEEE Trans. Inf. Theory, vol. 54, No. 9, Sep. 2008, pp. 3907-3928. |
Jiang et al., Iterative Soft-Decision Decoding of Reed-Solomon Codes Based on Adaptive Parity-Check Matrices, Proc. IEEE Int. Symp. Inf. Theory, Chicago, IL, Jun. 27-Jul. 2, 2004, 1 page. |
Jiang et al., Iterative Soft-Input Soft-Output Decoding of Reed-Solomon Codes by Adapting the Parity-Check Matrix, IEEE Trans. Inf. Theory, vol. 52, No. 8, Aug. 2006, 10 pages. |
Kasami et al., New Generations of the Reed-Muller Codes part I: Primitive codes, IEEE Trans. Inf. Theory, vol. 14, No. 2, Mar. 1968, pp. 189-199. |
Kasami et al., On Majority-Logic Decoding for Duals of Primitive Polynomial Codes, IEEE Trans. Inf. Theory, vol. 17, No. 3, May 1971, pp. 322-331. |
Kasami et al., Polynomial Codes, IEEE Trans. Inf. Theory, vol. 14, No. 6, Nov. 1968, pp. 807-814. |
Khamy et al., Iterative Algebraic Soft-Decision List Decoding of Reed-Solomon Codes, IEEE Journal on Selected Areas in Communications, vol. 24, No. 3, Mar. 2006, pp. 481-490. |
Koetter et al., Algebraic Soft-Decision Decoding of Reed-Solomon Codes, IEEE Trans. Inf. Theory, vol. 49, No. 11, Nov. 2003, pp. 2809-2825. |
Koetter et al., The Re-Encoding Transformation in Algebraic List-Decoding of Reed-Solomon Codes, IEEE Trans. Inf. Theory, vol. 57, No. 2, Feb. 2011, pp. 633-647. |
Kou et al., Low-Density Parity-Check Codes based on Finite Geometries: A Rediscovery and New Results, IEEE Trans. Inf. Theory, vol. 47, No. 7, Oct. 2001, pp. 2711-2738. |
Lan et al, Construction of Quasi-Cyclic LDPC Codes for AWGN and Binary Erasure Channels: A Finite Field Approach, IEEE Trans. Inf. Theory, vol. 53, No. 7, Jul. 2007, pp. 2429-2458. |
Leroux et al., Stochastic Chase Decoding of Reed-Solomon Codes, IEEE Communications letters, vol. 14, No. 9, Sep. 2010, pp. 863-865. |
Li et al, Algebraic Quasi-Cyclic LDPC codes: Construction, Low Error-Floor, Large Girth and a Reduced-Complexity Decoding Scheme, IEEE Trans. Commun., vol. 62, No. 8. Aug. 2014, pp, 2626-2637. |
Li et al., Decoding of Quasi-Cyclic LDPC Codes with Section-Wise Cyclic Structure, in Proc. Inf. Theory Applic. Workshop, San Diego, CA, Feb. 9-14, 2014, pp. 1-10. |
Lin et al., Error Control Coding: Fundamentals and Applications, 2nd Edition, Chapters 4.1, 5.1, 5.3, 5.5, 6.1, 6.2, 6.3, 6.4, 6.5, 7.2, 7.3. 7,4, 7.6, 17.1, 17,2, 17.6 and 17.12, 2005, 100 pages. |
Liu et al., A Revolving Iterative Algorithm for Decoding Algebraic Cyclic and Quasi-Cyclic LDPC Codes, IEEE Trans. Commun., vol. 61, No. 12, Dec. 2013, pp. 4816-4827. |
Liu et al, A Revolving Iterative Algorithm for Decoding Algebraic Quasi-Cyclic LDPC Codes, Proc. IEEE Int. Symp. Inf. Theory, Istanbul, Turkey, Jun. 7-12, 2013, pp. 2656-2660. |
Liu et al., Turbo Encoding and Decoding of Reed-Solomon Codes through Binary Decomposition and Self-Concatenation, IEEE Trans. Commun., vol. 52, No. 9, Sep. 2004, pp. 1484-1493. |
MacKay et al., Evaluation of Gallager Codes of Short Block Length and High Rate Applications, Proc. IMA Int. Conf. Mathematics and Its Applications: Codes, Systems and Graphical Models New York, Springer-Verlag, https://link.springer.com/chapter/10.1007%2F978-1-4613-0165-3_6, 2001, pp. 113-130, 18 pages. |
MACKAY, Good Error-Correcting Codes Based on Very Sparse Matrices, IEEE Trans. Inf Theor., vol. 45, No. 2, Mar. 1999. pp. 399-431. |
MASSEY, Shift-Register Synthesis and BCH Decoding, IEEE Trans. Inf. Theory, vol. 15, No. 1, Jan. 1969, pp. 122-127. |
Peterson et al., Error-Correcting Codes, 2nd Edition, Cambridge, 1972, pp. 170-308. |
PLESS, Introduction to the Theory of Error-Correcting Codes, 3rd Edition, Wiley http://www.wiley.com/WileyCDA/WileyTitle/productCd-0471190470.html, Jul. 1998, 214 pages. |
PONNAMPALAM, Soft Decision Decoding of Reed-Solomon codes, IEEE Trans. Commun., vol. 50, No. 11, Nov. 2002, pp. 1758-1768. |
Reed et al., Polynomial Codes Over Certain Finite Fields, Journal of the Society for Industrial and Applied Mathematics, vol. 8, No. 2, pp. 300-304 http://epubs.siam.org/doi/pdtplus/10.1137/0108018, Jun. 1960, 2 pages. |
ROTH, Introduction to Coding Theory, UK: Cambridge University Press, 2006, pp. v-xi. |
Ryan et al., Channel Codes: Classical and Modern, New York, NY: Cambridge Univ. Press, 2009, 710 pages. |
Shin et al., Soft Decision Decoding ot Reed-Solomon Codes Using Trellis methods, IEE Proc Commun., vol. 141, No. 5. Oct. 1994, pp. 303-308. |
SINGLETON, Maximum Distance q-nary codes, IEEE Trans. Inf. Theory, vol. 10, No. 2, Apr. 1964, pp. 116-118. |
Sugiyama et al., A Method for Solving Key Equation for Decoding Goppa codes, Information and Control, vol. 27, No. 12, Jan. 1975, pp. 87-99. |
Tang et al., On combining Chase-2 and GMD Decoding Algorithms for Nonbinary Block Codes, IEEE Commun. Lett., vol. 5, No. 5, May 2001, pp. 209-211. |
Tang et al., Progressive List-Enlarged Algebraic Soft Decoding of Reed-Solomon codes, IEEE Commun. Lett., vol. 16, No. 8, Jun. 2012, pp. 901-904. |
TANNER, A Recursive Approach to Low Complexity Codes, IEEE Trans. Inf. Theory, vol. 27, No. 5, Sep. 1981, pp. 533-547. |
TRIFONOV, Efficient Interpolation in the Guruswami-Sudan Algorithm, IEEE Trans. Inf. Theory, vol. 56, No. 9, Sep. 2010, pp. 4341-4349. |
Vardy et al., Bit-Level Soft-Decision Decoding of Reed-Solomon codes, IEEE Trans. Commun., vol. 39, No. 3, Mar. 1991, pp. 440-444. |
WU, New List Decoding Algorithms for Reed-Solomon and BCH codes, IEEE Trans. Inf. Theory, vol. 54, No. 8, Aug. 2008, pp. 3611-3630. |
Number | Date | Country | |
---|---|---|---|
20170149444 A1 | May 2017 | US |
Number | Date | Country | |
---|---|---|---|
62257587 | Nov 2015 | US |