Communication systems employ coding to ensure reliable communication across noisy communication channels. These communication channels exhibit a fixed capacity that can be expressed in terms of bits per symbol at certain signal to noise ratio (SNR), defining a theoretical upper limit (known as the Shannon limit). As a result, coding design has aimed to achieve rates approaching this Shannon limit. One such class of codes that approach the Shannon limit is Low Density Parity Check (LDPC) codes.
Traditionally, LDPC codes have not been widely deployed because of a number of drawbacks. One drawback is that the LDPC encoding technique is highly complex. Encoding an LDPC code using its generator matrix would require storing a very large, non-sparse matrix. Additionally, LDPC codes require large blocks to be effective; consequently, even though parity check matrices of LDPC codes are sparse, storing these matrices is problematic.
From an implementation perspective, a number of challenges are confronted. For example, storage is an important reason why LDPC codes have not become widespread in practice. Also, a key challenge in LDPC code implementation has been how to achieve the connection network between several processing engines (nodes) in the decoder. Further, the computational load in the decoding process, specifically the check node operations, poses a problem.
Therefore, there is a need for an LDPC communication system that employs simple encoding and decoding processes. There is also a need for using LDPC codes efficiently to support high data rates, without introducing greater complexity. There is also a need to improve performance of LDPC encoders and decoders. There is also a need to minimize storage requirements for implementing LDPC coding. There is a further need for a scheme that simplifies the communication between processing nodes in the LDPC decoder.
These and other needs are addressed by the present invention, wherein various approaches for encoding and decoding structured Low Density Parity Check (LDPC) codes is provided.
According to one aspect of an exemplary embodiment, a method comprises accessing memory storing information representing a structured parity check matrix of Low Density Parity Check (LDPC) codes. The information is organized in tabular form, such each row represents occurrences of one value within a first column of a group of columns of the parity check matrix, the rows correspond to groups of columns of the parity check matrix. Also, subsequent columns within each of the groups are derived according to a predetermined operation. Further, an LDPC coded signal is outputted based on the stored information representing the parity check matrix. Parity bit accumulators are to initialized to zero, and first information bit in the jth group of M information bits is accumulated in the ith parity bit accumulator if the ith entry in (jM)th column of the parity check matrix is 1, where j=0, 1, 2, 3, . . . kldpc/M−1, the remaining (M−1) information bits m=jM+1, jM+2, jM+3, . . . , (j+1)M−1 of the jth group being accumulated in the parity bit accumulators according to {x+m mod M×q} mod(nldpc−kldpc), wherein x denotes the address of the parity bit accumulator corresponding to the first bit, jM, in the group, and
is a code rate dependent constant, and after all of the information bits are exhausted, operations, starting with i=1 are performed according to p1=p1⊕p1−1, 2, . . . , nldpc−kldpc−1, wherein final content of p1=0, 1, . . . , nldpc−kldpc−1 is equal to the parity bit p1.
According to another aspect of an exemplary embodiment, an apparatus comprises a memory which is configured to store information representing a structured parity check matrix of Low Density Parity Check (LDPC) codes. The information is organized in tabular form, wherein each row represents occurrences of one values within a first column of a group of columns of the parity check matrix. Also, the rows correspond to groups of columns of the parity check matrix, where subsequent columns within each of the groups are derived according to a predetermined operation. The apparatus also comprises circuitry coupled to the memory and configured to output an LDPC coded signal using the stored information representing the parity check matrix. Parity bit accumulators are to initialized to zero, and first information bit in the jth group of M information bits is accumulated in the ith parity bit accumulator if the ith entry in (jM)th column of the parity check matrix is 1, where j=0, 1, 2, 3, . . . kldpc/M−1, the remaining (M−1) information bits m=jM+1, jM+2, jM+3, . . . , (j+1)M−1 of the jth group being accumulated in the parity bit accumulators according to {x+m mod M×q} mod(nldpc−kldpc), wherein x denotes the address of the parity bit accumulator corresponding to the first bit, jM, in the group, and
is a code rate dependent constant, and after all of the information bits are exhausted, operations, starting with i=1 are performed according to p1=p1⊕i=1, 2, . . . , nldpc−kldpc−1, wherein final content of p1, i=0, 1, . . . , nldpc−kldpc−1 is equal to the parity bit pi.
According to another aspect of an exemplary embodiment, a method comprises accessing memory storing edge information associated with a structured parity check matrix used to generate Low Density Parity Check (LDPC) coded signal. The edge information represent relationship between bit nodes and check nodes, and are stored according to a predetermined scheme that permits concurrent retrieval of a set of the edge information. Further, a decoded signal is outputted corresponding to the LDPC coded signal based on the stored edge information.
According to another aspect of an exemplary embodiment, a method comprises accessing memory storing edge information and a posteriori probability information associated with a structured parity check matrix used to generate Low Density Parity Check (LDPC) coded signal. The edge information represents relationship between bit nodes and check nodes, and are stored according to a predetermined scheme that permits concurrent retrieval of a set of the edge information. Further, a decoded signal is outputted corresponding to the LDPC coded signal based on the stored edge and a posteriori probability information.
Still other aspects, features, and advantages of the present invention are readily apparent from the following detailed description, simply by illustrating a number of particular embodiments and implementations, including the best mode contemplated for carrying out the present invention. The present invention is also capable of other and different embodiments, and its several details can be modified in various obvious respects, all without departing from the spirit and scope of the present invention. Accordingly, the drawing and description are to be regarded as illustrative in nature, and not as restrictive.
The present invention is illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:
A system, method, and software for efficiently encoding and decoding structured Low Density Parity Check (LDPC) codes are described. In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It is apparent, however, to one skilled in the art that the present invention may be practiced without these specific details or with an equivalent arrangement. In other instances, well-known structures and devices are shown in block diagram form in order to avoid unnecessarily obscuring the present invention.
The LDPC codes that are generated by the transmitter 101 enable high speed implementation without incurring any performance loss. These structured LDPC codes output from the transmitter 101 avoid assignment of a small number of check nodes to the bit nodes already vulnerable to channel errors by virtue of the modulation scheme (e.g., quadrature phase-shift keying (QPSK), offset quadrature phase-shift keying (OQPSK), 8-PSK, 16 amplitude and phase-shift keying (16-APSK), etc.).
Such LDPC codes have a parallelizable decoding algorithm (unlike turbo codes), which advantageously involves simple operations such as addition, comparison and table look-up. Moreover, carefully designed LDPC codes do not exhibit any sign of error floor.
According to one embodiment, the transmitter 101 generates, using a relatively simple encoding technique, LDPC codes based on parity check matrices (which facilitate efficient memory access during decoding) to communicate with the receiver 105. The transmitter 101 employs LDPC codes that can outperform concatenated turbo+RS (Reed-Solomon) codes, provided the block length is sufficiently large.
Encoder 203 generates signals from alphabet Y to a modulator 205 using a simple encoding technique that makes use of the parity check matrix by imposing structure onto the parity check matrix. According to certain embodiments, a restriction can be placed on the parity check matrix by constraining certain portion of the matrix to be triangular. The construction of such a parity check matrix is described more fully below in
Modulator 205 maps the encoded messages from encoder 203 to signal waveforms that are transmitted to a transmit antenna 207, which emits these waveforms over the communication channel 103. Accordingly, the encoded messages are modulated and distributed to a transmit antenna 207. In an exemplary embodiment, the modulation can include quadrature phase-shift keying (QPSK), offset quadrature phase-shift keying (OQPSK), 8-PSK, and/or 16 amplitude and phase-shift keying (16-APSK) modulation. The transmissions from the transmit antenna 207 propagate to a receiver, as discussed below.
The LDPC encoder 203 systematically encodes an information block of size kldpc, i=(i0, i1, . . . , ik
According to certain embodiments, for LDPC codes three different code rates can be defined: 1/2, 2/3, and 4/5. Also, for each code rate 1/2, 2/3, and 4/5, 22 different coded block sizes can be defined: 720, 960, 1200, 1440, 1680, 1920, 2160, 2400, 2640, 2880, 3120, 3360, 3600, 3840, 4080, 4320, 4560, 4800, 5040, 5280, 5520, and 5760 coded bits, corresponding to 3 up to 24 slots. Further, bursts longer than 24 slots can be obtained by coding multiple shorter LDPC codes of “almost equal” sizes. Among these 22 block sizes, eight of them correspond to mother LDPC codes, which include 720, 960, 1440, 2160, 2880, 3600, 4320, and 5760 bit blocks. The other 14 codes can be derived from the mother codes by shortening and puncturing, as discussed below.
According to an exemplary embodiment, shortening and puncturing process can be explained with respect to
For each block size which is not a mother code size, the parameters related to shortening and puncturing as well as the mother codes are given in Table 1 below. Note that if Kmother and Nmother denote the number of un-coded and coded bits of the mother code, respectively, for the derived code:
K=K
mother
−XS,
N=N
mother
−XS−XP.
Here, K and N denote the number of un-coded and coded bits of the derived code, respectively.
In this exemplary embodiment, for each mother LDPC code, the degree distribution of bit nodes is given in Table 2 below, where N denotes the total number of bit nodes, i.e., coded block size. For each code, all of the check nodes except one have the same degree, namely dc=7 for rate 1/2, dc=11 for rate 2/3, and dc=20 for rate 4/5. The remaining check node has degree one less.
According to certain embodiments, the task of the LDPC encoder 203 is to determine Nldpc−kldpc parity bits (p0, p1, . . . , pn
p83=p83⊕i0
p117=p117⊕i0
p156=p156⊕i0
p169=p169⊕i0
p231=p231⊕i0
p126=p126⊕i0
p112=p112⊕i0
p106=p106⊕i0
(All additions are in Galois field (2) (GF(2))).
Then, for the next M−1 information bits, im, m=1, 2, . . . , M−1, these bits are accumulated at parity bit addresses {x+m mod M×q} mod(nldpc−kldpc), where x denotes the address of the parity bit accumulator corresponding to the first bit i0, M is number of columns in edge RAM (Table 44), and
Continuing with the example, M=30 and q=8 for code rate 2/3 and coded block size 720. By way of example, for information bit i1 the following operations are performed:
p91=p91⊕i1
p125=p125⊕i1
p164=p164⊕i1
p177=p177⊕i1
p239=p239⊕i1
p134=p134⊕i1
p120=p120⊕i1
p114=p114⊕i1
For the (M+1)st information bit iM, the addresses of the parity bit accumulators are given in the second row of the Tables 3 through 26. In a similar manner the addresses of the parity bit accumulators for the following M−1 information bits im, m=M+1, M+2, . . . , 2M−1 are obtained using the formula {x+m mod M×q} mod(nldpc−kldpc), where x denotes the address of the parity bit accumulator corresponding to the information bit iM, i.e., the entries in the second row of the Tables 3-26. In a similar manner, for every group of M new information bits, a new row from Tables 3 through 26 are used to find the addresses of the parity bit accumulators.
After all of the information bits are exhausted, the final parity bits are obtained as follows. First, the following operations are performed, starting with i=1
p1=p1⊕p1−1,i=1,2, . . . ,nldpc−kldpc−1.
Final content of pi, i=0, 1, . . . , nldpc−kldpc−1 is equal to the parity bit pi.
Continuing with this scenario, the LDPC encoder 203 systematically encodes an information block of size kldpc, i=(i0, i1, . . . , ik
According to this exemplary embodiment, the LDPC encoder 203 determines nldpc−kldpc parity bits (p0, p1, . . . , pn
p853=p853⊕i0
p2958=p2958⊕i0
p3632=p3632⊕i0
p4515=p4515⊕i0
(All additions are in GF(2)).
Then, for the next 359 information bits, im, m=1, 2, . . . , 359, these bits are accumulated at parity bit addresses
where x denotes the address of the parity bit accumulator corresponding to the first bit i0. By way of example, for information bit i1, the following operations are performed:
p854=p854⊕i1
p2959=p2959⊕i1
p3633=p3633⊕i1
p4516=p4516⊕i1
For the 361st information bit i360, the addresses of the parity bit accumulators are given in the second row of the Tables 28 through 38. In a similar manner the addresses of the parity bit accumulators for the following 359 information bits im, m=361, 362, . . . , 719 are obtained using the formula
where x denotes the address of the parity bit accumulator corresponding to the information bit i360, i.e., the entries in the second row of the Tables 28 through 38. In a similar manner, for every group of 360 new information bits, a new row from Tables 28 through 38 are used to find the addresses of the parity bit accumulators.
After all of the information bits are exhausted, the following operations are preformed sequentially where M=360,
Final content of pi, i=0, 1, . . . , nldpc−kldpc−1 is equal to the parity bit pi.
In this exemplary embodiment, the outer BCH code can be a 12 bit error correcting, with nbch=kbch+192. It is noted that nbch=kldpc.
The generator polynomial of the t error correcting BCH encoder 209 is obtained by multiplying the first t polynomials in the following list of Table 39:
BCH encoding of information bits m=(mk
Continuing with this exemplary embodiment, the output of the LDPC encoder 203 can be bit interleaved using the interleaver 211. According to certain embodiments, the interleaving process can be performed for 8-PSK and/or 16-APSP modulations. Data can be serially written into the interleaver 211 column-wise (from the top to the bottom), and can be serially read out row-wise (from the left to the right except for the rate 3/5 8-PSK case where data is read out from the right to the left). The configuration of the interleaver 211 for each modulation format can be viewed in Table 40.
The above LDPC codes, in an exemplary embodiment, can be used to variety of digital video applications, such as MPEG (Motion Pictures Expert Group) packet transmission.
It is contemplated that the above transmitter 200 and receiver 300 can be deployed within a single wireless terminal, in which case a common antenna system can be shared. The wireless terminal can for example be configured to operate within a satellite communication, a cellular system, wireless local area network (WLAN), etc.
To appreciate the advantages offered by the present embodiments, it is instructive to examine how LDPC codes are generated, as discussed in
Returning to the receiver 300, the LDPC decoder 305 is considered a message passing decoder, whereby the decoder 305 aims to find the values of bit nodes. To accomplish this task, bit nodes and check nodes iteratively communicate with each other. The nature of this communication is described below.
From check nodes to bit nodes, each check node provides to an adjacent bit node an estimate (“opinion”) regarding the value of that bit node based on the information coming from other adjacent bit nodes. For instance, in the above example if the sum of n4, n5 and n8 “looks like” 0 to m1, then m1 would indicate to n1 that the value of n1 is believed to be 0 (since n1+n4+n5+n8=0); otherwise m1 indicate to n1 that the value of n1 is believed to be 1. Additionally, for soft decision decoding, a reliability measure is added.
From bit nodes to check nodes, each bit node relays to an adjacent check node an estimate about its own value based on the feedback coming from its other adjacent check nodes. In the above example n1 has only two adjacent check nodes m1 and m3. If the feedback coming from m3 to n1 indicates that the value of n1 is probably 0, then n1 would notify m1 that an estimate of n1 's own value is 0. For the case in which the bit node has more than two adjacent check nodes, the bit node performs a majority vote (soft decision) on the feedback coming from its other adjacent check nodes before reporting that decision to the check node it communicates. The above process is repeated until all bit nodes are considered to be correct (i.e., all parity check equations are satisfied) or until a predetermined maximum number of iterations is reached, whereby a decoding failure is declared.
H
(n−k)xn
=[A
(n−k)xk
B
(n−k)x(n−k)],
where B is lower triangular.
Any information block i=(i0, i1, . . . , ik−1) can be encoded to a codeword c=(i0, i1, ik−1, p0, p1, . . . , pn−k−1) using HcT=0, and recursively solving for parity bits; for example,
a
00
i
0
+a
01
i
1
+ . . . +a
0,k−1
i
k−1
+P
0
Solve p0
a
10
i
0
+a
11
i
1
+ . . . +a
1,k−1
i
k−1
b
10
p
0
+p
1=0Solve p1
Further, according to certain embodiments, the QPSK, 8-PSK, and 16-APSK modulation schemes of
According to certain embodiments, different decoding processes can be used. In one exemplary embodiment, a decoding process in accordance with Gray mapping can be used. In this example, outgoing messages from bit nodes are initialized, check nodes are updated, bit nodes are updated, a posteriori probability information is outputted, a determination is made whether all parity check equations are satisfied, and a hard decision is outputted. Otherwise, check node update, bit node update, and outputting a posteriori probability information is repeated. Alternatively, a decoding process in accordance with non-Gray mapping can used such that probability information can be exchanged back and forth (iteratively) between the decoder 305 and bit metric generator 307 (of
Referring to
Under the forward-backward approach, to compute these outgoing messages, forward variables, f1, f2, . . . , fdc, are defined as follows:
At step 801, these forward variables are computed (which can be performed in almost dc clock cycle), and stored, per step 803.
Similarly, backward variables, b1, b2, . . . , bdc, are defined by the following:
At step 805, these backward variables are then computed (which can be performed in almost dc clock cycles). Thereafter, the outgoing messages are computed, as in step 807, based on the stored forward variables and the computed backward variables. The outgoing messages are computed as follows (which can be performed in almost dc clock cycle):
wk→1=b2
w
k→i
=g(fi−1,bi+1) i=2, 3, . . . ,dc−1
wk→dc=fdc−1
Under this approach, only the forward variables, f2, f3, . . . , fdc, are required to be stored. As the backward variables bi are computed, the outgoing messages, wk→i, are simultaneously computed, thereby negating the need for storage of the backward variables.
The computation load can be further enhanced by using an enhanced layered belief decoding (LBD) approach, as discussed next. In one exemplary embodiment, use of the LBD approach can enhance the decoding technique by improving convergence of about half the standard belief decoding (SBD) iterations, therefore, increasing the speed of decoding by almost 2. Further, improvements discussed below with respect to
Referring to
|vn
Further, a second minimum of the absolute values of the incoming messages are determined, per step 813, as follows:
x=second min {|vn
At step 815, a sign of the incoming messages is determined as follows:
s=sign(vn
Thereafter, the outgoing messages are computed, as in step 817, based on the determined variables, according to one exemplary embodiment. The outgoing messages are computed as follows:
wk→n
w
k→n
≅(|vn
Alternatively, the outgoing messages can be computed, as per step 817, based on the determined variables, according to another exemplary embodiment, as follows:
w
k→n
≅(x−1)×s×sign(vn
w
k→n
≅(|vn
Based on the exemplary process 810 of the embodiment of
Incoming message vn
v
n
→k
=a
i
−w
k→n
i=1,2, . . . ,dc
Further, the a posteriori probabilities can be updated, after outgoing messages are computed, as follows:
a
i
=v
n
→k+w
k→n
i=1,2, . . . ,dc
These two processes usually can take dc clock cycles each. However, in an exemplary embodiment, two adders can be used for these processes such that each can be performed in dc/2 clock cycles and therefore, decrease the total iteration time from 5d, to 2dc clock cycles. Hence, use of the enhanced LBD improves decoding speed of the standard LBD by a factor of 2.5.
Two general approaches exist to realize the interconnections between check nodes and bit nodes: (1) a fully parallel approach, and (2) a partially parallel approach. In fully parallel architecture, all of the nodes and their interconnections are physically implemented. The advantage of this architecture is speed.
The fully parallel architecture, however, may involve greater complexity in realizing all of the nodes and their connections. Therefore with fully parallel architecture, a smaller block size may be required to reduce the complexity. In that case, for the same clock frequency, a proportional reduction in throughput and some degradation in FER versus Es/No performance may result.
The second approach to implementing LDPC codes is to physically realize only a subset of the total number of the nodes and use only these limited number of “physical” nodes to process all of the “functional” nodes of the code. Even though the LDPC decoder operations can be made extremely simple and can be performed in parallel, the further challenge in the design is how the communication is established between “randomly” distributed bit nodes and check nodes. The decoder 305 (of
In other words, the approach, according to certain embodiments, facilitates memory access during check node and bit node processing. The values of the edges in the bipartite graph can be stored in a storage medium, such as random access memory (RAM). It is noted that for a truly random LDPC code during check node and bit node processing, the values of the edges would need to be accessed one by one in a random fashion. However, such a conventional access scheme would be too slow for a high data rate application. The RAM of
As illustrated in
In an exemplary embodiment, a group of M bit nodes and M check nodes is processed at a time. According to another exemplary embodiment, for M check node processing, q=dc−2 consecutive rows from top edge RAM 1201 and 2 consecutive rows from bottom edge RAM 1203 is accessed, which the value of dc can depend on the code rate. In this exemplary embodiment, dc=7, 11, and 20 for rates 1/2, 2/3, and 4/5, respectively.
For bit node processing, if the group of M bit nodes are parity bit nodes, their edges are located in 2 consecutive rows of the bottom edge RAM 1303. If the bit nodes are information bit nodes with degree dv>2, their edges are located in some dv rows of the top edge RAM 1201. The address of these dv rows can be stored in non-volatile memory, such as Read-Only Memory (ROM). The edges in one of the rows correspond to the first edges of M bit nodes, the edges in another row correspond to the second edges of M bit nodes, etc. Moreover for each row, the column index of the edge that belongs to the first bit node in the group of M can also be stored in ROM. The edges that correspond to the second, third, etc. bit nodes follow the starting column index in a “wrapped around” fashion. For example, if the jth edge in the row belongs to the first bit node, then the (j+1)st edge belongs to the second bit node, (j+2)nd edge belongs to the third bit node, . . . , and (j−1)st edge belongs to the Mth bit node.
With the organization shown in
In this exemplary embodiment, check nodes are partitioned into groups of M=3 {0,2,4} and {1,3,5}. There is no need to store the groups since the computation engines do not care which nodes are processed. For the first group, dc−2=6 adjacent edges are e0 e1 e2 e3 e4 e5, e12 e13 e14 e15 e16 e17, e24 e25 e26 e27 e28 e29. All the above edges are stored in the first 6 rows of top edge RAM 1305. Therefore, it is only needed that those 6 rows be fetched. Moreover, in each row the jth element belongs to the jth check node in the group. The remaining adjacent edges are e36 e39 e40 e43 e44. These remaining edges are stored in the first two rows of bottom edge RAM 1307. Similar to top edge RAM 1305 case, in each row, the jth element belongs to the jth check node in the group. Similar structure exists for the other group of check nodes. Therefore, as far as check node processing is concerned, there is no need for any ROM storage. Simply, consecutive rows of edge RAM's are accessed and within each row, it is known exactly where to find the edge for a particular check node in the group.
Also, bit nodes of degree dv>2 are partitioned into groups of M=3: {0,1,2}, {3,4,5}, {6,7,8}, {9,10,11}. Bit nodes in the first group have degree 3 and their adjacent edges are e0 e14 e19, e12 e26 e31, e2 e7 e24. All these edges can appear in top edge RAM 1305, namely rows indexed 0, 2 and 7. These row indices can be stored in ROM. Row 0 carries e0 e12 e24; row 2 carries e2 e14 e26; and row 7 carries e7 e19 e31. Similar to the check node case, each row carries one adjacent edge per bit node. On the other hand they might not be in perfect order as in the case of check nodes. In other words the jth element in each row does not necessarily belong to the jth node. However, as it was explained before, if the jth element belongs to the first bit node, then (j+1)st element belongs to the second bit, (j+2)nd element belongs to the third bit node. etc., and finally (j−1)st element belongs to the Mth bit node (in a barrel shifted manner). Therefore, all needed to be stored in ROM is the index j of each row that belongs to the first bit node in the group. In this example, it can be noticed that for row indexed 0, e0 belongs to the first bit node (hence j=0); for row indexed 2, e14 belongs to the first bit node (hence j=1) and for row indexed 7, e19 belongs to the first bit node (hence j=1). Therefore, the first part of our ROM table reads 0/0 2/1 7/1. Going through the other groups in a similar manner, following ROM table can be obtained
Following with this example, bit nodes of degree dv=2 can be divided into the following groups: {12,14,16}, {13,15,17}. It can be verified that adjacent edges of each group occur in two consecutive rows of bottom edge RAM 1307 and moreover, in each row the jth element belongs to the jth bit node in the group (except for the last group where jth element of the bottom row and (j+1)mod M element of the top row belong to the same bit node). Therefore, for this case too, there is no need for ROM storage.
Next, a decoding scheme for LDPC codes with code rates 1/4, 1/3, 2/5, 1/2, 3/5, 2/3, 3/4, 4/5, 5/6, 8/9, and 9/10 with coded block size 64800, according to an exemplary embodiment, is described in accordance with
As mentioned before regarding LDPC codes with code rates 1/4, 1/3, 2/5, 1/2, 3/5, 2/3, 3/4, 4/5, 5/6, 8/9, and 9/10 with coded block size 64800, for QPSK modulation scheme, the I/Q demodulation output can be multiplied by a constant (quantized to the nearest integer between −15 and +15), and can be further fed to the LDCP decoder.
In the exemplary embodiment in accordance with LDPC codes with code rates 1/4, 1/3, 2/5, 1/2, 3/5, 2/3, 3/4, 4/5, 5/6, 8/9, and 9/10 with coded block size 64800, check nodes can have degree d More specifically all the check nodes in the code bipartite graph (as illustrated in, for instance,
Therefore (N−K)d−1 edges exist. These edges are denoted by e1, e2, e3, . . . , ed−1, ed+1, . . . , e(N−K)d. Here, denote the edges adjacent to the first check node, ed+1, ed+2, ed+3, . . . , e2d denote the edges adjacent to the second check node, etc. Furthermore, b0, b1, b2, . . . , bN−1 denote a posteriori probability values for all the LDPC coded bits (N=64800 for long codes and N=16200 for short codes).
According to certain embodiments, the edge values and a posteriori probability values, as discussed above, can be arranged in the edge storage medium (such as random access memory (RAM)) 1401 of
According to an exemplary embodiment, the first d rows of edge RAM 1401 can carry the edges adjacent to the first group of M check nodes. Further, the next d rows of the edge RAM 1401 can carry the edges adjacent to the second group of M check nodes, etc. Moreover in each row of the edge RAM 1401, the element in the ith column carries the edge adjacent to the ith check node in a group of M check nodes. It is noted that there is no edge value for the entry x in the edge RAM 1401. In one example, 9 bits can be allocated for the edge RAM 1401. Also, the edge RAM 1401 can be initialized with all zero entries. Moreover, a posteriori RAM 1403 includes a posteriori probability values and in one example, 10 bits are allocated for the entries a posteriori RAM 1403. The a posteriori RAM 1403 can, initially be filled with LDPC decoder soft inputs, which, for instance, can be 6 bits. As the iterations of the decoder, for instance, decoder 303, are run, new a posteriori probability values for each coded bit overwrites the pervious value in the a posteriori RAM 1403.
According to this exemplary embodiment, the LDPC decoder, such as decoder 303 of
The d inputs messages to the ith check node are then formed as follows:
v
k
=b
j
−e
id+k
k=1,2, . . . ,d .
The output message from the ith check node is also computed as follows:
z≡min {|v1|,|v2|, . . . |vd|}, let z=|vp|, for some p
x≡second min {|v1|,|v2|, . . . |vd|}
s≡sign(v1)×sign(v2)× . . . sign(vd)
Also, define t≡2 for code rate 1/4 and t≡1 for all the other code rates. The output message from the ith check node to d bit nodes is computed as follows:
w
p=(x−1)×s×sign(vp) if x≧1, otherwise wp=0
w
k=(z−t)×s×sign(vk)k=1,2, . . . ,p−1,p+1, . . . ,d if z≧1, otherwise wk=0
After all M check nodes are processed in parallel, the values wk, k=1, 2, . . . , Md overwrite ek in the edge RAM 1401. Further, the updated a posteriori values are computed by vk+wk and overwrite the old a posteriori values in the a posteriori RAM 1403. It is noted that, the new a posteriori values need to be circularly shifted back by the same amount before being written back to the memory. Also, it is noted that the first check node has one less degree (number of inputs) than the other check nodes.
The processes for encoding and decoding structured Low Density Parity Check (LDPC) codes are described herein may be implemented via software, hardware (e.g., general processor, Digital Signal Processing (DSP) chip, an Application Specific Integrated Circuit (ASIC), Field Programmable Gate Arrays (FPGAs), etc.), firmware or a combination thereof. Such exemplary hardware for performing the described functions is detailed below.
The computer system 1600 may be coupled via the bus 1601 to a display 1611, such as a cathode ray tube (CRT), liquid crystal display, active matrix display, or plasma display, for displaying information to a computer user. An input device 1613, such as a keyboard including alphanumeric and other keys, is coupled to the bus 1601 for communicating information and command selections to the processor 1603. Another type of user input device is a cursor control 1615, such as a mouse, a trackball, or cursor direction keys, for communicating direction information and command selections to the processor 1603 and for adjusting cursor movement on the display 1611.
According to an embodiment of the invention, the processes described herein are performed by the computer system 1600, in response to the processor 1603 executing an arrangement of instructions contained in main memory 1605. Such instructions can be read into main memory 1605 from another computer-readable medium, such as the storage device 1609. Execution of the arrangement of instructions contained in main memory 1605 causes the processor 1603 to perform the process steps described herein. One or more processors in a multi-processing arrangement may also be employed to execute the instructions contained in main memory 1605. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions to implement the embodiment of the invention. Thus, embodiments of the invention are not limited to any specific combination of hardware circuitry and software.
The computer system 1600 also includes a communication interface 1617 coupled to bus 1601. The communication interface 1617 provides a two-way data communication coupling to a network link 1619 connected to a local network 1621. For example, the communication interface 1617 may be a digital subscriber line (DSL) card or modem, an integrated services digital network (ISDN) card, a cable modem, a telephone modem, or any other communication interface to provide a data communication connection to a corresponding type of communication line. As another example, communication interface 1617 may be a local area network (LAN) card (e.g. for Ethernet™ or an Asynchronous Transfer Model (ATM) network) to provide a data communication connection to a compatible LAN. Wireless links can also be implemented. In any such implementation, communication interface 1617 sends and receives electrical, electromagnetic, or optical signals that carry digital data streams representing various types of information. Further, the communication interface 1617 can include peripheral interface devices, such as a Universal Serial Bus (USB) interface, a PCMCIA (Personal Computer Memory Card International Association) interface, etc. Although a single communication interface 1617 is depicted in
The network link 1619 typically provides data communication through one or more networks to other data devices. For example, the network link 1619 may provide a connection through local network 1621 to a host computer 1623, which has connectivity to a network 1625 (e.g. a wide area network (WAN) or the global packet data communication network now commonly referred to as the “Internet”) or to data equipment operated by a service provider. The local network 1621 and the network 1625 both use electrical, electromagnetic, or optical signals to convey information and instructions. The signals through the various networks and the signals on the network link 1619 and through the communication interface 1617, which communicate digital data with the computer system 1600, are exemplary forms of carrier waves bearing the information and instructions.
The computer system 1600 can send messages and receive data, including program code, through the network(s), the network link 1619, and the communication interface 1617. In the Internet example, a server (not shown) might transmit requested code belonging to an application program for implementing an embodiment of the invention through the network 1625, the local network 1621 and the communication interface 1617. The processor 1603 may execute the transmitted code while being received and/or store the code in the storage device 1609, or other non-volatile storage for later execution. In this manner, the computer system 1600 may obtain application code in the form of a carrier wave.
The term “computer-readable medium” as used herein refers to any medium that participates in providing instructions to the processor 1603 for execution. Such a medium may take many forms, including but not limited to computer-readable storage medium ((or non-transitory)—i.e., non-volatile media and volatile media), and transmission media. Non-volatile media include, for example, optical or magnetic disks, such as the storage device 1609. Volatile media include dynamic memory, such as main memory 1605. Transmission media include coaxial cables, copper wire and fiber optics, including the wires that comprise the bus 1601. Transmission media can also take the form of acoustic, optical, or electromagnetic waves, such as those generated during radio frequency (RF) and infrared (IR) data communications. Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium, a CD-ROM, CDRW, DVD, any other optical medium, punch cards, paper tape, optical mark sheets, any other physical medium with patterns of holes or other optically recognizable indicia, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, a carrier wave, or any other medium from which a computer can read.
Various forms of computer-readable media may be involved in providing instructions to a processor for execution. For example, the instructions for carrying out at least part of the embodiments of the invention may initially be borne on a magnetic disk of a remote computer. In such a scenario, the remote computer loads the instructions into main memory and sends the instructions over a telephone line using a modem. A modem of a local computer system receives the data on the telephone line and uses an infrared transmitter to convert the data to an infrared signal and transmit the infrared signal to a portable computing device, such as a personal digital assistant (PDA) or a laptop. An infrared detector on the portable computing device receives the information and instructions borne by the infrared signal and places the data on a bus. The bus conveys the data to main memory, from which a processor retrieves and executes the instructions. The instructions received by main memory can optionally be stored on storage device either before or after execution by processor.
While certain exemplary embodiments and implementations have been described herein, other embodiments and modifications will be apparent from this description. Accordingly, the invention is not limited to such embodiments, but rather to the broader scope of the presented claims and various obvious modifications and equivalent arrangements.