Low-density parity-check (LDPC) codes are a popular choice for stored data, such as data stored on solid state storage. As storage density increases, the number of errors included in the LDPC-encoded data likewise increases. To compensate for this, new and more powerful LDPC decoding techniques would be desirable. Furthermore, it would be desirable if these new LDPC decoding techniques could be easily implemented in existing storage systems, do not consume substantial amounts of resources (e.g., additional processing, memory, power, etc.), and/or do not introduce noticeable (additional) processing delays.
Various embodiments of the invention are disclosed in the following detailed description and the accompanying drawings.
The invention can be implemented in numerous ways, including as a process; an apparatus; a system; a composition of matter; a computer program product embodied on a computer readable storage medium; and/or a processor, such as a processor configured to execute instructions stored on and/or provided by a memory coupled to the processor. In this specification, these implementations, or any other form that the invention may take, may be referred to as techniques. In general, the order of the steps of disclosed processes may be altered within the scope of the invention. Unless stated otherwise, a component such as a processor or a memory described as being configured to perform a task may be implemented as a general component that is temporarily configured to perform the task at a given time or a specific component that is manufactured to perform the task. As used herein, the term ‘processor’ refers to one or more devices, circuits, and/or processing cores configured to process data, such as computer program instructions.
A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention encompasses numerous alternatives, modifications and equivalents. Numerous specific details are set forth in the following description in order to provide a thorough understanding of the invention. These details are provided for the purpose of example and the invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the invention is not unnecessarily obscured.
Various embodiments of a technique to perform LDPC decoding (e.g., by message passing or more specifically min-sum decoding) with de-saturation are described herein. For example, without de-saturation, LDPC systems that use fixed-point number representation can become trapped in some cases if the messages become saturated relative to the fixed-point number representation. In some embodiments, a saturation metric that represents a degree of saturation in a low-density parity-check (LDPC) decoding system that uses fixed-point number representation is determined. The saturation metric is compared against a saturation threshold and it is determined if the saturation metric exceeds the saturation threshold. If so, LDPC decoding data (e.g., one or more messages in the system) is more aggressively attenuated compared to when the saturation metric does not exceed the saturation threshold at the end of a decoding iteration. If not, the LDPC decoding data is less aggressively attenuated at the end of the decoding iteration. The attenuated message may then be passed from a first type of node to a second type of node in a Tanner Graph, assuming more decoding iterations are required to properly decode the LDPC data.
At 100 a saturation metric that represents a degree of saturation in a low-density parity-check (LDPC) decoding system that uses a fixed-point number representation is determined. In one example of step 100, the saturation metric is based on the number or percentage of messages (e.g., prior to any message attenuation) from check nodes to variable nodes that are saturated to the maximum fixed-precision magnitude. Alternatively, messages passed in the other direction (e.g., from variable nodes to check nodes) are used to determine the saturation metric at step 100. For example, lmin1(j) is the minimum variable node message to check node j and in some embodiments the saturation metric is based at least in part on the number of lmin1 messages (e.g., across all check nodes j in the system) that are saturated (e.g., to the maximum fixed precision magnitude) at the end of a decoding iteration.
In another example of step 100, the saturation metric is based on a number of saturated nodes in the system. For example, the saturation metric may be the number of saturated check nodes at the end of a decoding iteration. For example, check node j may be declared saturated if Lmin1(j) (i.e., the minimum attenuated variable node message to check node j where lmin1(j) is the message before attenuation and Lmin1(j) is the message after attenuation) reaches the maximum fixed-precision magnitude.
At 102, the saturation metric is compared against a saturation threshold. In one example, the saturation metric is a percentage and the saturation threshold is within a range of ¾ (75%) 15/16 (93.75%).
At 104, it is determined if the saturation metric exceeds the saturation threshold. If so, at the end of a decoding iteration, a message is more aggressively attenuated compared to when the saturation metric does not exceed the saturation threshold in order to produce an attenuated message at 106. If not, at the end of a decoding iteration, the message is less aggressively attenuated compared to when the saturation metric does exceed the saturation threshold in order to produce the attenuated message at 108.
In one example of step 106, the messages that are attenuated are the lmin1(j) and lmin2(j) messages, which are the minimum and second minimum variable node messages to check node j, respectively, before any attenuation or normalization. In this example of step 106, they are more aggressively attenuated using de-saturation attenuation factor αds and de-saturation attenuation rounding βds (e.g., Lmin1(j)=└αds·lmin1(j)+βds┘ and Lmin2(j)=└αds·lmin2(j)+βds┘).
In a corresponding example of step 108, the same messages (i.e., lmin1(j) and lmin2(j)) are less aggressively attenuated using attenuation factor α and attenuation rounding β (e.g., Lmin1(j)=└α·lmin1(j)+β┘ and Lmin2(j)=└α·lmin2(j)+β┘) wherein the desaturation attenuation parameters satisfy the following:
αds<α
└αds+βds┘≤└α+β┘
└αdsLmax+βds┘<└αLmax+β┘
where Lmax denotes the maximum message magnitude. Furthermore, the parameters α, β, αds, βds can be optimized through density evolution method.
By more aggressively attenuating the messages if the system is saturated (e.g., as measured or otherwise determined using the saturation threshold), the saturation of correct messages is facilitated and error correction over small trapping sets is facilitated, but only if/when needed (e.g., only when the saturation metric exceeds the saturation threshold). More generally, by de-saturating LDPC data if needed, the performance of the LDPC decoder is improved in cases where the data (e.g., the passed messages) would otherwise have saturated and the decoder would trend towards an uncorrectable state.
The attenuated message (e.g., output by step 106 or 108) may then pass from a first type of node (e.g., a variable node) to a second type of node (e.g., a check node) in a Tanner graph, assuming additional LDPC decoding is required. The second type of nodes in the Tanner graph may then send back updated messages to the first type of nodes and the process of
To provide more context for the decoding technique with de-saturation described herein, it may be helpful to discuss check nodes and variable nodes which perform LDPC decoding by passing messages between the nodes. The following figures describe some such examples.
Low-density parity-check (LDPC) codes are a type of error correction codes and are so named because they have a very low density of is in their parity-check matrix (H). This property makes it convenient to represent an LDPC code using a bipartite graph, called a Tanner Graph. There are two types of nodes in a Tanner graph: variable nodes (VNs) and check nodes (CNs). In the example shown here, vi (204a) is an example of a variable node and cj (200a) is an example of a check node. Each variable node (or check node) corresponds to a column (or row) of the parity-check matrix, H. As used herein, V={v1, . . . , vn} is the set of variable nodes and C={c1, . . . , cm} is the set of check nodes. Each row of H is indexed by C={1, 2, . . . , m} and each column of H is indexed by V={1, 2, . . . , n}. In the Tanner graph, variable node vi is connected to check node cj via an edge if Hj,i=1 and the set of edges on the Tanner graph is denoted by set E.
Quasi-cyclic low-density parity-check (QC-LDPC) codes are a special class of the LDPC codes with structured H matrix which can be generated by the expansion of an mb×nb base matrix. Each is element in the base matrix can be expanded by a circularly right-shifted b×b identity sub-matrix. QC-LDPC codes have advantages over other types of LDPC codes in terms of hardware implementations on both the encoding and decoding side. Encoding of a QC-LDPC code can be (more) efficiently implemented (e.g., in hardware) using simple shift registers. In a hardware implementation of a QC-LDPC decoder, the QC structure of the code simplifies the wire routing for message passing.
It may be helpful to describe message passing in more detail. The following figure shows an example of a variable node-to-check node message update, which is part of message passing (i.e., a type of LDPC decoding).
Message-passing (MP) is an efficient technique to achieve near-optimal decoding of LDPC codes. For notational conciseness, a variable node is referred to subsequently simply as i (instead of vi) and j (instead of cj) is used to denote a check node. As shown in this example, a variable node i (204b) receives an input message Lich (210) from the channel. For example, this message from the channel may be the log-likelihood ratio (LLR) of the corresponding channel output, defined as follows:
where ci∈{0,1} is the code bit and ri is the corresponding received symbol.
A conventional iterative message passing decoder alternates between two phases: a CN-to-VN phase (during which check nodes send messages to their adjacent variable nodes) and a VN-to-CN phase (during which variable nodes send messages to check nodes along their adjacent edges) which are depicted schematically in
Let Li→j and Lj→i represent the messages sent from variable node i to check node j and from check node j to variable node i, respectively. Let C(i) be the set of check nodes directly connected to variable node i and VU) be the set of variable nodes directly connected to check node j. Then, the message sent from variable node i to check node j in sum-product decoding is given by:
and the message from check node j to variable node i is computed as:
Let Pi be a posterior probability (APP) message of variable node i where:
In this example, a variable node receives the log-likelihood ratios of received information from the channel as an initial input message (i.e., Li→j=Lich) and the following equivalent check node update rule is employed:
where 0<α<1, β>0 is the attenuation factor and attenuation rounding, respectively, which can be either pre-fixed or dynamically adjusted. Herein we newly introduced the attenuation rounding parameter β, which satisfies the following:
1≤α+β<2
which prevents a minimum CN-to-VN message of 1 from being attenuated to zero (which would erase any information contained therein).
It is noted that channel LLR inputs may be conveniently scaled for min-sum decoding but preferably are precise for the original sum-product decoding. With that in mind, the following notations are used to simplify the above calculation. Let:
Si→j≙ sign(Li→j).
Let S(j) be the product sign of all variable nodes i to the check node j:
Let lmin1(j) and imin1(j) be the minimum variable node message to check node j and its associated index, respectively:
and lmin2(j) be the second minimum variable node message to check node j:
Furthermore, let Lmin1(j) and Lmin2(j) be the attenuated minimum and second minimum variable node message, respectively, to the check node j:
Lmin1(j)≙└α·lmin1(j)+β┘, Lmin2(j)≙└α·lmin2(j)+β┘ (11)
With the above notations, Equation (5) can be conveniently rewritten as:
Pseudocode 1 describes a hardware amenable min-sum decoding example along these lines. It is noted that Pseudocode 1 does not de-saturate (e.g., unlike
Initialization: Lmin1(j)=Lmin2(j)=0, ∀j∈C
Iteration:
Pseudocode 1: Example of Flooded Min-Sum Decoding without De-Saturation
When a QC-LDPC code with b×b circulants is in use, each circulant of b bits is updated independently and in parallel.
The paper “Propagation of LLR saturation and quantization error in LDPC min-sum iterative decoding” by KANISTRAS et al. (which does not describe the de-saturation approach described in
In “Quantized iterative message passing decoders with low error floor for LDPC codes” by ZHANG et al., a new non-uniform quantization method was proposed to extend the message quantization range by using an exponentially increased step size for large magnitudes while keeping a fixed step size for small magnitudes. However, the proposed exponentially increased step size design is difficult to implement in hardware. In contrast, the de-saturation decoding technique described in
In U.S. Pat. No. 9,755,666 by Yingquan Wu, CN-to-VN messages are halved (e.g., by the variable node in the middle of a decoding iteration) if a significant fraction of VN-to-CN messages are saturated. For example, this would correspond to having a new line between line 5 and line 6 in Pseudocode 1 (not shown) where Lj→i is halved (i.e., set Lj→i to Lj→i/2) if some saturation condition is flagged. However, halving the CN-to-VN messages at that stage (i.e., in the middle of a VN-to-CN message update) as opposed to at the end of a decoding layer/iteration results in longer critical (e.g., rate limiting) path and thus slower clock speed. For example, there may be a critical timing path that begins with reading or otherwise inputting some piece of data or variable (e.g., at line 4 in Pseudocode 1) and ending with computing new values for that piece of data or value (e.g., at lines 12 and/or 13 in Pseudocode 1) and the older technique of halving Lj→i (e.g., which would occur between lines 5 and 6 in Pseudocode 1 but which is not shown there) introduces additional delay into a critical timing path. Furthermore, halving is a crude way of attenuating information and it may not be necessary to apply that much attenuation even if saturation is detected in the system. By using two de-saturation attenuation parameters (e.g., αds as well as βds), sufficient attenuation can be achieved without losing as much information as halving does. In one example,
If iterative message-passing decoding is implemented in hardware, the decoding efficiency can be improved using a layered decoding approach. In layered decoding, check node messages are updated serially. That is, instead of sending all messages from variable nodes to check nodes, and then all messages from check nodes to variable nodes (i.e., flooding), the layered coding goes through the check nodes in sequential order such that, to each check node being updated, all messages are sent in and processed, and then sent out to neighboring variable nodes. Such scheduled serial updating of check nodes enables immediate propagation of the newly updated message, unlike the flooded scheme where the updated messages can propagate only in the next iteration.
As a result, layered decoding improves convergence speed by roughly twice compared to that of a flooded implementation. Moreover, it provides a good trade-off between speed and memory. This is achieved by iterating over dynamic CN-to-VN messages, denoted by Q≙[Q1, Q2, . . . , Qn]. Specifically, let variable node i∈V (j), then Qi over a layer j is defined as:
where the superscript (last) denotes the latest updated. It is worth noting that, in layered decoding, the VN-to-CN message updated at the last layer (all but the last are from the current iteration) is utilized to update the CN-to-VN Qi in the current layer, whereas in the flooded decoding updating a CN-to-VN message Lj→i utilizes the VN-to-CN messages each generated at the last iteration. The Q≙[Q1, Q2, . . . , Qn] memory is initialized with the channel messages Lch≙[L1ch, L2ch, . . . , Lnch] and no dedicated memory is needed to store Lch, whereas with flooded decoding, Lch is stored but not Q. Qi, i=1,2, . . . , n, is iteratively calculated as follows. Let j be the current layer and ji be the preceding layer associated with variable node i. A preceding layer is mathematically declared or otherwise defined as follows. Let j1<j2< . . . <jk be all check nodes directly connected to variable node i, then jl is the preceding layer of jl+1 for l=1, 2, . . . , k−1, and jk is the preceding layer of jl.
The APP (i.e., a posterior probability) message Pi at the layer j is calculated as:
Pi(j)=Qi(j)+Lj→inew (14)
where Lj
Qi(j)=Pi(j)−Lj→iold (15)
where Lj
Initialization: Lmin1(j)=Lmin2(j)=0, ∀j∈C; Qi=Lich, ∀i∈V; L=0
Iteration:
Pseudocode 2: Example of Layered Min-Sum Decoding without De-Saturation
When a QC-LDPC code with b×b circulants is in use, b quasi-cyclic rows of H are (naturally) treated as a layer. That is, a layer contains b check nodes, each being updated independently and in parallel. It is noted that convergence may occur within any layer for layered min-sum decoding (one example of which is shown in Pseudocode 2) whereas the convergence must occur at the end of an iteration for the flooded min-sum decoding (one example of which is shown in Pseudocode 1). Moreover, layered decoding enables or otherwise permits utilization of updated CN-to-VN messages within an iteration whereas it is not possible for flooded decoding. Consequently, layered decoding converges roughly twice as fast as flooded decoding.
As described above, in one alternate approach to
During simulations with Pseudocode 1 and 2 (or the like), it was observed that the range of messages passed between variable nodes and check nodes in the decoder has direct impact on the decoding performance in terms of both converge speed and error rate. When fixed-point magnitude was not enforced, correct messages typically grew faster than incorrect messages, with most errors due to small trapping sets correctable. However, given limited precision in practice (e.g., five bits of representation), after a certain number of iterations, messages tended to saturate to the maximum fixed-point magnitude. In such scenarios, correct messages are not able to outweigh incorrect messages, and the message in passing is gradually downgraded to bipolar messages.
From this observation, new and improved adaptive quantization methods have been developed (see, e.g.,
This enables an effective increase in the quantization range without (meaningfully) increasing complexity or memory. Pseudocode 3 shows an example of flooded decoding with de-saturation.
Initialization: Lmin1(j)=Lmin2(j)=0, ∀j∈C
Iteration:
Pseudocode 3: Example of De-Saturated Flooded Min-Sum Decoding
Pseudocode 3 shows one example of how the process of
The following figure shows an example system diagram which performs the process of Pseudocode 3.
The de-saturation signal controls a plurality of multiplexers, including the multiplexer shown here (302). If the de-saturation signal equals zero, then the multiplexer (302) selects the less aggressive attenuation parameters of α and β to pass on to the attenuator block (308) which uses the selected attenuation parameters to attenuate messages (e.g., lmin1(j) and) lmin2(j)). If the de-saturation signal instead equals one, then the multiplexer (302) selects the more aggressive attenuation parameters of αds and βds to output to the attenuator (308). In some embodiments, αds is within a range of
and
and βds is within a range of
and
The attenuated messages which are output by the attenuation block (308) are then passed from the variable node i (304) to the check node j (306) as VN-to-CN messages (e.g., Lmin1(j) and Lmin2(j)) assuming another iteration of decoding needs to be performed (e.g., because the syndrome is not all zeros which indicates that LDPC decoding has not yet successfully completed).
Returning briefly to
As described above, unlike some other techniques (e.g., a system which alternates between an exponentially increased step size and fixed step size), adding a saturation monitor (300) and a plurality of multiplexers (e.g., the multiplexer (302) shown here) to an existing (min-sum) LDPC decoding system is relatively easy to implement in hardware.
For completeness, Pseudocode 4 shows an example of layered min-sum decoding with de-saturation.
Initialization: Lmin1(j)=Lmin2(j)=0, ∀j∈C; Qi=Lich, ∀i∈V; L=0; Ids=0
Iteration:
Pseudocode 4: Example of De-Saturated Layered Min-Sum Decoding
Pseudocode 4 shows one example of how the process of
Although some of the examples described herein show flooded min-sum decoding and layered min-sum decoding, the techniques described herein may be extended to other variants of message-passing decoding of LDPC codes (e.g., shuffled decoding which has efficiency between that of flooded decoding and layered decoding).
In some embodiments, the decoder selects the degree of de-saturation applied (e.g., at step 106 in
and
at step 106 in
and
at step 106 because a larger degree of de-saturation and/or attenuation is called for given the state of the system.
In some embodiments, such a selection of de-saturation parameters (e.g., for use at step 106 in
Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, the invention is not limited to the details provided. There are many alternative ways of implementing the invention. The disclosed embodiments are illustrative and not restrictive.
Number | Name | Date | Kind |
---|---|---|---|
5572350 | Spanke | Nov 1996 | A |
7583945 | McCarthy | Sep 2009 | B2 |
7669106 | Farjadrad | Feb 2010 | B1 |
8433984 | Khandekar | Apr 2013 | B2 |
8457194 | Ali | Jun 2013 | B2 |
8739002 | Nakamura | May 2014 | B2 |
9246717 | Beidas | Jan 2016 | B2 |
9264073 | Malmirchegini | Feb 2016 | B2 |
9571168 | Moon | Feb 2017 | B2 |
9612903 | Tehrani | Apr 2017 | B2 |
9716602 | Beidas | Jul 2017 | B2 |
9755666 | Wu | Sep 2017 | B2 |
10164663 | Shin | Dec 2018 | B2 |
10236070 | Barndt | Mar 2019 | B2 |
Entry |
---|
Chen et al., Reduced-Complexity Decoding of LDPC Codes, IEEE Transactions on Communications, Aug. 2005, pp. 1288-1299, vol. 53, No. 8. |
Kanistras et al., Propagation of LLR Saturation and Quantization Error in LDPC Min-Sum Iterative Decoding, 2012 IEEE Workshop on Signal Processing Systems, pp. 276-281, 2012. |
Kim et al., A Reduced-Complexity Architecture for LDPC Layered Decoding Schemes, IEEE Transactions on Very Large Scale Integration (VLSI) Systems, Jun. 2011, pp. 1099-1103, vol. 19, No. 6. |
Mansour et al., High-Throughput LDPC Decoders, IEEE Transactions on Very Large Scale Integration Systems, Dec. 2003, pp. 976-996, vol. 11, No. 6. |
Richardson et al., The Capacity of Low-Density Parity-Check Codes Under Message-Passing Decoding, IEEE Transactions on Information Theory, Feb. 2001, pp. 599-618, vol. 47, No. 2. |
Zhang et al., Quantized Iterative Message Passing Decoders with Low Error Floor for LDPC Codes, IEEE Transactions on Communications, Jan. 2014, pp. 1-14, vol. 62, No. 1. |
Zhang et al., Shuffled Iterative Decoding, IEEE Transactions on Communications, Feb. 2005, pp. 209-213, vol. 53, No. 2. |