The present invention relates to error detection and correction of component-based codes and generalized product codes in communication systems.
In the transmission of information-bearing signals over communication channels, errors often arise in the data due to noise, or other interference, in the communication channels. These errors have been attempted to be corrected using various methods, such as forward error correction (FEC). FEC attempts to correct the errors by using an encoder to map the underlying data in the signals to code symbols. The corrupted signal is received by a decoder, which attempts to reconstruct the code symbols in order to recover the underlying data.
The number of errors that can be corrected by an FEC system is dependent on the structure of the code used. As such, various types of FEC codes have been proposed over time. One particular class of FEC codes, called component-based codes, comprise multiple ordered subsets of the code symbols (e.g. bits) that are themselves code-words of smaller component codes. These component-based codes include, but are not limited to, product codes, generalized product codes (GPCs), higher-dimensional product codes, half-product codes, and convolutional versions thereof which are commonly decoded by windowed decoding (e.g. staircase codes, braided codes, half-braided codes, or continuously-interleaved codes). Component-based codes can be decoded through an iterative process of decoding the individual component codes (e.g., using bounded-distance decoding (BDD)), and the process is generally effective at detecting and correcting noise-related errors.
However, in current systems, the decoder may occasionally miscorrect one or more code symbols. That is, the decoder outputs a valid sequence of a code symbol which is, however, not the transmitted code symbol. One way to mitigate this problem is to use a method of extrinsic message-passing. Extrinsic message-passing shows improvement over iterative BDD, at least in regards to avoiding miscorrections, but extrinsic message-passing can suffer from heightened data flow and storage requirements. Moreover, an idealized process of BDD, in which all miscorrections are avoided entirely, is more effective than the extrinsic method. Therefore, there is need for an improved process of iterative BDD that better detects and avoids instances of miscorrection.
The present disclosure relates, in part, to systems and methods for avoiding miscorrection during the decoding of component-based codes, such as generalized product codes (GPCs).
One aspect of the present disclosure provides a method of decoding forward error correction codes, the method includes decoding a number of component codes, wherein the number of component codes include code symbols, for which at least one code symbol is involved in multiple component codes. The method further includes analyzing the decoding of each of the number of component codes to generate an outcome. The analyzing of each of the number of component codes includes: estimating at least one possible error location; storing information related to the outcome; storing state information; and updating the state information based on the information related to the outcome and based on a previous version of the state information.
Another aspect of the present disclosure provides a method of windowed decoding, the method includes decoding a number of component codes within a specified window, and analyzing the decoding of each of the number of component codes to generate an outcome. The step of analyzing each of the number of component codes includes estimating at least one possible error location and storing information related to the at least one possible error location. The step of analyzing each of the number of component codes further includes storing state information, and updating the state information based on the information related to the outcome and based on a previous version of the state information and where the updating rules depend on the bit position of the component codes in the window.
In some embodiments of the above method, the step of decoding the number of component codes is performed by a component code decoder. The component code decoder is configured to output the information related to the at least one possible error location or else indicating a failure to decode. In another embodiment of the above method, the component code decoder is a soft decision decoder. In another embodiment of the above method, the component code decoder is a hard decision decoder. In still another embodiment of the above method, the component code decoder is a bounded-distance decoder.
In some embodiments, the state information for a component code includes a syndrome of a number of symbol estimates. In another embodiment, the state information for a component code includes a value indicating whether the component code is reliable. In other embodiments, the state information for a component code comprises an array of corrected error locations that have been corrected by the component code. In still other embodiments, updating the state information further includes using the information related to the outcome to change at least one symbol estimate and revert at least one previous change. In other embodiments, the updating of the state information further includes preventing a correction of at least one estimated error that coincides with a reliable code. In a still further embodiment, the updating of the state information further includes tracking a quantity of estimated miscorrections that coincide with a particular code, and if the quantity of estimated errors exceeds a threshold value, reverting the at least one previous change that coincides with the particular code, and marking the particular code as unreliable.
In another embodiment, the above decoding methods stop decoding when an established termination condition is satisfied. In other embodiments, the methods further include a post-processing step of correcting at least one remaining error. In further embodiments, the correcting of at least one remaining error includes: marking a set of component codes as unreliable based on the state information; locating at least one intersection point among the set of component codes marked as unreliable; changing the code symbols located at each of the at least one intersection points; and resuming the step of decoding the number of component codes. In yet another embodiment, the correcting the at least one remaining error includes marking a set of component codes as unreliable based on the state information, and using a specialized decoder to decode the set of component codes marked as unreliable.
Another aspect of the present disclosure discloses an apparatus for decoding forward error correction codes. The apparatus includes a memory and at least one processor coupled to the memory. The at least one processor is configured to decode a number of component codes, wherein the number of component codes include code symbols, for which at least one subset of code symbols is a code-word of a component code, and analyze the decoding of each of the number of component codes to generate an outcome. Analyzing each of the number of component codes includes: estimating at least one possible error location; storing information related to the at least one possible error location; storing state information; and updating the state information based on the information related to the outcome and based on a previous version of the state information.
Other aspects of the invention will become apparent by consideration of the detailed description and accompanying drawings.
One or more embodiments are described and illustrated in the following description and accompanying drawings. These embodiments are not limited to the specific details provided herein and may be modified in various ways. Furthermore, other embodiments may exist that are not described herein. Also, the functionality described herein as being performed by one component may be performed by multiple components in a distributed manner. Likewise, functionality performed by multiple components may be consolidated and performed by a single component. Similarly, a component described as performing particular functionality may also perform additional functionality not described herein. For example, a device or structure that is “configured” in a certain way is configured in at least that way, but may also be configured in ways that are not listed. Furthermore, some embodiments described herein may include one or more electronic processors configured to perform the described functionality by executing instructions stored in non-transitory, computer-readable medium. Similarly, embodiments described herein may be implemented as non-transitory, computer-readable medium storing instructions executable by one or more electronic processors to perform the described functionality. As used in the present application, “non-transitory computer-readable medium” comprises all computer-readable media but does not consist of a transitory, propagating signal. Accordingly, non-transitory computer-readable medium may include, for example, a hard disk, a CD-ROM, an optical storage device, a magnetic storage device, a ROM (Read Only Memory), a RAM (Random Access Memory), register memory, a processor cache, or any combination thereof.
In addition, the phraseology and terminology used herein is for the purpose of description and should not be regarded as limiting. For example, the use of “including,” “containing,” “comprising,” “having,” and variations thereof herein is meant to encompass the items listed thereafter and equivalents thereof as well as additional items. The terms “connected” and “coupled” are used broadly and encompass both direct and indirect connecting and coupling. Further, “connected” and “coupled” are not restricted to physical or mechanical connections or couplings and can include electrical connections or couplings, whether direct or indirect. In addition, electronic communications and notifications may be performed using wired connections, wireless connections, or a combination thereof and may be transmitted directly or through one or more intermediary devices over various types of networks, communication channels, and connections. Moreover, relational terms such as first and second, top and bottom, and the like may be used herein solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Articles “a” and “an” are used herein to refer to one or to more than one (at least one) of the grammatical object of the article. By way of example, “an element” means at least one element and can include more than one element. Unless otherwise defined, all technical terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs.
Embodiments are described herein with reference to flowchart illustrations and/or block diagrams and/or figures. The flowchart, block diagrams and other illustrations in the present disclosure illustrate the architecture, functionality, and operation of possible implementations of systems, methods, computer program products (non-transitory computer-readable medium storing instructions executable one electronic processors, such as a microprocessor, to perform a set of functions), and the like according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams, or accompanying figures herein may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block or figures may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration and/or figures and combinations of blocks in the block diagrams and/or flowchart illustration and/or figures can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.
The present disclosure provides, in part, an improved decoder for a class of forward error correction codes, such as GPCs, where a code-word of a GPC comprises a sequence of code symbols (e.g., bits) such that many ordered subsets of the code symbols also form code-words of similar component codes. Other types of forward error correction codes can include product codes, higher-dimensional product codes, etc. The systems and methods described herein avoid most miscorrections and backtracks miscorrections that are not avoided. The present disclosure may be a system, a method, and/or apparatus/medium (e.g., a computer program product, device, etc.). Examples of computer program products may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present disclosure.
One aspect of the present disclosure provides a process 100 of decoding forward error correction codes. As shown in
Examples of suitable component codes include, but are not limited to, block codes, such as Reed-Solomon, Golay, BCH, Multidimensional-parity, and Hamming Codes. Examples of suitable component decoders include, but are not limited to, syndrome, algebraic, Berlekamp-Massey, Viterbi, MAP, BCJR, iterative, and belief-propagation. Examples of suitable post-processing methods include, but are not limited to, bit-flip-and-iterate, algebraic-erasure decoding, algebraic-errors-and-erasures decoding, and MAP-erasures decoding.
Once the component codes have been decoded, an outcome is generated for each of the decoded component codes. At process block 104, outcomes are analyzed. In one embodiment, analyzing the decoded component code outcomes includes: (i) estimating at least one possible error location; (ii) storing information related to the at least one possible error location; and (iii) storing state information. The state information may include one or more of the following: separate bit reliability information for each bit of the decoded component code; a syndrome of the current symbol estimates for each component code-word; a value indicating whether the decoded component code is reliable; and/or the locations of the errors that have been previously corrected within the component code. In some embodiments, the locations of errors already corrected may be stored in an array.
At process block 106, the state information is updated. In one embodiment, the state information is updated based on the outcome generated at process block 104 and a previous version of the state information. The state information may be updated several ways. As shown in
Once the state information has been updated at process block 106, the process 100 determines if a termination condition has been satisfied at process block 118. If the termination conditions have not been met, the process 100 returns to process block 102 to continue decoding the component codes. If the termination conditions have been met, the process stops decoding at process block 120. The threshold is subjective and is based on the goals of the error correction process and therefore can be determined by one skilled in the art. For example, a suitable threshold may be that no one component code includes more than a specified number of errors (e.g., one, four, etc.); include an iteration limit, etc.
It is important to point out that the process described in
After the component codes have been decoded, post processing may be initiated at process block 122. Post-processing methods may be performed after the decoding methods provided herein are completed to help improve performance. Numerous post-processing methods have been described and are well-known to those skilled in the art, all of which are within the scope of the present disclosure.
Turning now to
Windowed Decoding
The systems and methods provided herein can further be applied to windowed decoding as commonly employed to decode, e.g. staircase codes, braided codes, half-braided codes, or continuously-interleaved codes. Similar to classical product codes, staircase codes are built from short component codes and decoded by iteratively applying bounded-distance decoding (BDD) to the component codes. As used herein, the term “anchor code” refers to those component codes that are determined to include all correct symbols during (and after) the decoding process. Turning to
Turning now to
In one embodiment, the processing circuit 402 is communicably connected to the communication interface 404. The processing circuit 402 includes one or more electronic processors 406 and a memory 408. The electronic processor 406 may be implemented as a programmed microprocessor, an application specific integrated circuit (ASIC), one or more field programmable gate arrays (FPGA), a group of processing components, or other suitable electronic processing components.
The memory 408 (e.g. memory, memory unit, storage device, etc.) may include one or more devices (e.g., RAM, ROM, Flash memory, hard disk storage, etc.) for storing data and/or computer code for completing or facilitating the various processes, layers and modules described herein. Memory 408 can be or include volatile memory or non-volatile memory. Memory 408 can include database components, object code components, script components, or any other type of information structure for supporting the various activities and information structure described in the present application. According to one example, the memory 408 is communicably coupled to the electronic processor 406 via the processing circuit 402, and can include computer code for executing (e.g., by the processing circuit 402 and/or the electronic processor 406), one or more processes described herein. For example, the memory 408 and the at least one electronic processor 406 can be configured to: (i) decoding a number of component codes, the number of component codes including code symbols, for which at least one subset of code symbols is a code-word of a component code; and (ii) analyzing the decoding of each of the component codes to generate an outcome. Analyzing each of the component codes can include: (i) estimating at least one possible error location; (ii) storing information related to the at least one possible error location; (iii) storing state information; and (iv) updating the state information based on the information related to the outcome and based on a previous version of the state information.
Yet another aspect of the present disclosure provides for a non-transitory processor-readable medium for decoding forward error correction codes, the non-transitory processor-readable medium configured to cause at least one processor to perform the steps comprising of, consisting of, or consisting essentially of: (i) decoding a plurality of component codes, the plurality of component codes comprising code symbols, for which at least one subset of code symbols is a code-word of a component code; and (ii) analyzing the decoding of each of the plurality of component codes to generate an outcome, the step of analyzing each of the plurality of component codes comprising: (i) estimating at least one possible error location; (ii) storing information related to the at least one possible error location; (iii) storing state information; and (iv) updating the state information based on the information related to the outcome and based on a previous version of the state information.
The following examples are provided by way of illustration and not by way of limitation. Although the invention has been described in detail with reference to certain preferred embodiments, variations and modifications exist within the scope and spirit of one or more independent aspects of the invention as described. Various features and advantages of the invention are set forth in the following claims.
Product Codes and Iterative Decoding
Simple Product Codes (“PC”)
Let H∈2(n−k)×n be the parity-check matrix of a binary linear (n, k, dmin) code , where n, k, and dmin are the code length, dimension, and minimum distance, respectively. A simple PC based on is defined as:
{X∈2n×n|HX=0,XHT=0}. Equaiton 1
Simple PC's are used for illustration purposes in one embodiment of the invention.
The code-words X can be represented as two-dimensional arrays. The two conditions in (1) enforce that the rows and columns in the array are valid code-words in . In the following, we use a pair (i, j) to identify a particular component code-word. The first parameter i∈{1, 2} refers to the code-word type which can be either a row (i=1) or a column (i=2). The second parameter j∈[n] enumerates the code-words of a given type.
Example 1—The code array for a PC where the component code has length n=15 is shown in
For PCs, the individual bits can be identified by their two coordinates within the PC array. However, this way of specifying bits does not generalize well to other code classes because the associated code array may have a different shape or there may not exist an array representation at all. Therefore, in order to keep the notation general we use the convention that a particular bit is specified by two component code-words (i,j) and (k, l), i.e., four parameters (i, j, k, l) in total.
Example 2—The two highlighted component code-words in
BCH Component Codes
In one embodiment, we use binary t-error-correcting BCH codes as component codes, as well as their extended versions. An extended BCH code is obtained through an additional parity bit, formed by adding (modulo 2) all coded bits c1, c2,. . . , c2 v−1 of the BCH code, where v is the Galois field extension degree. The overall component code has length n=2v−1+e, where e∈{0, 1} indicates if the code extension is used. The guaranteed code dimension is k=2v−1−vt. For e=0, the guaranteed minimum distance is dmin=2t+1. This is increased to dmin=2t+2 for e=1. We use a triple (v, t, e) to denote all BCH code parameters.
As an alternative to extending the code, one may also employ certain subcodes of the original BCH code. For example, the singly-extended BCH code behaves similarly to the even-weight subcode of a BCH code.
Bounded-Distance Decoding and Miscorrections
In one embodiment, we use bounded-distance decoding (BDD) to decode the BCH component codes. However, the description may be modified to account for other decoding algorithms as well.
Consider the transmission of a (component) code-word c∈ over the binary symmetric channel (BSC) with crossover probability p. The error vector introduced by the channel is denoted by n, i.e., the components of n are i.i.d. Bernoulli(p) random variables. Applying BDD to the received vector r=c+n results in
In practice, BDD is implemented by first computing the syndrome sT=H nT∈2n−k. Each of the 2n−k possible syndromes is then associated with either an estimated error vector n̂, where H(n̂)≤t, or a decoding failure. In the first case, the decoded output is computed as r+n̂.
The second case in (2) corresponds to an undetected component code error or miscorrection.
Example 3—Consider the component code-word (1, 4) in
Iterative Bounded-Distance Decoding
We now consider the transmission of a code-word X∈P() over the BSC with crossover probability p. The conventional iterative decoding procedure consists of applying BDD first to all row component code-words and then to all column component code-words. This may be repeated times or until a valid code-word in P() is found. Pseudocode for this iterative decoding procedure is given in Algorithm 1. The implementation of an exit-criterion is omitted for readability purposes.
Bit Error Rate Performance
In order to analyze the bit error rate (BER) of PCs under iterative BDD, the prevailing approach in the literature is to assume that no miscorrections occur in the BDD of the component codes, see, e.g., [4], [8]-[11]. To that end, we define BDD′(r) as shown in Equation 3, below.
Accordingly, BDD′(r) can be seen as an idealized version of BDD where a genie prevents miscorrections.
The asymptotic performance of genie-aided decoding can be computed using density evolution (DE) [4], [9], [10]. More-over, the error floor can be estimated by enumerating small-sized stopping sets, also known as stall patterns. Formally, a stopping set is a subset of bit positions such that every component code-word with at least one bit in the set must contain at least t+1 bits in the set. A minimal-size stopping set involves t+1 row code-words and t+1 column code-words and has size smin=(t+1)2. For example, the black crosses shown in
Thus, for sufficiently small p, where
is the total number of possible minimal-size stopping sets, also referred to as the stopping set's multiplicity.
Example 4—Consider a BCH code with parameters (7, 2, 1). The resulting PC P() has length n2=1282=16384 and code rate =k2/n2≈0.78. For =10 decoding iterations, the outcome of DE and the error floor analysis via (4) for this PC are shown in
In the next section, we give a detailed description of another example embodiment. The resulting BER performance for the code parameters considered in Example 4 is shown in
Iterative BDD can also be interpreted as a message-passing algorithm with binary “hard-decision” messages. The corresponding message-passing rule is intrinsic, in the sense that the outgoing message along some edge depends on the incoming message along the same edge. In some examples, an extrinsic message-passing algorithm based on BDD is proposed. The BER for this algorithm when applied to the PC in Example 4 is shown in
Proposed Anchor-Based Decoding
In the previous section, we have seen that there exists a significant performance gap between iterative BDD and idealized iterative BDD where a genie prevents miscorrections. The key observation we exploit to improve performance and close this gap is that miscorrections lead to inconsistencies. In particular, two component codes that protect the same bit may disagree on its value. In this section, we show how these inconsistencies can be used to (a) reliably prevent miscorrections and (b) identify miscorrected code-words in order to revert their decoding decisions.
The proposed method relies on so-called anchor code-words, which have presumably been decoded without miscorrections. Roughly speaking, the proposed method attempts to ensure that bit flips do not lead to inconsistencies with anchor code-words. Consequently, decoding decisions from code-words that are in conflict with anchors are not applied. On the other hand, a small number of anchor code-words may actually be miscorrected. Therefore the method allow for the decoding decisions of anchors to be overturned if too many other component code-words are in conflict with a particular anchor. In order to make this more precise, we start by introducing some additional concepts and definitions in this subsection.
Now, an embodiment for a simple PC is described. First, consider the BDD of a single component code-word (i, j). We explicitly regard this component decoding as a two-step process as follows. In the first step, the actual decoding is performed and the outcome is either an estimated error vector n̂ or a decoding failure. In the second step, error-correction is performed by flipping the bits corresponding to the error locations. The algorithm relies on separating these two steps in order to perform certain consistency checks (described below) before applying the error-correction step. It is also more convenient to assume that the estimated error vector n̂ is given in terms of a set of error locations. For component code-word (i, j), this set is denoted by εi,j, where |εi,j|≤t. The set comprises identifiers of the component code-words that are affected by the bit flips implied by n̂.
Example 5—Consider again the scenario described in Example 3, where the component code-word (1, 4) shown in
Furthermore, we use a set i,j for each component code-word in order to keep track of conflicts that may arise between code-words due to miscorrections. Lastly, each component code-word has an associated status to signify its current state in the iterative decoding. The status values range from 0 to 3 with the following meaning:
The precise use of the status and the transition rules between different status values are described in the following.
Main Algorithm Routine
The algorithm is initialized by performing the decoding step for all component code-words, i.e., all component code-words (i, j) for i∈{1, 2} and j∈[n] are decoded without applying any bit flips. Depending on the decoding outcome, the status of each component codeword is set to either 1 (if the decoding was successful) or 2 (if the decoding failed).
indicates data missing or illegible when filed
We then iterate times over the component code-words in the same fashion as in Algorithm 1, but replacing line 4 with lines 1-15 from Algorithm 2. Algorithm 2 represents the main routine of the proposed anchor-based decoding. It can be divided into four steps which are described in the following.
Step 1 (Line 1): If the decoding was successful for the component code-word (i, j), i.e., its status is 1, we proceed to the next step, otherwise, we skip to the next code-word.
Step 2 (Lines 2-9): For each found error location (k, l)∈εi,j, a consistency check is performed. That is, one checks if the implied component code-word (k, l) corresponds to an anchor. If so, note that |k,l| is the number of conflicts that this anchor is already involved in. This number is then compared against a fixed threshold δ. If |k,l|≥δ, the anchor (k, l) is deemed unreliable and it is marked for backtracking by adding it to the backtracking set . On the other hand, if |k,l|<δ, the code-word (i, j) is frozen by changing its status to 3. Moreover, the conflict between the (now frozen) code-word and the anchor is stored by modifying the respective sets j,k and k,l. Frozen code-words are always skipped (in the loop of Algorithm 1) for the rest of the decoding unless either the conflicting anchor is backtracked or any bits in the frozen code-word change.
Step 3 (Lines 10-12): If the component code-word (i, j) still has status 1, the bit flips are consistent with all reliable anchors, i.e., anchors that are involved in fewer than δ other conflicts. If that is the case, the algorithm proceeds by backtracking all anchor code-words in the set (if there are any). Generally, backtracking involves reversing all previously applied bit flips for a corresponding anchor. Moreover, the backtracked code-word loses its anchor status. The backtracking routine is implemented in Algorithm 3 and described in more detail below.
Step 4 (Lines 13-15): Lastly, the error-correction step for code-word (i, j) is applied, i.e., the bits (i, j, k, l) corresponding to all error locations (k, l)∈εi,j are flipped. The error-correction step is implemented in Algorithm 4 and described below. Finally, the code-word (i, j) becomes an anchor by changing its status to 0.
indicates data missing or illegible when filed
In the following, the above steps are illustrated with the help of two examples. For both examples, we assume that a component code with error-correcting capability t=2 is employed. Moreover, the conflict threshold is set to δ=1.
Example 6. Consider the scenario depicted in
Example 7. Consider the scenario depicted in
Backtracking
In Example 7, we have encountered a scenario that leads to the backtracking of a miscorrected anchor. The actual backtracking routine is implemented in Algorithm 3 and described in the following. First, all conflicts caused by the anchor are removed by modifying the respective conflict sets. All code-words (k, l)∈j for an anchor (i, j) necessarily have status 3, i.e., they are frozen. After removing conflicts, such code-words may be conflict-free, in which case their status is changed to 1. Next, all previously applied bit flips are reversed. In order to perform this operation, it is necessary to store the set εi,j for each anchor. Finally, the code-word loses its anchor status. The code-word is still decodable with the same error set εi,j because we have just flipped the bits. Thus, in principle the new code-word status can be chosen to be either 1 or 3. However, backtracked anchors are likely to have miscorrected. We therefore freeze the code-word by setting its status to 3 after the backtracking.
indicates data missing or illegible when filed
Error-Correction
The error-correction step is implemented in Algorithm 4 and described in the following. The algorithm input is a parameter tuple (i, j, k, l) where (i, j) is the code-word that initiated the bit flip and (k, l) is the corresponding code-word affected by it. Note that Algorithm 4 can be reached from both the main decoding routine (Algorithm 2, lines 13-14) and as part of the backtracking process (Algorithm 3, lines 6-7). If the routine is reached via backtracking, it is possible that the affected code-word (k, l) is now an anchor. In this case, we use the convention to trust the anchor's decision about the bit (i, j, k, l) and not apply any changes. In all other cases, apart from actually flipping the bit (i, j, k, l) (line 2), error-correction also triggers a re-decoding of the affected component code-word (k, l) (line 3) and a possible status change depending on the decoding outcome (lines 4-15). If the bit flip affects a frozen code-word, the code-word is unfrozen and we remove all conflicts that the code-word is involved in.
Post-Processing
In case the anchor-based decoding terminates unsuccessfully, one may use some form of post-processing (PP) in order to improve performance. For the conventional iterative BDD, several different PP methods have been studied in the literature [13]-[18]. In particular, let i denote the set of failed component code-words of type i after an unsuccessful decoding attempt, i.e., 1 and 2 are, respectively, row and column code-words that have nonzero syndrome. The intersection of these code-words defines a set of bit positions, as shown in Equation 5, below.
={(i,j, k,l)|(i,j)∈1 and (k,l)∈2} Equation 5
One popular PP method is bit-flip-and-iterate, where the bits in the intersection (5) are first flipped and then iterative decoding is resumed for one or more iterations. This method has been applied to PCs [13], [16], half-product codes [17], and staircase codes [18]. Another method is algebraic-erasure PP, where the bits in the intersection (5) are treated as erasures. An algebraic erasure decoder is then used to recover the bits. Assuming that there are no miscorrected code-words, algebraic-erasure PP succeeds as long as either |1|<dmin or |2|<dmin holds. This type of PP has been applied to braided codes in [14] and to half-product codes in [15].
In principle, the above PP methods can also be combined with the anchor-based decoding without change. However, it is possible to improve the effectiveness of PP by exploiting some additional information that is available for the anchor-based decoding. In particular, recall that it is necessary to keep track of the error locations of anchors in case they are backtracked. If the anchor-based decoding fails, these error locations can be used for the purpose of PP as follows. Assume that we have determined the sets 1 and 2. Then, we can check for anchors which satisfy the condition:
εi,j∈and |εi,j|=t Equation 6
Assuming in Equation that =1∪2, i.e., anchors whose error locations overlap entirely with the set of failed component code-words. For the algebraic-erasure decoding PP, it is beneficial to include such anchors into the respective sets 1 and 2 (even though they have zero syndrome), as long as |1|<dmin and |2|<dmin remain satisfied.
Simulation Results and Discussion
In this section, we present and discuss simulation results for the embodiment described in the previous section assuming different BCH code parameters. For the conflict threshold, we tested different values δ∈{0, 1, 2, 3} for a wide variety of component codes and BSC crossover probabilities. In all cases, δ=1 was found to give the best performance. Hence, δ=1 is assumed in the following.
Results Without Post-Processing
Recall that for Example 4 in Sec. III-E, a BCH component code with parameters (7, 2, 1) and =10 decoding iterations are assumed. For these parameters, the BER of the proposed method is shown in
Next, we provide a direct comparison with the work in [19], where the authors propose a hardware architecture for a PC based on a BCH code with parameters (8, 2, 1). The BCH code is further shortened by 61 bits, leading to an effective length of n=195 and dimension k=178. The shortening gives a desired code rate of R=k2/n2=1782/1952≈0.833. The number of decoding iterations is set to =4. For these parameters, BER results are shown in
Results with Post-Processing
The above examples further consider the application of bit-flip-and-iterate PP in order to reduce the error floor. The simulation data was provided and the results are reproduced for convenience in
It is shown in
We remark that without the modification to the error-correction example above, the PP performance for the proposed decoding would be slightly decreased. Indeed, we found that a dominant error event for the considered embodiment is such that 3 row (or column) code-words miscorrect towards the same estimated error pattern if of weight 6. This scenario is illustrated in
Hard-decision forward error correction (HD-FEC) can offer dramatically reduced complexity compared to soft-decision FEC, at the price of some performance loss. HD-FEC is used, for example, in regional/metro optical transport networks (OTNs) and has also been considered for other cost-sensitive applications such as optical data center interconnects. Our focus in on staircase codes, which provide excellent performance and have received considerable attention in the literature.
Similar to classical product codes, staircase codes are built from short component codes and decoded by iteratively applying bounded-distance decoding (BDD) to the component codes. For the purpose of this application, BDD of a t-error-correcting component code can be seen as a black box that operates as follows. Let r=c+e, where c, e ∈{0,1}n denote a component code-word and random error vector, respectively, and n is the code length. BDD yields the correct code-word c if H(r, c)=H(e)≤t, where H and H denote Hamming distance and weight, respectively. On the other hand, if H(e)>t, the decoding either fails or there exists another codeword c′ such that H(r, c′)≤t. In the latter case, BDD is technically successful but the decided codeword c′ is not the correct one. Such miscorrections are highly undesirable because they introduce additional errors into the iterative decoding process and significantly degrade performance.
In one example, a novel iterative HD decoding algorithm for staircase codes which reduces the effect of miscorrections on the performance is disclosed. The algorithm provides significant post-FEC bit error rate improvements, in particular when t is small (which is typically the case in practice). As an example, for t=2, the algorithm can improve performance by roughly 0.4 dB and reduce the error floor by over an order of magnitude, up to the point where the iterative decoding process is virtually miscorrection-free. Error floor improvements are particularly important for applications with stringent reliability constraints such as OTNs.
Staircase Codes and Iterative Decoding
Let be a binary linear component code with length n and dimension k. Assuming that n is even, a staircase code with rate =2k/n−1 based on is defined as the set of all matrix sequences k∈{0,1}a×a, k=0, 1, 2, . . ., such that the rows in [k−1T, k] for all k≥1 form valid codewords of , where α=n/2 is the block size and 0 is the all-zero matrix.
We use extended Bose-Chaudhuri-Hocquenghem (BCH) codes as component codes, i.e., a BCH code with an additional parity bit formed by adding (modulo 2) all 2v−1 coded bits of the BCH code, where υ is the Galois field extension degree. The overall extended code then has length n=2v and guaranteed dimension k−2v−ut−1. The extra parity bit increases the guaranteed minimum distance to dmin=2t+2.
We briefly review the conventional decoding procedure for staircase codes which uses a sliding window comprising W received blocks k, k+1, . . . , k+w−1. This is illustrated in
Analyzing the post-FEC bit error rate of staircase codes under the conventional decoding procedure is challenging. A major simplification is obtained by assuming that no miscorrections occur in the BDD of the component codes. In this case, it is possible to rigorously characterize the asymptotic performance as a →∞ using a technique called density evolution. Moreover, the error floor can be estimated by enumerating harmful stall patterns. However, if miscorrections are taken into account, both the asymptotic and error floor predictions are nonrigorous and become inaccurate.
Example 1: Let υ=8 and t=2, which gives a staircase code with α=128 and =0.867. For window decoding parameters W=8 and =7, the density evolution and error floor predictions are shown in
The performance degradation with respect to idealized decoding becomes less severe for larger values of t. Unfortunately, small values of t are commonly used in practice because BDD can be implemented very efficiently in this case. We note at this point that there exist several works that attempt to quantify the performance loss due to miscorrections. In terms of error floor, the work in introduces a heuristic parameter, whose value unfortunately has to be estimated from simulative data.
The main idea in order to improve performance is to systematically exploit the fact that miscorrections lead to inconsistencies, in the sense that two component codes that protect the same bit may disagree on its value. In the following, we show how these inconsistences can be used to (a) reliably prevent miscorrections and (b) identify miscorrected code-words in order to revert their decoding decisions.
Algorithm 5 relies on so-called anchor code-words, which have presumably been decoded without miscorrections. Roughly speaking, we want to make sure that bit flips do not lead to inconsistencies with anchor code-words. Consequently, decoding decisions from code-words that are in conflict with anchors are not applied. However, a small number of anchor code-words may be miscorrected and we allow for the decoding decisions of anchors to be reverted if too many other code-words are in conflict with a particular anchor.
In order to make this more precise, we regard the BDD of a component code (i, j) as a two-step process. In the first step, the decoding is performed and the outcome is either a set of error locations εi,j⊂{1, 2, . . ,n} where |εi,j|≤t, or a decoding failure. In the second step, error-correction is performed by flipping the bits corresponding to the error locations. Initially, we only perform the decoding step for all component codes, i.e., all component codes in the decoding window are decoded without applying any bit flips. We then iterate times over the component codes in the same fashion as in Algorithm 5, but replacing line 6 with the following four steps:
Step 1: If no decoding failure occurred for the component code-word (i, j), we proceed to step 2, otherwise, we skip to the next code-word.
Step 2: For each e∈εi,j, check if e corresponds to an anchor code-word. If so, let be the number of other conflicts that the anchor is involved in. If <T, where T is a fixed threshold, the code-word (i, j) is frozen and we skip to the next codeword. Frozen code-words are always skipped for the rest of the decoding unless any of their bits change. If ≥T, the anchor is marked for backtracking.
Step 3 Error-correction for codeword (i, j) is applied, i.e., the bits at all error locations in εi,j are flipped. We also apply the decoding step again for code-words that had their syndrome changed due to a bit flip. Finally, the codeword (i, j) becomes an anchor.
Step 4: Previously applied bit flips are reversed for all anchor code-words that were marked for backtracking during step 2. These code-words are no longer anchors and all frozen code-words that were in conflict with these code-words are unfrozen.
The following two examples illustrate the above steps for t=2 and T=1 with the help of
Example 2: Assume that we are at (i, j)=(3, 4), corresponding to a component code with three attached errors shown by the black crosses. The codeword is miscorrected with ε3,4={10, 12} shown by the red crosses. Assuming that the codeword (4, 4) is an anchor without any other conflicts, the code-word (3, 4) is frozen during step 2 and no bit flips are applied.
Example 3: Let the code-word (1, 3) in
The previous example shows how a miscorrected anchor is backtracked. Since we do not know if an anchor is miscorrected or not, it is also possible that we mistakenly backtrack “good” anchors. Fortunately, this is unlikely to happen for long component codes because the additional errors due to miscorrections are approximately randomly distributed within the codeword. This implies that errors of two (or more) miscorrected code-words rarely overlap.
For the algorithm to work well, a sufficiently large fraction of code-words at each position should be “good” anchors. However, when the decoding window shifts and a new block is added, no anchors exist at the last position W−1. We found that it is therefore beneficial to artificially restrict the error-correcting capability of these component codes in order to avoid anchoring too many miscorrected code-words. For example, for t=2, all component codes at position W−1 are treated as single-error-correcting. This restriction reduces the probability of miscorrecting a component code by roughly a factor of n, which is significant for long component codes. Note that due to the window decoding, we are merely gradually increasing the error-correction capability: once the decoding window shifts, the component codes shift as well and they are then decoded with their full error-correcting capability.
We remark that essentially the same algorithm can also be applied to product and other “product-like” code constructions, e.g., half-product or braided codes.
In terms of decoding complexity, one of the main advantages of iterative HD decoding of staircase codes compared to message-passing decoding of LDPC codes is the significantly reduced decoder data flow requirement'. While a thorough complexity analysis for the proposed algorithm is beyond the scope of this paper, we note that the algorithm can operate entirely in the syndrome domain, thereby leveraging the syndrome compression effect that is described in3. However, additional storage is needed compared to the conventional decoding to keep track of the error locations of anchor code-words (in case they are backtracked) and to store the conflicts between codes.
We consider the same parameters as in Example 1, i.e., υ=8, t=2, W=8, and =7. The conflict threshold is set to T=1 and we apply the error-correction capability restriction for component codes at position W−1 as described above. Simulation results for the proposed algorithm are show in
Note that the staircase code parameters were chosen such that the error floor is high enough to be within the reach of software simulations. For OTNs, post-FEC bit error rates below 10−15 are typically required. In this case, other code parameters should be used or one may apply post-processing techniques to reduce the error floor below the application requirements.
We have shown that the post-FEC performance of staircase codes can be significantly improved by adopting a modified iterative HD decoding algorithm that reduces the effect of miscorrections. For component codes with error-correcting capability t=2, an additional coding gain of around 0.4 dB can be achieved. Moreover, the error floor can be reduced by over an order of magnitude, giving virtually miscorrection-free performance.
This application claims priority from U.S. Provisional Patent Application No. 62/559,002 filed Sep. 15, 2017, the contents of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
62559002 | Sep 2017 | US |