The present inventions are related to systems and methods for data processing, and more particularly to systems and methods for scheduling in a data decoder.
Various data storage systems have been developed that include data decoding circuitry. Such data decoding circuitry generally schedules the processing of elements of a decoded message based upon a first in, first out scheduling. In some cases, such an approach to scheduling can use significant power.
Hence, for at least the aforementioned reasons, there exists a need in the art for advanced systems and methods for scheduling operations in a data processing system.
The present inventions are related to systems and methods for data processing, and more particularly to systems and methods for scheduling in a data decoder.
Various embodiments provide systems for decoding a data set that includes a layered data decoder circuit and a scheduler circuit. The layered data decoder circuit is operable to apply a data decoding algorithm to a data input to yield a decoded output. The data input includes a group of input elements. The scheduler circuit is operable to: identify layer based connections between individual elements of the group of input elements; assemble a first subset of the group of input elements that share a first set of layer based connections to yield a first processing group; assemble a second subset of the group of input elements that share a second set of layer based connections to yield a second processing group; and provide a layered processing order. The layered processing order includes the first processing group being processed by the layered data decoder circuit before the second processing group is processed by the layered data decoder circuit
This summary provides only a general outline of some embodiments of the invention. The phrases “in one embodiment,” “according to one embodiment,” “in various embodiments”, “in one or more embodiments”, “in particular embodiments” and the like generally mean the particular feature, structure, or characteristic following the phrase is included in at least one embodiment or one embodiment, and may be included in more than one embodiment. Importantly, such phases do not necessarily refer to the same embodiment. Many other embodiments of the invention will become more fully apparent from the following detailed description, the appended claims and the accompanying drawings.
A further understanding of the various embodiments may be realized by reference to the figures which are described in remaining portions of the specification. In the figures, like reference numerals are used throughout several figures to refer to similar components. In some instances, a sub-label consisting of a lower case letter is associated with a reference numeral to denote one of multiple similar components. When reference is made to a reference numeral without specification to an existing sub-label, it is intended to refer to all such multiple similar components.
a-10b shows an example element re-ordering in accordance with various embodiments.
The present inventions are related to systems and methods for data processing, and more particularly to systems and methods for synchronizing operations in a data storage system.
Various embodiments provide systems for decoding a data set that includes a layered data decoder circuit and a scheduler circuit. The layered data decoder circuit is operable to apply a data decoding algorithm to a data input to yield a decoded output. The data input includes a group of input elements. The scheduler circuit is operable to: identify layer based connections between individual elements of the group of input elements; assemble a first subset of the group of input elements that share a first set of layer based connections to yield a first processing group; assemble a second subset of the group of input elements that share a second set of layer based connections to yield a second processing group; and provide a layered processing order. The layered processing order includes the first processing group being processed by the layered data decoder circuit before the second processing group is processed by the layered data decoder circuit. In particular instances of the aforementioned embodiments, the system is implemented as part of an integrated circuit.
In some instances of the aforementioned embodiments, the layered processing order limits switching between layers during application of the data decoding algorithm. In particular cases, limiting switching between layers during application of the data decoding algorithm reduces power consumption by the layered data decoder circuit. In one or more instances of the aforementioned embodiments, the group of input elements is a group of circulants, the first subset of the group of input elements is a first subset of the group of circulants, and the second subset of the group of input elements is a second subset of the group of circulants. In various cases, the first subset of the group of circulants exhibits connections to a first previously processed layer, and the second subset of the group of circulants exhibits connections to a second previously processed layer. In other cases, the first subset of the group of circulants does not exhibit any connections to a previously processed layer, and the second subset of the group of circulants exhibits connections to the previously processed layer.
In one or more instances of the aforementioned embodiments, the system is implemented as part of a device selected from a group consisting of: a storage device, and a communication device. In various instances of the aforementioned embodiments, the layered data decoder circuit includes a variable node processor, and a check node processor. In some instances of the aforementioned embodiments, the data decoding algorithm is a low density parity check decoding algorithm.
Other embodiments provide methods for decoding a data set that includes: providing a layered data decoder circuit operable to apply a data decoding algorithm to a data input to yield a decoded output, where the data input includes a group of input elements; identifying layer based connections between individual elements of the group of input elements; assembling a first subset of the group of input elements that share a first set of layer based connections to yield a first processing group; assembling a second subset of the group of input elements that share a second set of layer based connections to yield a second processing group; and applying the data decoding algorithm by the data decoder circuit such that the first processing group is processed by the layered data decoder circuit before the second processing group is processed by the layered data decoder circuit.
In particular instances of the aforementioned embodiments, applying the data decoding algorithm by the data decoder circuit such that the first processing group is processed by the layered data decoder circuit before the second processing group is processed by the layered data decoder circuit limits switching between layers during application of the data decoding algorithm. In some cases, limiting switching between layers during application of the data decoding algorithm reduces power consumption by the layered data decoder circuit.
In various instances of the aforementioned embodiments, the group of input elements is a group of circulants, the first subset of the group of input elements is a first subset of the group of circulants, and the second subset of the group of input elements is a second subset of the group of circulants. In some cases, the first subset of the group of circulants exhibits connections to a first previously processed layer, and the second subset of the group of circulants exhibits connections to a second previously processed layer. In other cases, the first subset of the group of circulants does not exhibit any connections to a previously processed layer, and the second subset of the group of circulants exhibits connections to the previously processed layer.
Yet other embodiments provide storage devices that include: a storage medium; a data detector circuit operable to apply a data detection algorithm to a data input to yield a detected output, where the data input is derived from information accessed from the storage medium; a layered data decoder circuit operable to apply a data decoding algorithm to a data input to yield a decoded output, where the data input includes a group of input elements; and a scheduler circuit operable to: identify layer based connections between individual elements of the group of input elements; assemble a first subset of the group of input elements that share a first set of layer based connections to yield a first processing group; assemble a second subset of the group of input elements that share a second set of layer based connections to yield a second processing group; and provide a layered processing order, wherein the layered processing order includes the first processing group being processed by the layered data decoder circuit before the second processing group is processed by the layered data decoder circuit.
Turning to
In a typical read operation, read/write head 176 is accurately positioned by motor controller 168 over a desired data track on disk platter 178. Motor controller 168 both positions read/write head 176 in relation to disk platter 178 and drives spindle motor 172 by moving read/write head assembly 176 to the proper data track on disk platter 178 under the direction of hard disk controller 166. Spindle motor 172 spins disk platter 178 at a determined spin rate (RPMs). Once read/write head 176 is positioned adjacent the proper data track, magnetic signals representing data on disk platter 178 are sensed by read/write head 176 as disk platter 178 is rotated by spindle motor 172. The sensed magnetic signals are provided as a continuous, minute analog signal representative of the magnetic data on disk platter 178. This minute analog signal is transferred from read/write head 176 to read channel circuit 110 via preamplifier 170. Preamplifier 170 is operable to amplify the minute analog signals accessed from disk platter 178. In turn, read channel circuit 110 decodes and digitizes the received analog signal to recreate the information originally written to disk platter 178. This data is provided as read data 103 to a receiving circuit. A write operation is substantially the opposite of the preceding read operation with write data 101 being provided to read channel circuit 110. This data is then encoded and written to disk platter 178.
In operation, data accessed from disk platter 178 is processed using a layered decoding algorithm. The layered decoding algorithm utilizes an element scheduling algorithm that processes elements of each layer of a data set in an order that reduces power used by read channel circuit 110 by scheduling an order in which data elements of a current layer are processed to reduce the number of switches between other layers maintained in a decoder memory. The data processing including layered decoding may be implemented similar to that discussed below in relation to
It should be noted that storage system 100 may be integrated into a larger storage system such as, for example, a RAID (redundant array of inexpensive disks or redundant array of independent disks) based storage system. Such a RAID storage system increases stability and reliability through redundancy, combining multiple disks as a logical unit. Data may be spread across a number of disks included in the RAID storage system according to a variety of algorithms and accessed by an operating system as if it were a single disk. For example, data may be mirrored to multiple disks in the RAID storage system, or may be sliced and distributed across multiple disks in a number of techniques. If a small number of disks in the RAID storage system fail or become unavailable, error correction techniques may be used to recreate the missing data based on the remaining portions of the data from the other disks in the RAID storage system. The disks in the RAID storage system may be, but are not limited to, individual storage systems such as storage system 100, and may be located in close proximity to each other or distributed more widely for increased security. In a write operation, write data is provided to a controller, which stores the write data across the disks, for example by mirroring or by striping the write data. In a read operation, the controller retrieves the data from the disks. The controller then yields the resulting read data as if the RAID storage system were a single disk.
A data decoder circuit used in relation to read channel circuit 110 may be, but is not limited to, a low density parity check (LDPC) decoder circuit as are known in the art. Such low density parity check technology is applicable to transmission of information over virtually any channel or storage of information on virtually any media. Transmission applications include, but are not limited to, optical fiber, radio frequency channels, wired or wireless local area networks, digital subscriber line technologies, wireless cellular, Ethernet over any medium such as copper or optical fiber, cable channels such as cable television, and Earth-satellite communications. Storage applications include, but are not limited to, hard disk drives, compact disks, digital video disks, magnetic tapes and memory devices such as DRAM, NAND flash, NOR flash, other non-volatile memories and solid state drives.
In addition, it should be noted that storage system 100 may be modified to include solid state memory that is used to store data in addition to the storage offered by disk platter 178. This solid state memory may be used in parallel to disk platter 178 to provide additional storage. In such a case, the solid state memory receives and provides information directly to read channel circuit 110. Alternatively, the solid state memory may be used as a cache where it offers faster access time than that offered by disk platted 178. In such a case, the solid state memory may be disposed between interface controller 120 and read channel circuit 110 where it operates as a pass through to disk platter 178 when requested data is not available in the solid state memory or when the solid state memory does not have sufficient storage to hold a newly written data set. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize a variety of storage systems including both disk platter 178 and a solid state memory.
Turning to
In operation, data received via transfer medium 230 is processed by receiver 220 using a layered decoding algorithm. The layered decoding algorithm utilizes an element scheduling algorithm that processes elements of each layer of a data set in an order that reduces power used by read channel circuit 110 by scheduling an order in which data elements of a current layer are processed to reduce the number of switches between other layers maintained in a decoder memory. The data processing including layered decoding may be implemented similar to that discussed below in relation to
Turning to
Turning to
Analog to digital converter circuit 414 converts processed analog signal 412 into a corresponding series of digital samples 416. Analog to digital converter circuit 414 may be any circuit known in the art that is capable of producing digital samples corresponding to an analog input signal. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize a variety of analog to digital converter circuits that may be used in relation to different embodiments. Digital samples 416 are provided to an equalizer circuit 420. Equalizer circuit 420 applies an equalization algorithm to digital samples 416 to yield an equalized output 425. In some embodiments, equalizer circuit 420 is a digital finite impulse response filter circuit as are known in the art. It may be possible that equalized output 425 may be received directly from a storage device in, for example, a solid state storage system. In such cases, analog front end circuit 410, analog to digital converter circuit 414 and equalizer circuit 420 may be eliminated where the data is received as a digital data input. Equalized output 425 is stored to an input buffer 453 that includes sufficient memory to maintain a number of codewords until processing of that codeword is completed through a data detector circuit 430 and low density parity check (LDPC) decoding circuit 470 including, where warranted, multiple global iterations (passes through both data detector circuit 430 and LDPC decoding circuit 470) and/or local iterations (passes through LDPC decoding circuit 470 during a given global iteration). An output 457 is provided to data detector circuit 430.
Data detector circuit 430 may be a single data detector circuit or may be two or more data detector circuits operating in parallel on different codewords. Whether it is a single data detector circuit or a number of data detector circuits operating in parallel, data detector circuit 430 is operable to apply a data detection algorithm to a received codeword or data set. In some embodiments, data detector circuit 430 is a Viterbi algorithm data detector circuit as are known in the art. In other embodiments, data detector circuit 430 is a maximum a posteriori data detector circuit as are known in the art. Of note, the general phrases “Viterbi data detection algorithm” or “Viterbi algorithm data detector circuit” are used in their broadest sense to mean any Viterbi detection algorithm or Viterbi algorithm detector circuit or variations thereof including, but not limited to, bi-direction Viterbi detection algorithm or bi-direction Viterbi algorithm detector circuit. Also, the general phrases “maximum a posteriori data detection algorithm” or “maximum a posteriori data detector circuit” are used in their broadest sense to mean any maximum a posteriori detection algorithm or detector circuit or variations thereof including, but not limited to, simplified maximum a posteriori data detection algorithm and a max-log maximum a posteriori data detection algorithm, or corresponding detector circuits. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize a variety of data detector circuits that may be used in relation to different embodiments. In some cases, one data detector circuit included in data detector circuit 430 is used to apply the data detection algorithm to the received codeword for a first global iteration applied to the received codeword, and another data detector circuit included in data detector circuit 430 is operable apply the data detection algorithm to the received codeword guided by a decoded output accessed from a central memory circuit 450 on subsequent global iterations.
Upon completion of application of the data detection algorithm to the received codeword on the first global iteration, data detector circuit 430 provides a detector output 433. Detector output 433 includes soft data. As used herein, the phrase “soft data” is used in its broadest sense to mean reliability data with each instance of the reliability data indicating a likelihood that a corresponding bit position or group of bit positions has been correctly detected. In some embodiments, the soft data or reliability data is log likelihood ratio data as is known in the art. Detector output 433 is provided to a local interleaver circuit 442. Local interleaver circuit 442 is operable to shuffle sub-portions (i.e., local chunks) of the data set included as detected output and provides an interleaved codeword 446 that is stored to central memory circuit 450. Local interleaver circuit 442 may be any circuit known in the art that is capable of shuffling data sets to yield a re-arranged data set. Interleaved codeword 446 is stored to central memory circuit 450.
Once LDPC decoding circuit 470 is available, a previously stored interleaved codeword 446 is accessed from central memory circuit 450 as a stored codeword 486 and globally interleaved by a global interleaver/de-interleaver circuit 484. Global interleaver/de-interleaver circuit 484 may be any circuit known in the art that is capable of globally rearranging codewords. Global interleaver/De-interleaver circuit 484 provides a decoder input 452 into LDPC decoding circuit 470. LDPC decoding circuit 470 applies one or more iterations or a layered data decoding algorithm to decoder input 452 to yield a decoded output 471. In cases where another local iteration (i.e., another pass through LDPC decoding circuit 470) is desired (i.e., decoding failed to converge and more local iterations are allowed), LDPC decoding circuit 470 re-applies the layered data decoding algorithm to decoder input 452 guided by decoded output 471. This continues until either a maximum number of local iterations is exceeded or decoded output 471 converges (i.e., completion of standard processing).
In one embodiment, a power reduced decoder scheduler circuit 479 schedules the processing of LDPC decoding circuit 470. The scheduling involves selecting the next layer to be processed whenever a layer completes processing as indicated by a layer complete signal 474, and the order in which elements of the layer are to be processed. The scheduling information is provided from power reduced decoder scheduler circuit 479 to LDPC decoding circuit 470 as a scheduling input 473. In some cases, processing of layers by LDPC decoding circuit 470 overlap. In one particular embodiment, two layers may be processing at a given time. Turning to
In one example of operation of power reduced decoder scheduler circuit 479 identifies a current processing layer that has been processing through LDPC decoding circuit 470 is selected as a previous layer, and the next layer is selected as the current layer as the next layer begins processing through LDPC decoding circuit 470. Using
Connections between the selected current processing layer and all other layers are identified. Each of the layers include a number of elements to be processed by application of the data decoding algorithm and processing each of the elements may require access to elements from other layers within the data set being processed by the data decoding algorithm. In some cases, the elements are circulants as are known in the art. In one particular case, when processing circulants for a currently processing layer, corresponding messages (e.g., check node to variable node messages or variable node to check node messages) are accessed from other layers. Turning to
Power reduced decoder scheduling circuit 479 identifies one of the other layers (i.e., a layer other than the current layer) that exhibits the most connections with the current layer. Using
Power reduced decoder scheduling circuit 479 determines whether there are any elements of the current layer that have not yet been included in the updated group data set. Following the example in
Where decoded output 471 fails to converge (i.e., fails to yield the originally written data set) and a number of local iterations through LDPC decoding circuit 470 exceeds a threshold, but an allowable number of global iterations is not yet exceeded, the resulting decoded output is provided as a decoded output 454 back to central memory circuit 450 where it is stored awaiting another global iteration through a data detector circuit included in data detector circuit 430. Prior to storage of decoded output 454 to central memory circuit 450, decoded output 454 is globally de-interleaved to yield a globally de-interleaved output 488 that is stored to central memory circuit 450. The global de-interleaving reverses the global interleaving earlier applied to stored codeword 486 to yield decoder input 452. When a data detector circuit included in data detector circuit 430 becomes available, a previously stored de-interleaved output 488 is accessed from central memory circuit 450 and locally de-interleaved by a de-interleaver circuit 444. De-interleaver circuit 444 re-arranges decoder output 448 to reverse the shuffling originally performed by interleaver circuit 442. A resulting de-interleaved output 497 is provided to data detector circuit 430 where it is used to guide subsequent detection of a corresponding data set previously received as equalized output 425.
Alternatively, where the decoded output converges (i.e., yields the originally written data set), the resulting decoded output is provided as an output codeword 472 to a de-interleaver circuit 480 that rearranges the data to reverse both the global and local interleaving applied to the data to yield a de-interleaved output 482. De-interleaved output 482 is provided to a hard decision buffer circuit 428 buffers de-interleaved output 482 as it is transferred to the requesting host as a hard decision output 429.
As yet another alternative, where decoded output 471 fails to converge (i.e., fails to yield the originally written data set), a number of local iterations through LDPC decoding circuit 470 exceeds a threshold, and a number of global iterations through data detector circuit 430 and LDPC data decoding circuit 470 exceeds a threshold, the result of the last pass through LDPC decoding circuit 470 is provided as a decoded output along with an error indicator (not shown). It should be noted that power reduced decoder scheduler circuit 479 may also group elements of the current layer with connections to the previous layer toward the end of processing for the current layer to avoid latency similar to that discussed below in relation to
Turning to
LEH 510 yields stored Q values 512 or Qn(a) for the previous layers to the layer currently being processed, also referred to herein as the previous layer and the connected layer. An adder array circuit 514 adds Q values 512 to previous layer C2V messages 516 or R1,n(a) in array fashion to produce S messages 520 or Sn(a) containing total soft LLR values for the previous layer.
S messages 520 are provided to a normalization and permutation circuit 522, which converts the format of the S messages 520 from four soft LLR values to the equivalent content but different format of one hard decision and three soft LLR values (for a GF(4) embodiment), and which applies a permutation to rearrange the variable node updated values to prepare for the check node update and to apply the permutations specified by the non-zero elements of the H matrix. For example, in a GF(4) embodiment, the four elements 0-3 of the Galois Field are 0, 1, α, α2. The permutation applied by normalization and permutation circuit 522 is multiplication in the Galois Field. Element 2 (α) multiplied by element 1 (1) equals α×1 or α, which is element 2. Similarly, element 2×2=α×α=α2, which is element 3. Element 2×3=α×α2=1, which is element 1. Thus, element 2 multiplied by 1, 2 and 3 results in elements 2, 3, and 1, which are permutations of elements 1, 2 and 3. The normalization and permutation circuit 522 yields P messages 524 or Pn(a) for the previous layer. Normalization and permutation circuit 522 also yields soft LLR values 526 which are provided to a cyclic shifter 528. Cyclic shifter circuit 528 rearranges the soft LLR values 526 to column order, performs a barrel shift which shifts the normalized soft LLR values 526 from the previous layer to the current layer using the permutation specified by the H matrix, and which yields hard decisions 530 or an*, calculated as argmina Sn(a).
A parity check calculator 592 calculates parity checks based on the hard decisions 530, using any suitable technique. An example of the parity check calculations is provided in U.S. Pat. No. 8,656,249 filed on Sep. 7, 2011 for a “Multi-Level LDPC Layer Decoder” by Chen et al., which is incorporated by reference herein for all purposes. Parity check calculator 592 may calculate the parity checks on input data from LEH 510 before any local decoding iterations, or in other words, before any V2C (variable node to check node) or C2V messages are calculated in decoder 500.
P messages 524 from normalization and permutation circuit 522 are also provided to a shifter 532, a cyclic shifter or barrel shifter which shifts the symbol values in the normalized LLR P messages 524 to generate the next circulant sub-matrix, yielding current level P messages 534 which contain the total soft LLR values of the current layer. Current level P messages 534 are provided to a subtractor array 536 which subtracts the current layer C2V messages 538, or R2,n(a), from the current level P messages 534, yielding D messages 540, or Dn(a).
D messages 540 are provided to a normalization and saturation circuit 542 which converts the format of the D messages 540 from four soft LLR values to the equivalent content but different format of one hard decision and three soft LLR values, yielding new Q messages 544, or Q2,n(a), also referred to as V2C messages, for the current layer. Normalization and saturation circuit 542 also reduces the bit width of the Q messages 544, for example changing from 6-bit data words in section 582 of the decoder 500 to 5-bit data words to be stored in LEH 510 in section 580.
Q messages 544 are stored in LEH 510, overwriting previous channel or calculated values for the current layer, and are also provided to a scaler and saturation circuit 546 which scales the Q messages 544 to yield scaled V2C messages 548, or T2,n(a), and which lowers the bit-width further, for example to 4-bit data words to be used in section 584 of the decoder 500. The saturation functions used to lower the bit-width between sections 582 and 584 may be performed in multiple steps as shown, for example reducing from 6-bit words to 5-bit words in normalization and saturation circuit 542 and then from 5-bit words to 4-bit words in scaler and saturation circuit 546, or may be performed in separate single steps, for example reducing from 6-bit words to 5-bit words in normalization and saturation circuit 542 for storage in LEH 510 and reducing from 6-bit words to 4-bit words in scaler and saturation circuit 546 for use in section 584. Note that the bit-width of data words is increased in adder array circuit 514, which adds 5-bit words from LEH 510 to 4-bit words and yielding 6-bit S messages 520.
In some embodiments, V2C messages 548 are provided to min finder circuit 550 which calculates the minimum value min1(d), second or next minimum value min2(d) and the index of the minimum value idx(d). Min finder circuit 550 also calculates the signs of the V2C messages 548 and tracks the sign value of each non-zero element of the H matrix and the cumulative sign for the current layer. Min finder circuit 550 yields the current layer minimum, next minimum and index values with the sign values 552 to a previous layer C2V generator 554, which calculates the current layer C2V messages 538, or R2,n(a). Min finder circuit 550 also yields the previous layer minimum, next minimum and index values with sign values 556 to a current layer C2V generator 558, which calculates the previous layer C2V messages 516, or R1,n(a). Previous layer C2V generator 554 and current layer C2V generator 558 generate the C2V or R messages 538 and 516 based on the final state and current column index of the symbol. If the current column index is equal to the index of the minimum value, then the value of R is the second minimum value. Otherwise, the value of R is the minimum value of that layer. The sign of R is the XOR of the cumulative sign and the current sign of the symbol.
In summary, the variable node processor 504 and the check node processor 502 operate together to perform layered decoding of non-binary or multi-level data. Variable node processor 504 generates variable node to check node messages (V2C messages) and calculates perceived values based on check node to variable node messages (C2V messages). Check node processor 502 generates C2V messages and calculates checksums based on V2C messages, using a min finder circuit operable to identify a minimum, a next minimum and an index of minimum value in the V2C messages.
The saturation function applied in normalization and saturation circuit 542 and in scaler and saturation circuit 546 may be used in various embodiments to reduce bit-width from any source width to any target width. In the example embodiments disclosed herein, normalization and saturation circuit 542 and in scaler and saturation circuit 546 may be used to reduce the bitwidth of a data word from 6 bits to 5 bits, from 6 bits to 4 bits, and from 5 bits to 4 bits.
Turning to
Connections between the selected current processing layer and all other layers are identified (block 715). Each of the layers include a number of elements to be processed by application of the data decoding algorithm and processing each of the elements may require access to elements from other layers within the data set being processed by the data decoding algorithm. In some cases, the elements are circulants as are known in the art. In one particular case, when processing circulants for a currently processing layer, corresponding messages (e.g., check node to variable node messages or variable node to check node messages) are accessed from other layers.
Turning to
Power reduced decoder scheduling circuit 479 identifies one of the other layers (i.e., a layer other than the current layer) that exhibits the most connections with the current layer. Using
Returning again to
Returning to
Turning to
Following flow diagram 900, a current processing layer that has been processing through a data decoding circuit is selected as a previous layer (block 905), the next layer is selected as the current layer as the next layer begins processing through the data decoding circuit (block 910). Using
Connections between the selected current processing layer and all other layers are identified (block 915). Each of the layers include a number of elements to be processed by application of the data decoding algorithm and processing each of the elements may require access to elements from other layers within the data set being processed by the data decoding algorithm. In some cases, the elements are circulants as are known in the art. In one particular case, when processing circulants for a currently processing layer, corresponding messages (e.g., check node to variable node messages or variable node to check node messages) are accessed from other layers.
Turning to
Returning again to
One of the other layers (i.e., a layer other than the current layer) that exhibits the most connections with the current layer is identified (block 925). Using
It is determined whether there are any elements of the current layer that have not yet been included in the updated group data set (block 940). Following the example in
Once all of the elements have been incorporated in the updated group set (block 940), the order of the updated group set is fixed (block 945). The layered data decoding algorithm is then applied by the data decoding circuit to the updated grouped set (block 950). As shown in active layers 1005 and switches 1007 corresponding to layers 1020, the number of switches 1007 is reduced to fourteen (14) from thirteen (13). This reduction in the number of switches reduces the power usage of the data decoding circuit.
It should be noted that the various blocks discussed in the above application may be implemented in integrated circuits along with other functionality. Such integrated circuits may include all of the functions of a given block, system or circuit, or a subset of the block, system or circuit. Further, elements of the blocks, systems or circuits may be implemented across multiple integrated circuits. Such integrated circuits may be any type of integrated circuit known in the art including, but are not limited to, a monolithic integrated circuit, a flip chip integrated circuit, a multichip module integrated circuit, and/or a mixed signal integrated circuit. It should also be noted that various functions of the blocks, systems or circuits discussed herein may be implemented in either software or firmware. In some such cases, the entire system, block or circuit may be implemented using its software or firmware equivalent. In other cases, the one part of a given system, block or circuit may be implemented in software or firmware, while other parts are implemented in hardware.
In conclusion, the invention provides novel systems, devices, methods and arrangements for data processing. While detailed descriptions of one or more embodiments of the invention have been given above, various alternatives, modifications, and equivalents will be apparent to those skilled in the art without varying from the spirit of the invention. Therefore, the above description should not be taken as limiting the scope of the invention, which is defined by the appended claims.