An embodiment relates generally to a memory system using integrated parallel interleaved concatenation.
In a field of error correction, a phenomenon that a decreasing rate of an error rate is deteriorated in a high SNR region as compared to a low SNR (signal to noise power ratio) region. This phenomenon is referred to as an error-floor phenomenon. The error-floor phenomenon is a crucial obstacle for many communication systems and memory systems. In general, when two received words have errors completely independent from each other, it is possible to decrease the error rate in an error-floor by automatic repeat request. However, in some systems, an error position of a repeatedly transmitted received word has correlation with the error position of a first received word. In such system, it is difficult to decrease the error rate in the error-floor by the automatic repeat request.
According to the following embodiment, provided is a memory system provided with a non-volatile memory; and a memory controller that controls writing in/reading from the non-volatile memory, wherein the memory controller is provided with an encoding device that encodes a plurality of user data to be written in the non-volatile memory, and a memory interface that writes a result of encoding by the encoding device in the non-volatile memory, the encoding device is provided with a first encoder that encodes the plurality of user data to generate first parities of the plurality of user data, a first interleaver that sequentially interleaves the plurality of user data, a first XOR accumulator that sequentially executes bit-wise modulo-2 operation on the plurality of user data sequentially interleaved by the first interleaver, and a second encoder that encodes a result finally obtained by sequentially executing the bit-wise modulo-2 operation by the first XOR accumulator on the plurality of user data interleaved by the first interleaver to generate second parity, and the memory interface writes a code word sequence including the plurality of user data, the first parities and the second parity in the non-volatile memory as a result of the encoding.
In general, in a system in which an error included in a received word is not completely independent from an error in a previous received word or a future received word, a method other than automatic repeat request is considered to be effective in decreasing an error rate in an error-floor. As such method, following three methods (1) to (3) may be exemplified.
(2) Usage of trapping set information
(3) Improved decoding method
The methods may also be used in combination with one another. An embodiment belonging to the concatenation of method (1) is hereinafter described as an example.
The concatenation of method (1) relates to a decrease in the error rate in the error-floor and is expected to have a higher effect than other methods (2) and (3) although a loss of a code rate is relatively large. For example, serial concatenation is used in digital video broadcasting (DVB-S2) and the like; a BCH code is suitable for the concatenation as described in an example of the DVB-S2. This is because error bits forming a trapping set are randomly distributed across an entire code in general and the number of error bits is not so large. In addition, the fact that the BCH code has excellent error correcting capability in a high code rate is one of reasons why the BCH code is suitable for the concatenation.
In order to decrease the loss of the code rate and improve the correcting capability, it is effective to encode entire user data by using a long BCH code being an outer code, then divide the entire user data into a plurality of blocks, and encode to protect each block by using an inner code. In such a method, decoding failure probability due to the trapping set is low, so that it is sufficient to consider a case in which decoding of one of a plurality of inner codes fails in almost all the cases. However, this method has a disadvantage that encoding and decoding of the BCH code become complicated because of a long code length as a result of using the long outer BCH code.
Therefore, the following embodiment suggests a memory system designed based on an error correcting scheme in which a small amount of parity generated by using several interleaved code words is used. The small amount of parity generated in the error correcting scheme of this suggestion is referred to as shared parity in the following description.
According to the error correcting scheme suggested in a first embodiment, ideally, it is possible to decrease the decoding failure probability in an error-floor region to approximately the second power of original probability as in the automatic repeat request. The error correcting scheme of this suggestion also has an advantage that normal operation may be performed even when decoding of one or more constituent codes protected by the shared parity fails due to the trapping set.
This method is the same as integrated interleaving in using the concatenation and modulo-2 addition; however, although the modulo-2 addition is performed on a last constituent code in the integrated interleaving, an additional parity region for the shared parity is secured separate from the constituent codes considering an effect of decreasing the error-floor in this method. Although interleaving to change bit order of the constituent code is not performed in the integrated interleaving, the interleaving to change the bit order of the constituent code is essentially important in this method.
Finally, a point that one code is sufficient for the error correcting scheme of this suggestion although the integrated interleaving requires two or more different codes is one of advantages of the error correcting scheme of this suggestion over the existing integrated interleaving.
(Interleaver)
An interleaver according to the first embodiment is first described. In general, for an LDPC code in which a parallel turbo code and an accumulator are used, the interleaver is expected to improve a minimum distance by locating adjacent bits to bit positions away from one another. On the other hand, this property of extending is not so important for the interleaver of the error correcting scheme according to the first embodiment as is understood from the following study. In the error correcting scheme according to the first embodiment, the interleaver is required to shuffle error vectors so as to avoid being caught by the trapping set. In general, the bit positions regarding the trapping set are randomly distributed across all the code words. Therefore, possibility of escaping from the trapping set by shuffling is not related to a distance between the original position and the position after the shuffling. This suggests that there is a case in which a one-bit cyclic shift interleaver and a short bit length interleaver may be used as the interleaver of the error correcting scheme according to the first embodiment.
However, as for a quasi-cyclic code, there is the same code constraint structure within a range of a block size of the quasi-cyclic code. Therefore, the interleaver according to the first embodiment is preferably not the one-bit cyclic shift interleaver but a one-block size cyclic shift interleaver. As another representative plan, there is a block interleaver which does not change bit order in the block but changes order of blocks. For example, when the number of blocks is five and the block size is three, an example of the block interleaver is as follows.
Herein, a first row indicates an input index of bits and a second row indicates an interleaver output index. Although such block interleaver maintains the order of bits in the block, a block index is changed by
in which a first row indicates an input index of the block and a second row indicates an interleaver output index of the block.
(Error Correcting Scheme)
Next, the error correcting scheme according to the first embodiment is described. Meanwhile, a principal object of the first embodiment is to decrease the error rate in the error-floor, so that the following description focuses on a method of encoding and decoding to generate an error-floor phenomenon. In order to clarify the description, systematic encoding is used by using a linear code C characterized by a parity check matrix H.
In the following description, “m”, represents the number (positive integer) of code words of the constituent code. The user data being an encoding target of the constituent code are represented as u1, u2, . . . , um and parity of each of them is represented as p1, p2, . . . , pm. A superscript i (i is an integer not smaller than one and not larger than m) of each user data ui indicates the constituent code word. In the following description, each of p1, p2, . . . , pm is also referred to as first parity.
The parity check matrix H is also represented as HU|HP. Herein, HU represents a user data part of H and HP represents a parity part of H. “|” represents the concatenation of row vectors or matrices having the same row number. Therefore, u1, u2, . . . , um are encoded to u1|p1, u2|p2, . . . , um|pm, respectively, by the encoding by using H.
(Encoding)
In the encoding in the error correcting scheme according to the first embodiment, following equation (1) is first calculated. Herein, “⊕” represents bit-wise modulo-2 addition. In the following description, the bit-wise modulo-2 addition is also referred to as “XOR”.
u
T
=u
1
⊕u
2
⊕ . . . ⊕u
m (1)
uT defined by equation (1) is interleaved according to interleave operation Π. The interleaved uT is represented as uT,I.
uT,I is encoded to uT,I|pT,I according to the parity check matrix H. Herein, pT,I is the shared parity. A code word sequence {u1|p1, u2|p2, . . . , um|pm, pT,I} obtained by the encoding is output as a result of the encoding in the error correcting scheme according to the first embodiment. That is to say, in the first embodiment, uT,I is not stored but the code word sequence ({u1|p1, u2|p2, . . . , um|pm, pT,I} formed of the code word sequence {u1|p1, u2|p2, . . . , um|pm} obtained by the encoding and the shared parity pT,I is output as the result of the encoding to be stored in a predetermined storage region. In the following description, the shared parity is also referred to as second parity for distinguishing the shared parity pT,I from the first parity pi.
As illustrated in
As illustrated in
u1, u2, . . . , um are sequentially input to the interleaver 111 to be interleaved for calculating uT,I and input to the XOR accumulator 112.
When describing operation by the XOR accumulator 112 focusing on ith user data ui, ui is interleaved to be input to the XOR 113 of the XOR accumulator 112. Data stored in the buffer 114 of the XOR accumulator 112 is also input to the XOR 113. Herein, the data stored in the buffer 114 is a result obtained by executing the bit-wise modulo-2 addition on the interleaved u1, . . . , ui−1. This result is output from the XOR 113 to be stored in the buffer 114. By executing the above-described operation on u1 to um, the result obtained by executing the bit-wise modulo-2 addition on all the interleaved user data is stored in the buffer 114. Meanwhile, when processing the user data u1 at the top, the result of the modulo-2 addition is not stored in the buffer 114. In this case, it is possible to input a predetermined bit pattern such as ‘0’ with the same data length to the XOR 113 in place of this or store the interleaved u1 in the buffer 114 without processing by the XOR 113.
When uT,I is stored in the buffer 114 in this manner, the data stored in the buffer 114 is input to the encoder 15 through the multiplexer 14 as illustrated in
As illustrated in
As a result of outputting the user data u1, u2, . . . , um, the first parity p1, p2, . . . , pm, and the second parity pT,I in the above-described manner, {u1|p1, u2|p2, . . . , um|pm, pT,I} is stored in a predetermined storage region as a result of final encoding.
Meanwhile, although the interleaver 111 is arranged upstream of the XOR accumulator 112 in
(Decoding)
Subsequently, decoding in the error correcting scheme according to the first embodiment is described.
Hereinafter, a case in which the number of decoding failures is one and a case in which this is plural are described by giving examples. Meanwhile, a received word string is represented as {v1|q1, v2|q2, . . . , vm|qm, qT,I} and a corresponding log-likelihood ratio (LLR) string is represented as {Lv
In a decoding process in the error correcting scheme according to the first embodiment, the decoding of all the constituent code words is first executed. A log-likelihood ratio (LLR) of a decoded result obtained by this and a hard decision result of the LLR are represented as ({Lu
(When there is One Decoding Failure)
Subsequently, a flow of the decoding process when the decoding of one constituent code word fails as a result of the above-described decoding is described. Hereinafter, a case in which the decoding of an ith constituent code word fails as a result of decoding all the constituent code words is described.
When the decoding of one constituent code word fails, prescriptions for the user data including an error which causes the decoding failure are executed. In the prescriptions, following procedures 1 to 5 are executed.
1. First, the bit-wise modulo-2 addition by following equation (2) is executed for m−1 user data u1, . . . , ui−1, ui+1, . . . , um which are successfully decoded.
u
T\i
=u
1
⊕ . . . ⊕u
i−1
⊕u
i+1
. . . ⊕u
m (2)
2. Next, uT\i is interleaved by the same replacing rule as that of the interleaver 111 in the above-described encoding and a result obtained by the interleaving is encoded by using the same code as that used in the above-described encoding. According to this, third parity HP−1HUΠuT\i is calculated. Subsequently, binary-to-bipolar conversion to convert bit values 0 and 1 to +1 and −1, respectively, is performed for each bit on HP−1HUΠuT\i. A result thereof is a row vector with a positive or negative parity length and is represented as si.
3. Next, LT,Isi is calculated. Herein, “
” represents component-wise multiplication of the row vector.
4. Next, user data Lvsi is made a parity part to be input to the decoder for decoding the code defined by HU|HP to be decoded. When the decoding is successful, interleaved user data wu
5. Next, by executing de-interleaving Π−1 on wu
As illustrated in
As illustrated in
In contrast, the decoding illustrated in
The decoder 27 makes the input interleaved Lv
Meanwhile, although the interleaver 111 is arranged upstream of the XOR accumulator 112 in
In the graph illustrated in
As is clear from
When the decoder may be provided with more memories, it is also possible to use the decoder having more powerful
in place of HU|HP as the parity check matrix. In this case, the input to the decoder is Lusi). In such a configuration, the interleaver may be used, so that a logic circuit for a second row block in the parity check matrix may be omitted. However, it is necessary to store a message corresponding to the second row block in the memory of the decoder.
(When there are Plural Decoding Failures)
Subsequently, the following two methods are described about the prescriptions when there is a plurality of decoding failures as a result of the above-described decoding. Although a case in which there are two decoding failures of the constituent code words is described as an example for simplification in the following description, the similar method may also be applied when there are three or more decoding failures unless the number thereof is not so large. Meanwhile, an upper limit of the number of the constituent code words correctable by the error correcting scheme described in the first embodiment depends on the systematic linear code C and the number m of the inner codes.
(First Prescription)
A first prescription is first described. One of advantages of the first prescription is that one interleaver may correspond to m code words. In the first prescription, following procedures 1 to 5 are executed. Hereinafter, a case in which the decoding of ith and jth constituent code words fails as a result of decoding all the constituent code words is described.
1. First, the bit-wise modulo-2 addition by following equation (3) is executed on m−2 user data u1, . . . , ui−1, ui+1, . . . , uj−1, uj+1, . . . , um successfully decoded.
u
T\i
=u
1
⊕ . . . ⊕u
i−1
⊕u
i+1
. . . ⊕u
j−1
⊕u
j+1
. . . ⊕u
m (3)
2. Next, uT\i,j is interleaved by the same replacing rule as that of the interleaver 111 in the above-described encoding and a result obtained by the interleaving is encoded by using the same code as the code used in the above-described encoding. According to this, third parity HP−1HUΠuT\i,j is calculated. Subsequently, the bit-wise binary-to-bipolar conversion is executed on HP−1HUΠuT\i,j. As a result, si,j is obtained.
3. Next, the component-wise multiplication LT,Isi,j is calculated.
4. Next, Lvsi,j of the component-wise multiplication in procedure 3 are input to the decoder which uses the code defined by a parity check matrix HU|HU|HP. When the decoding is successful, wu
5. Next, error-corrected user data ui (=Π−1wu
When the decoder may be provided with more memories, it is also possible to use the decoder having a more powerful parity check matrix
in place of HU|HU|HP as the parity check matrix. Herein, “I” is a unit matrix of the same size as the user data. In this case, the input to the decoder is Lusi,j).
(Second Prescription)
A second prescription is next described. The second prescription has an advantage that an additional memory size is not required although different interleavers are required for each user data. In the second prescription, following procedures 0 to 6 are executed. Hereinafter, a case in which the decoding of the ith and jth constituent code words fails as a result of decoding all the constituent code words is described.
0. First, uI,T is calculated by following equation (4). Herein, the second measure is different from the first prescription in that the interleaving is executed before each bit-wise modulo-2 addition. Then, second parity pI,T is calculated from uI,T by using HU|HP. Meanwhile, a corresponding LLR is represented by LI,T.
u
I,T=Π1u1⊕Π2u2⊕ . . . ⊕Πmum (4)
1. Next, the bit-wise modulo-2 addition by following equation (5) is executed on m−2 user data u1, . . . , ui−1, ui+1, . . . , uj−1, uj+1, . . . , um successfully decoded. Herein, uI,T\i,j is a sum of interleaved user data except Πiui and Πjuj.
u
I,T\i,j=Π1u1⊕ . . . ⊕Πi−1ui−1⊕Πi+1ui+1 . . . ⊕Πj−1uj−1⊕Πj+1uj+1 . . . ⊕Πmum (5)
2. Next, uI,T\i,j is encoded by using the same code as that used in the above-described encoding. According to this, third parity HP−1HUuI,T\i,j is calculated. Subsequently, the bit-wise binary-to-bipolar conversion is executed on the calculated HP−1HUuI,T\i,j. As a result, sI,i,j is obtained.
3. Next, the component-wise multiplication LI,TsI,i,j is calculated.
4. Next, (ΠiLvΠjLv
sI,i,j) is input to a HU|HP decoder. Herein, “
” represents component-wise box-plus operation. For example, the box-plus operation of x and y is defined as x
y=max(0,x+y)−max(x,y)+log(1+e−|x+y|)−log(1+e−|x−y|). “max” is a function which outputs a larger input argument value the approximation of which is sign(x)·sign(y)·min(|x|,|y|). Herein, sign(x)=x/|x|. “min” is a function which outputs a smaller input argument value. ΠiLu
ΠjLu
ΠjLu
ΠjLv
5. Next, when the decoding in procedure 4 is successful, an estimated word wI,uΠjLv
6. Next, (Lvti is input to the decoder corresponding to the code defied by HU|HP regarding the ith constituent code word, and when the decoding is successful, the error-corrected user data ui is obtained. Meanwhile, the same also applies to the jth constituent code word.
Herein, additional description is supplied for procedures 4 to 6 in the second prescription. A reason why the box-plus operation is used in procedure 4 is described. When the user data ui and uj are independent from each other and a communication path is memoryless, an LLR of ui⊕uj is represented by the box-plus operation of LLRs of ui and uj. Procedure 4 is the decoding of the code word ui⊕uj by using the shared parity. Since it is supposed that the two user data are independently generated and the communication path is memoryless, the box-plus operation is used for obtaining the input LLR to the decoder. In procedure 5, Πi−1eI,u
As described above, according to the error correcting scheme according to the first embodiment, it is possible to decrease the error rate in the error-floor even when the decoding failure occurs in one of a plurality of code words protected by the shared parity (second parity).
According to the error correcting scheme according to the first embodiment, it is possible to decrease the error rate in the error-floor by using an additional memory even when the decoding failure occurs in a plurality of a plurality of code words protected by the shared parity (second parity).
Furthermore, according to the error correcting scheme according to the first embodiment, it is possible to decrease the error rate in the error-floor by changing the method of interleaving in each code word for a plurality of code words even when the decoding failure occurs in a plurality of a plurality of code words protected by the shared parity (second parity).
Furthermore, in the error correcting scheme according to the first embodiment, it is possible to decrease the error rate in the error-floor also by using the interleaver having a simple configuration.
(Memory System)
Subsequently, an example of the memory system on which the encoding device and the decoding device designed based on the error correcting scheme according to the first embodiment are mounted is hereinafter described in detail with reference to the drawings.
The non-volatile memory 9 being a storage region in which the data is stored in a non-volatile manner is formed of one or more NAND flash memory, for example. The non-volatile memory other than the NAND flash memory may also be used. In the NAND flash memory, the data is written and read for each write unit data generally referred to as a page.
The memory controller 2 controls writing in the non-volatile memory 9 according to a write command from a host 30. The memory controller 2 controls reading from the non-volatile memory 9 according to a read command from the host 30. The memory controller 2 has a configuration in which a host interface (host I/F) 7, a memory interface (memory I/F) 5, a controller 3, an ECC (error correcting code) unit 6, and a data buffer 4 are connected to one another through an internal bus 8, for example.
The host interface 7 outputs a command received from the host 30, user data (write data) and the like to the internal bus 8. The host interface 7 transmits the user data read from the non-volatile memory 9 and a response from the controller 3 to the host 30. The host 30 may be an electronic device such as a personal computer and a portable terminal, for example.
The memory interface 5 controls a process of writing the user data and the like in the non-volatile memory 9 and a process of reading the data from the non-volatile memory 9 based on an instruction of the controller 3.
The controller 3 generally controls the memory system 1. The controller 3 is a CPU (central processing unit), an MPU (micro processing unit) and the like, for example. When the controller 3 receives the command from the host 30 through the host interface 7, this controls according to the command. For example, the controller 3 instructs the memory interface 5 to write the user data and the parity in the non-volatile memory 9 according to the command from the host 30. The controller 3 instructs the memory interface 5 to read the user data and the parity from the non-volatile memory 9 according to the command from the host 30.
The controller 3 determines a storage region (memory region) on the non-volatile memory 9 for the user data accumulated in the data buffer 4. The user data is stored in the data buffer 4 through the internal bus 8. The controller 3 determines the memory region for the data in a page unit being a writing unit (page data).
In this description, the user data stored in one page of the non-volatile memory 9 is defined as unit data. In the first embodiment, the code word is generated by the above-described encoding by using the unit data and the code word is stored in one page of the non-volatile memory 9 as one page data.
In this description, memory cells commonly connected to one word line are defined as a memory cell group. When the memory cell is a single-level cell, the memory cell group corresponds to one page. When the memory cell is a multi-level cell, the memory cell group corresponds to a plurality of pages. For example, when the multi-level cell capable of storing two bits is used, the memory cell group corresponds to two pages.
The controller 3 determines the memory region of the non-volatile memory 9 being a written destination for each unit data. A physical address is assigned to the memory region of the non-volatile memory 9. The controller 3 manages the memory region being the writing destination of the unit data by using the physical address. The controller 3 instructs the memory interface 5 to write the user data in the non-volatile memory 9 while specifying the determined memory region (physical address).
The controller 3 manages correspondence between a logic address (logic address managed by the host 30) of the user data and the physical address. When receiving the read command including the logic address from the host 30, this specifies the physical address corresponding to the logic address and instructs the memory interface 5 to read the user data while specifying the physical address.
The ECC unit 6 having a configuration for realizing the error correcting scheme according to the above-described first embodiment encodes the unit data stored in the data buffer 4 to generate the code word. The ECC unit 6 includes the encoding device 10 illustrated in
The data buffer 4 temporarily stores the user data received from the host 30 before storing the same in the non-volatile memory 9. The data buffer 4 temporarily stores the data read from the non-volatile memory 9 before transmitting the same to the host 30. The data buffer 4 is formed of a general-purpose memory such as a SRAM (static random access memory) and a DRAM (dynamic random access memory), for example.
Although a configuration in which the memory controller 2 is provided with the ECC unit 6 and the memory interface 5 separate from each other is illustrated in
Meanwhile, a SSD (flash solid state drive) is described above as an example of the memory system 1, the system is not limited thereto. For example, the error correcting scheme according to the above-described first embodiment may be applied to a system in which the error-floor phenomenon might occur such as a cSSD (client SSD) card and an eSSD (enterprise SSD).
In the error correcting scheme in the integrated parallel interleaved concatenation illustrated in the first embodiment, the generated code word is formed of a plurality of constituent codes and the shared parity common to all the constituent codes, and after the decoding (hereinafter, referred to as first-stage decoding) by using the first parities p1, p2, . . . , pm generated for the user data u1, u2, . . . , um, respectively, is executed, the decoding (hereinafter, referred to as second-stage decoding) by using the shared parity pT,I being the second parity is executed as needed. Therefore, in the first embodiment, when the decoding of at least one constituent code is unsuccessful as a result of decoding each constituent code in the first-stage decoding, the shared parity is appropriately corrected and the second-stage decoding is executed by using the corrected shared parity.
In order to realize the second-stage decoding by using the shared parity, an interleaver for shuffling error vectors so as to avoid being caught by a trapping set is required as described above. For example, when a decoding device 20 mounted on a memory controller of a NAND flash memory adopts a QC (quasi-cyclic)-LDPC code, an effect sufficient for decreasing an error rate in an error-floor may be obtained by using the interleaver which changes order of a circulant matrix. This means that the decoding device 20 used in the first-stage decoding can be reused in the second-stage decoding. Therefore, in the second embodiment, an example of a configuration of reusing the decoding device 20 for the first-stage decoding in the second-stage decoding is described. According to the second embodiment, complexity of the second-stage decoding may be decreased.
When a BP (belief propagation) decoding algorithm is used as a decoding algorithm of binary QC-LDPC decoding, an LLR (log-likelihood ratio) is conveniently used. The LLR is a value defined by
wherein x represents a value of a transmission bit, y represents a reception value, and P(y|x=0) represents a conditional probability that y is received when x=0 is transmitted.
The decoding device 20 which executes the BP decoding is required to store the LLRs having a length of a code length in a predetermined memory. Check node processing to decode in a row direction of a parity check matrix and variable node processing to decode in a column direction are executed on the stored LLR according to a position of a non-zero portion of a parity check matrix (H=HU|HP) (hereinafter, simply referred to as the check matrix). Meanwhile, the binary QC-LDPC code is the code having a structure, and the check matrix is the matrix which may be described using a matrix obtained by shifting a unit matrix referred to as the circulant matrix as a unit.
A check matrix 210 illustrated in
Subsequently, a decoding device according to the second embodiment is described in detail with reference to the drawings.
Herein, an example of the LLR memory 204 illustrated in
It is described with reference to
For example, when the check node processing is performed on the LLR block in the first row of the matrix, information of second, third, fifth, and seventh LLR blocks 2042, 2043, 2045, and 2047 and shift values 3, 2, 0, and 1 characterizing the LLR blocks, respectively, are required. Therefore, when the row block number (=1) of the LLR block in the first row is input, the check matrix manager 201 specifies the information of 2, 3, 5, and 7 as the LLR block numbers associated with the row block number and the information of 3, 2, 0, and 1 as the shift values characterizing the LLR blocks, respectively, from the management table 207 and outputs the specified information of the LLR block numbers (2, 3, 5, and 7) and shift values (3, 2, 0, and 1) to the LLR block number converter 202.
The LLR block number converter 202 corresponding to an interleaver 111 in the first embodiment is shared by an encoding device 10 and the decoding device 200 illustrated in
On the other hand, at the time of decoding, the LLR block number converter 202 is turned off in the first-stage decoding and turned on in the second-stage decoding. In the second-stage decoding in which the LLR block number converter 202 is turned on, the order of the input LLR block numbers is changed with a rule determined in advance (for example, refer to
Meanwhile, LLR block order random interleaving and LLR block cyclic shift interleaving are examples of a method of the interleaving by the LLR block number converter 202. In the LLR block order random interleaving, only the LLR block numbers are randomly changed without changing the order of the LLRs in each data LLR block. In the LLR block cyclic shift interleaving, the LLR block numbers are cyclically shifted by a method similar to that of the LLR block order interleaving. The two methods are simple and an effect of significantly decreasing the error rate in the error-floor may also be expected.
It is described with reference to
The row processing operator 205 performs row processing operation on the LLR block by using the access information and the shift value received from the LLR memory I/O 203 and stores a result of the row processing in the row processing result memory 206.
Herein, the row processing on the check matrix 210 is described with a focus on a flow of the row processing on the LLR block in the first row of the check matrix 210.
In a matrix illustrated in
Subsequently, a flow of a decoding process in the second embodiment is described with reference to a flowchart illustrated in
When the first-stage decoding at step S102 is successful (step S103; YES), the decoding result obtained by the first-stage decoding is output (step S107) and this operation is finished. On the other hand, when the first-stage decoding at step S102 is unsuccessful (step S103; NO), the LLR block number converter 202 turned into a stopped state at step S101 is put into an operating state (step S104) and the decoding (second-stage decoding) is executed in this state (step S105). In the state in which the LLR block number converter 202 is stopped, the LLR block numbers and the shift values input to the LLR block number converter 202 are interleaved with the replacing rule illustrated in
Thereafter, when the second-stage decoding at step S105 is successful (step S106; YES), the decoding result obtained by the second-stage decoding is output (step S107) and this operation is finished. On the other hand, when the second-stage decoding at step S105 is unsuccessful (step S106; NO), an error message indicating decoding failure is returned (step S108) and this operation is finished.
As described above, according to the second embodiment, it becomes possible to reuse the decoding device 200 used in the first-stage decoding in the second-stage decoding, so that the complexity of the second-stage decoding may be decreased. Other configuration, operation and effect may be similar to those described in the first embodiment, so that the detailed description thereof is herein omitted.
Meanwhile, in the integrated parallel interleaved concatenation, decoding performance may be further improved by decoding by using the parity check matrix of a wider range according to the decoding algorithm. In the simplest generalization, a data LLR block group the decoding of which is unsuccessful, a corresponding parity LLR block group, and a shared parity LLR block group are simultaneously BP-decoded.
While certain embodiments have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the inventions. Indeed, the novel embodiments described herein may be embodied in a variety of other forms; furthermore, various omissions, substitutions and changes in the form of the embodiments described herein may be made without departing from the spirit of the inventions. The accompanying claims and their equivalents are intended to cover such forms or modifications as would fall within the scope and spirit of the inventions.
This application is based upon and claims the benefit of priority from U.S. Provisional Application No. 62/281,790, filed on Jan. 22, 2016; the entire contents of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
62281790 | Jan 2016 | US |