The field generally relates to unidirectional or write only memories (WOM), which have cells which may be individually changed in a single direction, but are erased as a group, and more specifically but not exclusively to rewriting write only memories with error correction.
Unless otherwise indicated herein, the materials described in this section are not prior art to the claims in this application and are not admitted to be prior art by inclusion in this section.
Coding for rewriting is a technology used for flash memories. Coding has the potential to substantially increase the longevity, speed, and power efficiency of flash memories. Coding for rewriting has been proposed recently. See, e.g. V. Bohossian, A. Jiang, and J. Bruck, “Buffer coding for asymmetric multi-level memory,” in Proc. IEEE Internatio. Nal Symposium on Information Theory, June 2007, pp. 1186-1190, and A. Jiang, V. Bohossian, and J. Bruck, “Floating codes for joint information storage in write asymmetric memories,” in Proc. IEEE International Symposium on Information Theory, June 2007, pp. 1166-1170, which are incorporated herein by reference. Since the time coding was initially proposed, many works have appeared in this area. See, e.g. Y. Wu, “Low complexity codes for writing a write-once memory twice,” in Proc. IEEE International Symposium on Information Theory, June 2010, pp. 1928-1932, Y. Wu and A. Jiang, “Position modulation code for rewriting write-once memories,” IEEE Trans. Inf. Theor., vol. 57, no. 6, pp. 3692-3697, June 2011, E. Yaakobi, S. Kayser, P. H. Siegel, A. Vardy, and J. K. Wolf, “Codes for write-once memories,” IEEE Trans. Inf. Theor., vol. 58, no. 9, pp. 5985-5999, September 2012, and E. Yaakobi, S. Kayser, P. Siegel, A. Vardy, and J. Wolf, “Efficient two-write wom-codes,” in Proc. IEEE Information Theory Workshop, September 2010, pp. 1-5, which are incorporated herein by reference.
A model for rewriting is a write-once memory (WOM) model. See. e.g. R. L. Rivest and A. Shamir, “How to reuse a write-once memory,” Information and Control, vol. 55, no. 1-3, pp. 1-19, 1982, which is incorporated herein by reference. In the WOM model, a set of binary cells are used to store data, and the cell charge levels are increased when the data are rewritten. For flash memories, this constraint implies that the rewriting operation may delay the expensive block erasure, which leads to better preservation of cell quality and higher writing performance.
There have been many techniques for the design of WOM codes. They include linear code, tabular code, codes based on projective geometry, coset coding, etc. See, e.g. G. Cohen, P. Godlewski, and F. Merkx, “Linear binary code for write-once memories,” IEEE Trans. Inf. Theor., vol. 32, no. 5, pp. 697-700, September 1986, Merkx, “Womcodes constructed with projective geometries” Traitement du Signal, vol. 1, no. 2-2, pp. 227-231, 1984, and R. L. Rivest and A. Shamir, “How to reuse a write-once memory,” Information and Control, vol. 55. no. 1-3, pp. 1-19, 1982, which are incorporated herein by reference. Codes with substantially higher rates were discovered in recent years. See, e.g. V. Bohossian, A. Jiang, and J. Bruck, “Buffer coding for asymmetric multi-level memory,” in Proc. IEEE International Symposium on Information Theory, June 2007, pp. 1186-1190, and A. Jiang, V. Bohossian, and J. Bruck, “Floating codes for joint information storage in write asymmetric memories,” in Proc. IEEE International Symposium on Information Theory, June 2007, pp. 1166-1170, which are incorporated herein by references. Since the time such coding was initially proposed, many works have appeared in this area. See, e.g. Y. Wu, “Low complexity codes for writing a write-once memory twice,” in Proc. IEEE International Symposium on Information Theory, June 2010, pp. 1928-1 2, E. Yaakobi, S. Kayser, P. H. Siegel, A. Vardy, and J. K. Wolf, “Codes for write-once memories,” IEEE Trans. Inf. Theor., vol. 58, no. 9, pp. 5985-5999, September 2012, which are incorporated herein by reference. In 2012, WOM codes that achieved capacity were discovered by Shpilka et al. and Burshtein et al. See, e.g. A. Shpilka, “Capacity achieving multiwrite wom codes,” CoRR, vol. abs/1207.1128, 2012, “Capacity achieving two-write wom codes,” in LATIN 2012: Theoretical Informatics, ser. Lecture Notes in Computer Science, vol. 7256. Springer Berlin Heidelberg, 2012, pp. 631-642, E. Yaakobi and A. Shpilka, “High sum-rate three-write and non-binary wom codes,” in Proc. IEEE International Symposium on Information Theory, July 2012, pp. 1386-1390, and D. Burshtein and A. Strugatski, “Polar write once memory codes,” in Proc. IEEE International Symposium on Information Theory, July 2012, pp. 1972-1976, which are incorporated herein by reference. The latter code used a construction based on polar coding.
Compared to the large amount of work on WOM codes, the work on WOM codes that also correct errors has been much more limited. Existing works are mainly on correcting a few errors (for example, 1, 2, or 3 errors). See, e.g. E. Yaakobi, P. Siegel, A. Vardy, and J. Wolf, “Multiple error-correcting wom-codes,” IEEE Trans. Inf. Theor., vol. 58, no. 4, pp. 2220-2230, April 2012, and G. Zemor and G. D. Cohen, “Error-correcting wom-codes,” IEEE Trans. Inf. Theor., vol. 37, no. 3, pp. 730-734, May 1991, which are incorporated herein by reference.
Coding schemes that combine rewriting with error correction are discussed. In some embodiments, the schemes may support any number of rewrites and can correct a substantial number of errors. The code construction may use polar coding. In some embodiments, an analytical technique may be used which is based on frozen sets corresponding to the WOM channel and the error channel, respectively, including their common degrading and common upgrading channels. In some embodiments, lower bounds to the sum-rate may be achieved. The actual sum-rates may be further computed for various parameters. The analytical technique may focus on the binary symmetric channel (BSC). In practice, for relatively small error probabilities, the frozen set for BSC may be often contained in the frozen set for the WOM channel, which enables some embodiments to have a nested structure. The coding scheme can be further extended to multi-level cells (MLC) and more general noise models.
One aspect is a method of rewriting a memory. The method includes determining a current cell charge level of each of a plurality of cells of the memory, and generating a plurality of next cell charge levels according to a linear transformation, where each next cell charge level is generated based on a corresponding one of the current cell charge levels and based on input data. Each next cell charge level is greater than or equal to the corresponding current cell charge level, the plurality of next cell charge levels represent the input data, and the plurality of next cell charge levels include redundancy for error correction. The method also includes storing the next cell charge levels in the memory.
Another aspect is a method of reading a memory. The method includes determining current cell charge levels of a plurality of cells of the memory, where the current cell charge levels represent data and error correction information. The method also includes generating a plurality of data values, where the data values are generated based on a linear transformation of the current cell charge levels, and transmitting the data values to a data destination.
Another aspect is a memory system. The system includes a memory that includes a plurality of cells, a processor coupled to the memory and configured to determine current cell charge levels of the plurality of cells of the memory, and an encoder coupled to the memory and to the processor, and configured to generate a plurality of next cell charge levels according to a linear transformation, where each next cell charge level is generated based on a corresponding one of the current cell charge levels and based on input data. Each next cell charge level is greater than or equal to the corresponding previous cell charge level, the plurality of next cell charge levels represent the input data, and the plurality of next cell charge levels include redundancy for error correction. The system also includes a write device, configured to store the next cell charge levels in the memory.
Another aspect is a memory system. The system includes a memory that includes a plurality of cells, a processor coupled to the memory and configured to determine current cell charge levels of the plurality of cells of the memory, where the current cell charge levels represent data and error correction information. The system also includes a decoder coupled to the processor and to the memory, and configured to generate a plurality of data values, where the data values are generated based on a linear transformation of the current cell charge levels. The system also includes a transmitter coupled the decoder and configured to transmit the data values to a data destination.
The foregoing summary is illustrative and is not intended to be in any way limiting. In addition to the illustrative aspects, embodiments, and features described above, further aspects, embodiments, and features will become apparent by reference to the drawings and the following detailed description.
The foregoing and other features of this disclosure will become more fully apparent from the following description and appended claims, taken in conjunction with the accompanying drawings. Understanding that these drawings depict only several embodiments in accordance with the disclosure and are, therefore, not to be considered limiting of its scope, the disclosure will be described with additional specificity and detail through use of the accompanying drawings, in which:
FIG. illustrates an example of: Degrading channel
I. Introduction
This disclosure presents a code construction for error-correcting WOM codes. The code construction scheme may support any number of rewrites and can correct a substantial number of errors.
For example, a memory may have N binary cells, where every cell has two charge levels: 0 and 1. The cell-programming method can only change a cell from 0 to 1, but not from 1 to 0. However, noise can change a cell in both directions: from 0 to 1, and from 1 to 0.
The coding for the cell programming has two functions: (1) Rewrite data, which means to store new data that replace old data, and (2) Correct errors. In some embodiments, a coding method may support t writes, where each write is also called a rewrite, and where t is an integer parameter greater than or equal to two.
After a write of data, noise may appear in the cell levels, and certain embodiments of cell programming can correct errors and recover the data stored by the latest write. In some embodiments, each of the writes may use substantially the same process.
As an example, let (s1, s2, . . . , sN) be a binary vector that represent the levels of the N binary cells before a write. Let (s′1, s′2, . . . , s′N) be a binary vector that represent cell levels of the N binary cells after this write. In addition, let (b1, b2, . . . , bM) be a binary vector that represents the data to be written, where the data has M bits. Note the following constraints or conditions:
To write the data (b1, b2, . . . , bM), the method may have the following inputs:
Based on the inputs from (1) and (2), the method computes the output (s′1, s′2, . . . , s′N), and programs the cells to the new levels with the following constraints:
Data to be stored in the memory is included with the frozen set for WOM channel 26, and data of the frozen set for WOM channel 26 is provided to the polar encoder 27. In addition, additional data 31 is provided to the polar encoder 27. In the embodiment of
The polar codeword 28 is provided to a set of WOM channel devices 29, which determine cell levels to be written into the WOM. The WOM channel devices 29 determine the cell levels such that the new state of the WOM includes the data of the frozen set and such that the only cell level changes in the WOM are from the 0 state to the 1 state. Embodiments of WOM channel devices are described below in further detail.
The method of
The method shown in
In the first operation, represented by the flow diagram block 1002, current cell charge levels are determined. As discussed above, writing data into memory causes cell levels to change unidirectionally, for example, 0 to 1, but not 1 to 0. Accordingly, to determine which memory cells to change from 0 to 1, the current cell levels are determined.
In the next encoding operation, indicated by block 1004, next cell charge levels are generated based on the data to be written and on the current cell levels. Examples of techniques for determining next cell levels are discussed in further detail below. At block 1006, the next cell charge levels are written into the memory. For example, the next cell levels may be stored into memory cells of a data device, or maybe provided to a memory of a host device, or may be transmitted to an external destination.
As a result, the rewriting method enable new data to be stored in cells (to replace old data) two or more times without decreasing cell levels. In addition, an error correction method enables the most recently written data to be recovered even though noise may have changed the levels of some of the cells.
In particular, the rewriting operation takes the current cell levels (s1, . . . , sN) and the new data (b1, . . . , bM) as input, and computes new cell levels (s′1, . . . , s′N), where For i=1, . . . , N, s′i is greater than or equal to si. In addition, the new cell-level vector (s′1, . . . , s′N) represents the new data (b1, b2, . . . , bM), according to the mapping specified by the method. Accordingly, given the new cell levels (s′1, . . . , s′N), the decoding algorithm can uniquely determine the value of the new data (b1, . . . , bM).)
In some embodiments, to generate the next cell levels such as in block 1004, the method may use a binary matrix AN×N=(ai,j)N×N of N rows and N columns, whose value is fixed and known to the system. For example, for i=1, . . . ,N and j=1, . . . ,N, ai,j∈{0,1} is the bit in the i-th row and j-th column of the matrix AN×N. Furthermore, the method may use a subset of {1, 2, . . . ,N} denoted by FWOM, and another subset of {1,2, . . . ,N} denoted by FC to generate the next cell levels, such that:
FC contains at least one element. That is, |FC| is greater than or equal to 1,
There are M elements in FWOM but not in FC, that is, |FWOM−FC|=M,
At least one element of {1, 2, . . . ,N} is neither in FWOM nor in FC. That is, |{1, 2, . . . ,N}−FWOM−FC| is greater than or equal to 1.
FC is a subset of FWOM. That is, FC ∈FWOM.
Furthermore, in some embodiments, to generate the next cell levels, the method may use a binary vector (g1, g2, : : : , gN) called a dither, whose value is known to the system. The value of the dither can be the same or different for each of the rewrites. Two other binary vectors (v1, v2, . . . , vN) and (v′1, v′2, . . . , v′N ) may be determined as follows:
For i=1, . . . ,N, vi=si⊕gi is called the value of the i-th cell before the write. Here ⊕ is the exclusive-OR operation.
For i=1, . . . ,N, v′i=s′i⊕gi is called the value of the i-th cell after the write.
Furthermore, let A−1N×N be the inverse of the matrix AN×N, and let (u1, u2, . . . , uN) be the binary vector (u1, u2, . . . , uN)=(v′1, v′2, . . . , v′N) A−1N×N. In addition, the vector (u1, u2, . . . , uN) may have the following properties:
The value of the bits {ui|1≦i≦N, i∈FWOM−FC}—namely those bits in (u1, . . . , uN) whose indices are in FWOM FC—equals the value of the data (b1, , bM).
The value of the bits {ui|i∈FC} is a constant (such as all 0s or any other fixed value) known to the system.
The value of the bits {ui|1≦i≦N, i≠FWOM−FC} is computed by the method for rewriting.
In some embodiments, the number of cells whose levels are change from 0 to 1 by the rewriting method is minimized or otherwise reduced. Accordingly, the method may minimize or otherwise reduce the value of Σi=1N s′i−si.
In some embodiments, the method may use the following 1st probabilistic model to describe the relationship between cell levels (s1, . . . , sN) and (s′1, . . . s′N). Let
In some embodiments, the method may use the new cell levels (s′1, . . . , s′N) as the vector that maximizes or otherwise increases the value Pr{(s′1, . . . , s′N)|(s1, . . . , sN)}.
In some embodiments, the method may choose the new cell levels (s′1, . . . , s′N) as the vector that maximizes or otherwise increases the value Pr{(s1, . . . , sN)|(s′1, . . . , s′N)}.
In some embodiments, to generate the next cell levels, the method may compute the value of bits u1, u2, . . . , uN sequentially, and subsequently computes the value of cell levels (s′1, . . . , s′N) as (s′1, . . . , s′N)=(v′1, . . . , v′N)⊕(g1, . . . , gN)=(u2, . . . , uN)·A−1N×N⊕(g1, . . . , gN). For i=1, . . . , N, the method computes the following:
1: If i∈FWOM−FC and i is the j-th element in FWOM−FC ( for j∈{1, 2, . . . M}), let ui=bj (the j-th bit in the new data). Otherwise, go to step 2.
2. If i∈FC, let μi be the constant value specified previously (such as the value 0). Otherwise, go to step 3.
3. If i∈{1, 2, . . . , N}−FWOM, let μi take the value that maximizes the value of Pr{μi|(s1, ν1), . . . , (sN, νN), μ1, . . . , μi−1} assuming that each of μi+1, . . . , μN is equally likely to be 0 and 1.
In some embodiments, the method may use the following 2nd probabilistic model to describe the relationship between vectors (v′1, . . . , v′N) and ((s1,v1), . . . , (sn,vn)). For I=1, . . . , N.
In some embodiments, the method may choose the new cell levels (s′1, . . . , s′N) as the vector that maximizes or otherwise increases the value Pr{(s′1, . . . , s′N)|(s1, . . . , sN)}.
In some embodiments, the method may choose the new cell levels (s′1, . . . , s′N) as the vector that maximizes or otherwise increases the value Pr{(s1, . . . , sN)|(s′1, . . . , s′N)}.
In some embodiments, the method may compute the value of bits u1, u2, . . . , uN sequentially, and subsequently computes the value of levels (s′1, . . . , s′N) as (s′1, . . . , s′N)=(v′1, . . . , v′N) ⊕(g1, . . . , gN)=(u2, . . . , uN)·A−1N×N⊕⊕(g1, . . . , gN). For i=1, . . . , N the method may compute the following:
1: If i∈FWOM−FC and i is the j-th element in FWOM−FC (for j∈{1,2, . . . , M}), let ui=bj (the j-th bit in the new data). Otherwise, go to step 2.
2: If i∈FC, let ut be the constant value specified previously (such as the value 0. Otherwise, go to step 3.
3: If i∈{1,2, . . . , N}−FWOM, let μi take the value that maximizes the value of Pr{μi|(s1, ν1), . . . , (sN, νN), μ1, . . . , μi−1} assuming that each of μi+1, . . . , μN is equally likely to be 0 and 1.
In some embodiments, the matrix AN×N, may be specified as follows. Let m be a positive integer, and let N=2m, Let G2=
In some embodiments, the dither may be a pseudo-random number, which can be generated by a pseudo-random-number generator using a seed number as input. The seed number may be stored in the system.
In some embodiments, the method may compute the value of bits u1, u2, . . . ,uN sequentially, and subsequently computes the value of cell levels (s′1, . . . , s′N) as (s′1, . . . s′N) ′1, . . . , v′N)⊕(g1, . . . ,gN)=(u2, . . . , uN)·A−1N×N⊕⊕(g1, . . . , gN). For i=1, . . . , N, the method may compute the following::
W11((1,0), (*)|0)=1−α
W11((0,0), (*)|0)=α(1−∈)
W11((0,1), (*)|0)=α∈
W11((1,1), (*)|0)=0
W11((1,0), (*)|1)=0
W11((0,0), (*)|1)=α∈
W11((0,1), (*)|1)=α(1−∈)
W11((1,1), (*)|1)=1−α. (3)
In some embodiments, the matrix AN×N may be the generating matrix of a polar code, and FWOM may be contained in the frozen set of the polar code corresponding to the channel described by the transition 2nd probabilistic model, where the channel has v′i as the channel input and (si, vi) as the channel output.
In some embodiments, FWOM may be determined as follows:
For i=1,2, ,N, let z denote ((y1, y2, , yN), (u1, u2, , ui−1)), and compute a value
In some embodiments, the matrix AN×N may be the generating matrix of a polar code.
After the rewrite described above, which changes cell levels to (s′1, s′2, . . . , s′N) to store the data (b1, b2, . . . , bM), noise may appear in the cells and change the cell levels to (c1, c2, . . . , CM), where ci is a bit that is not necessarily equal to s′i. The error-correction method takes the noisy cell levels (c1, c2, . . . , CM) as input, and outputs the stored data (b1, b2, . . . , bM).
In the first operation, represented by the flow diagram block 2002, a set of cell level values to be decoded is retrieved from the WOM.
In the next decoding operation, indicated by block 2004, the set of values is decoded to generate the corresponding data value. Examples of techniques for decoding are discussed in further detail below. At block 2006, the data value is provided to a data destination, such as a host computer. For example, the data value may be transmitted by a transmitter to the data destination.
In particular, to generate the data value, the decoding or error correction function may use a binary vector (u′1, . . . , 1′N) and has the following properties:
The probability may be determined by the errors' probabilistic characteristic. Accordingly, for every i∈FWOM−FC a data bit bj is recovered as u′i, where bj was assigned to ui for the most recent rewrite.
In some embodiments, to generate the data value, the method may use a binary matrix AN×N=(ai,j)N×N of N rows and N columns, whose value is fixed and known to the system. For example, for i=1, . . . ,N and j=1, ,N ai,j∈{0,1} is the bit in the i-th row and j-th column of the matrix AN×N. Furthermore, the method may use a subset of {1,2, . . . , N} denoted by FWOM, and another subset of {1,2, ,N} denoted by FC, such that:
FC contains at least one element. That is, |FC| is greater than or equal to 1,
There are M elements in FWOM but not in FC, that is, |FWOM−FC|=M,
At least one element of {1,2, . . . ,N} is neither in FWOM nor in FC. That is, |{1, 2, . . . ,N}−FWOM−FC| is greater than or equal to 1.
FC is a subset of FWOM. That is, FC∈FWOM.
Furthermore, in some embodiments, the errors in the cells may have the following distribution:
In such embodiments, binary vector ((u′1, . . . , 1′N) may have the following properties:
In some embodiments, the matrix AN×N may be the generating matrix of a polar code, and p is a real number parameter between zero and one. The method may correct errors and may recover the data values (b1, b2, . . . , bM) as follows:
1: Recovering the value of (ν′1, . . . , ν′N) using a decoding algorithm of an error-correcting polar code based on the following setting:
2: Let (u′1, . . . , u′N)=(ν′1, . . . , ν′N)AN×N−1. For every i∈FWOM−FC, —say bj was assigned to ui for the most recent rewrite, —we recover the data bit bj as u′i.
In some embodiments, to generate the data value, the decoding algorithm also may include computing the value of ui for i=1, . . . , N as follows:
In some embodiments, the decoding algorithm may be a list-decoding algorithm that is further specified as follows: Let L be a positive integer parameter that indicates the size of a list, where each clement in the list is an assignment of value to (u1, u2, . . . , ui−1) before the value of ui is computed. For i=1, . . . , N, the decoding algorithm computes the value of ui with the following:
The most likely element in the list may then be chosen as the recovered value for (u1, . . . , uN).
In some embodiments, the stored data (b1, . . . , bM) may itself be an error-correcting code C and therefore may contain redundancy. The decoding algorithm, in 1 may be further specified as follows:
If i∈FC, let ui be the constant value that was assigned to it by the previous write.
If i∉FC, for each element in the list (which is an assignment of value to (u1, u2, . . . , ui−1)), append to it the value assignment ui=0, and also append to it the value assignment ui=0, thus converting that element into two new elements in the list. If the list now contains more than L elements, keep only the L elements that are most likely in the list. The likelihood of an element in the list may be computed as Pr{c1, . . . , cN|u1, . . . , ui} assuming that each of ui+1, . . . , uN is equally likely to be 0 and 1.
Then, for each element in the list, the method checks if its value assignment to (ui|i∈FWOM−FC), which should equal (b1, . . . , bM), is a valid codeword in the error-correcting code C; and if not, remove it from the list. Then, the method may choose the most likely element in the remaining list as the recovered value for (u1, . . . , uN).
In some embodiments, the decoding algorithm may be based on a belief-propagation decoding algorithm for polar error-correcting codes.
In some embodiments, the matrix AN×N may be specified as follows. Let m be a positive integer, and let N=2m. Let G2=
In some embodiments, recovering the value action discussed above may further include the following:
1: Let y=(y1, y2, . . . , yN)=((c1⊕g1), (c2⊕g2) . . . , (cN ⊕gN)) be a vector of N elements, where every element yi=ci⊕gi is a bit.
2: For every i∈FC, let ui32 0.
3: Compute the values of WNi(y,(u1, u2, . . . , u1−1)|ui=0) and WNi(y, (u1, u2, . . . , ui−1)|ui=1) for i=1,2, . . . N as follows:
W11((0), (.)|0)=1−p
W11((1), (.)|0)=p
W11((0), (.)|1)=p
W11((1), (.)|1)=1−p (7)
5: Let {ui|i∈FWOM−FC} be the recovered value for the data (b1, . . . , bM), where the mapping from the bits in {ui|i∈FWOM−FC} to the bits in (b1, . . . , bM) is as specified in the previous rewrite operation.
In some embodiments, the method may be a list-decoding algorithm further specified as follows:
1: Let y=(y1y2, . . . , yn)=((c1⊕g1), (c2⊕g2), . . . , (cN⊕gN)) be a vector on N elements, where every element yi=ci⊕gi is a bit.
2: For every i∈FC, let ui=0.
3: Compute the values of WNi(y,(u1, u2, . . . , ui−1)|ui=0) and WNi(y,(u1, u2, . . . , ui−1)|ui=1) for i=1,2, . . . , N as follows:
W11((0), (.)|0)=1−p
W11((1), (.)|0)=p
W11((0), (.)|1)=p
W11((1), (.)|1)=1−p (10)
4: For i=1,2, . . . , N, do:
5: Choose the most likely element in the list as the recovered value for (u1, . . . uN). The likelihood of an element in the list here may be computed as
In some embodiments, the stored data (b1, . . . , bM) may itself be an error-correcting code C and therefore may contain redundancy and the method may include:
1: Let y=(y1y2, . . . , yn)=((c1⊕g1), (c2⊕g2), . . . , (cN⊕gN)) be a vector of N elements, where every element yi=ct⊕gt is a bit.
2: For every i∈FC, let ui=0.
3: Compute the values of WNi(y,(u1, u2, . . . , ui−1)|ui=0) and WNi(y,(u1, u2, . . . , ui−1)|ui=1) for i=1,2, . . . , N as follows:
W11((0),(.)|0)=1−p
W11((1),(.)|0)=p
W11((0),(.)|1)=p
W11((1),(.)|1)=1−p (13)
4: For i=1,2, . . . , N, do:
5: Then, for each element in the list, check if its value assignment to (ui|i∈FWOM−FC), which should equal (b1, . . . , bM), is a valid codeword in the error-correcting code C; and if not, remove it from the list. Then, choose the most likely element in the remaining list as the recovered value for (u1, . . . uN). The likelihood of an element in the list here may be computed as WNN ((y1. . . , uN), (u1, . . . , uN−1). Then let {ui|i∈FWOM−FC} be the recovered value for the data (bi, . . . , bM), where the mapping from the bit in {ui|i∈FWOM−FC} to the bits in (b1, . . . , bM) is as specified in the previously rewrite operation.
In some embodiments, FC may be determined as follows:
In some embodiments, FC ⊂FWOM. In alternative embodiments, FC is not a subset of FWOM. That is, FC−FWOM is not empty.
In some embodiments, the rewriting or encoding methods described above are modified to include storing the value of the bits {ui|i∈FC−FWOM} in Nadditional additional cells, using an error-correcting code of length Nadditional that encodes |FC−FWOM| bits. (Those additional cells are disjoint for the rewrites.)
In some embodiments, the error correction of decoding methods described above are modified to include recovering the value of the bits in {ui|i∈FC−FWOM} by decoding the error-correcting code stored in the Nadditional additional cells. In addition, the error correction or decoding method described above may be such that the value of the bits in {ui|i∈FC−FWOM} is fixed as the above recovered value.
In some embodiments, for j=1, 2, . . . , t, FCj denotes the set FC for the j-th rewrite. FWOMj denotes the set FWOM for the j-th rewrite, and Nadditional additional cells are used for a code that supports t rewrites and error-correction processes. In such, embodiments, for j=1, . . . , t, M=|FCj−FWOMj| bits are stored in the j-th rewrite. The code can be any code described herein.
In some embodiments, methods of encoding discussed above may be modified such that for j=1, 2, .. . . , t, and for the j-th rewrite, the value of the bits {ui|i∈FC,j−FWOM,j} in the Nadditional additional cells, are stored using the code specified above for the Nadditional additional cells.
In some embodiments, for j=1, . . . , t, M=|FC,j−FWOMj| bits are stored in the j-th rewrite, and methods of decoding discussed above may be modified such that the value of the bits in {ui|i∈FC,j−FWOM,j} are recovered by decoding the rewrite-and error-correction code stored in the Nadditional additional cells. In addition, methods of decoding discussed above may be modified such that the value of the bits in {ui|i∈FC,j−FWOM, j} is fixed as the above value.
In some embodiments, the WOM cells are binary. In alternative or additional embodiments, every cell has q-levels: level 0, level 1, . . . , and level q-1,
In such embodiments, methods for rewriting and error-correction or encoding may use the cells in the following level-by-level approach:
In some embodiments, methods for rewriting and error correction or encoding may be modified such that each ui, si, gi, vi, s′i or v′i takes its value from the set {0,1, . . . , q-1}, and the exclusive-OR operation may be generalized to the modular-q summation.
An example, as discussed with reference to
Consider a rewrite (encode) operation and the subsequent error-correction (decode) operation. For the rewrite, let the dither be (g1, . . . , gN)=(g1, . . . , g8)=(1,0,1,0,1,0,0,1). Let a=0:87. Let e=0:29. Let FWOM={1, 2, 4, 5, 6}. Let FC={2, 6}. Let M=|FWOM−FC|=|{1, 4, 5}|=3. And let the data to be stored be (b1, b2, . . ., bM)=(b1, b2, b3)=(1, 0, 1). Assume the cell levels before this rewrite are (s1, . . . , sN)=(s1, . . ., sN)=(0, 0, 1, 0, 0, 0, 0, 0). As a result (v1, . . . , vn)=(v1, . . . , v8)=(s1 ⊕g1, . . . , s8⊕g8)=(1, 0, 0, 0, 1, 0, 0, 1).
Consider the rewrite where (v1, . . . , v8)=(u1, . . . , u8) A8×8 and the transition probabilities of the WOM channels 42 are shown in
Next, consider the error correction (decode).
The remainder of the description is organized as follows. Section II includes a basic model and notations. Section III includes an embodiment of code construction. Section IV includes embodiments of code. Section V includes embodiments of code extensions. Section VI includes an analysis of actual sum-rates achieved by a code embodiment. Section VII includes further example embodiments. Section VIII includes concluding remarks.
II. Basic Model
A memory may have N(=2m) WOM cells that are used to store data, where each cell has two levels; 0 and 1. The cells can change only from level 0 to level 1, but not vice versa.
A sequence of t messages M1, M2, . . . , Mt may be written into the WOM cells, and when Mi is written, the value of the previous messages need not be remembered. (Let Mj denote the number of bits in the message Mj, and let Mj∈{0,1}M
After cells are programmed, noise may appear in the cell levels. Noise may be considered to be a BSC with error probability p, denoted by BSC(p). These errors may be hard errors, namely, they physically change the cell levels from 0 to 1 or from 1 to 0. For flash memories, such errors can be caused by read/write disturbs, interference and charge leakage, for example, and may be quite common.
A. The Model for Rewriting
A code for rewriting and error correction may comprise of t encoding functions E1, E2, . . . , Et and t decoding functions D1,D2, . . . , Dt. For i=1,2, . . . , N and j=1,2, . . . , t, let si,j∈{0,1} and s′i,j∈{0,1} denote the level of the i-th cell right before and after the j-th write, respectively. As discussed above, s′i,j≧si,j. Let ci,j∈{0,1} denote the level of the i-th cell at any time after the j-th write and before the (j+1)-th write, when reading of the message Mj can happen. The error ci,j⊕s′i,j∈{0,1} is the error in the i-th cell caused by the noise channel BSC(p). (Here ⊕ is an XOR function.) For j=1,2, . . . , t, the encoding function
Ej:{0,1}N×{0,1}M
Dj:{0,1}N→{0,1}M
For j=1, . . . ,t,
B. Polar Codes
A short introduction to polar codes follows due to its relevance to the code construction embodiments discussed herein. A polar code is a linear block error correcting code proposed by Arkan, See, e.g. E. Arkan, “Channel polarization: A method for constructing capacity-achieving codes for symmetric binary-input memoryless channels,” IEEE Trans. Inf. Theor., vol. 55, no. 7, pp. 3051-3073, July 2009. It is the first known code with an explicit construction that provably achieves the channel capacity of symmetric binary-input discrete memoryless channels (B-DMC). The encoder of a polar code transforms N input bits u=(u1, u2, . . . , uN) to N codeword bits x=(x1, x2, . . . , xN) through a linear transformation. (e.g. x=u where G2=
and is the m-th Kronecker product of G2.) The N codeword bits (x1, x2, . . . , xN) are transmitted through N independent copies of a B-DMC. For decoding, N transformed binary input channels {WN(1), WN(2), . . . , WN(N)} can be synthesized for u1, u2, . . . , uN, respectively. The channels are polarized such that for large N, the fraction of indices i for which I(WN(t)) is nearly 1 approaches the capacity of the B-DMC, while the values of I(WN(i)) for the remaining indices i are nearly 0. The latter set of indices are called the frozen set. For error correction, the ui′s with i in the frozen set take fixed values, and the other ui′s are used as information bits. A successive cancellation (SC) decoding algorithm achieves diminishing block error probability as N increases.
Polar code can also be used for optimal lossy source coding, which has various applications. See, e.g. S. Korada and R. Urbanke, “Polar codes are optimal for lossy source coding,” IEEE Trans. Inf. Theor. , vol. 56, no. 4, pp. 1751-1768, April 2010. In particular the idea may be used to build capacity achieving WOM codes. See, e.g., D. Burshtein and A. Strugatski, “Polar write once memory codes,” in Proc. IEEE international Symposium on Information Theory, July 2012, pp., 1972-1976.
The presented code analysis uses the concept of upgrading and degrading channels, defined based on frozen sets. As taught by Tal and Vardy, a channel W′: X→Z is called “degraded with respect to a channel W: X→Y” if an equivalent channel of W′ can be constructed by concatenating W with an additional channel Q: Y→Z, where the inputs of Q are linked with the outputs of W. See, e.g. I. Tal and A. Vardy, “How to construct polar codes,” CoRR, vol. abs/1105.6164, 2011. That is,
W′(z|x)=Σy∈yW(y|x)Q(z|y)
It may be denoted as W′≦W. Equivalently, the channel W is called “an upgrade with respect to W′”, denoted by W≧W′.
III. Code Construction
In this section, embodiments of code construction are described that combine rewriting with error correction.
A. Basic Code Construction with a Nested Structure
1) General: First, consider a single rewrite (namely, one of the t writes). Let s=(s1, s2, . . . , sN)∈{0,1}N and s′=(s′1, s′2, . . . , s′N)∈{0,1}N denote the cell levels right before and after this rewrite, respectively. Let g=(g1, g2, . . . , gn) be a pseudo-random, bit sequence with independent and identically distributed bits that are uniformly distributed. The value of g is known to both the encoder and the decoder, and g is called a dither.
For i=1,2, . . . , N, let νi=si⊕gi⊕gi∈{0,1} and ν′i=s′i⊕gi∈{0,1} be the value of the i-th cell before and after the rewrite, respectively. The WOM channel is built, such as shown in
Here α∈[0,1] and
For the noise channel BSC(p), let FBSC(p)⊂{1,2, . . . , N} be the frozen set of the polar code corresponding to the channel BSC(p). It is known that
In this subsection, FBSC(p)⊂FWOM(α,ε). This case is as illustrated in
2) The encoder: Let E: {0,1}N×{0,1}M→{0,1}N be the encoder for this rewrite. Namely, given the current cell state s and the message to write M∈{0,1}M, the encoder is configured to find a new cell state s′=E(s, M) that represents M and is above s (that is, cell levels only increase).
The encoding function in Algorithm 1 follows. Here y and u are two vectors of length
.
3) The decoder: The decoder D: {0,1}N→{0,1}M M now follows. Let c=(c1, c2, . . . , cN)∈{0,1}N be the noisy cell levels after the message is written. Given c, the decoder should recover the message as D(c)=M.
The decoder may be configured to operate the same way as a polar error correcting code. The decoding operation of one embodiment is presented as Algorithm 2.
)−1)F
)−1
4) Nested code for t writes: In the above, the operations of the encoder and the decoder for one rewrite have been described. The operations can be applied to a t-write error correcting WOM code as follows. For j=1,2, . . . , t, for the j-th write, replace α, ε, s, s′, v, v′, M, M, E, D, c, {circumflex over (M)}, {circumflex over (v)}, by αj−1, εj, sj, s′j, vj, v′j, Mj, Mj, Ej, Dj, cj, {circumflex over (M)}j, {circumflex over (v)}j, respectively, and apply the above encoder and decoder.
Note that when N→∞, the values of α1, α2, . . . , αt−1 can be computed using ε1, ε2, . . . , εt−1; for BSC(p), αj=αj−1(1−εj)(1−p)+(1−αj−1(1−εj))p. Improving the code may include choosing other values for ε1,ε2, . . . , εt that maximize or otherwise increases the sum-rate.
B. Extended Code Construction
Code for the case FBSC(p)⊂FWOM(α,ε) has so far been discussed. For relatively small p and typical values of (α0, ε1), (α1, ε2, . . . , (αt−1, εt) , the above condition holds. Now consider the general case where FBSC(p) is not necessarily a subset of FWOM(α,ε).
The encoder in Algorithm 1 may be revised as follows. The bits are then stored in uF
Therefore,
The decoder in Algorithm 2 may be revised as follows. First recover the bits in uF
IV. Code Analysis
In this section, the correctness of the above code construction is shown, and its performance is analyzed.
A. Correctness of the Code
The correctness of the code is addressed first. The encoder in Algorithm 1 works similarly to the WOM code encoder in D. Burshtein and A. Strugatski, “Polar write once memory codes.” In Proc. IEEE International Symposium on Information Theory, July 2012, pp. 1972-1976, with an exception that the bits in FWOM(α,ε) are not all occupied by the message M, instead, the bits in its subset FWOM(α,ε)∩FBSC(p) are set to be constant values: all 0s. Therefore, the encoder successfully rewrites data in the same way as the code in D. Burshtein and A. Strugatski, “Polar write once memory codes,” in Proc. IEEE International Symposium on Information Theory, July 2012, pp. 1972-1976. Next, the decoder in Algorithm 2 recovers the cell values from noise in the same way as the standard polar ECC. Then, the stored message M is extracted from it.
One item to note is that although the physical noise may act on the cell levels s=(s1, s8, . . . , sN), the error correcting code used in the construction is actually for the cell values v=(ν1, ν2, . . . , νn)=(s1⊕g1, s2⊕g2, . . . , sN⊕gN). However, the pseudo-random dither g has independent and uniformly distributed elements, and so when the noise channel for s is BSC(p), the corresponding noise channel for v is also BSC(p).
B. The Size of FWOM(α,ε)∩FBSC(p)
It may be seen that if FBSC(p)⊂FWOM(α,ε), the code has a nested structure. In general, it is observed how large the intersection FWOM(α,ε)∩FBSC(p) can be. For convenience of presentation, consider one rewrite as in Section III-A, where the parameters are α and ε (instead of αj−1, εj).
Proof: (1) In
(2) In
(3) In
(4) Since p≦αε, clearly BSC(αε)≦BSC(p). Therefore FBSC(p)⊂FBSC(αε).
The meaning of Lemma 2 is illustrated in
Proof: |FWOM(α,ε)∩FBSC(p)|=FWOM(α,ε)|+|FBSC(p)|−|FWOM(α,ε)∪FBSC(p)|≧|FWOM(α,ε)|+FBSC(p)|−FBSC(α, ε)| (by Lemma 2).
C. Lower Bound to Sum-Rate
The sum-rate of the general code construction is analyzed as N→∞. Let
Therefore, the sum-rate is
Lemma 5. Let 0<p≦αj−1εj. Then xj≧γj.
Proof: By Lemma 3,
Theorem 6 Let 0<p≦αj−1εj for j=1,2, . . . , t. If Σj−1t αj−1H(εj)≧1−H(p)+H (p)t, then the sum-rate Rsum is lower bounded by
If Σj=1t αj−1H(εj)<1−H(p)+H(p)t, and H(p)≦αj−1H(εj) for j=1,2, . . . , t, then Rsum is lower bounded by
(Σj=1tαj−1H(εj))−H(p)t.
Proof: If Σj=1t αj−1H(εj)≧1−H(p)+H(p)t, the sum-rate is minimized when xj (j=1,2, . . . , t) takes the minimum value, and xj≧γj. Otherwise, the sum-rate is minimized when xj takes the maximum value 1.
Some numerical results of the lower bound are shown to sum-rate Rsum in
V. Extensions
More general noise models are now described. For simplicity, an erasure channel is described. But the principles or other features can be extended to other noise models. Let the noise be a binary erasure channel (BEC) with erasure probability p, denoted by BEC(p). After a rewrite, noise may appear in some cell levels (both level 0 and level 1) and may change the cell levels to erasures. An erasure represents a noisy cell level between 0 and 1. Erasures may be handled as follows: before a rewrite, all the erased cell levels are increased to 1, and then rewriting is performed as before.
Note that although the noise for cell levels is BEC(p), when rewriting happens, the equivalent noise channel for the cell value v=s⊕g is a
The code can also be extended to multi-level cells (MLC), by using q-ary polar codes.
VI. Additional Information
In this section, example achievable rates of the error correcting WOM code are discussed, using polar codes of finite lengths. In the following, assume the noise channel is BSC(p), and search is performed for parameters ε1,ε2, . . . , εt that achieve high sum-rate for rewriting. When the code can have a nested structure is also discussed. This simplifies the code construction.
A. Finding BSCs Satisfying FBSC(p)⊂FWOM(αε)
The first consideration is when BSC(p) satisfies the condition FBSC(p)⊂FWOM(α,ε), which leads to a particular nested code structure. Let N=8192. Let the polar codes be constructed using the method(s) described herein. To obtain the frozen sets, let |FWOM(αε)|=N(αH(ε)−ΔR), where ΔR=0.025 is a rate loss considered for the polar code of the WOM channel, and let FBSC(p) be chosen with the target block error rate 10−5.
Example results are shown in
B. Achievable Sum-Rates for Nested Code
Achievable sum-rates of codes with a nested structure may be found. Accordingly, the condition FBSC(p)⊂FWOM(α
Results for t-write error-correcting WOM codes—for t=2,3,4,5—are shown in
C. Achievable Sum-Rate for General Code
The achievable sum-rates of the general code, when FBSC(p) is not necessarily a subset of FWOM (α
The lower bound to sum-rate Rsum is discussed with reference to
The memory controller 1604 operates under control of a microcontroller 1610, which manages communications with the memory 1602 via a write device, such as memory interface 1612 and manages communications with the host device via a host interface 1614. Thus, the memory controller supervises data transfers from the host 1606 to the memory 1602 and from the memory 1602 to the host 1606. The memory controller 1604 also includes a data buffer 1616 in which data values may be temporarily stored for transmission over the data channel controller 1617 between the memory 1602 and the host 1606. They memory controller also includes an Error Correcting code (ECC) block 1618 in which data for the ECC is maintained. For example, the ECC block 1618 may comprise data and program code to perform error correction operations for a an embodiment of a encoding and decoding scheme as discussed herein. Some embodiments of such error correction operations are described, for example, in U.S. Patent Application entitled “Error Correcting Codes for Rank Modulation” by Anxiao Jiang et al. filed Nov. 20, 2008, and incorporated herein by reference. The ECC block 1618 may contain parameters for the error correction code to be used for the memory 1602, such as programmed operations for translating between received symbols and error-corrected symbols, or the ECC block may contain lookup tables for codewords or other data, or the like. The memory controller 1604 of one embodiment may perform the operations described above for decoding data and for encoding data.
The operations described above for operating a data storage device, for reading data from a device, for programming a data storage device, and encoding and decoding, can be carried out by the operations discussed above which can be performed by the microcontroller 1610 and associated components of the data storage device 1608.
The processing components such as the controller 1604 and microcontroller 1610 to perform the various operations described herein may be implemented in some embodiments in the form of control logic in software (or other computer-readable instruction stored on a non-transitory computer-readable storage medium and executable by one or more processors) or hardware or a combination of both, and may comprise the processor(s) that execute software program instructions from program memory, or as firmware, or the like. The host device 1606 may comprise a computer apparatus. A computer apparatus also may carry out the operations depicted in the various figures herein or otherwise described above.
Depending on the desired configuration, processor 910 can be of any type including but not limited to a microprocessor (μP), a microcontroller (μC), a digital signal processor (DSP), or any combination thereof. Processor 910 can include one more levels of caching, such as a level one cache 911 and a level two cache 912, a processor core 913, and registers 914. The processor core 913 can include an arithmetic logic unit (ALU), a floating point unit (FPU), a digital signal processing core (DSP Core), or any combination thereof. A memory controller 915 can also be used with the processor 910, or in some implementations the memory controller 915 can be an internal part of the processor 910.
Depending on the desired configuration, the system memory 920 can be of any type including but not limited to volatile memory (such as RAM), non-volatile memory (such as ROM, flash memory, etc.) or any combination thereof. System memory 920 typically includes an operating system 921, one or more applications 922, and program data 924. Applications 922 include encoding and decoding algorithms 923 that is arranged to encode and decode, for example, program data 924 as discussed below. In some embodiments, applications 922 can be arranged to operate with program data 924 on an operating system 921. This described basic configuration is illustrated in
Computing device 900 can have additional features or functionality, and additional interfaces to facilitate communications between the basic configuration 901 and any required devices and interfaces. For example, a bus/interface controller 940 can be used to facilitate communications in accordance with the present disclosure between the basic configuration 901 and one or more data storage devices 950 via a storage interface bus 941. The data storage devices 950 can be removable storage devices 951, non-removable storage devices 952, or a combination thereof. Examples of removable storage and non-removable storage devices include magnetic disk devices such as flexible disk drives and hard-disk drives (HHDs), optical disk drives such as compact disk (CD) drives or digital versatile disk (DVD) drives, solid state drives (SSDs), and tape drives to name a few. Example computer storage media can include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data.
System memory 920, removable storage 951 and non-removable storage 952 are all examples of computer storage media. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVDs) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computing device 900. Any such computer storage media can be part of device 900.
Computing device 900 can also include an interface bus 942 for facilitating communication from various interface devices (e.g., output interfaces, peripheral interfaces, and communication interfaces) to the basic configuration 901 via the bus/interface controller 940. Example output devices 960 include a graphics processing unit 961 and an audio processing unit 962, which can be configured to communicate to various external devices such as a display or speakers via one or more A/V ports 963. Example peripheral interfaces 970 include a serial Interface controller 971 or a parallel interlace controller 972, which can be configured to communicate with external devices such as input devices (e.g., keyboard, mouse, pen, voice input device, touch input device, etc.) or other peripheral devices (e.g. printer, scanner, etc.) via one or more I/O ports 973. An example communication device 980 includes a network controller 981, which can be arranged to facilitate communications with one or more other computing devices 990 over a network communication via one or more communication ports 982. The communication connection is one example of a communication media. Communication media may typically be embodied by computer readable instructions, data structures, program modules,, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and includes any information delivery media. A “modulated data signal” can be a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media can include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, radio frequency (RF), infrared (IR) and other wireless media. The term computer readable media as used herein can include both storage media and communication media.
Computing device 900 can be implemented as a portion of a small-form factor portable (or mobile) electronic device such as a cell phone, a personal data assistant (PDA), a personal media player device, a wireless web-watch device, a personal headset device, an application specific device, or a hybrid device that include any of the above functions. Computing device 900 can also be implemented as a personal computer including both laptop computer and non-laptop computer configurations.
The operations of encoding and decoding data according to the error correction encoding schemes discussed above can be illustrated ad in
The information values 1306 comprise a physical representation of the data values and codewords. For example, the information values 1306 may represent charge levels of memory cells, such that multiple cells are configured to operate as virtual cell in which charge levels of the cells determine a representation of the data values with error correction. Data values are received and encoded to represent the data values with error correction and charge levels of cells are adjusted accordingly, and codewords are determined according to cell charge levels, from which a corresponding data value is determined. Alternatively, the information values 1306 may represent features of a transmitted signal, such as signal frequency, magnitude, or duration, such that the cells or bins are defined by the signal features and represent a data value having error correction. For example, frequency changes over time can determine a data value having error correction. Other schemes for physical representation of the cells are possible in view of the description herein.
For information values 1306 in the case of cell charge levels, the source/destination 1310 comprises memory cells in which n memory cells provide n cell values whose charge levels define data values having error correction. For storing a codeword, the memory cells receive an encoded codeword and comprise a destination, and for reading a code word, the memory cells provide a codeword for decoding and comprise a source. In the case of data transmission, the source/destination 1310 may comprise a write device, such as a transmitter/receiver that processes a signal with signal features such as frequency, magnitude, or duration that define cells or bins such that the signal features determine a data value having error correction. For example, signal components comprising signal frequency, magnitude, or duration may be controlled and modulated by the transmitter such that data values having error correction are represented. When the source/destination 1310 receives a codeword from the controller 1304, the source/destination comprises a transmitter of the device 1302 for sending an encoded signal. When the source/destination provides a codeword to the controller 1304 from a received signal, the source/destination comprises a receiver of the device for receiving an encoded signal. Signal components of the transmitted signal can be suitable modulated to define data values having error correction, in view of the description herein.
The various aspects discussed herein can be implemented in the form of control logic in software or hardware or a combination of both. The control logic may be stored in an information storage medium as a plurality of instructions adapted to direct an information-processing device to perform the operations discussed herein. Based on the disclosure and teachings provided herein, other ways and/or methods to implement the discussed features are possible.
The methods and operations described herein can be implemented in a variety of systems for encoding and decoding data for transmission and storage. For example, codewords may be received from a source over an information channel according to a memory writing scheme and are decoded into their corresponding data values and provided to a destination, such as a memory or a processor, and data values for storage or transmission are received from a source over an information channel and are encoded according to the memory writing scheme.
The coding scheme can be further extended to multi-level cells (MLC). For example, using aspects of the coding scheme described above, multilevel cell memories can be written level by level. Alternatively or additionally, aspects of the coding scheme described above may be used for writing multilevel cell memories using a multilevel polar code.
VIII. Concluding Remarks
This application presents a code construction for error-correcting WOM codes. The embodiment(s) described herein supports any number of rewrites and can correct a substantial number of errors. In some embodiments, construction may be based on polar coding to achieve improved performance for both rewriting and error correction.
The present disclosure is not to be limited in terms of the particular embodiments described in this application, which are intended as illustrations of various aspects. Many modifications and variations can be made without departing from its spirit and scope, as will be apparent to those skilled in the art. Functionally equivalent methods and apparatuses within the scope of the disclosure, in addition to those enumerated herein, will be apparent to those skilled in the art from the foregoing descriptions. Such modifications and variations are intended to fall within the scope of the appended claims. The present disclosure is to be limited only by the terms of the appended claims, along with the full scope of equivalents to which such claims are entitled. It is to be understood that this disclosure is not limited to particular methods, systems, or devices which can, of course, vary. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting.
With respect to the use of substantially any plural and/or singular terms herein, those having skill in the art can translate from the plural to the singular and/or from the singular to the plural as is appropriate to the context and/or application. The various singular/plural permutations may be expressly set forth herein for sake of clarity.
It will be understood by those within the art that, in general, terms used herein, and especially in the appended claims (e.g., bodies of the appended claims) are generally intended as “open” terms (e.g., the term “including” should be interpreted as “including but not limited to,” the term “having” should be interpreted as “having at least,” the term “includes” should be interpreted as “includes but is not limited to,” etc.). It will be further understood by those within the art that if a specific number of an introduced claim recitation is intended, such an intent will be explicitly recited in the claim, and in the absence of such recitation no such intent is present. For example, as an aid to understanding, the following appended claims may contain usage of the introductory phrases, “at least one” and “one or more” to introduce claim recitations. However, the use of such phrases should not be construed to imply that the introduction of a claim recitation by the indefinite articles “a” or “an” limits any particular claim containing such introduced claim recitation to embodiments containing only one such recitation, even when the same claim includes the introductory phrases “one or more” or “at least one” and indefinite articles such as “a” or “an” (e.g., “a” and/or “an” should be interpreted to mean “at least one” or “one or more”); the same holds true for the use of definite articles used to introduce claim recitations. In addition, even if a specific number of an introduced claim recitation is explicitly recited, those skilled in the art will recognize that such recitation should be interpreted to mean at least the recited number (e.g., the bare recitation of “two recitations,” without other modifiers, means at least two recitations, or two or more recitations). Furthermore, in those instances where a convention analogous to “at least one of A, B, and C, etc.” is used, in general such a construction is intended in the sense one having skill in the art would understand the convention (e.g., “a system having at least one of A, B, and C” would include but not be limited to systems that have A along, B alone, C alone, A and B together, A and C together, B and C together, and/or A, B, and C together, etc.). In those instances where a convention analogous to “at least one of A, B, or C, etc.” is used, in general such a construction is intended in the sense one having skill in the art would understand the convention (e.g., “a system having at least one of A, B, or C” would include but not be limited to systems that have A along, B alone, C alone, A and B together, A and C together, B and C together, and/or A, B, and C together, etc.). It will be further understood by those within the art that virtually any disjunctive word and/or phrase presenting two or more alternative terms, whether in the description, claims, or drawings, should be understood to contemplate the possibilities of including one of the terms, either of the terms, or both terms. For example, the phrase “A and B” will be understood to include the possibilities of “A” or “B” or “A and B.”
In addition, where features or aspects of the disclosure are described in terms of Markush groups, those skilled in the art will recognize that the disclosure is also thereby described in terms of any individual member or subgroup of members of the Markush group.
As will be understood by one skilled in the art, for any and all purposes, such as in terms of providing a written description, all ranges disclosed herein also encompass any and all possible subranges and combinations of subranges thereof. Any listed range can be easily recognized as sufficiently describing and enabling the same range being broken down into at least equal halves, thirds, quarters, fifths, tenths, etc. As a non-limiting example, each range discussed herein can be readily broken down into a lower third, middle third and upper third, etc. As will also be understood by one skilled in the art all language such as “up to,” “at least,” “greater than,” “less than,” and the like include the number recited and refer to oranges which can be subsequently broken down into subranges as discussed above. Finally, as will be understood by one skilled in the art, a range includes each individual member. Thus, for example, a group having 1-3 cells refers to groups having 1, 2, or 3 cells. Similarly, a group having 1-5 cells refers to groups having 1, 2, 3, 4, and 5 cells, and so forth.
While various aspects and embodiments have been disclosed herein, other aspects and embodiments will be apparent to those skilled in the art. The various aspects and embodiments disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims.
This application claims the benefit of U.S. Provisional Patent Application No. 61/756,324, titled “JOINT REWRITING AND ERROR CORRECTION IN WRITE-ONCE MEMORIES,” filed Jan. 24, 2013, which is hereby incorporated by reference in its entirety and for all purposes.
The subject matter described herein was made at least in part with government support under CIF1218005 awarded by the National Science Foundation. The United States government may have certain rights in the subject matter disclosed herein.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US13/49480 | 7/5/2013 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
61756324 | Jan 2013 | US |