Embodiments of the present disclosure are directed to methods of repairing corrupted digital data.
In coding theory, Reed-Solomon (RS) codes form a class of error-correcting codes (ECCs) that are constructed using finite fields. RS codes can be defined as follows. Let q be a prime power, and let q be the finite field of q elements. For a subset I ⊆
q, an enumeration I:={x1, . . . , xn} of I, with n:=|l|, and a positive integer k≤n, the RS code with evaluation set l and dimension k is defined as the set of length-n vectors in
qn obtained by evaluating all polynomials f(X) ∈
q[X] of degree up to k−1 on the points of I. Exlplicitly, the RS code with evaluation set I and dimension k is defined as
{(f(x1), . . . , f(xn)|f ∈ q[x], deg(f)≤k−1}.
RS codes allow a precise control over the number of symbol erasures correctable by the code during code design. In particular, it is possible to design RS codes that can correct multiple symbol erasures.
A generalized RS (GRS) code is obtained by coordinate-wise multiplication of all codewords of an RS code of length n by the entries of some pre-defined vector (a1, . . . , an) with non-zero entries in the same field over which the RS code is defined. Erasure decoding of a GRS code can be reduced to that of an RS code by the following steps: (1) Each surviving node divides its content by the corresponding ai; (2) The system performs RS erasure decoding, possibly with low-bandwidth in case of a single erasure; and (3) The reconstructed RS symbols are multiplied by the corresponding ai in order to be converted to the required reconstructed US symbols.
An exemplary application of RS codes is encoding stored data. For example, a host may write incoming messages to a storage array of unreliable “nodes”. According to an embodiment, a node includes memory and some processing capability. In a flash array, a node may include an individual SSD with or without an “augmented” controller for performing, e.g., the encoding calculations on top of the usual tasks performed by an individual SSD controller. Alternatively, in cloud storage, a node may be, e.g., a computer with one or more drives.
The incoming messages are the data that users would like to store, and are free to take any value. An actual stored message, such as a movie, can be of various sizes, but from a coding point of view, what is important are the encoded chunks, sometimes called “stripes”. A stripe includes information symbols and parity symbols.
To achieve reliable storage, the host encodes the data before storing it. Then, if one or more storage nodes fails, the host has to restore (decode) the lost data. Note that whether or not the user wishes to read his/her stored data, the host may correct erasures as they are discovered, by, e.g., a scanning mechanism that constantly searches for failed nodes. In a typical case, the encoding process introduces enough redundancy to overcome more than a single failure. However, the most common case of failure is that of a single failure (erasure), for which embodiments of disclosure provide a low-bandwidth decoding algorithm as described below.
An approximate conceptual description of an exemplary system operation is as follows. Note that to keep this example concrete, it will consider a cloud storage setup, but there are also other potential scenarios, e.g., that of a flash array, i.e., a “super disk” constructed from several individual solid-state, drives, SSDs, typically placed on a rack.
According: to an embodiment of the disclosure, there is provided a method for repairing a single erasure in an RS code over a finite field F2mr with an evaluation set V={u0=0, u1, . . . , un−1} that is a subspace of size n=2d, for d<mr and divisible by m, wherein a normalized repair bandwidth
wherein w(i)j(i,j of r bits; reconstructing, by the controller, content of the failed storage node from the reduced representation of each surviving node's content; and outputting the reconstructed content of the failed storage node.
According to a further embodiment of the disclosure, j′ is calculated by the controller, and transmitted to the surviving :storage nodes.
According to a further embodiment of the disclosure, each surviving storage node calculates j′ based on an index of each surviving storage node and an index of the failed storage node.
According to a further embodiment of the disclosure, j′ is pre-calculated for each storage node.
According to a further embodiment of the disclosure, multiplying, by each surviving storage node i, the content ci of each said node i by a j-th component of vector w(i) comprises calculating, by each node i, i≠j0, wherein j0 is the identifier of the failed node, outi:=Tr(w(i)j(i,j
According to a further embodiment of the disclosure, convening the result from an m×r bit representation into a reduced representation of r bits comprises, when elements of field F2
According to a further embodiment of the disclosure, converting the result from an m×r bit representation into a reduced representation of r bits comprises, when elements of field F2
According to a further embodiment of the disclosure, matrix M ∈ F2r×mr is r rows of a matrix T ∈ F2r×mr with indices supporting an r×r identity matrix in TM1, wherein M1 ∈ F2mr×r is a matrix whose j-th column, j=0, . . . , m−1 is a length-mr binary representation of βj according to a basis {1, α, . . . , αmr−1} of field E over field F2, E:=F2
According to a further embodiment of the disclosure, reconstructing content of the failed storage node from the reduced representation of each surviving node's content includes the steps of calculating t,T=(t1, . . . , tm)T:=A′zT ∈ (F′)m using a matrix A′∈ (F′)m×(n−1), wherein F′:=F2[X]/(P1(X)), wherein p1(X) is a minimal polynomial of β, wherein vector z:=(z1, . . . , zn−1) ∈ (F′)n−1 wherein za;=i for all a=1, . . . , n−1 and for i ≠j0 for which j′=a, wherein n is a total number of storage nodes and
i is the reduced representation of surviving node i's content; wherein matrix A′ is a matrix obtained by replacing each entry aij of A by Maij when aij is regarded as a column vector in F2mr according a basis {1, α, . . . , αmr−1} of F2
are elements of vector v, and (u1, . . . , un−1) are non-zero elements of evaluation set V, and wherein matrix M ∈ F2r×mr is r rows of a matrix T ∈ F2mr×mr with indices supporting an r×r is identity matrix in TM1, wherein M1 ∈ F2mr×r is a matrix whose j-th column, j=0, . . . , m−1, is the length-mr binary representation of βj according to the basis {1, α, . . . , αmr−1} of E over F2, wherein β is a primitive element of field Fi
According to a further embodiment of the disclosure, embedding elements of t ∈ (F′)m in E to obtain a vector w ∈ Fm includes calculating wα:=M1·tα for α=1, . . . , m, wherein each coordinate tα of t is represented as a length-r binary column vector according to the basis 1, β′, . . . , (β′)r−1 of F′ over F2, and outputting w;=(w1, . . . , wm), wherein each entry of w is a vector representation of an element of E, according to the basis {1, α, . . . , αmr−1}.
According to another embodiment of the disclosure, there is provided a non-transitory program storage device readable by a computer, tangibly embodying a program of instructions executed by the computer to perform the method steps for repairing a single erasure in an RS code over a finite field F2mr with an evaluation set V={u0,, u1, . . . , un−1} that is a subspace of size n-2d, for d<mr and divisible by m, wherein a normalized repair bandwidth
Exemplary embodiments of the invention as described herein generally provide systems and methods for repairing corrupted Reed-Solomon codes. While embodiments are susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that there is no intent to limit the invention to the particular forms disclosed, but on the contrary, the invention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the invention. Herein, when two or more elements are described as being about the same as each other, such as A≈B, it is to be understood that the elements are identical to each other, indistinguishable from each other, or distinguishable from each other but functionally the same as each other as would be understood by a person having, ordinary skill in the art.
It was recently shown by V. Guruswami and M. Wootters, “Repairing Reed-Solomon. Codes”, arXiv:1509.04764v1, the contents of which are herein incorporated by reference in their entirety, hereinafter Guruswami, that if C is a Reed-Solomon code of length n:=qm over q
, and n−dim(C)≥n/q, then any erased coordinate in C can be repaired by “transmitting” only one
q-element from each one of the non-erased coordinates. Embodiments of the disclosure consider such simple repair schemes:
q-linear repair schemes in which repairing a single erased coordinate in an
q
q-symbol from each non-erased coordinate.
According to embodiments, a necessary and sufficient condition for a code to have a simple repair scheme is provided. A condition according to an embodiment is closely related to the task of finding the dimension of subfield-subcodes. In the case where the code to be repaired is a Reed-Solomon code, a condition according to an embodiment is closely related to the task of calculating the dimension of alterant codes. As a first application of the condition, it is used to re-prove the above result of Guruswami.
According to embodiments, the case of Reed-Solomon codes over 2
2-subspace of
2
2-subspace U of
2
Reed-Solomon code defined on U has a simple repair scheme.
For the special case where d=4, it is shown that for all r≥2, there exists an appropriate 2-subspace U ⊆
2
Reed-Solomon code defined on U has a simple repair scheme. Shortening such a
code, a
shortened Reed-Solomon code is obtained that requires reading a total of only 52 bits for repairing any single erased coordinate. Note that in the best existing repairing scheme for the
Hadoop distributed file system (HDFS) code, repairing a single erased coordinate requires reading a total of 64 or 60 bits, depending on the erased coordinate, from the non-erased coordinates. In addition, the “Hitchhiker-nonXOR” code of Rashmi et al., “A ‘Hitchhiker's’ Guide to Fast and Efficient Data Reconstruction in Erasure coded Data Centers”, SIGCOMM'14, Aug. 17-22, 2014, Chicago, USA, the contents of which are herein incorporated by reference in their entirety, achieves the same saving in repair bandwidth as a code according to embodiments, but only for the 10 information coordinates, and at the cost of coupling pairs of bytes,
In addition to the above case of of Reed-Solomon codes over ♯22
with evaluation set U, there is a simple 2
Throughout this disclosure, q is a prime power and m ∈ +. If F is a field and M ∈ Fr×n, for r, n ∈
+, then ker(M) represents the kernel of the map Fn→Fr defined by x
Mx. If F is a finite field, then for k, n ∈
, an [n, k]F code is an F-linear code of dimension k in Fn.
where τ:xxq is the Frobenius automorphism of
q
1. For x, Y ∈
q
q
q
+, write
x·
Tr
y;=(x·Try0, . . . , x·Tryn−1) ∈qn,
and similarly for column vectors,
Since a polynomial of degree qm−1 cannot have qm roots, the trace is not identically zero, and hence (x, y)Tr(xy)=x·Try is a non-degenerate symmetric
q-bilinear form. As usual in this case of finite dimension, the map φ:x
x·Try) is an isomorphism of
q-vector spaces between
q
Let b1, . . . , bm be a basis of q
q. The elements b′1, . . . , b′m that map under φ to the dual basis of b1, . . . , bm are referred to as the trace-dual basis of b1, . . . , bm. By definition, b1·Trb′j=1(i=j), so that if x ∈
q
q for all i, then for all i, ai=x·Trb′i.
Linear repair schemes are defined in Guruswami, incorporated by reference above, as follows.
The repair bandwidth bi of the above repair scheme for coordinate i of C is defined as bi:=log2(q)·Σj≠i|Lj|. This is the total number of bits that have to be read from the non-erased coordinates to restore the erased coordinate i. The normalized repair bandwidth
If for all j≠i, |Lj|≤1, then the linear repair scheme for coordinate i is simple.
So, in a simple linear repair scheme for coordinate i, at most 1 sub-symbol, i.e., a symbol from the subfield q, has to be read from each non-erased coordinate to restore coordinate i. Hence, the repair bandwidth is not larger than (n−1)log2(q).
According to an embodiment of the disclosure, a necessary and sufficient condition for the existence of a simple q-linear repair scheme is presented. A result according to an embodiment of the disclosure is Theorem 3.2, which specifies a condition in terms of the dimensions of some sub field subcodes.
For a vector v, write diag(v) for the diagonal matrix with v on its main diagonal.
Then the code C has a simple q-linear repair scheme for the first coordinate if and only if the following condition holds:
c=(c0, c1, . . . , cn−1):=(x0, x1, . . . , xk−1)G
for some x0, . . . , xk−1 ∈ q
A general q-linear functional of the form vj ·Tr(−), for some vi ∈
q
v
j·Trcj=(vjg0j)·Trx0+Σi=1k−1(vjgij)·Trxi. (2)
Collecting EQS. (2) for all j ∈ (1, . . . , n−1):
{vj·Trcj}j=1n−1=x0·Tr{vjg0j}j=1n=1+Σi=1k−1xi·Tr{vjgij}j=1n−1. (3)
Because q
qm as an
q-vector space, the map f of Definition 2.3 might as well have a codomain of
qm. So, let A ∈ γqm×(n−1), and consider the vector
A{v
j·Trcj}j=1n−1=x0·Tr(A{vjg0j}j=1n−1)+Σi=1k−1xi·Tr(A{vjgij}j=1n−1),
where the equality follows from EQ. (3) and the q-bilinearity of (-)·Tr(-),which implies that, for all i and all l:
Because the xi are arbitrary, a necessary condition for reconstruction is that
(i) ∀i ∈ {1, . . . , k−1}: A{vjgij}j=1n−1=0, and
(ii) The m elements of the vector A{vjg0j}j=1n−1=A(g0diag(v))T ∈ q
q-linearly independent, that is, these elements form a basis for
q
q.
But (i) and (ii) is clearly also a sufficient condition: given (i), then
A{v
j·Trcj}j=1n−1=x0·Tr(A{vjg0j}j=1 n−1),
and given (ii), the elements of this vector are the coefficients in the description of x0 as a linear combination of the elements of the dual of the basis in (ii). So, there exists a simple repair scheme for the first coordinate of C if there exist v and A such that conditions (i) and (ii) bold.
Note that condition (i) is equivalent to
Recalling that H′:=G′ diag(v), it can be seen that condition (i) is equivalent to H′AT=0.
To, complete the proof, it will be verified that condition (*) ⇔ there exists v and A such that conditions (i) and (ii) hold.
Suppose that condition (*) holds. Let the m rows of A be a basis of the code D. Then by the definition of D, it follows that H′AT=0, so that condition (i) holds. For condition (ii), suppose that there exists a row vector t ∈ qm such that t(A(g0diag(v))T)=0. Then ATtT is in D ∩ ker(g0diag (v)), and hence must be the zero vector. Since the rows of A are independent, this implies that t is the zero vector, as required.
Suppose that there exist v and A such that conditions (i) and (ii) hold. Because of condition (ii), the rows of A must be q-linearly independent. Let D be the
q-span of the rows of A. Then dim(D)=m. In addition, since by condition (i) H′AT=0, it holds that D ⊆ C′. Finally, a vector in D is of the form tA for some row vector t ∈
qm. If (t A)T ∈ ker(g0diag (v)), then
(g0diag(v)AT)tT=0.
Because the elements of the vector g0diag(v)AT q-linearly independent by condition (ii), it can be seen that t is the zero vector, as required.
Writing C′ ⊆ qn−1 for the code with parity-check matrix H′:=G′diag(v) and C″ ⊆
qn−1 for the code with parity-check matrix
it follows that dim(C″)≤dim(C′)−m, which is equivalent to dim(C″)=dim(C′)−m. Note that adding one line to a check matrix over q
q can decrease the dimension by at most m.
Suppose that (**) holds. Then it is possible to choose c1, . . . , cm ∈ qn−1 that complete a basis of C″ to a set of dim(C″)+m linearly independent elements in C′. Let D be the
q-span of c1, . . . , cm). Then by construction D ⇔ C′, and dim(D))=m. Also,
where the last equality follows front the way the basis of D as constructed.
Suppose that Condition (*) holds. Then
C′ ⊇ (C′ ∩ ker(g0diag(v)))⊕D.
But C′∩ ker(g0diag(v)) is just C″, so that C′ ⊇ C″ ⊕D. Since dim(D)=m by assumption, it follows that dim(C″)≤dim(C′)−m.
4. Reed-Solomon Codes of Length qm over q
As a first example for an application of Theorem 3.2, Theorem 1 of Guruswami, incorporated above, will be re-proven. According to embodiments of the disclosure, from this to point on, a Reed-Solomon code of dimension k ∈ over θq
q
q
Now, suppose that k=qm−qm−1, and let α ∈ q
where g0=1, the all ones row vector of length qm−1.
For j ∈ {1, . . . , qm31 1}, let vj:=α−(j−1). Then H′=G′diag(v) is a check matrix of the cyclic code over q of length qm−1 and zeros the elements of Z′:={1, α, . . . , αq
q, while
is a check matrix of the cyclic code over q with zeros the elements of Z″:=Z′ ∪ {α−1} and their conjugates.
By Theorem 3.2, it is sufficient to, show that Z′ does, not contain an element of the orbit of α−1 under the action of Gal(q
q)
α
, and that the orbit of α−1 has m elements. In this regard, recall that for a cyclic code C of length n′ with generator polynomial g, it follows that dim(C)=n′−deg(g). Translating to the language of cyclotomic cosets, it should he shown that, the cyclotomic cosets modulo qm−1 the numbers in {0,1, . . . , qm−qm−1−2} do not include the cyclotomic coset of (−1)≡qm−2, and that this last cyclotomic coset has m elements.
According to embodiments of the disclosure, it will be convenient to represent numbers in {0, . . . , qm−2} as their length-m base-q expansions. So, for i ∈ {0, . . . , qm−2}, write [i]q=(im−1, im=2, . . . , i0) for its base-q expansion, so that i=Σj=0m−1 ijqj.
Now,
[qm−2]q=(q−1, q−1, . . . , q−1, q−2). (4)
Since all m cyclic shifts of this vector are distinct, the cyclotomic coset of qm−2 indeed has m elements.
Also, since qm−qm−11=(qm−1)−qm−1, then
[qm−qm−11]q=(q−2, q−1, . . . , q−1, q−1). (5)
It follows from EQS. (4) and (5) that qm−qm−1−1 is the smallest element in the cyclotomic coset of qm−2. Hence this cyclotomic coset does not contain any number in {0,1, . . . , qm−qm−1−2}.
5. Reed-Solomon Codes on 2-Subspaces of
2
According to embodiments, this section presents simple repair schemes for Reed-Solomon codes defined on an 2-subspace of
q
2-subspace U ⊆
2
Reed-Solomon code defined on U has a simple repair scheme. Moreover, the probability that a randomly selected subspace results in a simple repair scheme is 1−O(1/2r).
In general, to check if Condition (**) holds, there is a need for an exact formula for the dimension of subfield subcodes, as lower bounds are insufficient. For this, some results from H. Stichtenoth, “On the dimension of Subfield Subcodes”, IEEE Trans. Inform. Theory, Vol. 36, No. 1, January 1990, the contents of which are herein incorporated by reference in their entirety, will be used, as follows. In what follows, let σ
operate on vectors coordinate-wise.
It will be convenient to state explicitly a special case of Proposition 5.1.
Before considering Reed-Solomon codes defined on 2-vector subspaces of
q
From this point on, q is always a power of 2. Turn now to Reed-Solomon codes defined on 2-vector subspaces of
q
x+y permutes U , and is therefore an automorphism of the Reed-Solomon code defined on U, for the reason that the variable change X
X+y does not increase the degree of a polynomial.
Let U be an 2-subspace of
q
, l≤n′−1, let C=C(U, l) be the Reed-Solomon code obtained by evaluating polynomials of degree≤l from
q
From this point on, according to embodiments, it will be assumed that the vector v of Theorem 3.2 is defined by vu:=1/u for all u ∈ U\{0}, assuming that coordinates are labeled by elements of U. With this choice of v, the rows of the matrix H′ of Theorem 3.2 can be considered as a basis for the space of polynomial functions U\{0}→q
q
Let fU(X):=Πx∈U (X−x). Then by Theorem 5.4, there are some a0, . . . , ad−1 ∈ q
f
U(X)=X2
and with a0≠0 (since f separable by definition). There is an isomorphism of q
μ:q
q
g+(fU/X)(x
g(x)).
For a polynomial g(X) ∈ q
q
Let D′ ⊆ q
Similarly, let D″ ≣ q
{
Recalling that fU(X)/X=X2
−1
=a
0
−1
2
−2
a
0
−1
a
d−1
2
−2
+ . . . +a
0
−1
a
1,
and
(
This implies the following lemma.
Spa{
{
and
Spa{
{
r:=max{([log2(j1)],s}+1,
so that r>s and ar=0. If as=0, then clearly wt2(j1+j2)=wt2(j1)+wt2(j2) Otherwise, let s′>s be the smallest such that ax′=0. The single 1 from the binary expansion of j2 propagates as a carry, so that for all i ∈ {s,s+1, . . . , s′−1}, the i-th digit of the binary expansion of ji+j2 is 0, the s′-th digit of this binary expansion is 1, and all other digits of this expansion are the same as those of the expansion of j1. All in all, the total, number of 1's is wt2(j1)+1−(s′−s)wt2(j1)+wt2(j2)−(s′−s)<wt2(j1)+wt2(j2), as required
For the general case, suppose that wt2(j2)≤2, and assume the assertion true for all j2 with smaller weight (note that the case j2=0 is obvious). Then, write j2=j′2+j″2 with wt2(j′2)=1 and wt2(j″2)=wt2(j2)−1. Then
as required.
({
2
∈ Spa{
for which induction on r will be used. If r<d the assertion is clear. Suppose, therefore, that r≥d and that there exist b0, . . . , bd−1 ∈ such that
2
=Σn=0d−1 bu
Squaring this equation:
2
=Σu=0d−2 bu2
But since fU(X)=X2
For the induction step, suppose that wt2(j)≤2 and the assertion is true for all j′ with wt2(j′)<wt2(i). Write j=2j
for some α0(1), . . . , αd−1(1), . . . , α0(w) ∈ q
2
+ . . . +2
11)
Two cases can be distinguished as follows,
Case I. If there are some l1, l2 such that il
2
+ . . . +2
=
2
·
Σ
2
. (12)
By Lemma 5.7, wt2(Σr≠l
for some coefficients {βi}, {γk}. This last expression is itself an q
Case 2. All the il are distinct. Since by assumption w=wt2(j)≤d−1 and wt2(2d−1)=d, 2i
where the rightmost implication holds because fU, and hence also
is separable. Hence
For the second assertion, suppose that Σi=1r αiq
(Σi=1r σ−1(αi)
so that Σi=1r σ−1(αi)
Then for C′ of Theorem 3.2, dim(C′)≤2.
1. Informally, this proposition states that “half” of Condition (**) is satisfied for any subspace of dimension d if l≤l0. Note that l0 depends on d, but not on q. For example, for even d and for all q with q2≥2d, the proposition shows that half of Condition (**) is satisfied for codes of length n=2 and dimension up to n−√{square root over (n)} in q
Reed-Solomon codes.
2. Note that for odd d, the codimension is the same as that for d+1. Hence, the parameters of an odd d can be achieved by shortening a code designed for the even dimension d+1. Therefore, apart from the current proposition, only even d will be considered.
3. Note that for a choice according to an embodiment of v, there do exist some subspaces U ⊆ q
q ⊂
q
q in the check matrix cannot decrease the
q-dimension by more than 1. Similarly, the condition does not hold for subspaces of the form z·
q for z ∈
*q
where dim(D′⊆)=l has been used, and recalling that that
dim(D′⊆ ∩ σ(D′⊆))≤2l+3−2d.
In view of EQ. (8), it should be shown that
dim(Spa{
{
In other words, it should be shown that
Spa{
contains 2l+3−2d independent elements that can be written as linear combinations of q
Write
A:=|{i ∈ {0, . . . , l−1}|wt2(i)>wt2(l−1)}|
and
B:=|{i ∈ {l, . . . , 2d−2}|wt2(i)≤wt3(l−1)}|.
It follows from Proposition 5.8 that for each of the l−A values of i in {0, . . . , l−1} with wt2(i)≤wt2(l−1), (
Theorem 5.11 Let d≥4 be even. Suppose that q=2r and that d elements x0, . . . , xd−1 are drawn uniformly and independently from q
code. Suppose that the vector v, is given by vu=1/u for all u ∈ U\{0}. Then the probability that both dim(U)=d and Condition (**) holds is at least
and is positive for r≥d−1+log2(d).
1. The case of r=d/2 is a special case of Theorem 4.1, above.
2. Some examples (see below) seem to suggest that the result should hold for all r≥d/2, and that the probability of drawing a “good” subspace U is much higher than in the theorem.
3. For d=4, the theorem proves the existence of a good subspace for all r≥5, and Theorem 4.1 covers the case r=2. Since the two remaining cases of r=3, 4 are considered in the examples below, it can be seen that for d=4, a good subspace exists for all r≥d/2.
Before proving the theorem, some auxiliary results are needed.
β1{
or, equivalently, to
2
−2 ∉ Spa{
and
({
Since D″⊆D′⊆+Spa{v}, then
D″
⊆+σ(D″⊆)=D′⊆+σ(D′⊆)+Spa{v,σ(v)}.
Hence Condition (**) holds iff for β1, β2 ∈ q
β1v+β2σ(v) ∈ D′⊆+σ(D′⊆) ⇒ β1=β2=0.
Applying the isomorphism μ−1, it can be seen that Condition (**) holds iff for β1, β2 ∈ q
β1{
which, from Lemma 5.5, is equivalent to
β1{
This concludes the proof
2d−1−1=i0<i1< . . . <id‘12d−2
are all the numbers I ∈ {0, . . . , 2d−2} with wt2(I)=d−1, and i0, . . . , id/2−2 are all the numbers i ∈ {0, . . . , l0−1} with wt2(i)=d−1.
Note that each column of Mr includes some of the coefficients corresponding to (
Define also a matrix {tilde over (M)}r ∈ q
(i) q
(ii) to the projection of (q
Consider (i) above. By Proposition 5.8, if q
Similarly, by Proposition 5.8 if the projection of (
To continue, parametrize the problem. Instead of working in q
2[X, T0, . . . , Td−1] for free variables T0, . . . , Td−1. For this, let T:=(T0, . . . , Td−1) be a vector of free variables over
2[X], and let f (X, T) ∈
2[X, T] be defined by
f(X, T):=X2
Re-define (g2[X, T]→
2[X, T]/(f/X). Since
2[T] ⊆
2[X, T]→
2[X, T]/(f/X) is an injection, which follows from degX(uv)=degX(u)+degX(v),
2[T] can be thought of as a subring of
2[X, T]/(f/X). In addition,
2[X, T]/(f/X) is a free
2[T]-module with basis
2[X, T]/(f/X) as an
2[T]-module, since the coefficient of X2
The idea is that for any finite dimensional 2-subspace U of any extension field K of
2 (e.g., K=
q
Next, it will be shown that for r≥d/2, det(r) is not the zero polynomial in
2[T]. To this end, begin with the following lemma.
For all i and r, and for the ij's of Definition 5.14, the αi,j(r)=αi,j(r)(T) of Definition 5.17 satisfy
From Proposition 5.8, only j's with wt2(j)=d−1 can contribute to αi,k(r+1) for k With wt2(k)=d−1. In the sum S1, there is only such j, namely 2d−1−1. In the sum S2, the relevant j's are those of the form j=2d−1−2j′ for j′ ∈ {0,1, . . . , d−2}, from Definition 5.14, Hence, the partial sum
S
3:=(αi,2
includes all contributions to the αi,k(r+1) for k with wt2(k)=d−1.
Now
S
3:=(αi,2
Once again, only those exponents with wt2(·)=d−1 can contribute to αi,k(r+1) for k with wt2(k)=d−1 and should be kept. For j+1 ∈ {1, . . . , d−1}, the binary representation of 2d−2−2j+1=(2d−1)−1−2j+1 has exactly two zeros (the coefficients of 20 and 2j+1), and so wt2(2l+2d−2−2j+1)=d−1 iff l ∈ {0, j+1}. It follows that all contributions to the αi,k(r+1) for k with wt2(k)=d−1 appear in
Recalling that 2d−1−2j=id−1−j and using a variable change j↔d−1−j in the last sum, the above expression reads
2
−2((αi,i
This concludes the proof.
in particular, using 0,1, . . . , to number rows, rows number d/2−1, . . . , d−2 are zero, and
and for r≥1,
r
=A·A
⊙2
·A
⊙4
. . . A
⊙2
·A
1
⊙2
, (21)
a total of r multiplied matrices. Consequently.
r=2A·A⊙2·A⊙4 . . . A⊙2
It can now be proven that det(r) is not the zero polynomial.
N:=B
⊙2
. . . B
⊙2
=B
d/2−2
represents the cyclic shift left by d/2−2 operator. Therefore,
N′:=N·A
1(1,0, . . . , 0)⊙2
is given by
It follows that d/2(1,0, . . . , 0)=2 A(1,0, . . . , 0)N′=Id/2, and
(det(d/2)) (1,0, . . . , 0)=1. (23)
For d/2=2, where d/2=2A·A1⊙2, it can be verified directly that
d/2(1,0, . . . , 0)=Id/2, so that EQ. (23) holds also for this case. This shows that det(
d/2) ≠ 0 ∈
2[T] and establishes the induction basis.
For the induction step, assume that r≥d/2+1 and the assertion is true for r−1. Write 2A′:=2 A/T0 ∈ 2d/2×d, and note that 2A′=(2A′)⊙2. When by EQ. (22),
Now, using 0,1, . . . to index rows:
Cyclically shifting the rows results in the matrix
Write 3A for the matrix in EQ. (25), and let
r
:=
3
A·A
⊙2
·A
⊙2
. . . A
⊙2
·A
1
⊙2
.
It follows from EQ. (24) that det(r) ≠ 0 iff det(
r) ≠ 0. By the
2[T]-linearity of the determinant in the first row, it is shown that
Note that the following fact was used: for rows r0, r′0, r1, r2, . . . and for a matrix. M,
Note also that the first determinant in the above sum equals det(r−1)/T0, and is therefore non-zero by the induction hypothesis. In addition, the first summand (Td/2·det( . . . )) is the only one in which Td/2 may appear in an odd degree, and as the first determinant is non-zero and Td/2 appears in all its monomials in an even degree, the first summand must contain monomials whose Td/2-degree is odd. As such monomials cannot be canceled by monomials in the second determinant, whose Td/2-degree is essentially even, it can be concluded that det(
r) ≠ 0, so that det(
r) ≠ 0. This completes the proof that det(
r) ≠ 0 for all r≥d/2.
As for the assertion regarding the total degree, note first that combining EQ. (22), the fact that all polynomials in A are either 1 or some Tj, and the additivity of the total degree, it can be seen that the total degree of each entry in r is at most 1+2+ . . . +2r−1=2r−1. As the determinant of a d/2×d/2 matrix is a sum of products of d/2 entries, it follows that tot. deg(
r)≥(d/2) (2r−1), as required.
Theorem 5.11 can now be proven.
g(X0, . . . , Xd−1):=(X0+X1) (X0+X2) . . . (Xd−2+Xd−1) . . .(X0+ . . . +Xd−1),
a product of 2d−d−1 linear polynomials. In addition, assuming that x0, . . . , xd−1 are , 2-linearly independent and writing U for the subspace that they span, the coefficients a0, . . . , ad−1 of fU=Πx∈U (X+x)=X2
2[X0, . . . , Xd−1]. Substituting these polynomials from
2[X0, . . . , Xd−1] for the Ti in det(
r) with r=log2(q), results in a polynomial h ∈
2[X0, . . . , Xd−1] of total degree at most
where EQ. (26) follows from Lemma 5.24 and the fact that for a field K, a polynomial u ∈ K[Y0, . . . , Yn
tot. deg(u(v0, . . . , vn
To see the validity of the previous equation, take a monomial Y0i
Write X:=(X0, . . . , Xd−1). According to an embodiment, it should be shown that (gh) (X) has a non-zero in q
, and δ ∈
), and if S ⊆ K is a finite set, then the number of (y0, . . . , ym−1) ∈ Sm for which f (y0, . . . , ym−1)=0 is at most δ·|S|m−1. However, it first needs to be shown that gh ≠ 0. Since g(X) is obviously non-zero, it only needs to be shown that h(X) ≠ 0. If h(X) is the zero polynomial in
2[X], then for any extension field K of
2 and for any d-dimensional
2-subspace U of K , the result of substituting the αi coming from Πx∈U (X−x)=X2
r) is 0 ∈ K. Suppose that |K|=Q where Q≥2d is some power of 2. Noting that distinct subspaces result in distinct vectors (a0, . . . , ad−1), this means that the non-zero polynomial det(
r) has at least
zeroes. However, since r ≠ 0 and tot. deg(det(
r))≤d/2) (q−1), it follows from the Schwartz-Zippel lemma that the number of zeros of det(
r) in Kd is at most Qd−1(d/2)(q−1), which is in O(Qd−1) for fixed d and q. As N1=O(Qd), a contradiction is obtained for large enough Q. This shows, that h(X) ≠ 0.
As gh ∈ 2[X] is a non-zero polynomial and
it follows from the Schwartz-Zippel lemma that the probability that a randomly drawn point (x0, . . . , xd−1) from (q
For fixed d, p>0 if
Note that since L is monotonically decreasing with q while R is monotonically increasing with q, if L(2r)<R(2r) for some r, then the same is true for all r′≥r.
Now, for even d≥4 and r=log2(d2d−1), it follows that
as required.
6. The Case d=4 and Additional Examples
For d=4, it follows that l0=11, and the only integers j≤2d−2=14 with wt2(j)=d−1=3 are j=7, 11, 13, 14, of which only 7 is <l0. Hence, by definition
recalling that a0 ≠ 0 is the free coefficient of fU/X, where (*) follows from Lemma 5.18 by noting that by definition, α2i,j(r)=αi,j(r+1), as (
(
so that with the above notation, α13=0, α11 ≠ 0, and α14 ≠ 0, and it follows that α133+α11α142 ≠ 0. Hence, Condition (**) holds, and the [16,12] Reed-Solomon code defined on U has a simple repair scheme.
C can be shortened to obtain a [14, shortened Reed-Solomon code. For this shortened code, repairing any single erased coordinate requires reading ,a total of only 13·4=52 bits from the remaining coordinates. Note that in the best existing repairing scheme for the [14,
Facebook FUNS code, repairing a single erased coordinate requires reading a total of 64 or 60 bits, depending on the erased coordinate, from the non-erased coordinates.
While the proofs given in the previous sections consider only the case of extensions 2
2
2
2
2-subspaces U ⊆
2
2
Recall that in such a case, if a single coordinate is erased, then each of the non-erased cc ordinates has to transmit only 1/m of its content to reconstruct the erased coordinate.
defined on U. This gives a [256,192]2
defined on U. This gives a [64, RS code with a simple
2
defined on U. This gives a [64, RS code with a simple
2
A repair algorithm according n an embodiment of the disclosure includes two parts: an offline pre-processing part, and an online part. In the offline part, the various blocks used by the online part are prepared, In the online part, data is encoded using the RS code constructed in the offline part, and erasures are decoded if necessary.
If there is a single erasure, then, using the blocks constructed in the of part, the single erased coordinate can, be corrected with a low repair bandwidth, in that only a small, part of the content of the non-erased coordinates is transmitted to reconstruct the erased coordinate, where “small part” will be defined below.
If there are two or more erasures, then standard erasure-decoding techniques for RS codes are used, and there is no gain in repair bandwidth. However, since single erasures are much more common in practice than multiple erasures, there is it in supporting low-bandwidth repairs for the case of a single erasure.
In what follows, for a field E to be defined below, 1:=(1, . . . , 1) ∈ En−1 is the vector of n−1 ones, 0:=(0, . . . , 0)∈ En−1 is the vector of n−1 zeros, and (·)T stands for vector and matrix transposition. In addition, for a field L, a subset I ⊆ L, and a positive integer k≤|l|, the RS code with evaluation set I and dimension k is defined as the set of length-|l| vectors in L|l| obtained by evaluating all polynomials f ∈ L[X] of degree up to k−1 on the points of I. Explicitly, letting I={x1, . . . , xn}, with n:=|I|, then the above RS code can be defined as
{(f(x1), . . . , f(xn))|f ∈ L[X], deg(f)≤k−1{.
In addition, a stripe according to an embodiment can have up to
symbols of information, and
symbols of parity, where d, m are as defined above. For example, when d=4 and m=2, there can be up to 12 symbols of information, and 4 symbols of parity.
The inputs of an offline part of an algorithm according to an embodiment are as follows,
Note that lengths 2d−1<n0<2d can be obtained by shortening For simplicity, embodiments will only consider the case of n0=n=2d.
The outputs of an offline part of an algorithm according to an embodiment are as follows.
With the inputs and outputs defined as above, an offline part of an, algorithm according to an embodiment is as follows, with reference to the steps of the flowchart of
Step 111: Let BF/E={x1, . . . , xm} ⊂ E be the dual basis of {y1, . . . , ym} with respect to the trace map E→F, and output BF/E. The trace tap Tr:E→F, defined above in Def. 2.1, can equivalently be defined by uΣi=0m−1 u|f|
Step 115: Obtain a matrix of the form (Ik|G1) from matrix G ∈ Ek×n of Step 105, where G1 ∈ Ek×(n−k). According to an embodiment, such a matrix can be obtained by e.g., Gaussian elimination. Output G1.
According to embodiments, during online operation, incoming messages are encoded and stored in individual unreliable storage nodes. When it is found that one or more nodes have failed and its data is erased, erasure-correction algorithms are used to restore the erased data. In the common case of a single node failure, algorithms according to embodiments of the disclosure can be used. Otherwise, if more than one node fails, standard erasure-correction methods for RS codes can be used.
According to embodiments, a systematic encoding algorithm is presented, because it is the one most likely to be used in practice. Note however, that single-erasure-decoding algorithms according to other embodiments are not limited to systematic encoding, and any encoding for the code CRS with generator matrix (Ik|G1) may be used. For example the generator matrix G from Step 105 of
i. Systematic Encoding
According to an embodiments, it is assumed that the entries yj of y are stored in n different storage nodes, labeled 0, . . . , n−1. If exactly one node j0 fails, a following algorithm according to an embodiment can be used for restoring yj
The inputs of an online part of a single erasure decoding algorithm according to an embodiment are the failed node index, j0. It should be noted that U can learn the failed node index j0, by, e.g., periodically scanning nodes and checking their status.
The outputs of an online part of a single-erasure decoding algorithm according to an embodiment are a restored value for yj
Each storage node can store a different vector of length n−1, call it w(i) for node i. The j-th entry of w(i) is used by node i in the process of reconstructing the j0-th failed node in a predefined enumeration of all nodes but i. In such a setup, the controller can send surviving node i the index j0 of the failed node, and node i then multiplies its content by w(i)j for
and then takes the trace of the result, converts the result to a compact representation and transmits the compact representation to the controller. The vector w(i) ∈ (F2
With the inputs and outputs defined as above, an online part of an algorithm according to an embodiment is as follows, with reference to the steps of the flowchart of
Note that it is only possible to re-store the value in node j0 if the failed node had a temporary failure. According to embodiments, if the node is considered lost, then the reconstructed data can be stored on a different node. According to other embodiments, the reconstructed data is not re-stored at all, but is rather re-calculated (erasure decoding) and sent to the user each time the user wishes to access the data stored on the failed node. In such a case, each time the user wishes to access the data, much more data than the required lost block has to be read from the non-failed nodes for reconstruction.
Alternatively, according to other embodiments of the disclosure, the overall action of steps 22 and 23 can be represented by an r×mr binary matrix Vj ∈ F2r×mr that can be pre-computed, depending on the basis of F2mr over F2 used for representing elements. In detail, one can pre-compute “overall” matrices V1, . . . , Vn−1 ∈ F2r×mr, and instead of the current steps 22 and 23, use a single combined step in which node j multiplies its content, regarded as a length-mr binary vector according to the above basis, by the binary matrix Vj′ to obtain the required length-r binary vector i.
In other embodiments of the disclosure instead of storing the vector v and using its j′-th entry, vj′, for the j′ defined by Step 21, above, it is possible to use a different method, in which node i stores a different vector w(i), depending on i, and uses its j=j(i,j0)-th entry, for
In an alternative embodiment, all nodes store the vector v. When node j0 fails, the controller tells surviving node i, for all i ≠ j0: “node j0 failed, calculate j′ by yourself”. Then, each surviving node calculates j′, depending on its own index i and the failed node index j0, by using the above formula, and, as before, uses w(i)j=vj′. In another alternative embodiment, instead of letting node i calculate j′, the right permutation of v is pre-calculated and stored: its first entry is the entry that has to be used when node 0 fails, . . . its (i−1)-th entry is the entry that has to be used when node i−1 fails, its i-th entry is the entry that has to be used when node i+1 fails, etc. Explicitly, define, for all j0 ∈ {0, . . . , i−1, i+1, . . . , n}, w(i)j(i,j
There are n=2d different vectors w(i), one for each node. These vectors are pre-computed in the offline stage, and stored in the respective n nodes, i.e., w(i) is stored on node i for all i. In some cases, this may be a favorable method, because it avoids the online permutation calculation by both the controller and the surviving nodes.
It is to be understood, however, that the above scenario, in which an algorithm according to an embodiment is performed by a controller U and of the n−1 surviving nodes, is exemplary and non-limiting, and the situation can differ in actual practice. It is possible, for example, that one node coordinates the encoding and reconstruction processes but does not actually performs them, or performs only part of the encoding/reconstruction.
For example, in Hadoop File System (HDFS), developed by the Apache Software Foundation, which is a distributed file system there is a module called HDFS RAID (RAID stands for Redundant Array of independent Disks, and is typically, related to erasure coding) that works roughly as follows. In HDFS RAID, there is a node called RaidNode that coordinates the encoding and decoding related to the erasure code, but need not actually perform them.
For encoding the RaidNode periodically scans all paths that should be erasure encoded, and selects large enough files that have not been recently modified. Once it selects a source file, it iterates over all its stipes and creates the appropriate number of parity blocks for each stripe. The actual calculation of the parity blocks may be performed locally at the RaidNode (“LocalRaidNode”), or distributed to other machines (“DistributedRaidNode”). The LocalRaidNode configuration is similar to what is described in the present disclosure.
For one case of decoding, BlockFixer, a program that runs at the RaidNode, periodically inspects the health of all relevant paths. When a file with missing or corrupt blocks is encountered, these blocks are decoded after a (parallel) read of a sufficient amount of data from the surviving nodes. Once again, the actual computation of the missing blocks may be performed either locally at the RaidNode (“LocalBlockFixer”), or distributed to one or more other machines (“DistBlockFixed”). Embodiments of the disclosure can be incorporated into the distributed case.
Note that the above decoding is a result of a periodic scan. Recall that there is also a case where the user (“client”) itself finds out that there are erasures when trying to read a file. In HDFS RAID, the client implements the DRFS (Distributed RAID File System) client software, which can locally reconstruct such erased blocks, and acts as “U”. In this case, the client does not re-store the reconstructed blocks: these reconstructed blocks are sent to higher layers and discarded. In this case, the client has to re-calculate the missing blocks every time the file is read.
It is to be understood that embodiments of the present disclosure can be implemented in various forms of hardware, software, firmware, special purpose processes, or a combination thereof. In one embodiment, the present disclosure can be implemented in hardware as an application-specific integrated circuit (ASIC), or as a field programmable gate array (FPGA). In another embodiment, the present disclosure can be implemented in software as an application program tangible embodied on a computer readable program storage device. The application program can be uploaded to, and executed by, a machine comprising any suitable architecture.
The computer system 41 also includes an operating system and micro instruction code. The various processes and functions described herein can either be part of the micro instruction code or part of the application program (or combination thereof) which is executed via the operating system. In addition, various other peripheral devices can be connected to the computer platform such as an additional data storage device and a printing device.
It is to be further understood that, because some of the constituent system components and method steps depicted in the accompanying figures can be implemented in software, the actual connections between the systems components (or the process steps) may differ depending upon the manner in which the present invention is programmed. Given the teachings of the present invention provided herein, one of ordinary skill in the related art will be able to contemplate these and similar implementations or configurations of the present invention.
While the present invention has been described in detail with reference to exemplary embodiments, those skilled in the art will appreciate that various modifications and substitutions can be made thereto without departing from the spirit and scope of the invention as set forth in the appended claims.
it should be shown that
2d−3−l0≥A+B. (28)
Suppose first that d is even. Then the binary expansion of l0−1=2d−2d/2−2 is, with the MSB on the left:
In particular, wt2(l0−1)=d−2.
Note that the usual ordering of the integers corresponds to the lexicographic ordering of the binary expansions. Hence an expansion represents a number k≤l0−1 with wt2(k)=d−1, which is the only possibility of a weight>wt2(l0−1) for a number≤l0−1, iff the single 0 in its binary expansion is to the left of the 0 in position d/2 in the expansion of l0−1. Hence, there are exactly d/2−1 such numbers k, so that. A=d/2−1.
Similarly, a number k is strictly larger than l0−1 iff its binary expansion has 1 in the first coordinate from the left in which it differs from the binary expansion of l0−1. Since only numbers k with wt2(k)≤wt2(l0−1) are of interest, k=l0 is excluded, and the remaining numbers are exactly the numbers k with binary expansion of the form
where the sub-vector with question marks may be any binary vector of length d/2 and weight≤d−2−d/2=d/2−2. The total number of such vectors is B=2d/2−d/2−1. There are now explicit expressions for both A and B, and it is straightforward to verify that EQ. (28) holds with equality.
The case where d is odd is similar: now the binary expansion of l0−1 is
The numbers A and B are determined as above: there are A=(d−1)/2−1 ways to place a single zero to the left of the zero at index (d+1)/2 of the binary expansion of l0−1, and there are B=2(d+1)/2−(d+1)/2−1 binary vectors of length (d+1)/2 with weight≤d−2−(d−1)/2=(d+1)/2−2. So now the left-hand side of EQ. (28) equals
2d−3−(2d−2(d+1)/2−1)=2(d+1)/2−2,
while the right-hand side equals A+B=2(d+1)/2−3.
f(X,T):=X22[X,T].
Let K be an extension field of 2, let a0, . . . , ad−1 ∈ K, and write a:=(a0, . . . , ad−1). Then the map on the dashed arrow at the bottom of the following diagram is a well-defined homomorphism of rings, and the unique homomorphism making the diagram commutative.
Since ϕ(a) ⊆ b, the kernel of the left vertical arrow, which is surjective, is contained in the kernel of the path →↓. Hence there exists a unique dashed homomorphism making the square commutative, and this homomorphism is given by a+aϕ(a)+b. Now the assertion of the lemma follows as a special case.