The present application claims priority to Russian Patent App. No. 2014104571 entitled “Systems and Methods for Area Efficient Data Encoding”, and filed Feb. 10, 2014 by Panteleev et al. The entirety of the aforementioned patent application is incorporated herein by reference for all purposes.
The present inventions are related to systems and methods for data processing, and more particularly to systems and methods for data encoding.
Various data transfer systems have been developed including storage systems, cellular telephone systems, and radio transmission systems. In each of the systems data is transferred from a sender to a receiver via some medium. For example, in a storage system, data is sent from a sender (i.e., a write function) to a receiver (i.e., a read function) via a storage medium. Encoding may involve vector multiplication by a quasi-cyclic matrices. Such vector multiplication is complex both in terms of circuit design and the area required to implement the circuits. Such significant area requirements increase the costs of encoding devices.
Hence, for at least the aforementioned reasons, there exists a need in the art for advanced systems and methods for data processing.
The present inventions are related to systems and methods for data processing, and more particularly to systems and methods for data encoding.
Various embodiments of the present invention provide data processing systems that include an encoder circuit. The encoder circuit includes a cyclic convolution circuit and an encoded output circuit. The cyclic convolution circuit is operable to multiply a vector input derived from a user data input by a portion of a circulant matrix to yield a convolved output. The encoded output circuit is operable to generate an encoded data set corresponding to the user data input and based at least in part on the convolved output.
This summary provides only a general outline of some embodiments of the invention. The phrases “in one embodiment,” “according to one embodiment,” “in various embodiments”, “in one or more embodiments”, “in particular embodiments” and the like generally mean the particular feature, structure, or characteristic following the phrase is included in at least one embodiment of the present invention, and may be included in more than one embodiment of the present invention. Importantly, such phases do not necessarily refer to the same embodiment. Many other embodiments of the invention will become more fully apparent from the following detailed description, the appended claims and the accompanying drawings.
A further understanding of the various embodiments of the present invention may be realized by reference to the figures which are described in remaining portions of the specification. In the figures, like reference numerals are used throughout several figures to refer to similar components. In some instances, a sub-label consisting of a lower case letter is associated with a reference numeral to denote one of multiple similar components. When reference is made to a reference numeral without specification to an existing sub-label, it is intended to refer to all such multiple similar components.
The present inventions are related to systems and methods for data processing, and more particularly to systems and methods for data encoding.
Various embodiments of the present invention provide data processing systems that include an encoder circuit. The encoder circuit includes one or more area efficient quasi-cyclic matrix multiplication circuit(s). Such quasi-cyclic matrix multiplication circuit(s) are designed as a number of cyclic convolutions. Using such an approach, it is possible to implement a encoder circuit for quasi-cyclic low density parity check (LDPC) codes that is smaller and offering several times higher throughput compared with an encoder circuit relying exclusively on shift registers and/or barrel shifters to perform quasi-cyclic matrix multiplications. In some cases, the quasi-cyclic matrix multiplication circuit(s) designed as a number of cyclic convolutions may use a combination of Winograd and Agarwal-Cooley fast convolution algorithms, though many other fast convolution algorithms can be used as well. Such Winograd and Agarwal-Cooley algorithms are discussed in detail in Richard E. Blahut, “Fast Algorithms for Digital Signal Processing,” Addison-Wesley, Reading, Mass. 1985. The entirety of the aforementioned reference is incorporated herein by reference for all purposes.
Most encoding algorithms for quasi-cyclic LDPC codes can be roughly divided into two main categories: generator matrix based (G-based) and parity-check matrix based (H-based). In a G-based encoder a systematic quasi-cyclic generator matrix G=(I|Gp) is used, where Gp is a quasi-cyclic matrix, which is usually dense. The parity bits vector p is obtained by formula p=uGp, where u is a user bits vector. In an H-based encoder we usually represent a quasi-cyclic parity-check matrix of the code as H=(Hu|Hp), where Hu, Hp are its quasi-cyclic sub-matrices corresponding to the user and parity parts of the codeword. Subsequently, the vector sT=HuuT is calculated, and based thereon the parity vector p is determined as a solution of the equation HppT=sT. As it can be seen from the above description both categories of encoders involve a vector by a quasi-cyclic matrix multiplication step. As such, embodiments of the present invention offering improved quasi-cyclic multiplication circuits offer improved encoding.
Various embodiments of the present invention provide data processing systems that include an encoder circuit. The encoder circuit includes a cyclic convolution circuit and an encoded output circuit. The cyclic convolution circuit is operable to multiply a vector input derived from a user data input by a portion of a circulant matrix to yield a convolved output. The encoded output circuit is operable to generate an encoded data set corresponding to the user data input and based at least in part on the convolved output. In some cases, the data processing system is implemented as part of a storage device, or a communication device. In various cases, the data processing system is implemented as part of an integrated circuit.
In some instances of the aforementioned embodiments, the encoded output circuit includes: a vector adder circuit operable to sum instances of the convolved output with instances of a cyclic convolution output to yield a corresponding instance of a vector sum, and a shift register circuit operable to shift instances of the vector sum to yield the instances of the cyclic convolution output. In some cases, the encoded data set generated based at least in part on the cyclic convolution output. In various cases, the number of instances of the vector sum is l, where l corresponds to the number of sub-vectors into which the user data input is divided.
In various instances of the aforementioned embodiments, the cyclic convolution circuit includes: a first cyclic convolution circuit and a second cyclic convolution circuit. In such instances, the first cyclic convolution circuit operates in parallel with the second cyclic convolution circuit, and the first cyclic convolution circuit operates on a first portion of the vector input and the second cyclic convolution circuit operates on a second portion of the vector input. In some cases, the first portion of the vector input is a 3×1 portion of the vector input, and wherein the second portion of the vector input is a 3×4 portion of the vector input. In other cases, the first portion of the vector input is a 3×4 portion of the vector input, and wherein the second portion of the vector input is a 3×8 portion of the vector input.
In one or more instances of the aforementioned embodiments, the systems further include a transformation circuit operable to transform a first number of bits of the user data input into a second number of bits of the vector input. In some such instances, the first number of bits is 128, and the second number of bits is 255. In various such instances, the cyclic convolution circuit includes: a first cyclic convolution circuit, a second cyclic convolution circuit, and a combining circuit. In such instances, the first cyclic convolution circuit operates in parallel with the second cyclic convolution circuit, and the first cyclic convolution circuit operates on a first portion of the vector input and the second cyclic convolution circuit operates on a second portion of the vector input. The combining circuit is operable to combine at least the first sub-output and the second sub-output to yield a non-transformed output. In some cases, the system further includes an inverse transformation circuit operable transform the second number of bits of the non-transformed output to the first number of bits of a cyclic convolution output.
Other embodiments of the present invention provide methods for data encoding that include: receiving a user data input; using a cyclic convolution circuit to multiply a vector input derived from a user data input by a portion of a circulant matrix to yield a convolved output; and generating an encoded data set corresponding to the user data input and based at least in part on the convolved output. In some instances of the aforementioned embodiments, the methods further include transforming a first number of bits of the user data input into a second number of bits to yield the vector input. In some cases, the first number of bits is 128, and the second number of bits is 255.
In one or more instances of the aforementioned embodiments, the cyclic convolution circuit includes: a first cyclic convolution circuit and a second cyclic convolution circuit. The first cyclic convolution circuit operates in parallel with the second cyclic convolution circuit. The first cyclic convolution circuit operates on a first portion of the vector input and the second cyclic convolution circuit operates on a second portion of the vector input. In some cases, the methods further include: adding instances of the convolved output with instances of a cyclic convolution output to yield a corresponding instance of a vector sum; and shifting instances of the vector sum to yield the instances of the cyclic convolution output.
Turning to
In a typical read operation, read/write head 176 is accurately positioned by motor controller 168 over a desired data track on disk platter 178. Motor controller 168 both positions read/write head 176 in relation to disk platter 178 and drives spindle motor 172 by moving read/write head assembly 176 to the proper data track on disk platter 178 under the direction of hard disk controller 166. Spindle motor 172 spins disk platter 178 at a determined spin rate (RPMs). Once read/write head 176 is positioned adjacent the proper data track, magnetic signals representing data on disk platter 178 are sensed by read/write head 176 as disk platter 178 is rotated by spindle motor 172. The sensed magnetic signals are provided as a continuous, minute analog signal representative of the magnetic data on disk platter 178. This minute analog signal is transferred from read/write head 176 to read channel circuit 110 via preamplifier 170. Preamplifier 170 is operable to amplify the minute analog signals accessed from disk platter 178. In turn, read channel circuit 110 decodes and digitizes the received analog signal to recreate the information originally written to disk platter 178. This data is provided as read data 103 to a receiving circuit. A write operation is substantially the opposite of the preceding read operation with write data 101 being provided to read channel circuit 110. This data is then encoded and written to disk platter 178.
In operation, data stored to disk platter 178 is encoded using an area efficient encoder circuit to yield an encoded data set. The encoded data set is then written to disk platter 178, and later accessed from disk platter and decoded using a decoder circuit. In some cases, the area efficient encoder circuit may be implemented to include quasi-cyclic matrix multiplication circuit(s) designed as a number of cyclic convolutions such as that discussed below in relation to
It should be noted that storage system 100 may be integrated into a larger storage system such as, for example, a RAID (redundant array of inexpensive disks or redundant array of independent disks) based storage system. Such a RAID storage system increases stability and reliability through redundancy, combining multiple disks as a logical unit. Data may be spread across a number of disks included in the RAID storage system according to a variety of algorithms and accessed by an operating system as if it were a single disk. For example, data may be mirrored to multiple disks in the RAID storage system, or may be sliced and distributed across multiple disks in a number of techniques. If a small number of disks in the RAID storage system fail or become unavailable, error correction techniques may be used to recreate the missing data based on the remaining portions of the data from the other disks in the RAID storage system. The disks in the RAID storage system may be, but are not limited to, individual storage systems such as storage system 100, and may be located in close proximity to each other or distributed more widely for increased security. In a write operation, write data is provided to a controller, which stores the write data across the disks, for example by mirroring or by striping the write data. In a read operation, the controller retrieves the data from the disks. The controller then yields the resulting read data as if the RAID storage system were a single disk.
A data decoder circuit used in relation to read channel circuit 110 may be, but is not limited to, a low density parity check (LDPC) decoder circuit as are known in the art. Such low density parity check technology is applicable to transmission of information over virtually any channel or storage of information on virtually any media. Transmission applications include, but are not limited to, optical fiber, radio frequency channels, wired or wireless local area networks, digital subscriber line technologies, wireless cellular, Ethernet over any medium such as copper or optical fiber, cable channels such as cable television, and Earth-satellite communications. Storage applications include, but are not limited to, hard disk drives, compact disks, digital video disks, magnetic tapes and memory devices such as DRAM, NAND flash, NOR flash, other non-volatile memories and solid state drives.
In addition, it should be noted that storage system 100 may be modified to include solid state memory that is used to store data in addition to the storage offered by disk platter 178. This solid state memory may be used in parallel to disk platter 178 to provide additional storage. In such a case, the solid state memory receives and provides information directly to read channel circuit 110. Alternatively, the solid state memory may be used as a cache where it offers faster access time than that offered by disk platter 178. In such a case, the solid state memory may be disposed between interface controller 120 and read channel circuit 110 where it operates as a pass through to disk platter 178 when requested data is not available in the solid state memory or when the solid state memory does not have sufficient storage to hold a newly written data set. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize a variety of storage systems including both disk platter 178 and a solid state memory.
Turning to
Turning to
Turning to
Encoded output 439 is provided to a transmission circuit 430 that is operable to transmit the encoded data to a recipient via a medium 440. Transmission circuit 430 may be any circuit known in the art that is capable of transferring encoded output 439 via medium 440. Thus, for example, where data processing circuit 400 is part of a hard disk drive, transmission circuit 430 may include a read/write head assembly that converts an electrical signal into a series of magnetic signals appropriate for writing to a storage medium. Alternatively, where data processing circuit 400 is part of a wireless communication system, transmission circuit 430 may include a wireless transmitter that converts an electrical signal into a radio frequency signal appropriate for transmission via a wireless transmission medium. Transmission circuit 430 provides a transmission output to medium 440. Medium 440 provides a transmitted input that is the transmission output augmented with one or more errors introduced by the transference across medium 440.
Of note, original data input 405 may be any data set that is to be transmitted. For example, where data processing system 400 is a hard disk drive, original data input 405 may be a data set that is destined for storage on a storage medium. In such cases, a medium 440 of data processing system 400 is a storage medium. As another example, where data processing system 400 is a communication system, original data input 405 may be a data set that is destined to be transferred to a receiver via a transfer medium. Such transfer mediums may be, but are not limited to, wired or wireless transfer mediums. In such cases, a medium 440 of data processing system 400 is a transfer medium.
Data processing circuit 400 includes an analog processing circuit 450 that applies one or more analog functions to the transmitted input. Such analog functions may include, but are not limited to, amplification and filtering. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize a variety of pre-processing circuitry that may be used in relation to different embodiments of the present invention. In addition, analog processing circuit 450 converts the processed signal into a series of corresponding digital samples. Data processing circuitry 460 applies data detection and/or data decoding algorithms to the series of digital samples to yield a data output 465. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize a variety of data processing circuitry that may be used to recover original data input from the series of digital samples.
As background to understanding an area efficient quasi-cyclic matrix multiplication circuit used to implement the area efficient encoder circuit 420, an l×l matrix over GF(q) is called a circulant if it has the following form:
Such a circulant matrix can be uniquely represented by its first column (a0,a1, . . . ,al-1)T, and it can be seen that a vector can be re-written by a circulant matrix multiplication in the following way:
The aforementioned multiplication may be represented in the following way:
The vector c=(c0, . . . , cl-1)T is referred to herein as a cyclic convolution of the vectors a=(a0, . . . , al-1)T and b=(b0, . . . , bl-1)T, and for simplicity is denoted as a*b.
A quasi-circulant matrix may be represented as follows:
where each block Aij, i=1 to m, j=1 to n, is an l×l circulant matrix over a finite field GF(q). Using a column vector u=(u1, . . . , un)T, where sub-vectors u1, . . . , un are of length l, multiplying u by the aforementioned quasi-circulant matrix yields:
where each sub-vector vi of length l is given by the following formula:
vi=Ai1u1+ . . . +Ainun; for i=1 to m.
Applying cyclic convolution, the preceding formula for each sub-vector vi of length l may be re-written as:
vi=ai1*u1+ . . . +ain*un; for i=1 to m.
where aij is the first column of the aforementioned circulant matrix Aij;
for i=1 to m, and j=1 to n. Thus, quasi-cyclic multiplication can be obtained by performing m×n cyclic convolutions and m×(n −1) vector additions over GF(q).
Turning to
Original data input 405 (i.e., uj) and the first columns of circulant matrices 478 (i.e., aij) are provided to a cyclic convolution circuit 485 that applies cyclic convolution to the received inputs to yield a convolved output 482 (i.e., aij*uj). Convolved output 482 is provided to a vector addition circuit 490 that is operable to calculate the sum of two vectors of length l over GF(q). In some embodiments of the present invention, vector addition circuit 490 is implemented using XOR gates as is known in the art. In particular, vector addition circuit 490 calculates the sum of convolved output 482 and an accumulated cyclic convolution output 497 over a length l. A resulting vector sum 492 is stored to a shift register circuit 495 where it is shifted over the length l with the final shift yielding the final value of cyclic convolution output 497. Initially, all of the values in shift register circuit 495 are zeros. The final value of cyclic convolution output 497 may be represented by the following equation:
cyclic convolution output 497=ai1*u1+ . . . +ain*un; for i=1 to m.
The approach used in area efficient quasi-cyclic matrix multiplication circuit 470 operates over m×n clock cycles plus the delay of cyclic convolution circuit 485. Original data input 405 (uj) and the first columns of circulant matrices 478 (aij) should be in the following order:
Turning to
Vector element 902 is provided to a multiplier circuit 922 where it is multiplied by vector element 908 to yield a product 942; vector element 902 is provided to a multiplier circuit 928 where it is multiplied by vector element 910 to yield a product 948; and vector element 902 is provided to a multiplier circuit 938 where it is multiplied by vector element 912 to yield a product 958. Vector element 904 is provided to a multiplier circuit 924 where it is multiplied by vector element 912 to yield a product 944; vector element 904 is provided to a multiplier circuit 930 where it is multiplied by vector element 908 to yield a product 950; and vector element 904 is provided to a multiplier circuit 936 where it is multiplied by vector element 910 to yield a product 956. Vector element 906 is provided to a multiplier circuit 926 where it is multiplied by vector element 910 to yield a product 946; vector element 906 is provided to a multiplier circuit 932 where it is multiplied by vector element 912 to yield a product 952; and vector element 906 is provided to a multiplier circuit 934 where it is multiplied by vector element 908 to yield a product 954.
Product 942, product 944, and product 946 are provided to an adder circuit 962 where they are summed to yield a vector component 972 (c0). Product 948, product 950, and product 952 are provided to an adder circuit 964 where they are summed to yield a vector component 974 (c1). Product 954, product 956, and product 958 are provided to an adder circuit 966 where they are summed to yield a vector component 976 (c2).
Where the length l of convolved output 482 is small, implementation of area efficient quasi-cyclic matrix multiplication circuit 470 using blocks similar to that discussed in
Area efficient quasi-cyclic matrix multiplication circuit 500 includes a register circuit 510 that holds a number of bits of an original data input 505 in parallel. In one embodiment of the present invention, the number of bits is one-hundred twenty-eight (128) bits. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize other bit widths that may be used in relation to different embodiments of the present invention. The registered data is accessed in parallel from register circuit 510 as a registered vector 515. Registered vector 515 is provided to a transformation circuit 520 where the number of bits in registered vector 515 are increased to yield a transformed vector 525. The operation of transformation circuit 520 is more fully discussed below. In one embodiment of the present invention, the number of bits in transformed vector 525 is two-hundred fifty-five (255) bits. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize other bit widths that may be used in relation to different embodiments of the present invention. Transformed vector 525 is stored to a register circuit 530 that provides the registered data as a registered vector 535 (a′).
Similarly, area efficient quasi-cyclic matrix multiplication circuit 500 includes a register circuit 511 that holds a number of bits of an original data input 506 in parallel. In one embodiment of the present invention, the number of bits is one-hundred twenty-eight (128) bits. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize other bit widths that may be used in relation to different embodiments of the present invention. The registered data is accessed in parallel from register circuit 511 as a registered vector 516. Registered vector 516 is provided to a transformation circuit 521 where the number of bits in registered vector 516 are increased to yield a transformed vector 526. The operation of transformation circuit 521 is more fully discussed below. In one embodiment of the present invention, the number of bits in transformed vector 526 is two-hundred fifty-five (255) bits. Based upon the disclosure provided herein, one of ordinary skill in the art will recognize other bit widths that may be used in relation to different embodiments of the present invention. Transformed vector 526 is stored to a register circuit 531 that provides the registered data as a registered vector 536 (b′).
Assuming the width of registered vector 535 and registered vector 536 is 255, parallel cyclic convolution circuit 540 that splits each of registered vector 535 and registered vector 536 into chunks (s0(1), . . . , s0(12), s1(1), . . . , s1(12), s2(1), . . . , s2(12)), where 1-bit chunks s0(1), s1(1), s2(1) are considered as elements of GF(2); 4-bit chunks s0(2), s1(2), s2(2) are considered as elements of GF(24) ; 8-bit chunks s0(3), s1(3), s23, . . . , s0(12), s1(12), s2(12) are considered as elements of GF(28).
The aforementioned chunks are distributed between twelve cyclic convolution blocks 550, 560, 570, 580 over the finite fields GF(2), GF(24) , and GF(28) as shown on
Returning to
In order to define the matrices T and T−1 the following 3×3 block matrix (TF) with 85 bits per column is defined:
where T85 is itself an 85×85 matrix by the following row permutations: for all i=1 to 255 move row number 1+85((i−1)mod3)+(i−1)mod85 to the place number i. The transformation matrix T is then obtained from TF by removing the last 127 columns. Using the notation indicating that TF−1 is the inverse of TF, and ri is the ith row of TF−1, then the inverse matrix T−1 is obtained as follows:
The aforementioned T85 matrix is obtained by factoring the polynomial x85+1 to irreducible factors (i.e., primes) over GF(2):
x85+1=ƒ(1)(x) . . . ƒ(12)(x),
where
ƒ(1)(x)=x+1,
ƒ(2)(x)=x4+x3+x2+1,
ƒ(3)(x)=x8+x7+x6+x4+x2+x+1,
ƒ(4)(x)=x8+x7+x5+x+1,
ƒ(5)(x)=x8+x7+x3+x+1,
ƒ(6)(x)=x8+x5+x4+x3+1,
ƒ(7)(x)=x8+x5+x4+x3+x2+x+1,
ƒ(8)(x)=x8+x6+x5+x4+x2+x+1,
ƒ(9)(x)=x8+x6+x5+x4+x3+x+1,
ƒ(10)(x)=x8+x7+x6+x4+x3+x2+1,
ƒ(11)(x)=x8+x7+x5+x4+x3+x2+1, and
ƒ(12)(x)=x8+x7+x6+x5+x4+x3+1.
Let di=degƒ(i)(x) for i=1 to 12, di×85 matrix Ti such that its jth column is equal to (c0, . . . , cd
c0+c1x+ . . . +cd
Each irreducible polynomial ƒ(i)(x) defines the finite field F(i)=GF(2)[x]/(ƒ(i)(x)) of polynomials over GF(2) modulo ƒ(i)(x). The field F(1)is isomorphic to the field GF(2), the field F(2) is isomorphic to the field GF(24) defined by the irreducible polynomial x4+x +1, the fields F(2), . . . , F(12) are isomorphic to the field GF(28) defined by the irreducible polynomial x8+x4+x3+x+1. Let Bi be the di×di transition matrix from the field F(i) to the corresponding isomorphic field. It means that if a binary column vector a representing an element from the field F(i) then the vector Bia represents the corresponding element in the isomorphic field. Then the matrix T85 can be calculated by the following formula:
The resulting matrix T85 is as follows:
It should be noted that the various blocks discussed in the above application may be implemented in integrated circuits along with other functionality. Such integrated circuits may include all of the functions of a given block, system or circuit, or a subset of the block, system or circuit. Further, elements of the blocks, systems or circuits may be implemented across multiple integrated circuits. Such integrated circuits may be any type of integrated circuit known in the art including, but are not limited to, a monolithic integrated circuit, a flip chip integrated circuit, a multichip module integrated circuit, and/or a mixed signal integrated circuit. It should also be noted that various functions of the blocks, systems or circuits discussed herein may be implemented in either software or firmware. In some such cases, the entire system, block or circuit may be implemented using its software or firmware equivalent. In other cases, the one part of a given system, block or circuit may be implemented in software or firmware, while other parts are implemented in hardware.
In conclusion, the invention provides novel systems, devices, methods and arrangements for data processing. While detailed descriptions of one or more embodiments of the invention have been given above, various alternatives, modifications, and equivalents will be apparent to those skilled in the art without varying from the spirit of the invention. Therefore, the above description should not be taken as limiting the scope of the invention, which is defined by the appended claims
Number | Date | Country | Kind |
---|---|---|---|
2014104571 | Feb 2014 | RU | national |
Number | Name | Date | Kind |
---|---|---|---|
4553221 | Hyatt | Nov 1985 | A |
4805174 | Kubota | Feb 1989 | A |
5278703 | Rub | Jan 1994 | A |
5278846 | Okayama | Jan 1994 | A |
5317472 | Schweitzer, III | May 1994 | A |
5325402 | Ushirokawa | Jun 1994 | A |
5351274 | Chennakeshu | Sep 1994 | A |
5392299 | Rhines | Feb 1995 | A |
5406593 | Chennakeshu | Apr 1995 | A |
5417500 | Martinie | May 1995 | A |
5450253 | Seki | Sep 1995 | A |
5513192 | Janku | Apr 1996 | A |
5523903 | Hetzler | Jun 1996 | A |
5550810 | Monogioudis | Aug 1996 | A |
5550870 | Blaker | Aug 1996 | A |
5612964 | Haraszti | Mar 1997 | A |
5696504 | Oliveros | Dec 1997 | A |
5710784 | Kindred | Jan 1998 | A |
5717706 | Ikeda | Feb 1998 | A |
5719871 | Helm | Feb 1998 | A |
5802118 | Bliss | Sep 1998 | A |
5844945 | Nam | Dec 1998 | A |
5898710 | Amrany | Apr 1999 | A |
5923713 | Hatakeyama | Jul 1999 | A |
5978414 | Nara | Nov 1999 | A |
5983383 | Wolf | Nov 1999 | A |
6005897 | Mccallister | Dec 1999 | A |
6023783 | Divsalar | Feb 2000 | A |
6029264 | Kobayashi | Feb 2000 | A |
6065149 | Yamanaka | May 2000 | A |
6097764 | McCallister | Aug 2000 | A |
6145110 | Khayrallah | Nov 2000 | A |
6175588 | Visotsky | Jan 2001 | B1 |
6216249 | Bliss | Apr 2001 | B1 |
6216251 | McGinn | Apr 2001 | B1 |
6266795 | Wei | Jul 2001 | B1 |
6317472 | Choi | Nov 2001 | B1 |
6351832 | Wei | Feb 2002 | B1 |
6377610 | Hagenauer | Apr 2002 | B1 |
6381726 | Weng | Apr 2002 | B1 |
6393074 | Mandyam | May 2002 | B1 |
6412088 | Patapoutian et al. | Jun 2002 | B1 |
6473878 | Wei | Oct 2002 | B1 |
6535553 | Limberg et al. | Mar 2003 | B1 |
6625775 | Kim | Sep 2003 | B1 |
6643814 | Cideciyan et al. | Nov 2003 | B1 |
6697441 | Bottomley | Feb 2004 | B1 |
6747827 | Bassett et al. | Jun 2004 | B1 |
6748034 | Hattori et al. | Jun 2004 | B2 |
6757862 | Marianetti, II | Jun 2004 | B1 |
6785863 | Blankenship et al. | Aug 2004 | B2 |
6807238 | Rhee | Oct 2004 | B1 |
6810502 | Eidson | Oct 2004 | B2 |
6839774 | Ahn et al. | Jan 2005 | B1 |
6948113 | Shaver | Sep 2005 | B1 |
6970511 | Barnette | Nov 2005 | B1 |
6975692 | Razzell | Dec 2005 | B2 |
6986098 | Poeppelman | Jan 2006 | B2 |
7035327 | Nakajima et al. | Apr 2006 | B2 |
7047474 | Rhee | May 2006 | B2 |
7058853 | Kavanappillil et al. | Jun 2006 | B1 |
7058873 | Song et al. | Jun 2006 | B2 |
7073118 | Greenberg | Jul 2006 | B2 |
7093179 | Shea | Aug 2006 | B2 |
7117427 | Ophir | Oct 2006 | B2 |
7130875 | Abe | Oct 2006 | B2 |
7133228 | Fung | Nov 2006 | B2 |
7136244 | Rothberg | Nov 2006 | B1 |
7149239 | Hudson | Dec 2006 | B2 |
7184486 | Wu | Feb 2007 | B1 |
7191378 | Eroz | Mar 2007 | B2 |
7203887 | Eroz | Apr 2007 | B2 |
7230550 | Mittal | Jun 2007 | B1 |
7237181 | Richardson | Jun 2007 | B2 |
7308061 | Huang | Dec 2007 | B1 |
7310768 | Eidson | Dec 2007 | B2 |
7313750 | Feng | Dec 2007 | B1 |
7370258 | Iancu | May 2008 | B2 |
7415651 | Argon | Aug 2008 | B2 |
7499490 | Divsalar et al. | Mar 2009 | B2 |
7502189 | Sawaguchi | Mar 2009 | B2 |
7523375 | Spencer | Apr 2009 | B2 |
7587657 | Haratsch | Sep 2009 | B2 |
7590168 | Raghavan | Sep 2009 | B2 |
7596196 | Liu et al. | Sep 2009 | B1 |
7646829 | Ashley | Jan 2010 | B2 |
7702986 | Bjerke | Apr 2010 | B2 |
7738202 | Zheng | Jun 2010 | B1 |
7752523 | Chaichanavong | Jul 2010 | B1 |
7779325 | Song | Aug 2010 | B2 |
7802172 | Vila Casado | Sep 2010 | B2 |
7913149 | Gribok | Mar 2011 | B2 |
7952824 | Dziak | May 2011 | B2 |
7957251 | Ratnakar Aravind | Jun 2011 | B2 |
7958425 | Chugg | Jun 2011 | B2 |
7996746 | Livshitz | Aug 2011 | B2 |
8018360 | Nayak | Sep 2011 | B2 |
8020069 | Feng | Sep 2011 | B1 |
8020078 | Richardson | Sep 2011 | B2 |
8065598 | Gunnam et al. | Nov 2011 | B1 |
8095859 | Peterson et al. | Jan 2012 | B1 |
8161361 | Song et al. | Apr 2012 | B1 |
8201051 | Tan | Jun 2012 | B2 |
8219868 | Chaichanavong et al. | Jul 2012 | B1 |
8225168 | Yu et al. | Jul 2012 | B2 |
8237597 | Liu | Aug 2012 | B2 |
8255765 | Yeo | Aug 2012 | B1 |
8261171 | Annampedu | Sep 2012 | B2 |
8276055 | Gunnam et al. | Sep 2012 | B1 |
8291284 | Savin | Oct 2012 | B2 |
8291299 | Li et al. | Oct 2012 | B2 |
8295001 | Liu | Oct 2012 | B2 |
8296637 | Varnica | Oct 2012 | B1 |
8370711 | Alrod | Feb 2013 | B2 |
8381069 | Liu | Feb 2013 | B1 |
8413032 | Song | Apr 2013 | B1 |
8429498 | Anholt | Apr 2013 | B1 |
8443267 | Zhong et al. | May 2013 | B2 |
8458555 | Gunnam | Jun 2013 | B2 |
8464142 | Gunnam | Jun 2013 | B2 |
8495462 | Liu | Jul 2013 | B1 |
8516339 | Lesea | Aug 2013 | B1 |
8527849 | Jakab | Sep 2013 | B2 |
8560900 | Bellorado | Oct 2013 | B1 |
8611033 | Li et al. | Dec 2013 | B2 |
8630053 | Yang et al. | Jan 2014 | B2 |
20010010089 | Gueguen | Jul 2001 | A1 |
20010016114 | Van Gestel et al. | Aug 2001 | A1 |
20020021519 | Rae | Feb 2002 | A1 |
20020067780 | Razzell | Jun 2002 | A1 |
20020168033 | Suzuki | Nov 2002 | A1 |
20030031236 | Dahlman | Feb 2003 | A1 |
20030123364 | Nakajima et al. | Jul 2003 | A1 |
20030126527 | Kim et al. | Jul 2003 | A1 |
20030138102 | Kohn et al. | Jul 2003 | A1 |
20030147168 | Galbraith et al. | Aug 2003 | A1 |
20030188252 | Kim | Oct 2003 | A1 |
20040042436 | Terry et al. | Mar 2004 | A1 |
20040194007 | Hocevar | Sep 2004 | A1 |
20040228021 | Yamazaki | Nov 2004 | A1 |
20040264284 | Priborsky et al. | Dec 2004 | A1 |
20050047514 | Bolinth | Mar 2005 | A1 |
20050149842 | Kyung | Jul 2005 | A1 |
20050210367 | Ashikhmin | Sep 2005 | A1 |
20050243456 | Mitchell et al. | Nov 2005 | A1 |
20060002689 | Yang et al. | Jan 2006 | A1 |
20060159355 | Mizuno | Jul 2006 | A1 |
20060195730 | Kageyama | Aug 2006 | A1 |
20070185902 | Messinger et al. | Aug 2007 | A1 |
20070297496 | Park et al. | Dec 2007 | A1 |
20080037676 | Kyun et al. | Feb 2008 | A1 |
20080069373 | Jiang et al. | Mar 2008 | A1 |
20080140686 | Hong | Jun 2008 | A1 |
20080304558 | Zhu et al. | Dec 2008 | A1 |
20090003301 | Reial | Jan 2009 | A1 |
20090092174 | Wang | Apr 2009 | A1 |
20090106633 | Fujiwara | Apr 2009 | A1 |
20090125780 | Taylor | May 2009 | A1 |
20090132893 | Miyazaki | May 2009 | A1 |
20090150745 | Langner et al. | Jun 2009 | A1 |
20090177852 | Chen | Jul 2009 | A1 |
20090185643 | Fitzpatrick | Jul 2009 | A1 |
20090216942 | Yen | Aug 2009 | A1 |
20090273492 | Yang et al. | Nov 2009 | A1 |
20100077276 | Okamura et al. | Mar 2010 | A1 |
20100088575 | Sharon et al. | Apr 2010 | A1 |
20100150252 | Camp | Jun 2010 | A1 |
20100172046 | Liu et al. | Jul 2010 | A1 |
20100241921 | Gunnam | Sep 2010 | A1 |
20100268996 | Yang | Oct 2010 | A1 |
20100322048 | Yang et al. | Dec 2010 | A1 |
20100325511 | Oh | Dec 2010 | A1 |
20110041040 | Su | Feb 2011 | A1 |
20110043938 | Mathew | Feb 2011 | A1 |
20110066768 | Brittner | Mar 2011 | A1 |
20110167227 | Yang | Jul 2011 | A1 |
20110258508 | Ivkovic | Oct 2011 | A1 |
20110264987 | Li | Oct 2011 | A1 |
20110307760 | Pisek | Dec 2011 | A1 |
20110320902 | Gunnam | Dec 2011 | A1 |
20120020402 | Ibing | Jan 2012 | A1 |
20120038998 | Mathew | Feb 2012 | A1 |
20120063023 | Mathew | Mar 2012 | A1 |
20120079353 | Liikanen | Mar 2012 | A1 |
20120124118 | Ivkovic | May 2012 | A1 |
20120182643 | Zhang | Jul 2012 | A1 |
20120185744 | Varnica | Jul 2012 | A1 |
20120203986 | Strasser et al. | Aug 2012 | A1 |
20120207201 | Xia | Aug 2012 | A1 |
20120212849 | Xu | Aug 2012 | A1 |
20120236428 | Xia | Sep 2012 | A1 |
20120262814 | Li | Oct 2012 | A1 |
20120265488 | Sun | Oct 2012 | A1 |
20120317462 | Liu et al. | Dec 2012 | A1 |
20130024740 | Xia | Jan 2013 | A1 |
20130031440 | Sharon | Jan 2013 | A1 |
20130120169 | Li | May 2013 | A1 |
20130194955 | Chang | Aug 2013 | A1 |
20130198580 | Chen | Aug 2013 | A1 |
20130238955 | D'Abreu | Sep 2013 | A1 |
20130254616 | Yang | Sep 2013 | A1 |
20130254619 | Zhang | Sep 2013 | A1 |
Number | Date | Country |
---|---|---|
2001319433 | Nov 2001 | JP |
WO 2010059264 | May 2010 | WO |
WO 2010126482 | Nov 2010 | WO |
Entry |
---|
Casado et al., Multiple-rate low- density parity-check codes with constant blocklength, IEEE Transations on communications, Jan. 2009, vol. 57, pp. 75-83. |
Cui et al., “High-Throughput Layered LDPC Decoding Architecture”, IEEE Transactions on Very Large Scale Integration (VLSI) Systems, vol. 17, No. 4 (Apr. 2009). |
Fan et al., “Constrained coding techniques for soft iterative decoders” Proc. IEEE Global Telecommun. Conf., vol. 1b, pp. 631-637 (1999). |
Fossorier, Marc P.C. “Quasi-Cyclic Low-Density Parity-Check Codes From Circulant Permutation Maricies” IEEE Transactions on Information Theory, vol. 50, No. 8 Aug. 8, 2004. |
Gross, “Stochastic Decoding of LDPC Codes over GF(q)”, HDPCC Workshop, Tel Aviv (Mar. 2, 2010). |
Gunnam et al., “VLSI Architectures for Layered Decoding for Irregular LDPC Codes of WiMax”, IEEE ICC Proceedings (2007). |
Hagenauer, J. et al A Viterbi Algorithm with Soft-Decision Outputs and its Applications in Proc. IEEE Globecom, pp. 47. 11-47 Dallas, TX Nov. 1989. |
Han and Ryan, “Pinning Techniques for Low-Floor Detection/Decoding of LDPC-Coded Partial Response Channels”, 5th International Symposium on Turbo Codes &Related Topics, 2008. |
Kautz, “Fibonacci Codes for Synchronization Control”, IEEE Trans. Info. Theory, vol. 11, No. 2, pp. 284-292 (Apr. 1965). |
Kschischang et al., “Factor Graphs and the Sum-Product Algorithm”, IEEE Transactions on Information Theory, vol. 47, No. 2 (Feb. 2001). |
Leduc-Primeau et al., “A Relaxed Half-Stochastic Iterative Decoder for LDPC Codes”, IEEE Communications Society, IEEE Globecom proceedings (2009). |
Lee et al., “Partial Zero-Forcing Adaptive MMSE Receiver for DS-CDMA Uplink in Multicell Environments” IEEE Transactions on Vehicular Tech. vol. 51, No. 5, Sep. 2002. |
Li et al “Efficient Encoding of Quasi-Cyclic Low-Density Parity Check Codes” IEEE Transactions on Communications on 53 (11) 1973-1973, 2005. |
Lim et al. “Convergence Analysis of Constrained Joint Adaptation in Recording Channels” IEEE Trans. on Signal Processing vol. 54, No. 1 Jan. 2006. |
Lin et al “An efficient VLSI Architecture for non binary LDPC decoders”—IEEE Transaction on Circuits and Systems II vol. 57, Issue 1 (Jan. 2010) pp. 51-55. |
Moon et al, “Pattern-dependent noise prediction in signal-dependent Noise,” IEEE JSAC, vol. 19, No. 4 pp. 730-743, Apr. 2001. |
Moon et al., “Maximum transition run codes for data storage systems”, IEEE Trans. Magn., vol. 32, No. 5, pp. 3992-3994 (Sep. 1996). |
Patapoutian et al “Improving Re-Read Strategies by Waveform Averaging” IEEE Transactions on Mag. vol. 37 No. 6, Nov. 2001. |
Planjery et al “Finite Alphabet Iterative Decoders, pt 1: Decoding Beyond Beliver Propogation on BSC” 7/12, printed from the internet Apr. 21, 2014 http://arxiv.org/pdf/1207.4800.pd. |
Richardson, T “Error Floors of LDPC Codes” Flarion Technologies Bedminster NJ 07921, tjr@flarion.com (not dated). |
Shokrollahi “LDPC Codes: An Introduction”, Digital Fountain, Inc. (Apr. 2, 2003). |
Spagnol et al, “Hardware Implementation of GF(2^m) LDPC Decoders”, IEEE Transactions on Circuits and Systems{hacek over (s)}i: Regular Papers, vol. 56, No. 12 (Dec. 2009). |
Tehrani et al., “Fully Parallel Stochastic LDPC Decoders”, IEEE Transactions on Signal Processing, vol. 56, No. 11 (Nov. 2008). |
Todd et al., “Enforcing maximum-transition-run code constraints and low-density parity check decoding”, IEEE Trans. Magn., vol. 40, No. 6, pp. 3566-3571 (Nov. 2004). |
U.S. Appl. No. 13/302,119, filed Nov. 22, 2011, Lei Chen, Unpublished. |
U.S. Appl. No. 13/227,544, filed Sep. 8, 2011, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/239,683, filed Sep. 22, 2011, Changyou Xu, Unpublished. |
U.S. Appl. No. 13/186,234, filed Jul. 19, 2011, Haitao Xia, Unpublished. |
U.S. Appl. No. 14/025,104, filed Sep. 12, 2013, Bruce Wilson, Unpublished. |
U.S. Appl. No. 13/545,833, filed Jul. 10, 2012, Zhi Bin Li, Unpublished. |
U.S. Appl. No. 13/340,974, filed Dec. 30, 2011, Dan Liu, Unpublished. |
U.S. Appl. No. 13/445,848, filed Apr. 12, 2012, Bruce Wilson, Unpublished. |
U.S. Appl. No. 13/340,951, filed Dec. 30, 2011, Lei Chen, Unpublished. |
U.S. Appl. No. 13/369,468, filed Feb. 9, 2012, Zongwang Li, Unpublished. |
U.S. Appl. No. 13/283,549, filed Oct. 27, 2011, Wu Chang, Unpublished. |
U.S. Appl. No. 13/171,615, filed Jun. 29, 2011, Bradley D. Seago, Unpublished. |
U.S. Appl. No. 13/300,078, filed Nov. 18, 2011, Chung-Li Wang, Unpublished. |
U.S. Appl. No. 13/305,510, filed Nov. 28, 2011, Lei Chen, Unpublished. |
U.S. Appl. No. 13/227,416, filed Sep. 7, 2011, Lei Chen, Unpublished. |
U.S. Appl. No. 13/305,551, filed Nov. 28, 2011, Yang Han, Unpublished. |
U.S. Appl. No. 13/296,022, filed Nov. 14, 2011, Victor Krachkovsky, Unpublished. |
U.S. Appl. No. 13/445,878, filed Apr. 12, 2012, Yu Liao, Unpublished. |
U.S. Appl. No. 13/174,537, filed Jun. 30, 2011, Anantha Raman Krishnan, Unpublished. |
U.S. Appl. No. 13/174,453, filed Jun. 30, 2011, Johnson Yen, Unpublished. |
U.S. Appl. No. 13/269,832, filed Oct. 10, 2011, Haitao Xia, Unpublished. |
U.S. Appl. No. 13/213,751, filed Aug. 19, 2011, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/180,495, filed Jul. 11, 2011, Chung-Li Wang, Unpublished. |
U.S. Appl. No. 13/853,711, filed Mar. 29, 2013, Shu Li, Unpublished. |
U.S. Appl. No. 13/483,105, filed May 30, 2012, Xuebin Wu, Unpublished. |
U.S. Appl. No. 13/426,693, filed Mar. 22, 2012, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/284,767, filed Oct. 28, 2011, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/746,301, filed Jan. 21, 2013, Shu Li, Unpublished. |
U.S. Appl. No. 13/327,279, filed Dec. 15, 2011, Wei Feng, Unpublished. |
U.S. Appl. No. 13/766,891, filed Feb. 14, 2013, Shu Li, Unpublished. |
U.S. Appl. No. 13/875,357, filed May 2, 2013, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/945,787, filed Jul. 18, 2013, Shu Li, Unpublished. |
U.S. Appl. No. 13/945,777, filed Jul. 18, 2013, Shu Li, Unpublished. |
U.S. Appl. No. 13/483,100, filed May 30, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/868,779, filed Apr. 23, 2013, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/652,012, filed Oct. 15, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/918,510, filed Jun. 14, 2013, Shu Li, Unpublished. |
U.S. Appl. No. 13/770,008, filed Feb. 19, 2013, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/912,059, filed Jun. 6, 2013, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/954,573, filed Jul. 30, 2013, Kaitlyn T. Nguyen, Unpublished. |
U.S. Appl. No. 14/072,604, filed Nov. 5, 2013, Shu Li, Unpublished. |
U.S. Appl. No. 14/047,441, filed Oct. 7, 2013, Haitao Xia, Unpublished. |
U.S. Appl. No. 14/101,368, filed Dec. 10, 2013, Yequn Zhang, Unpublished. |
U.S. Appl. No. 14/047,319, filed Oct. 7, 2013, Shaohua Yang, Unpublished. |
U.S. Appl. No. 14/026,722, filed Sep. 13, 2013, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/944,966, filed Jul. 18, 2013, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/284,730, filed Oct. 28, 2011, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/372,580, filed Feb. 14, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/422,986, filed Mar. 16, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/474,660, filed May 17, 2012, Zongwang Li, Unpublished. |
U.S. Appl. No. 13/433,693, filed Mar. 29, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/596,819, filed Aug. 28, 2012, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/596,947, filed Aug. 28, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/766,911, filed Feb. 14, 2013, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/412,520, filed Mar. 5, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/705,407, filed Dec. 5, 2012, Lingyan Sun, Unpublished. |
U.S. Appl. No. 13/426,714, filed Mar. 22, 2012, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/316,858, filed Dec. 12, 2011, Zongwang Li, Unpublished. |
U.S. Appl. No. 13/989,583, filed Oct. 15, 2012, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/415,326, filed Mar. 8, 2012, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/596,978, filed Aug. 28, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/597,001, filed Aug. 28, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/619,907, filed Sep. 14, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/621,341, filed Sep. 17, 2012, Shaohua Yang, Unpublished. |
U.S. Appl. No. 13/316,741, filed Dec. 12, 2011, Yang Han, Unpublished. |
U.S. Appl. No. 13/295,150, filed Nov. 14, 2011, Zongwang Li, Unpublished. |
U.S. Appl. No. 13/415,430, filed Mar. 8, 2012, Nayak Ratnakar Aravind, Unpublished. |
U.S. Appl. No. 13/362,409, filed Jan. 31, 2012, Fan Zhang, Unpublished. |
U.S. Appl. No. 13/269,852, filed Oct. 10, 2011, Haitao Xia, Unpublished. |
U.S. Appl. No. 13/113,219, filed May 23, 2011, Yang Han, Unpublished. |
Vasic, B., “High-Rate Girth-Eight Codes on Rectangular Integer Lattices”, IEEE Trans. Communications, vol. 52, Aug. 2004, pp. 1248-1252. |
Vasic, B., “High-Rate Low-Density Parity-Check Codes Based on Anti-Pasch Affine Geometries,” Proc ICC 2002, pp. 1332-1336. |
Weon-Cheol Lee et al., “Vitierbi Decoding Method Using Channel State Info. In COFDM System” IEEE Trans. on Consumer Elect., IEEE Service Center, NY, NY vol. 45, No. 3 Aug. 1999. |
Xiao, et al “Nested Codes With Multiple Interpretations” retrieved from the Internet URL: http://www.ece.nmsu.edu/˜jkliewer/paper/XFKC—CISS06 (retrieved on Dec. 5, 2012). |
Yeo et al., “VLSI Architecture for Iterative Decoders in Magnetic Storage Channels”, Mar. 2001, pp. 748-755, IEEE trans. Magnetics, vol. 37, No. 2. |
Zhang et al., “Analysis of Verification-Based Decoding on the q-ary Symmetric Channel for Large q”, IEEE Trans. on Information Theory, vol. 57, No. 10 (Oct. 2011). |
Zhong et al., “Design of VLSI Implementation-Oriented LDPC Codes”, IEEE, pp. 670-673, 2003. |
Zhong et al., “High-Rate Quasi-Cyclic LDPC Codes for Magnetic Recording Channel with Low Error Floor”, ISCAS, IEEE pp. 3546-3549, May 2006. |
Zhong et al., “Joint Code-Encoder Design for LDPC Coding System VLSI Implementation”, ISCAS, IEEE pp. 389-392, May 2004. |
Zhong et al., “Quasi Cyclic LDPC Codes for the Magnetic Recording Channel: Code Design and VSLI Implementation”, IEEE Transactions on Magnetics, v. 43, pp. 1118-1123, Mar. 7. |
Zhong, “Block-LDPC: A Practical LDPC Coding System Design Approach”, IEEE Trans. on Circuits, Regular Papers, vol. 5, No. 4, pp. 766-775, Apr. 2005. |
U.S. Appl. No. 13/426,714, filed Mar. 22, 2012, Shaohua Yang. |
Number | Date | Country | |
---|---|---|---|
20150229331 A1 | Aug 2015 | US |