The present invention relates to encoding and decoding data in communications systems and more specifically to communication systems that encode and decode data to account for errors and gaps in communicated data. Communication is used in a broad sense, and includes but is not limited to transmission of digital data of any form through space and/or time.
Transmission of files and streams between a sender and a recipient over a communications channel has been the subject of much literature. Preferably, a recipient desires to receive an exact copy of data transmitted over a channel by a sender with some level of certainty. Where the channel does not have perfect fidelity (which covers most all physically realizable systems), one concern is how to deal with data lost or garbled in transmission. Lost data (erasures) are often easier to deal with than corrupted data (errors) because the recipient cannot always tell when corrupted data is data received in error. Many error-correcting codes have been developed to correct for erasures and/or for errors. Typically, the particular code used is chosen based on some information about the infidelities of the channel through which the data is being transmitted and the nature of the data being transmitted. For example, where the channel is known to have long periods of infidelity, a burst error code might be best suited for that application. Where only short, infrequent errors are expected a simple parity code might be best.
Data transmission is straightforward when a transmitter and a receiver have all of the computing power and electrical power needed for communications and the channel between the transmitter and receiver is clean enough to allow for relatively error-free communications. The problem of data transmission becomes more difficult when the channel is in an adverse environment or the transmitter and/or receiver has limited capability.
One solution is the use of forward error correcting (FEC) techniques, wherein data is coded at the transmitter such that a receiver can recover from transmission erasures and errors. Where feasible, a reverse channel from the receiver to the transmitter allows for the receiver to communicate about errors to the transmitter, which can then adjust its transmission process accordingly. Often, however, a reverse channel is not available or feasible or is available only with limited capacity. For example, where the transmitter is transmitting to a large number of receivers, the transmitter might not be able to handle reverse channels from all those receivers. As another example, the communication channel may be a storage medium and thus the transmission of the data is forward through time and, unless someone invents a time travel machine that can go back in time, a reverse channel for this channel is infeasible. As a result, communication protocols often need to be designed without a reverse channel or with a limited capacity reverse channel and, as such, the transmitter may have to deal with widely varying channel conditions without a full view of those channel conditions.
The problem of data transmission between transmitters and receivers is made more difficult when the receivers need to be low-power, small devices that might be portable or mobile and need to receive data at high bandwidths. For example, a wireless network might be set up to deliver files or streams from a stationary transmitter to a large or indeterminate number of portable or mobile receivers either as a broadcast or multicast where the receivers are constrained in their computing power, memory size, available electrical power, antenna size, device size and other design constraints. Another example is in storage applications where the receiver retrieves data from a storage medium which exhibits infidelities in reproduction of the original data. Such receivers are often embedded with the storage medium itself in devices, for example disk drives, which are highly constrained in terms of computing power and electrical power.
In such a system, considerations to be addressed include having little or no reverse channel, limited memory, limited computing cycles, power, mobility and timing. Preferably, the design should minimize the amount of transmission time needed to deliver data to potentially a large population of receivers, where individual receivers and might be turned on and off at unpredictable times, move in and out of range, incur losses due to link errors, mobility, congestion forcing lower priority file or stream packets to be temporarily dropped, etc.
In the case of a packet protocol used for data transport over a channel that can lose packets, a file, stream or other block of data to be transmitted over a packet network is partitioned into equal size input symbols, encoding symbols the same size as the input symbols are generated from the input symbols using an FEC code, and the encoding symbols are placed and sent in packets. The “size” of a symbol can be measured in bits, whether or not the symbol is actually broken into a bit stream, where a symbol has a size of M bits when the symbol is selected from an alphabet of 2M symbols. In such a packet-based communication system, a packet oriented erasure FEC coding scheme might be suitable. A file transmission is called reliable if it allows the intended recipient to recover an exact copy of the original file even in the face of erasures in the network. A stream transmission is called reliable if it allows the intended recipient to recover an exact copy of each part of the stream in a timely manner even in the face of erasures in the network. Both file transmission and stream transmission can also be somewhat reliable, in the sense that some parts of the file or stream are not recoverable or for streaming if some parts of the stream are not recoverable in a timely fashion. Packet loss often occurs because sporadic congestion causes the buffering mechanism in a router to reach its capacity, forcing it to drop incoming packets. Protection against erasures during transport has been the subject of much study.
In the case of a protocol used for data transmission over a noisy channel that can corrupt bits, a block of data to be transmitted over a data transmission channel is partitioned into equal size input symbols, encoding symbols of the same size are generated from the input symbols and the encoding symbols are sent over the channel. For such a noisy channel the size of a symbol is typically one bit or a few bits, whether or not a symbol is actually broken into a bit stream. In such a communication system, a bit-stream oriented error-correction FEC coding scheme might be suitable. A data transmission is called reliable if it allows the intended recipient to recover an exact copy of the original block even in the face of errors (symbol corruption, either detected or undetected in the channel). The transmission can also be somewhat reliable, in the sense that some parts of the block may remain corrupted after recovery. Symbols are often corrupted by sporadic noise, periodic noise, interference, weak signal, blockages in the channel, and a variety of other causes. Protection against data corruption during transport has been the subject of much study.
Chain reaction codes are FEC codes that allow for generation of an arbitrary number of output symbols from the fixed input symbols of a file or stream. Sometimes, they are referred to as fountain or rateless FEC codes, since the code does not have an a priori fixed transmission rate. Chain reaction codes have many uses, including the generation of an arbitrary number of output symbols in an information additive way, as opposed to an information duplicative way, wherein the latter is where output symbols received by a receiver before being able to recover the input symbols duplicate already received information and thus do not provide useful information for recovering the input symbols. Novel techniques for generating, using and operating chain reaction codes are shown, for example, in Luby I, Luby II, Shokrollahi I and Shokrollahi II.
One property of the output symbols produced by a chain reaction encoder is that a receiver is able to recover the original file or block of the original stream as soon as enough output symbols have been received. Specifically, to recover the original K input symbols with a high probability, the receiver needs approximately K+A output symbols. The ratio A/K is called the “relative reception overhead.” The relative reception overhead depends on the number K of input symbols, and on the reliability of the decoder.
It is also known to use multi-stage chain reaction (“MSCR”) codes, such as those described in Shokrollahi I and/or II and developed by Digital Fountain, Inc. under the trade name “Raptor” codes. Multi-stage chain reaction codes are used, for example, in an encoder that receives input symbols from a source file or source stream, generates intermediate symbols from the input symbols and encodes the intermediate symbols using chain reaction codes. More particularly, a plurality of redundant symbols is generated from an ordered set of input symbols to be communicated. A plurality of output symbols are generated from a combined set of symbols including the input symbols and the redundant symbols, wherein the number of possible output symbols is much larger than the number of symbols in the combined set of symbols, wherein at least one output symbol is generated from more than one symbol in the combined set of symbols and from less than all of the symbols in the combined set of symbols, and such that the ordered set of input symbols can be regenerated to a desired degree of accuracy from any predetermined number, N, of the output symbols. It is also known to use the techniques described above to encode and decode systematic codes, in which the input symbols are includes amongst the possible output symbols of the code. This may be achieved as described in Shokrollahi II by first applying a transformation to the input symbols followed by the steps described above, said enhanced process resulting in the first output symbols generated by the code being equal to the input symbols. As will be clear to those of skill in the art of error and erasure coding, the techniques of Shokrollahi II may be applied directly to the codes described or suggested herein.
For some applications, other variations of codes might be more suitable or otherwise preferred.
The MSCR codes and chain reaction codes described above are extremely efficient in terms of their encoding and decoding complexity. One of the reasons for their efficiency is that the operations that are performed are linear operations over the field GF(2), i.e., the simple field over one bit where the operation of adding two field elements is simply the logical XOR operation, and the operation of multiplying two field elements is simply the logical AND operation. Generally these operations are performed over multiple bits concurrently, e.g., 32 bits at a time or 4 bytes at a time, and such operations are supported natively on all modern CPU processors. On the other hand, when used as erasure FEC codes, because the operations are over GF(2), it turns out that the chance that the receiver can decode all the input symbols goes down by at most approximately one-half for each additional symbol received beyond the first K, where K is the number of original input symbols. For example, if K+A encoding symbols are received then the chance that the recover process fails to recover the K original input symbols is at least 2−A. What would be a more desirable behavior is if the chance of decoding failure decreased more rapidly as a function of A.
There are other erasure and error-correcting FEC codes that operate over larger fields, for example Reed-Solomon codes that operate over GF(4), or over GF(8), or over GF(256), or more generally over GF(2L) for any L>1, and also LDPC codes that operate over larger fields. The advantage of such FEC codes is that, for example in the case of erasure FEC codes, the chance of decoding failure decreases much more rapidly as a function of A than FEC codes over GF(2). On the other hand, these FEC codes are typically much less efficient in terms of encoding and decoding complexity, and one of the primary reasons for that is because the operations over larger fields are much more complex and/or are not natively supported on modern CPUs, and the complexity typically grows as the field size grows. Thus, the FEC codes that operate over larger finite fields are often much slower or impractical compared to FEC codes that operate over GF(2).
Thus, what is needed are erasure and error-correcting FEC codes that are extremely efficient in terms of their encoding and decoding complexity with the property that the chance of decoding failure decreases very rapidly as a function of the number of symbols received beyond the minimal number needed by an ideal FEC code to recover the original input symbols.
According to one embodiment of the invention, a method of encoding data for transmissions from a source to a destination over a communications channel is provided. The method operates on an ordered set of input symbols and may generate zero or more redundant symbols from the input symbols, each redundant symbol being equal to a linear combination of a number of the input symbols with coefficients taken from one or more finite fields, wherein the finite field used may differ as between different input symbols and between different redundant symbols. The method includes generation of a plurality of output symbols from the combined set of symbols including the input symbols, and the redundant symbols if there are any redundant symbols, wherein each output symbol may be generated from one or more of the combined input and redundant symbols, wherein each output symbol is generated as a linear combination of a number of the input and redundant symbols with coefficients taken from one or more finite fields wherein the finite field used may differ as between different input and redundant symbols, between different output symbols and between the output symbols and the redundant symbols and such that the ordered set of input symbols can be regenerated to a desired degree of accuracy from any predetermined number of the output symbols.
The methods can also be used to generate output symbols, wherein the number of possible output symbols that can be generated from a fixed set of input symbols may be much larger than the number of input symbols.
According to another embodiment of the invention, the method includes receiving at a destination at least some of the output symbols sent from a source over a communications channel, where the transmission over the channel may result in the loss or corruption of some of the sent symbols, and where some of the received symbols may be known to be correctly received and information about the degree of corruption of symbols may also be provided. The method includes regenerating at the destination the ordered set of input symbols to a desired degree of accuracy that depends on how many symbols are received and the knowledge of the corruption of the received symbols.
This embodiment can also include receiving at a destination at least some of the output symbols, wherein the number of possible output symbols that can be received may be much larger than the number of input symbols.
According to another embodiment of the invention, a method of encoding data for transmission from a source to a destination over a communications channel is provided. The method operates on an ordered set of input symbols and includes generating a plurality of redundant symbols from the input symbols. The method also includes generating a plurality of output symbols from a combined set of symbols including the input symbols and the redundant symbols, wherein the operation applied in the generation of output symbols is over a small finite field (for example GF(2)) and such that the ordered set of input symbols can be regenerated to a desired degree of accuracy from any predetermined number of the output symbols. The plurality of redundant symbols is generated from the ordered set of input symbols, wherein the operations to generate the redundant symbols is over a finite field that is not GF(2) (for example, GF(256)) or is over a mix of more than one finite field (for example, some operations over GF(2), some operations over GF(256)).
According to still another embodiment of the invention, a system for receiving data transmitted from a source over a communications channel is provided using similar techniques. The system comprises a receive module coupled to a communications channel for receiving output symbols transmitted over the communications channel, wherein each output symbol is generated from at least one symbol in the combined set of symbols including the input symbols and the redundant symbols, wherein the operation applied in the generation of output symbols is over a small finite field (for example GF(2)) and such that the ordered set of input symbols can be regenerated to a desired degree of accuracy from any predetermined number of the output symbols, wherein the input symbols are from an ordered set of input symbols, wherein the redundant symbols are generated from the input symbols and wherein the plurality of redundant symbols is generated from the ordered set of input symbols, wherein the operations to generate the redundant symbols is over a finite field that is not GF(2) (for example, GF(256)) or is over a mix of more than one finite field (for example, some operations over GF(2), some operations over GF(256)).
According to yet another embodiment of the invention, a computer data signal embodied in a carrier wave is provided.
Numerous benefits are achieved by way of the present invention. For example, in a specific embodiment, the computational expense of encoding data for transmission over a channel is reduced. In another specific embodiment, the computational expense of decoding such data is reduced. In yet another specific embodiment, the error probability of the decoder is reduced, while keeping the computational expense of encoding and decoding low. Depending upon the embodiment, one or more of these benefits may be achieved. These and other benefits are provided in more detail throughout the present specification and more particularly below.
A further understanding of the nature and the advantages of the inventions disclosed herein may be realized by reference to the remaining portions of the specification and the attached drawings.
a illustrates a block decomposition of the matrix X that can be used for decoding.
b illustrates a block decomposition of the matrix X after several steps of the first phase of the decoding process.
The detailed description is followed by three appendices: Appendix A contains example values for systematic indices J(K); Appendix B.1 contains example values for table V0; and Appendix B.2 contains example values for table V1.
The inventions described herein make use of mathematical operations for encoding and decoding based on operations in one or more finite fields. Finite fields are finite algebraic structures for which the four arithmetic operations are defined, and which form a field with respect to these operations. Their theory and their construction are well understood by those of skill in the art.
In the description that follows we shall require a multiplication process to be defined between the elements of a finite field and symbols which represent or are derived from the data to be encoded or decoded. Three distinct types of symbols are considered in this description: input symbols comprise information known to the sender which is to be communicated to the receiver, redundant symbols comprise symbols which are derived from the input symbols and output symbols comprise symbols which are transmitted by the sender to the receiver. Of the many possibilities for defining such a multiplication process, we concentrate on two particular ones: simple transformations, and interleaved transformations.
Simple Transformations
In this case, the multiplication process is defined between an element a from a finite field GF(2M) and a symbol S that is M bits in length. As used herein, “symbol” refers to a piece of data that is typically smaller than the source block. The size of a symbol can often be measured in bits, where a symbol has the size of M bits and the symbol is selected from an alphabet of 2M symbols. In applications of reliable transmission of information over packet networks, for example, the size of a symbol could be equal to the packet size, or it could be smaller, so that each packet contains one or more symbols.
In the case of simple transformation, the symbol S is interpreted as an element of GF(2M), and the multiplication a*S is defined as the normal multiplication in the field GF(2M). The operation performed on the symbol is called a “simple transformation” of the symbol. As an illustrative example, consider the field GF(4). Elements of GF(4) can be represented with 2 bits, for example according to their binary expansion. The field GF(4) has four field elements 00, 01, 10, 11, wherein addition is the normal exclusive-or of bit strings, and multiplication is defined via the table:
According to the above multiplication table the result of 10*01 would be 10, since 01 is the multiplicative neutral element (sometimes called the identity element) of the field.
Interleaved Transformations
To illustrate interleaved transformations, we will make use of the mathematical concept of a ring. As is well-known to those of ordinary skill in the art, a ring is a set on which two operations, addition and multiplication, are defined such that these operations satisfy the distributive laws. Moreover, the set considered with addition alone forms an abelian group, i.e., the result of an addition is independent of the ordering of the summands, there is a neutral element 0 for addition, and for each element there is another element such that the sum of these elements is 0. The other requirement is that the multiplication has a neutral element 1, such that multiplication of any element with 1 does not change the value of that element. For a general ring, we do not require that any nonzero element has a multiplicative inverse, nor do we require that multiplication is commutative. When both these conditions are satisfied, however, then we call the ring a “field.” This notation is a standard one in the area of algebra.
A mapping (symbol-wise sum) is a logical construct implementable in hardware, software, data storage, etc. that maps pairs of symbols of the same size to another symbol of that size. We denote this mapping by ⊕, and the image of this map on the pair (S,T) of symbols by S⊕T. An example of such a mapping is the bit-wise exclusive-or (XOR).
Another construct used here is that of the “action” of a special type of sets on symbols. Suppose that A is a set equipped with a commutative addition operation “+” that has a neutral element and that, for every element, contains its additive inverse. Such a set is also commonly called an abelian group. An “action” of this group on the set of symbols is a mapping that maps a pair, comprising a group element r and a symbol S, to another symbol. We denote the image by r*S where this mapping respects addition in the group, i.e., for every pair of elements a and b in the group A, (a+b)*S=a*S⊕b*S. If A is a ring and the action also respects multiplication in A, where the multiplication operator in A is •, i.e., (a•b)*S=a*(b*S), then this action is the desired multiplication process between elements of a finite field and symbols. In this setting we say that the field “operates” on the set of symbols. The operation performed on symbols in this way is called an “interleaved transformation.”
There are abundant examples of such multiplication processes. A few examples are mentioned below. This list of examples is meant for illustrative purposes only, and should not be considered an exhaustive list, nor should it be construed to limit the scope of this invention.
The field GF(2) with field elements 0 and 1, with addition being exclusive-or (XOR) and multiplication being the logical operation AND, operates on the set of symbols by defining 1*S=S, and 0*S=0, wherein S denotes an arbitrary symbol and 0 denotes the symbol that is entirely zeros.
The field GF(4) can operate on symbols of even size in the following way: for such a symbol S we denote by S[0] and S[1] its first and second half, respectively, so that S=(S[0],S[1]) is the concatenation of S[0] and S[1]. Then, we define
00*S=0
01*S=S
10*S=(S[1], S[0]⊕S[1])
11*S=(S[0]⊕S[1],S[0]).
It can be verified quickly that this is indeed a valid operation. It can be seen that the multiplication table of the field describes an operation that coincides with the operation defined above in the case of 2-bit symbols.
Alternatively, the field GF(4) can operate on symbols of even size in the following way: for such a symbol S we denote by S[0] the concatenation of the bits at even positions within S and similarly we denote by S[1] the concatenation of the bits at odd positions within S (where positions are numbered sequentially starting with zero). For two equal length bit strings A and B, let (A|B) be defined to be the bit string C of twice the length where the bit in position 2*i of C is the bit in position i of A and the bit in position 2*i+1 of C is the bit in position i+1 of B. Then, we define
00*S=0
01*S=S
10*S=(S[1]|S[0]⊕S[1])
11*S=(S[0]⊕S[1]|S[0]).
It can be verified quickly that this is indeed a valid operation. It can be seen that all the operations defined above are the same in the case of 2-bit symbols.
The interleaved transformations described above can be viewed as a particular case of an interleaved transformation in which the binary length of an element of the field coincides with the length of the symbols in bits, and the operation of field elements on symbols is the same as the multiplication in the finite field.
More generally, if K is an extension field of GF(2) of degree d, then an operation of the field can be defined on symbols whose size is divisible by d. Such an operation is described in the paper “An XOR-based erasure resilient coding scheme”, by Bloemer, Kalfane, Karpinksi, Karp, Luby, and Zuckerman, published as Technical Report Number TR-95-048 of the International Computer Science Institute in Berkeley, 1995. This scheme uses the so-called “regular representation” of the field K as d×d matrices with binary entries.
For these generalizations, the first interleaved transformation partitions S, a string that is d*I bits in length, into d equal-size parts, where the first part S[0] is the first I bits of S, S[1] is the next I bits of S, and S[d−1] is the last I bits of S. The transformation operates on the d parts of S and produces d parts that are concatenated together to form the result of the operation. Alternatively, the second interleaved transformation partitions S into d equal-size parts, where the first part S[0] is the concatenation of each dth bit of S starting at position 0 in S, the second part S[1] is the concatenation of each dth bit of S starting at position 1 in S, the dth part S[d−1] is the concatenation of each dth bit of S starting at position L−1 in S. This second transformation operates on the d parts of S (exactly the same as the first transformation) and produces d parts that are interleaved together to form the result of the operation.
Note that the first interleaved transformation can be computed by XORing consecutive bits of the original string S together, and this is a benefit for software implementations where typically a CPU supports such operations natively. On the other hand, the values of the bits in particular positions in the result of the operation depend on the length of the original string S, and this is somewhat of a disadvantage if one wants to implement the operation in hardware that supports variable length symbols, as the operation of the hardware needs to be different depending on the symbol length. Note that the second interleaved transformation involves XORing non-consecutive bits of the original string together, and this is somewhat of a disadvantage for software implementations where typically a CPU does not support such XORs as a native operation. Nevertheless, software operations that work on the finite field elements of the symbol directly can be implemented rather efficiently in software, and thus the software implementations of the second interleaved transformation are possible. Furthermore, for the second interleaved transformation the values of the bits in particular positions in the result of the operation does not depend on the length of the original string S, and this is a benefit if one wants to implement the operation in hardware that supports variable length symbols, as the operation of the hardware can be independent of the symbol length. Thus, the second interleaved transformation does have some overall advantages over the first interleaved transformation.
Linear Transformations
The concept of a “linear transformation” can be defined with reference to the simple or interleaved transformations. For given integers m and n, a linear transformation induced by the operation maps vectors of n symbols into vectors of m symbols using the space of matrices with entries in the specified field. A matrix over the field F is a 2-dimensional collection of entries, whose entries belong to F. If a matrix has m rows and n columns, then it is commonly referred to as an m×n matrix. The pair (m,n) is called the “format” of the matrix. Matrices of the same format can be added and subtracted, using the addition and subtraction in the underlying field or ring. A matrix of format (m,n) can be multiplied with a matrix of format (n,k) as is commonly known.
In operation, if B denotes a matrix with format (m,n), and B[j/k] denotes the entry of B at position (j,k), and if S denotes the column vector comprising the symbols S1, S2, . . . , Sn, and X denotes the column vector comprising the symbols X1, X2, . . . , Xm, then the transformation can be expressed as
X=B{circle around (×)}S.
Thus, the following relationship is valid:
for all j from 1 to m, Xj=B[j,1]*S1⊕B[j,2]*S2⊕ . . . ⊕B[j,n]*Sn
The above formula describes a process for calculating X from B and S in an encoder or decoder, referred to as a “simple transformation process” that can be performed by the steps of:
1. Set j to 1, and Xj to 0.
2. For values of k from 1 to n do Xj=Xj⊕B[j,k]*Sk.
3. Increment j by 1. If j is larger than m, then stop, otherwise go to step 2.
Such linear transformations are commonplace in a variety of applications. For example, when using a linear code to encode a piece of data, or source block, S could be the source symbols of the source block to be encoded, X could be the encoded version of S, and B could be a generator matrix for the code. In other applications, for example where the code used is systematic, X could be the redundant symbols of the encoding of S, while B could be the matrix describing the dependency of the redundant symbols on the source symbols.
As will be known to those of skill to in the art, methods are known to perform the operations described above either through the provision of instructions executed within a general-purpose processor, through hardware designed specifically to perform such operations or through a combination of both. In all cases, the cost of the operations, in terms of the number of instructions required, the amount of hardware required, the cost of the hardware, the electrical power consumed by the operation and/or the time required to perform the operation is generally larger when larger finite fields are used. In particular, in the case of the field GF(2), the operations required are equivalent to bit-wise AND and XOR operations which are widely provided within general-purpose processors and simple, fast and inexpensive to implement in hardware where required. By contrast, operations using larger finite fields than GF(2) are rarely provided directly in general-purpose processors and require either specialized hardware or a number of processor instructions and memory operations to perform.
Multi-Field Erasure and Error Correction Codes
Numerous specific embodiments of multi-field erasure and error correction codes are described herein by reference to a generalized matrix description. This approach is adopted as a descriptive tool only and does not represent a unique way to describe the embodiments described herein, nor should it be construed to limit the scope of this invention. In the generalized description, a matrix is constructed whose elements are taken from one or more finite fields. Different elements may be taken from different finite fields, with the property that there is a single field in which all the fields can be embedded and specific such embeddings are chosen. Some or all of the output symbols may be identical to some of the input or redundant symbols, or may be distinct from the input and redundant symbols depending on the particular embodiment chosen as will be illustrated further below.
A one-to-one correspondence is made between the input symbols of the code and some of the columns of the matrix. A further one-to-one correspondence is made between the redundant symbols of the code and the remaining columns of the matrix. Furthermore, a number of rows of the matrix equal to the number of redundant symbols are designated as static rows. Remaining rows of the matrix are designated as dynamic rows. A one to one correspondence is made between the dynamic rows of the matrix and the output symbols of the code. In this description, static rows represent constraints which are required to hold between the input and the redundant symbols and the static rows fully define the relationship between input and redundant symbols such that knowledge of the input symbols and the static rows is sufficient to construct the redundant symbols. Dynamic rows represent the output symbols which are actually sent on the channel. In many codes, the input and/or redundant symbols themselves are sent and this is represented in this description by adding a dynamic row for each input and redundant symbol that is to be transmitted, said dynamic row having a non-zero entry in the column corresponding to the required input or redundant symbol and zero entries in the remaining columns. In some embodiments, the non-zero entry is the identity. In other embodiments, this non-zero entry need not be the identity element.
A matrix of the form described above may be used to determine a method of encoding data for transmission from a source to a destination over a communications channel, the method comprising generating a plurality of redundant symbols from an ordered set of input symbols, wherein each redundant symbol is generated based on a set of linear constraints over one or more of the input symbols and other redundant symbols with coefficients over finite fields, said linear constraints corresponding to the static rows of the matrix description, generating a plurality of output symbols from the combined set of input and redundant symbols, wherein each output symbol is generated as a linear combination of one or more of the combined set of input and redundant symbols with coefficients chosen from finite fields, said linear constraints corresponding to the dynamic rows of the matrix description and sending at least some of the plurality of generated output symbols.
Conversely, a method comprising the above steps may be described in terms of a matrix of the kind described above in which the static rows correspond to the linear constraints over one or more of the input symbols and redundant symbols and the dynamic rows correspond to the linear combinations of the input and redundant symbols which are used to form the output symbols. In practice, embodiments of the method described above may not involve explicit or implicit representation or construction of the matrix described.
As is well-known, in the case that all elements of the matrix are taken from the field GF(2), then a large class of well-known error-correction and erasure-corrections codes can be described in this way. For example, for the case of Low-Density Parity Check (LDPC) codes, including for example those described in the paper entitled “Design, Evaluation and Comparison of Four Large Block FEC Codecs, LDPC, LDGM, LDGM Staircase and LDGM Triangle, plus a Reed-Solomon Small Block FEC Codec” by V. Roca and C. Neumann published as INRIA Research Report RR-5225, June 2004, available at www.inrialpes.fr (referred to hereinafter as “Roca”), the generalized matrix can be constructed from the parity check matrix by designating every row of the parity check matrix as a static row and adding a further dynamic row for each input and redundant symbol as described above. Another example might use the single-stage chain reaction codes described in Luby I and Luby II, in which the number of static rows in the matrix is zero and the dynamic rows comprise a standard chain reaction matrix. Another example is the use of MSCR codes, in which case the generalized description here is equivalent to the standard matrix presentation of such codes.
Other codes over larger fields can also be described in this way. For example, Reed-Solomon codes such as those derived from Vandermonde matrices in which the input symbols are the source symbols, the generalized matrix is equal to the Vandermonde matrix and all rows are dynamic, where in this case each entry is a finite field element from a field that has at least as many elements in its multiplicative group as there are rows and columns in total, e.g., the finite field GF(256) when the number of rows and columns in total is less than 256. Another example is systematic Reed-Solomon codes over a finite field such as GF(256) which are derived from Vandermonde matrices in which case the input symbols are the source symbols, the redundant symbols are the parity symbols, and the matrix is the rows corresponding to the parity symbols within the systematic form of the Vandermonde matrix with all such rows considered static and additional dynamic rows are added for each source and parity symbol as described above since these are exactly the symbols sent over the channel
As is well-known to those of skill in the art of error and erasure correcting codes, desirable properties of error and erasure correcting include low encoding complexity, low decoding complexity, low decoding error probability and low error floor. The complexity of a code is a measure of the computational resources required to encode or decode the code. Low complexity is of especial value in applications where encoding or decoding is to be performed by resource constrained devices such as mobile terminals, consumer electronics devices, storage devices or devices which may process many encoding or decoding operations simultaneously. Computational complexity is a function in part of the density of the matrix used to encode and decode the code and of the size of the finite field from which the matrix elements are taken. Dense matrices generally result in higher complexity and this has led to many designs of codes based on sparse matrices, for example Low Density Parity Check codes and chain reaction codes. Larger finite fields also result in higher complexity, which has led to many designs of code based on small fields, most commonly GF(2).
Error probability in this context is the probability that completely successful decoding is not possible. Error probability for a given error correcting or erasure correcting code is a function of the information received over the channel, and the specific algorithm used for decoding. In the case of erasure correction codes the error probability is one whenever fewer symbols are received than the number of input symbols. Ideal erasure codes have the property that the error probability is zero whenever the number of symbols received is greater than or equal to the number of input symbols. Other codes have non-zero probability of failure in this case.
It is known that ideal erasure codes can be constructed using dense matrices, in particular Reed-Solomon codes. However, in the case of Reed-Solomon codes the size of the field required is a function of the code size, which is the sum of the number of input and redundant symbols, and this fact, together with the density of the matrix results in generally high computational complexity, especially as the code size grows. Furthermore, in the case of low density codes, it is known that larger finite fields can be used to reduce error probability for error correction codes (as is demonstrated for example in the paper “Low Density Parity Check Codes over GF(q)” by M. C. Davey and D. J. C. MacKay, which has appeared in the IEEE Communications Letters, volume 2, number 6, pages 165-167, 1998) and for erasure codes. Additionally, it is known that introduction of a small number of high density matrix rows or columns into a low density code can improve the error probability, providing a compromise between error probability and complexity [MSCR codes and chain reaction codes]. However, a disadvantage of all such codes is that there is always a significant trade-off between low complexity and low error probability.
For many FEC codes, i.e., LDPC codes and chain reaction codes and MSRC codes, as more output symbols than the number of input symbols are received, the error probability for successful decoding decreases exponentially at some rate. The error floor of such a code is the error probability at which receipt of additional output symbols decreases the error probability at a much slower rate than when the number of received output symbols first exceeds the number of input symbols. It is known that use of a small number of high density rows or columns and/or the use of a larger finite field for the matrix can result in lower error floor at the cost of higher computational complexity. A disadvantage of many known error and erasure correction codes with low complexity is that the error floor is higher than desirable.
Herein, novel methods are described for construction of error correction and erasure correction codes which address some of the disadvantages mentioned above. Methods for efficient encoding and decoding of such codes are presented with relation to specific embodiments described herein by way of example.
The choice of fields for the matrix elements from a set of more than one possible field as described herein permits the design of codes which retain the low computational complexity of codes over small fields with the low error probability and error floor of codes over larger fields and thus represents a significant advantage over the state of the art.
In one preferred embodiment which will be described in more detail below, for the majority of the rows the entries are chosen from GF(2) and for the remainder of the rows the entries are chosen from GF(256). In another embodiment, for each row exactly one entry is chosen from GF(256) and the remaining elements are chosen from GF(2).
There are many other possible embodiments of use of elements from more than one field that result in an improvement in the trade-off between computational complexity and error probability and error floor compared to codes known in the art in which all elements are selected from the same field.
As used herein, the term “file” refers to any data that is stored at one or more sources and is to be delivered as a unit to one or more destinations. Thus, a document, an image, and a file from a file server or computer storage device, are all examples of “files” that can be delivered. Files can be of known size (such as a one megabyte image stored on a hard disk) or can be of unknown size (such as a file taken from the output of a streaming source). Either way, the file is a sequence of input symbols, where each input symbol has a position in the file and a value.
As used herein, the term “stream” refers to any data that is stored or generated at one or more sources and is delivered at a specified rate at each point in time in the order it is generated to one or more destinations. Streams can be fixed rate or variable rate. Thus, an MPEG video stream, AMR audio stream, and a data stream used to control a remote device, are all examples of “streams” that can be delivered. The rate of the stream at each point in time can be known (such as 4 megabits per second) or unknown (such as a variable rate stream where the rate at each point in time is not known in advance). Either way, the stream is a sequence of input symbols, where each input symbol has a position in the stream and a value.
Transmission is the process of transmitting data from one or more senders to one or more recipients through a channel in order to deliver a file or stream. A sender is also sometimes referred to as the encoder. If one sender is connected to any number of recipients by a perfect channel, the received data can be an exact copy of the input file or stream, as all the data will be received correctly. Here, we assume that the channel is not perfect, which is the case for most real-world channels. Of the many channel imperfections, two imperfections of interest are data erasure and data incompleteness (which can be treated as a special case of data erasure). Data erasure occurs when the channel loses or drops data. Data incompleteness occurs when a recipient does not start receiving data until some of the data has already passed it by, the recipient stops receiving data before transmission ends, the recipient chooses to only receive a portion of the transmitted data, and/or the recipient intermittently stops and starts again receiving data. As an example of data incompleteness, a moving satellite sender might be transmitting data representing an input file or stream and start the transmission before a recipient is in range. Once the recipient is in range, data can be received until the satellite moves out of range, at which point the recipient can redirect its satellite dish (during which time it is not receiving data) to start receiving the data about the same input file or stream being transmitted by another satellite that has moved into range. As should be apparent from reading this description, data incompleteness is a special case of data erasure, since the recipient can treat the data incompleteness (and the recipient has the same problems) as if the recipient was in range the entire time, but the channel lost all the data up to the point where the recipient started receiving data. Also, as is well known in communication systems design, detectable errors can be considered equivalent to erasures by simply dropping all data blocks or symbols that have detectable errors.
In some communication systems, a recipient receives data generated by multiple senders, or by one sender using multiple connections. For example, to speed up a download, a recipient might simultaneously connect to more than one sender to transmit data concerning the same file. As another example, in a multicast transmission, multiple multicast data streams might be transmitted to allow recipients to connect to one or more of these streams to match the aggregate transmission rate with the bandwidth of the channel connecting them to the sender. In all such cases, a concern is to ensure that all transmitted data is of independent use to a recipient, i.e., that the multiple source data is not redundant among the streams, even when the transmission rates are vastly different for the different streams, and when there are arbitrary patterns of loss.
In general, a communication channel is that which connects the sender and the recipient for data transmission. The communication channel could be a real-time channel, where the channel moves data from the sender to the recipient as the channel gets the data, or the communication channel might be a storage channel that stores some or all of the data in its transit from the sender to the recipient. An example of the latter is disk storage or other storage device. In that example, a program or device that generates data can be thought of as the sender, transmitting the data to a storage device. The recipient is the program or device that reads the data from the storage device. The mechanisms that the sender uses to get the data onto the storage device, the storage device itself and the mechanisms that the recipient uses to get the data from the storage device collectively form the channel. If there is a chance that those mechanisms or the storage device can lose data, then that would be treated as data erasure in the communication channel.
When the sender and recipient are separated by a communication channel in which symbols can be erased, it is preferable not to transmit an exact copy of an input file or stream, but instead to transmit data generated from the input file or stream (which could include all or parts of the input file or stream itself) that assists with recovery of erasures. An encoder is a circuit, device, module or code segment that handles that task. One way of viewing the operation of the encoder is that the encoder generates output symbols from input symbols, where a sequence of input symbol values represents the input file or a block of the stream. Each input symbol would thus have a position, in the input file or block of the stream, and a value. A decoder is a circuit, device, module or code segment that reconstructs the input symbols from the output symbols received by the recipient. In multi-stage coding, the encoder and the decoder are further divided into sub-modules each performing a different task.
In embodiments of multi-stage coding systems, the encoder and the decoder can be further divided into sub-modules, each performing a different task. For instance, in some embodiments, the encoder comprises what is referred to herein as a static encoder and a dynamic encoder. As used herein, a “static encoder” is an encoder that generates a number of redundant symbols from a set of input symbols, wherein the number of redundant symbols is determined prior to encoding. Examples of static encoding codes include Reed-Solomon codes, Tornado codes, Hamming codes, Low Density Parity Check (LDPC) codes, etc. The term “static decoder” is used herein to refer to a decoder that can decode data that was encoded by a static encoder.
As used herein, a “dynamic encoder” is an encoder that generates output symbols from a set of input symbols and possibly a set of redundant symbols. In one preferred embodiment described here, the number of possible output symbols is orders of magnitude larger than the number of input symbols, and the number of output symbols to be generated need not be fixed. One example of such a dynamic encoder is a chain reaction encoder, such as the encoders described in Luby I and Luby II. The term “dynamic decoder” is used herein to refer to a decoder that can decode data that was encoded by a dynamic encoder.
Embodiments of multi-field coding need not be limited to any particular type of input symbol. Typically, the values for the input symbols are selected from an alphabet of 2M symbols for some positive integer M. In such cases, an input symbol can be represented by a sequence of M bits of data from the input file or stream. The value of M is often determined based on, for example, the uses of the application, the communication channel, and/or the size of the output symbols. Additionally, the size of an output symbol is often determined based on the application, the channel, and/or the size of the input symbols. In some cases, the coding process might be simplified if the output symbol values and the input symbol values were the same size (i.e., representable by the same number of bits or selected from the same alphabet). If that is the case, then the input symbol value size is limited when the output symbol value size is limited. For example, it may be desired to put output symbols in packets of limited size. If some data about a key associated with the output symbols were to be transmitted in order to recover the key at the receiver, the output symbol would preferably be small enough to accommodate, in one packet, the output symbol value and the data about the key.
As an example, if an input file is a multiple megabyte file, the input file might be broken into thousands, tens of thousands, or hundreds of thousands of input symbols with each input symbol encoding thousands, hundreds, or only few bytes. As another example, for a packet-based Internet channel, a packet with a payload of size of 1024 bytes might be appropriate (a byte is 8 bits). In this example, assuming each packet contains one output symbol and 8 bytes of auxiliary information, an output symbol size of 8128 bits ((1024−8)*8) would be appropriate. Thus, the input symbol size could be chosen as M=(1024−8)*8, or 8128 bits. As another example, some video distribution systems use the MPEG packet standard, where the payload of each packet comprises 188 bytes. In that example, assuming each packet contains one output symbol and 4 bytes of auxiliary information, an output symbol size of 1472 bits ((188−4)*8), would be appropriate. Thus, the input symbol size could be chosen as M=(188−4)*8, or 1472 bits. In a general-purpose communication system using multi-stage coding, the application-specific parameters, such as the input symbol size (i.e., M, the number of bits encoded by an input symbol), might be variables set by the application.
As another example, for a stream that is sent using variable size source packets, the symbol size might be chosen to be rather small so that each source packet can be covered with an integral number of input symbols that have aggregate size at most slightly larger than the source packet.
Each output symbol has a value. In one preferred embodiment, which we consider below, each output symbol also has associated therewith an identifier called its “key.” Preferably, the key of each output symbol can be easily determined by the recipient to allow the recipient to distinguish one output symbol from other output symbols. Preferably, the key of an output symbol is distinct from the keys of all other output symbols. There are various forms of keying discussed in previous art. For example, Luby I describes various forms of keying that can be employed in embodiments described herein.
Multi-field Multi-stage coding is particularly useful where there is an expectation of data erasure or where the recipient does not begin and end reception exactly when a transmission begins and ends. The latter condition is referred to herein as “data incompleteness.” Regarding erasure events, multi-stage coding shares many of the benefits of chain reaction coding described in Luby I. In particular, multi-stage codes may be fountain codes, or rateless codes, in which case many times more distinct output symbols than there are input symbols can be generated for a set of fixed-value input symbols, and any suitable number of distinct output symbols can be used to recover the input symbols to a desired degree of accuracy. These conditions do not adversely affect the communication process when multi-field multi-stage coding is used, because the output symbols generated with multi-field multi-stage coding are information additive. For example, if a hundred packets are lost due to a burst of noise causing data erasure, an extra hundred packets can be picked up after the burst to replace the loss of the erased packets. If thousands of packets are lost because a receiver did not tune into a transmitter when it began transmitting, the receiver could just pickup those thousands of packets from any other period of transmission, or even from another transmitter. With multi-field multi-stage coding, a receiver is not constrained to pickup any particular set of packets, so it can receive some packets from one transmitter, switch to another transmitter, lose some packets, miss the beginning or end of a given transmission and still recover an input file or block of a stream. The ability to join and leave a transmission without receiver-transmitter coordination helps to simplify the communication process.
In some embodiments, transmitting a file or stream using multi-field multi-stage coding can include generating, forming or extracting input symbols from an input file or block of a stream, computing redundant symbols, encoding input and redundant symbols into one or more output symbols, where each output symbol is generated based on its key independently of all other output symbols, and transmitting the output symbols to one or more recipients over a channel. Additionally, in some embodiments, receiving (and reconstructing) a copy of the input file or block of a stream using multi-field multi-stage coding can include receiving some set or subset of output symbols from one of more data streams, and decoding the input symbols from the values and keys of the received output symbols.
Suitable FEC erasure codes as described herein can be used to overcome the above-cited difficulties and would find use in a number of fields including multimedia broadcasting and multicasting systems and services. An FEC erasure code hereafter referred to as “a multi-field multi-stage chain reaction code” has properties that meet many of the current and future requirements of such systems and services.
Some basic properties of multi-field multi-stage chain reaction codes are that, for any packet loss conditions and for delivery of source files of any relevant size or streams of any relevant rate: (a) reception overhead of each individual receiver device (“RD”) is minimized; (b) the total transmission time needed to deliver source files to any number of RDs can be minimized (c) the quality of the delivered stream to any number of RDs can be maximized for the number of output symbols sent relative to the number of input symbols, with suitable selection of transmission schedules. The RDs might be handheld devices, embedded into a vehicle, portable (i.e., movable but not typically in motion when in use) or fixed to a location.
The amount of working memory needed for decoding is low and can still provide the above properties, and the amount of computation needed to encode and decode is minimal. In this document, we provide a simple and easy to implement description of some variations of multi-field multi-stage chain reaction codes.
Multi-field Multi-stage chain reaction codes are fountain codes, i.e., as many encoding packets as needed can be generated on-the-fly, each containing unique encoding symbols that are equally useful for recovering a source file or block of a stream. There are many advantages to using fountain codes versus other types of FEC codes. One advantage is that, regardless of packet loss conditions and RD availability, fountain codes minimize the number of encoding packets each RD needs to receive to reconstruct a source file or block of a stream. This is true even under harsh packet loss conditions and when, for example, mobile RDs are only intermittently turned-on or available over a long file download session.
Another advantage is the ability to generate exactly as many encoding packets as needed, making the decision on how many encoding packets to generate on-the-fly while the transmission is in progress. This can be useful if for example there is feedback from RDs indicating whether or not they received enough encoding packets to recover a source file or block of a stream. When packet loss conditions are less severe than expected the transmission can be terminated early. When packet loss conditions are more severe than expected or RDs are unavailable more often than expected the transmission can be seamlessly extended.
Another advantage is the ability to inverse multiplex. Inverse multiplexing is when a RD is able to combine received encoding packets generated at independent senders to reconstruct a source file or block of a stream. One practical use of inverse multiplexing is described in below in reference to receiving encoding packets from different senders.
Where future packet loss, RD availability and application conditions are hard to predict, it is important to choose an FEC solution that is as flexible as possible to work well under unpredictable conditions. Multi-stage chain reaction codes provide a degree of flexibility unmatched by other types of FEC codes.
A further advantage of multi-field multi-stage codes is that the error probability and error floor of the codes is much lower than those of previously known codes with equivalent computational complexity. Equally, the computational complexity of multi-field multi-stage chain reaction codes is much lower than that of previously known codes with equivalent error probability and/or error floor.
Another advantage of multi-field multi-stage chain reaction codes is that parameters such as symbol size and field sizes can be chosen flexibly to achieve any desired balance between computational complexity and error probability and/or error floor.
Aspects of the invention will now be described with reference to the figures.
System Overview
Static key generator 130 produces a stream of static keys S0, S1, . . . . The number of the static keys generated is generally limited and depends on the specific embodiment of encoder 115. The generation of static keys will be subsequently described in more detail. Dynamic key generator 120 generates a dynamic key for each output symbol to be generated by the encoder 115. Each dynamic key is generated so that a large fraction of the dynamic keys for the same input file or block of a stream are unique. For example, Luby I describes embodiments of key generators that can be used. The outputs of dynamic key generator 120 and the static key generator 130 are provided to encoder 115.
From each key I provided by dynamic key generator 120, encoder 115 generates an output symbol, with a value B(I), from the input symbols provided by the input symbol generator. The operation of encoder 115 will be described in more detail below. The value of each output symbol is generated based on its key, on some function of one or more of the input symbols, and possibly on or more redundant symbols that had been computed from the input symbols. The collection of input symbols and redundant symbols that give rise to a specific output symbol is referred to herein as the output symbol's “associated symbols” or just its “associates”. The selection of the function (the “value function”) and the associates is done according to a process described in more detail below. Typically, but not always, M is the same for input symbols and output symbols, i.e., they both code for the same number of bits.
In some embodiments, the number K of input symbols is used by the encoder 115 to select the associates. If K is not known in advance, such as where the input is a streaming file, K can be just an estimate. The value K might also be used by encoder 115 to allocate storage for input symbols and any intermediate symbols generated by encoder 115.
Encoder 115 provides output symbols to a transmit module 140. Transmit module 140 is also provided the key of each such output symbol from the dynamic key generator 120. Transmit module 140 transmits the output symbols, and depending on the keying method used, transmit module 140 might also transmit some data about the keys of the transmitted output symbols, over a channel 145 to a receive module 150. Channel 145 is assumed to be an erasure channel, but that is not a requirement for proper operation of communication system 100. Modules 140, 145 and 150 can be any suitable hardware components, software components, physical media, or any combination thereof, so long as transmit module 140 is adapted to transmit output symbols and any needed data about their keys to channel 145 and receive module 150 is adapted to receive symbols and potentially some data about their keys from channel 145. The value of K, if used to determine the associates, can be sent over channel 145, or it may be set ahead of time by agreement of encoder 115 and decoder 155.
As explained above, channel 145 can be a real-time channel, such as a path through the Internet or a broadcast link from a television transmitter to a television recipient or a telephone connection from one point to another, or channel 145 can be a storage channel, such as a CD-ROM, disk drive, Web site, or the like. Channel 145 might even be a combination of a real-time channel and a storage channel, such as a channel formed when one person transmits an input file from a personal computer to an Internet Service Provider (ISP) over a telephone line, the input file is stored on a Web server and is subsequently transmitted to a recipient over the Internet.
Because channel 145 is assumed to be an erasure channel, communications system 100 does not assume a one-to-one correspondence between the output symbols that exit receive module 150 and the output symbols that go into transmit module 140. In fact, where channel 145 comprises a packet network, communications system 100 might not even be able to assume that the relative order of any two or more packets is preserved in transit through channel 145. Therefore, the key of the output symbols is determined using one or more of the keying schemes described above, and not necessarily determined by the order in which the output symbols exit receive module 150.
Receive module 150 provides the output symbols to a decoder 155, and any data receive module 150 receives about the keys of these output symbols is provided to a dynamic key regenerator 160. Dynamic key regenerator 160 regenerates the dynamic keys for the received output symbols and provides these dynamic keys to decoder 155. Static key generator 163 regenerates the static keys S0, S1, . . . and provides them to decoder 155. The static key generator has access to random number generator 135 used both during the encoding and the decoding process. This can be in the form of access to the same physical device if the random numbers are generated on such device, or in the form of access to the same algorithm for the generation of random numbers to achieve identical behavior. Decoder 155 uses the keys provided by dynamic key regenerator 160 and static key generator 163 together with the corresponding output symbols, to recover the input symbols (again IS(0), IS(1), IS(2), . . . ). Decoder 155 provides the recovered input symbols to an input file reassembler 165, which generates a copy 170 of input file 101 or input stream 105.
One property of the output symbols produced by a chain reaction encoder is that a receiver is able to recover the original file or block of the original stream as soon as enough output symbols have been received. Specifically, to recover the original K input symbols with a high probability, the receiver needs approximately K+A output symbols. The ratio A/K is called the “relative reception overhead.” The relative reception overhead depends on the number K of input symbols, and on the reliability of the decoder. Luby I, Luby II and Shokrollahi I provide teachings of systems and methods that can be employed in certain embodiments. It is to be understood, however, that these systems and methods are not required of the present invention, and many other variations, modifications, or alternatives can also be used.
An Encoder
Dynamic encoder receives the input symbols and the redundant symbols, and generates output symbols as will be described in further detail below. In one embodiment in which the redundant symbols are stored in input symbol buffer 205, dynamic encoder 220 receives the input symbols and redundant symbols from input symbol buffer 205.
Redundancy calculator 230 computes the number R of redundant symbols from the number K of input symbols. This computation is described in further detail below.
Overview of Static Encoder
The general operation of static encoder 210 is shown with reference to
Referring again to
Referring now to
Overview of Multi-Stage Encoder
Referring again to
Decoder 900 comprises a dynamic decoder 905 and a static decoder 910. Input symbols and redundant symbols recovered by dynamic decoder 905 are stored in a reconstruction buffer 915. Upon completion of dynamic decoding, static decoder 910 attempts to recover any input symbols not recovered by dynamic decoder 905, if any. In particular, static decoder 910 receives input symbols and redundant symbols from reconstruction buffer 915.
In step 1010, dynamic decoder 905 regenerates input symbols and redundant symbols from the Q received output symbols. It is to be understood, that steps 1005 and 1010 can be performed substantially concurrently. For example, dynamic decoder 905 can begin regenerating input symbols and redundant symbols prior to the decoder receiving Q output symbols.
After dynamic decoder 905 has processed Q output symbols, then it is determined whether the input symbols have been recovered to a desired degree of accuracy. The desired degree of accuracy may be, for example, all of the input symbols, or some number, percentage, etc., less than all of the input symbols. If yes, then the flow ends. If no, then the flow proceeds to step 1020. In step 1020, static decoder 910 attempts to recover any input symbols that dynamic decoder 905 was unable to recover. After static encoder 910 has processed the input symbols and redundant symbols recovered by dynamic encoder 905, then the flow ends.
In step 1065, if it is determined that dynamic decoding is not to be stopped, then the flow proceeds back to step 1055. But, if in step 1065, it is determined to end dynamic decoding, then the flow proceeds to step 1070. In step 1070, it is determined whether the input symbols have been recovered to a desired degree of accuracy. If yes, then the flow ends. If no, then the flow proceeds to step 1075. In step 1075, static decoder 910 attempts to recover any input symbols that dynamic decoder 905 was unable to recover. After static encoder 910 has processed the input symbols and redundant symbols recovered by dynamic encoder 905, the flow ends.
Many variations of LDPC decoders and HDPC decoders are well known to those skilled in the art, and can be employed in various embodiments according to the present invention. In one specific embodiment, HDPC decoder is implemented using a Gaussian elimination algorithm. Many variations of Gaussian elimination algorithms are well known to those skilled in the art, and can be employed in various embodiments according to the present invention.
A Variation of HDPC Coding
Another type of HDPC encoding is now described. In this embodiment of HDPC encoding, the mathematical operation for creating redundant symbols from a given set of data is based on operations in a finite field.
In this embodiment of HDPC coding, the elements of a finite field are used to obtain redundant symbols HD[0], . . . , HD[D−1]. These symbols are obtained by defining a multiplication process between the symbols IS[0], . . . ,IS[K−1],LD[0], . . . , LD[E−1] and elements of the finite field as described above.
HDPC Coding
When using an HDPC code, the code might be described by a generator matrix over a finite field GF(2M). Where the code is systematic, which is the case in a preferred embodiment, the generator matrix can be described using only the relationship between the K+E input symbols IS[0], . . . ,IS[K−1],LD[0], . . . ,LD[E−1] and the redundant symbols HD[0], . . . ,HD[D−1]. This matrix, called G, is of format Dx(K+E). If X denotes the column vector comprising the symbols HD[0], . . . ,HD[D−1] and S denotes the column vector comprising the symbols IS[0], . . . ,IS[K−1],LD[0], . . . ,LD[E−1], then we have X=G{circle around (×)}S. More specific embodiments for the matrix G and various methods for efficient computation of the symbols are described below.
Variations
Multi-stage chain reaction codes as described above are not systematic codes, i.e., all of the original source symbols of a source block are not necessarily among the encoding symbols that are sent. However, systematic FEC codes are useful for a file download system or service, and very important for a streaming system or service. As shown in the implementation below, a modified code can be made to be systematic and still maintain the fountain code and other described properties.
One reason why it is easy to construct a variety of supplemental services using multi-stage codes is that it can combine received encoding symbols from multiple senders to reconstruct a source file or stream without coordination among the senders. The only requirement is that the senders use differing sets of keys to generate the encoding symbols that they send in encoding packets to the code. Ways to achieve this include designating different ranges of the key space to be used by each such sender, or generating keys randomly at each sender.
As an example of the use of this capability, consider providing a supplemental service to a file download service that allows multi-stage chain reaction codes that did not receive enough encoding packets to reconstruct a source file from the file download session to request additional encoding packets to be sent from a make-up sender, e.g., via a HTTP session. The make-up sender generates encoding symbols from the source file and sends them, for example using HTTP, and all these encoding symbols can be combined with those received from the file download session to recover the source file. Using this approach allows different senders to provide incremental source file delivery services without coordination between the senders, and ensuring that each individual receiver need receive only a minimal number of encoding packets to recover each source file.
Decoding of multi-stage chain reaction codes as described above may require a relatively large overhead when the number of source symbols is small, for example in the order of hundreds to a few thousands source symbols. In such a case, a different decoder is preferred, for example a decoder disclosed in Shokrollahi III. As shown in the implementation below, a modified decoding algorithm can be designed for the class of codes disclosed herein that uses features of the codes and concepts disclosed in Shokrollahi III, and provides low decoding error probability for very small numbers of source symbols, while maintaining efficiency in the decoding.
Implementations of Various Stages of Multi-Field Multi-Stage Codes
FEC Scheme Definition
A packet using these techniques might be represented with header information such as an FEC Payload ID of four octets comprising a Source Block Number (SBN) (16 bit integer identifier for the source block that the encoding symbols within the packet relate to) and an Encoding Symbol ID (ESI) (16 bit integer identifier for the encoding symbols within the packet). One suitable interpretation of the Source Block Number and Encoding Symbol Identifier is defined in Sections B below. FEC Object Transmission information might comprise the FEC Encoding ID, a Transfer Length (F) and the parameters T, Z, N and A defined in below. The parameters T and Z are 16 bit unsigned integers, N and A are 8 bit unsigned integers. If needed, other integer sizes might be used.
An FEC encoding scheme for forward error correction is defined in the sections below. It defines two different FEC Payload ID formats, one for FEC source packets and another for FEC repair packets, but variations for nonsystematic codes are also possible.
The Source FEC payload ID might comprise a Source Block Number (SBN) (16 bit integer identifier for the source block that the encoding symbols within the packet relate to) and an Encoding Symbol ID (ESI) (16 bit integer identifier for the encoding symbols within the packet), while the Repair FEC Payload ID might comprise a Source Block Number (SBN) (16 bit integer identifier for the source block that the repair symbols within the packet relate to), an Encoding Symbol ID (ESI) (16 bit integer identifier for the repair symbols within the packet), and a Source Block Length (SBL) (16 bits, representing the number of source symbols in the source block. The interpretation of the Source Block Number, Encoding Symbol Identifier and Source Block Length is defined below.
FEC Object Transmission information might comprise the FEC Encoding ID, the maximum source block length, in symbols, and the symbol size, in bytes. The symbol size and maximum source block length might comprise a four octet field of Symbol Size (T) (16 bits representing the size of an encoding symbol, in bytes), and a Maximum Source Block Length (16 bits representing the maximum length of a source block, in symbols).
The sections below specify the systematic multi-field MSCR forward error correction code. Multi-field MSCR codes are fountain codes, i.e., as many encoding symbols as needed can be generated by the encoder on-the-fly from the source symbols of a block. The decoder is able to recover the source block from any set of encoding symbols only slightly more in number than the number of source symbols. The code described in this document is a systematic code, that is, the original source symbols are sent unmodified from sender to receiver, as well as a number of repair symbols.
B.1 Definitions, Symbols and Abbreviations
B.1.1 Definitions
For the purposes of this description, the following terms and definitions apply.
B.1.3 Abbreviations
For the purposes of the present document, the following abbreviations apply:
B.2. Overview
The MSCR forward error correction code can be applied to both file delivery and streaming applications. MSCR code aspects which are specific to each of these applications are discussed in Sections B.3 and B.4 of this document.
A component of the systematic MSCR code is the basic encoder described in Section B.5. First, it is described how to derive values for a set of intermediate symbols from the original source symbols such that knowledge of the intermediate symbols is sufficient to reconstruct the source symbols. Secondly, the encoder produces repair symbols which are each the exclusive OR of a number of the intermediate symbols. The encoding symbols are the combination of the source and repair symbols. The repair symbols are produced in such a way that the intermediate symbols and therefore also the source symbols can be recovered from any sufficiently large set of encoding symbols.
This document defines the systematic MSCR code encoder. A number of possible decoding algorithms are possible. An efficient decoding algorithm is provided in Section B.6.
The construction of the intermediate and repair symbols is based in part on a pseudorandom number generator described in Section B.5. This generator is based on a fixed set of 512 random numbers that are available to both sender and receiver. An example set of numbers are those provided in Appendices B.1 and B.2.
Finally, the construction of the intermediate symbols from the source symbols is governed by a “systematic index”. An example set of values for the systematic index is shown in Appendix A for source block sizes from 4 source symbols to KMAX=8192 source symbols.
B.3. File Download
B.3.1. Source Block Construction
B.3.1.1. General
In order to apply the MSCR encoder to a source file, the file may be broken into Z≧1 blocks, known as source blocks. The MSCR encoder is applied independently to each source block. Each source block is identified by a unique integer Source Block Number (SBN), where the first source block has SBN zero, the second has SBN one, etc. Each source block is divided into a number, K, of source symbols of size T bytes each. Each source symbol is identified by a unique integer Encoding Symbol Identifier (ESI), where the first source symbol of a source block has ESI zero, the second has ESI one, etc.
Each source block with K source symbols is divided into N≧1 sub-blocks, which are small enough to be decoded in the working memory. Each sub-block is divided into K sub-symbols of size T′.
Note that the value of K is not necessarily the same for each source block of a file and the value of T′ may not necessarily be the same for each sub-block of a source block. However, the symbol size T is the same for all source blocks of a file and the number of symbols, K is the same for every sub-block of a source block. Exact partitioning of the file into source blocks and sub-blocks is described in B.3.1.2 below.
B.3.1.2 Source Block and Sub-Block Partitioning
The construction of source blocks and sub-blocks is determined based on five input parameters, F, A, T, Z and N and a function Partition[ ]. The five input parameters are defined as follows:
These parameters might be set so that ceil(ceil(F/T)/Z)≦KMAX. An example of some suitable derivations of these parameters are provided in Section B.3.4.
The function Partition[ ] takes a pair of integers (I, J) as input and derives four integers (IL, IS, JL, JS) as output. Specifically, the value of Partition[I, J] is a sequence of four integers (IL, IS, JL, JS), where IL=ceil(I/J), IS=floor(I/J), JL=I−IS·J and JS=J−JL. Partition[ ] derives parameters for partitioning a block of size I into J approximately equal sized blocks. Specifically, JL blocks of length IL and JS blocks of length IS.
The source file might be partitioned into source blocks and sub-blocks as follows:
Let,
Then, the file might be partitioned into Z=ZL+ZS contiguous source blocks, the first ZL source blocks each having length KL·T bytes and the remaining ZS source blocks cach having KS·T bytes.
If Kt·T>F then for encoding purposes, the last symbol might be padded at the end with Kt·T−F zero bytes.
Next, each source block might be divided into N=NL+NS contiguous sub-blocks, the first NL sub-blocks each comprising K contiguous sub-symbols of size of TL·A and the remaining NS sub-blocks each comprising K contiguous sub-symbols of size of TS·A. The symbol alignment parameter A ensures that sub-symbols are always a multiple of A bytes.
Finally, the mth symbol of a source block comprises the concatenation of the mth sub-symbol from each of the N sub-blocks.
B.3.2. Encoding Packet Construction
B.3.2.1. General
Each encoding packet contains a Source Block Number (SBN), an Encoding Symbol ID (ESI) and encoding symbol(s). Each source block is encoded independently of the others. Source blocks are numbered consecutively from zero. Encoding Symbol ID values from 0 to K−1 identify the source symbols. Encoding Symbol IDs from K onwards identify repair symbols.
B.3.2.2 Encoding Packet Construction
Each encoding packet preferably either contains source symbols (source packet) or contains repair symbols (repair packet). A packet may contain any number of symbols from the same source block. In the case that the last symbol in the packet includes padding bytes added for FEC encoding purposes then these bytes need not be included in the packet. Otherwise, only whole symbols might be included.
The Encoding Symbol ID, X, carried in each source packet is the Encoding Symbol ID of the first source symbol carried in that packet. The subsequent source symbols in the packet have Encoding Symbol IDs, X+1 to X+G−1, in sequential order, where G is the number of symbols in the packet.
Similarly, the Encoding Symbol ID, X, placed into a repair packet is the Encoding Symbol ID of the first repair symbol in the repair packet and the subsequent repair symbols in the packet have Encoding Symbol IDs X+1 to X+G−1 in sequential order, where G is the number of symbols in the packet.
Note that it is not necessary for the receiver to know the total number of repair packets. The G repair symbol triples (d[0], a[0], b[0]), . . . , (d[G−1], a[G−1], b[G−1]) for the repair symbols placed into a repair packet with ESI X are computed using the Triple generator defined in B.5.3.4 as follows:
The G repair symbols to be placed in repair packet with ESI X are calculated based on the repair symbol triples as described in Section B.5.3 using the intermediate symbols C and the LT encoder LTenc[K, C, (d[i], a[i], b[i])].
B.3.3. Transport
This section describes the information exchange between the MSCR encoder/decoder and any transport protocol making use of MSCR forward error correction for file delivery.
The MSCR encoder and decoder for file delivery require the following information from the transport protocol: the file size, F, in bytes, the symbol alignment parameter, A, the symbol size, T, in bytes, which is a multiple of A, the number of source blocks, Z, the number of sub-blocks in each source block, N. The MSCR encoder for file delivery additionally requires the file to be encoded, F bytes.
The MSCR encoder supplies the transport protocol with encoding packet information comprising, for each packet, the SBN, the ESI and the encoding symbol(s). The transport protocol might communicate this information transparently to the MSCR decoder.
B.3.4. Details OF Specific Examples for Parameters
B.3.4.1 Parameter Derivation Algorithm
This section provides examples for the derivation of the four transport parameters, A, T, Z and N that provide good results. These are based on the following input parameters:
Based on the above inputs, the transport parameters T, Z and N are calculated as follows:
Let,
G=min{ceil(P·KMIN/F), P/A, GMAX}−the approximate number of symbols per packet
T=floor(P/(A·G))·A
Kt=ceil(F/T)−the total number of symbols in the file
Z=ceil(Kt/KMAX)
N=min{ceil(ceil(Kt/Z)·T/W), T/A}
The values of G and N derived above should be considered as lower bounds. It may be advantageous to increase these values, for example to the nearest power of two. In particular, the above algorithm does not guarantee that the symbol size, T, divides the maximum packet size, P, and so it may not be possible to use the packets of size exactly P. If, instead, G is chosen to be a value which divides P/A, then the symbol size, T, will be a divisor of P and packets of size P can be used.
Suitable values for the input parameters might be W=256 KB, A=4, KMIN=4, and GMAX=1.
B.3.4.2 Examples
The above algorithm leads to transport parameters as shown in
B.4. Streaming
B.4.1. Source Block Construction
A source block is constructed by the transport protocol, for example as defined in this document, making use of the Systematic MSCR Forward Error Correction code. The symbol size, T, to be used for source block construction and the repair symbol construction are provided by the transport protocol. The parameter T might be set so that the number of source symbols in any source block is at most KMAX.
An example of parameters that work well are presented in section B.4.4.
B.4.2. Encoding Packet Construction
As described in B.4.3., each repair packet contains the SBN, ESI, SBL and repair symbol(s). The number of repair symbols contained within a repair packet is computed from the packet length. The ESI values placed into the repair packets and the repair symbol triples used to generate the repair symbols are computed as described in Section B.3.2.2.
B.4.3. Transport
This section describes the information exchange between the MSCR encoder/decoder and any transport protocol making use of MSCR forward error correction for streaming. The MSCR encoder for streaming might use the following information from the transport protocol for each source block: the symbol size, T, in bytes, the number of symbols in the source block, K, the Source Block Number (SBN) and the source symbols to be encoded, K·T bytes. The MSCR encoder supplies the transport protocol with encoding packet information comprising, for each repair packet, the SBN, the ESI, the SBL and the repair symbol(s). The transport protocol might communicate this information transparently to the MSCR decoder.
B.4.4. Selection of Parameters
A number of methods for parameter selection can be used. Some of those are described below in detail.
B.4.4.1 Parameter Derivation Algorithm
This section explains a derivation of the transport parameter T, based on the following input parameters:
A requirement on these inputs is that ceil(B/P)≦KMAX. Based on the above inputs, the transport parameter T is calculated as follows:
Let G=min{max{ceil(P·KMIN/B), floor(Px/Pmax)}, P/A, GMAX}−the number of symbols per SPI
T=floor(P/(A·G))·A
The value of T derived above should be considered as a guide to the actual value of T used. It may be advantageous to ensure that T divides into P, or it may be advantageous to set the value of T smaller to minimize wastage when full size repair symbols are used to recover partial source symbols at the end of lost source packets (as long as the maximum number of source symbols in a source block does not exceed KMAX). Furthermore, the choice of T may depend on the source packet size distribution, e.g., if all source packets are the same size then it is advantageous to choose T so that the actual payload size of a repair packet P′, where P′ is a multiple of T, is equal to (or as few bytes as possible larger than) the number of bytes each source packet occupies in the source block.
Suitable values for the input parameters might be A=16, KMIN=4 and GMAX=4.
B.4.4.2 Examples
The above algorithm leads to transport parameters as shown in
B.5. Systematic Multi-Field MSCR Encoder
B.5.1. Encoding Overview
The systematic MSCR encoder is used to generate repair symbols from a source block that comprises K source symbols.
Symbols are the fundamental data units of the encoding and decoding process. For each source block (sub-block) all symbols (sub-symbols) are the same size. The atomic operation performed on symbols (sub-symbols) for both encoding and decoding is the exclusive-or operation.
The first step of encoding is to generate a number, L>K, of intermediate symbols from the K source symbols. In this step, K source triples (d[0], a[0], b[0]), . . . , (d[K−1], a[K−1], b[K−1]) are generated using the Trip[ ] generator as described in Section B.5.4.4. The K source triples are associated with the K source symbols and are then used to determine the L intermediate symbols C[0], . . . , C[L−1] from the source symbols using an inverse encoding process. This process can be can be realized by a MSCR decoding process.
Certain “pre-coding relationships” preferably hold within the L intermediate symbols. Section B.5.2 describes these relationships and how the intermediate symbols are generated from the source symbols.
Once the intermediate symbols have been generated, repair symbols are produced and one or more repair symbols are placed as a group into a single data packet. Each repair symbol group is associated with an Encoding Symbol ID (ESI) and a number, G, of encoding symbols. The ESI is used to generate a triple of three integers, (d, a, b) for each repair symbol again using the Trip[ ] generator as described in Section B.5.4.4. This is done as described in Sections B.3 and B.4 using the generators described in Section B.5.4. Then, each (d,a,b)-triple is used to generate the corresponding repair symbol from the intermediate symbols using the LTEnc [K, C[0], . . . , C[L−1], (d,a,b)] generator described in Section B.5.4.3.
B.5.2. First Encoding Step: Intermediate Symbol Generation
B.5.2.1 General
The first encoding step is a pre-coding step to generate the L intermediate symbols C[0], . . . , C[L−1] from the source symbols C′[0], . . . , C′[K−1]. The intermediate symbols are uniquely defined by two sets of constraints:
These are defined in Section B.5.2.3. The generation of the L intermediate symbols is then defined in Section 5.2.4.
B.5.2.2 Source Symbol Triples
Each of the K source symbols is associated with a triple (d[i], a[i], b[i]) for 0≦i<K. The source symbol triples are determined using the Triple generator defined in Section B.5.4.4 as:
The pre-coding relationships amongst the L intermediate symbols are defined by expressing the last L−K intermediate symbols in terms of the first K intermediate symbols.
The last L−K intermediate symbols C[K], . . . ,C[L−1] comprise SLDPC symbols and H HDPC symbols The values of S and H are determined from K as described below. Then L=K+S+H.
Let
The S LDPC symbols are defined to be the values of C[K], . . . , C[K+S−1] at the end of the following process:
For the construction of the HHDPC symbols, the system uses the field GF(256). The field can be represented with respect to the irreducible polynomial f=x8+x4+x3+x2+1 over the field GF(2). Let a denote the element x modulo f. As is well-known to those of ordinary skill in the art, the element a is primitive, i.e., the 255 first powers of a coincide with the 255 nonzero elements of GF(256). In one embodiment, the system choose K+S integers a[0], . . . ,a[K+S−1], and denote by β[0], . . . , β[K+S−1] the elements αα[0], . . . ,αα[K+S−1]. Further, we choose H further integers b[0], . . . ,b[H−1] and denote by Γ[0], . . . ,Γ[H−1] the elements αb[0], . . . ,αb[H−1]. Further preferred embodiments of the present invention will specify specific choices for these integers. However, it should be noted that are many equivalent choices of these integers. Let g[i]=i^(floor(i/2)) for all positive integers i. Note that g[i] is the Gray sequence, in which each element differs from the previous one in a single bit position. Furthermore, let g[j,k] denote the jth element,j=0, 1, 2, . . . , of the subsequence of g[i] whose elements have exactly k non-zero bits in their binary representation. As is well-known to those of skill in the art, the sequence g[j,k] has the property that the binary representations of g[j,k] and g[j+1,k] differ in exactly two positions. We denote these positions by p[j,k,1] and p[j,k,2].
The values of the HDPC symbols are defined as the values of C[K+S], . . . , C[L−1] after the following process.
We initialize a symbol U as 0. The size of this symbol is the same as the common size of source, LDPC, and HDPC symbols.
Next, for a variable h ranging from 0 to K+S−2, we perform the following: The variable U is updated as U=U*β[h]^C[h]. At the same time, we set C[K+S+p[j,H′,1]]=C[K+S+p[j,H′,1]]^U, and C[K+S+p[j,H′,2]]=C[K+S+p[j,H′,2]]^U.
In a further step, we transform U into U*β[K+S−1]^C[K+S−1].
Next, for a variable h ranging from 0 to H−1 we update C[K+S+h]=C[K+S+h]^Γ[h]*U. This completes the description of the HDPC coding process.
In a preferred embodiment, the system chooses the following integers a[0], . . . ,a[K+S−1], and b[0], . . . ,b[H−1]: a[0]=a[1]= . . . =a[K+S−1]=1 and b[0]=1, b[1]=2, . . . b[i]=i+1, etc. Advantageously, in this preferred embodiment, the construction of the HDPC symbols can be performed using only the action of the primitive element, α, along with bit-wise exclusive OR operations between symbols. The choice of irreducible polynomial give above admits highly efficient implementation of the action of α, thereby reducing the computational complexity of the HDPC construction algorithm. As will be apparent to those of skill in the art, the construction algorithm described above can easily be adapted to perform the required decoding operations within a multi-stage code decoder, thus realizing the above mentioned reduction in computational complexity at the decoder as well.
B.5.2.4 Intermediate Symbols
B.5.2.4.1 Definitions
Given the K source symbols C′[0], C′[1], . . . , C′[K−1] the L intermediate symbols C′[0], C[1], . . . , C[L−1] are the uniquely defined symbol values that satisfy the following conditions:
This subsection describes a possible method for calculation of the L intermediate symbols C[0], C[1], . . . , C[L−1] satisfying the constraints in B.5.2.4.1
The generator matrix G for a code which generates N output symbols from K input symbols is an N×K matrix over GF(2), where each row corresponds to one of the output symbols and each column to one of the input symbols and where the ith output symbol is equal to the sum of those input symbols whose column contains a non-zero entry in row i.
Then, the L intermediate symbols can be calculated as follows:
Let
The matrix A is depicted in
C=A−1·D
The source triples are generated such that for any K matrix A has full rank and is therefore invertible. This calculation can be realized by applying a MSCR decoding process to the K source symbols C′[0], C′[1], . . . , C′[K−1] to produce the L intermediate symbols C[0], C[1], . . . , C[L−1].
To efficiently generate the intermediate symbols from the source symbols, an efficient decoder implementation such as that described in Section B.6 might be used. The source symbol triples are designed to facilitate efficient decoding of the source symbols using that algorithm.
B.5.3. Second Encoding Step: Chain Reaction Encoding
In the second encoding step, the repair symbol with ESI X is generated by applying the generator LTEnc[K, (C[0], C[1], . . . , C[L−1]), (d, a, b)] defined in Section B.5.4 to the L intermediate symbols C[0], C[1], . . . , C[L−1] using the triple (d, a, b)=Trip[K,X] generated according to Sections B.3.2.2 and B.4.2.
B.5.4. Generators
B.5.4.1 Random Generator
The random number generator Rand[X, i, m] is defined as follows, where X is a non-negative integer, i is a non-negative integer and m is a positive integer and the value produced is an integer between 0 and m−1. Let V0 and V1 be arrays of 256 entries each, where each entry is a 4-byte unsigned integer. Suitable arrays of random numbers are provided in Appendices B.1 and B.2 by way of example only and should not be construed to limit the scope of the invention. Given those assumptions, Rand[X, i, m]=(V0[(X+i) % 256]^V−1[(floor(X/256)+i) % 256]) % m. As used herein, unless otherwise indicated, “random” should be assumed to include “pseudorandom” and “essentially random”.
B.5.4.2 Degree Generator
The degree generator Deg[v] is defined as follows, where v is an integer that is at least 0 and less than 220=1048576.
In
The encoding symbol generator LTEnc[K, (C[0], C[1], . . . , C[L−1]), (d, a, b)] takes the following inputs:
K is the number of source symbols (or sub-symbols) for the source block (sub-block). Let L be derived from K as described in Section B.5.2, and let L′ be the smallest prime integer greater than or equal to L.
(C[0], C[1], . . . , C[L−1]) is the array of L intermediate symbols (sub-symbols) generated as described in Section B.5.2
(d, a, b) is a source triple determined using the Triple generator defined in Section B.5.3.4, whereby d is an integer denoting an encoding symbol degree, a is an integer between 1 and L′−1 inclusive and b is an integer between 0 and L′−1 inclusive.
The encoding symbol generator produces a single encoding symbol as output, according to the following algorithm:
The triple generator Trip[K,X] takes the following inputs:
The output of the triple generator is a triples, (d, a, b) determined as follows:
This section describes an efficient decoding algorithm for the MSCR codes described in this specification. Note that each received encoding symbol can be considered as the value of an equation amongst the intermediate symbols. From these simultaneous equations, and the known pre-coding relationships amongst the intermediate symbols, any algorithm for solving simultaneous equations can successfully decode the intermediate symbols and hence the source symbols. However, the algorithm chosen has a major effect on the computational efficiency of the decoding.
B.6.2 Decoding A Source Block
B.6.2.1 General
It is assumed that the decoder knows the structure of the source block it is to decode, including the symbol size, T, and the number K of symbols in the source block.
From the algorithms described in Sections B.5, the MSCR decoder can calculate the total number L=K+S+H of pre-coding symbols and determine how they were generated from the source block to be decoded. In this description it is assumed that the received encoding symbols for the source block to be decoded are passed to the decoder. Furthermore, for each such encoding symbol it is assumed that the number and set of intermediate symbols whose exclusive-or is equal to the encoding symbol is passed to the decoder. In the case of source symbols, the source symbol triples described in Section B.5.2.2 indicate the number and set of intermediate symbols which sum to give each source symbol.
Let N≧K be the number of received encoding symbols for a source block and let M=S+H+N. The following M×L matrix A can be derived from the information passed to the decoder for the source block to be decoded. Let C be the column vector of the L intermediate symbols, and let D be the column vector of M symbols with values known to the receiver, where the last S+H of the M symbols are zero-valued symbols that correspond to LDPC and HDPC symbols (these are check symbols for the LDPC and HDPC symbols, and not the LDPC and HDPC symbols themselves), and the remaining N of the M symbols are the received encoding symbols for the source block. Then, A is the matrix that satisfies A·C=D, where here · denotes matrix multiplication over G(256). The matrix A has a block structure, as shown in
Decoding a source block is equivalent to decoding C from known A and D. It is clear that C can be decoded if and only if the rank of A over GF(256) is L. Once C has been decoded, missing source symbols can be obtained by using the source symbol triples to determine the number and set of intermediate symbols which are exclusive-ORed to obtain each missing source symbol.
The first step in decoding C is to form a decoding schedule. In this step A is converted, using Gaussian elimination (using row operations and row and column reorderings) and after discarding M−L rows, into the L by L identity matrix. The decoding schedule comprises the sequence of row operations and row and column re-orderings during the Gaussian elimination process, and only depends on A and not on D. The decoding of C from D can take place concurrently with the forming of the decoding schedule, or the decoding can take place afterwards based on the decoding schedule.
The correspondence between the decoding schedule and the decoding of C is as follows. Let c[0]=0, c[1]=1 . . . ,c[L−1]=L−1 and d[0]=0, d[1]=1 . . . ,d[M−1]=M−1 initially.
Each time row i of A is exclusive-ORed into row i′ in the decoding schedule then in the decoding process symbol D[d[i]] is exclusive-ORed into symbol D[d[i′]]. We call this operation a GF(2)-row operation.
Each time a multiple α (for some α in GF(256)) of row i of A is exclusive-ORed into row i′ in the decoding schedule, then in the decoding process symbol α*D[d[i]] is exclusive-ORed into symbol D[d[i′]]. We call this operation a GF(256)-row operation. Note that a GF(2)-row operation is a particular case of a GF(256)-row operation in which the element α is 1.
Each time row i is exchanged with row i′ in the decoding schedule then in the decoding process the value of d[i] is exchanged with the value of d[i′].
Each time column j is exchanged with column j′ in the decoding schedule then in the decoding process the value of c[j] is exchanged with the value of c[j′].
From this correspondence it is clear that the total number of exclusive-ORs of symbols in the decoding of the source block is related to the number of row operations (not exchanges) in the Gaussian elimination. Since A is the L by L identity matrix after the Gaussian elimination and after discarding the last M−L rows, it is clear at the end of successful decoding that the L symbols D[d[0]], D[d[0]], . . . , D[d[L−1]] are the values of the L symbols C[c[0]], C[c[1]], . . . , C[c[L−1]].
The order in which Gaussian elimination is performed to form the decoding schedule has no bearing on whether or not the decoding is successful. However, the speed of the decoding depends heavily on the order in which Gaussian elimination is performed. (Furthermore, maintaining a sparse representation of A is crucial, although this is not described here). It is also clear that it is more efficient to perform GF(2)-row operations rather than GF(256)-row operations. Therefore, when performing the Gaussian elimination, it is better to pivot on rows of the matrix A which with elements taken from the field GF(2). It is also advantageous to leave the elimination of the rows of the matrix corresponding to the HDPC symbols to the end of the Gaussian elimination process. The remainder of this section describes an order in which Gaussian elimination could be performed that is relatively efficient.
B.6.2.2 First Phase
Referring to
The first phase of the Gaussian elimination the matrix X is conceptually partitioned into submatrices. The submatrix sizes are parameterized by non-negative integers i and u which are initialized to 0. The submatrices of X are:
There are at most L steps in the first phase. The phase ends when V either disappears or becomes the zero matrix-. In each step, a row of X is chosen as follows:
If all entries of V are zero then no row is chosen and the first phase ends.
therwise, let r be the minimum integer such that at least one row of X has exactly r ones in V.
After the row is chosen in this step the first row of X that intersects V is exchanged with the chosen row so that the chosen row is the first row that intersects V. The columns of X among those that intersect V are reordered so that one of the r ones in the chosen row appears in the first column of V and so that the remaining r−1 ones appear in the last columns of V. Then, the chosen row is exclusive-ORed into all the other rows of X below the chosen row that have a one in the first column of V. In other words, we perform a GF(2)-row operation in this step. Finally, i is incremented by 1 and u is incremented by r−1, which completes the step.
Let v denote the number of columns of the matrix V at the end of this phase. After permuting the columns of the matrix B so that the columns of V correspond to the last v columns of X, the matrix X will have the form given in
B.6.2.3 Second Phase
We modify the matrix U so it comprises additionally the last v rows of the matrix X, and we replace u accordingly by u+v. The submatrix U is further partitioned into the first i rows, Uupper, and the remaining N+S−i rows, Ulower, as depicted in
After this step, the matrix A has the form given in
B.6.2.4 Third Place
After the second phase the portions of A which need to be zeroed out to finish converting A into the L by L identity matrix are W and all u columns of Uupper, in the case that the first method of zeroing out B1 and B2 has been followed, or W and the last u−s columns of Uupper, in the case that the second method of zeroing out B1 and B2 has been followed. In the former case, since the matrix W is generally of small size, it can be zeroed out using elementary GF(2)-row operations. After this step, the matrix A has the form given in
The number of rows i′ of the remaining submatrix Û is generally much larger than the number of columns u′. There are several methods which may be used to zero out Û efficiently. In one method, the following precomputation matrix U′ is computed based on, the last u rows and columns of A, which we denote Iu and then U′ is used to zero out Û. The u rows of Iu are partitioned into ceil(u/z) groups of z rows each, for some integer z. Then, for each group of z rows all non-zero combinations of the z rows are computed, resulting in 2z−1 rows (this can be done with 2z−z−1 exclusive-ors of rows per group, since the combinations of Hamming weight one that appear in Iu do not need to be recomputed). Thus, the resulting precomputation matrix U′ has ceil(u/z)·2z−1 rows and u columns. Note that U′ is not formally a part of matrix A, but will be used subsequently to zero out Uupper. In a preferred embodiment, z=8.
For each of the i′ rows of Û, for each group of z columns in the Û submatrix of this row, if the set of z column entries in Û are not all zero then the row of the precomputation matrix U′ that matches the pattern in the z columns is exclusive-ORed into the row, thus zeroing out those z columns in the row at the cost of exclusive-oring one row of U′ into the row.
After this phase A is the L by L identity matrix and a complete decoding schedule has been successfully formed. Then, the corresponding decoding comprising exclusive-ORing known encoding symbols can be executed to recover the intermediate symbols based on the decoding schedule.
The triples associated with all source symbols are computed according to B.5.2.2. The triples for received source symbols are used in the decoding. The triples for missing source symbols are used to determine which intermediate symbols need to be exclusive-ORed to recover the missing source symbols.
Multi-Field, Single-Stage Chain Reaction Encoders/Decoders
Multi-field, single-stage (MFSS) codes have useful properties that are disclosed or suggested herein. Novel arrangements for MFSS codes, encoders and decoders are described herein. In one embodiment, data is encoded for transmission from a source to a destination in which each output symbol is generated as a linear combination of one or more of the input symbols with coefficients taken from finite fields and, for each output symbol:
The random process for selecting the degrees of the output symbols may be a process described in Luby I and Luby II in which the degree is selected according to a degree distribution. The random process for selecting the input symbols to associate with each output symbol may be a process described in Luby I and Luby II in which the input symbols are selected randomly and uniformly. As used herein “random” may include “pseudorandom”, “biased random” and the like.
The set of possible finite fields may be the set {GF(2), GF(256)}.
The process for selecting the finite field may be based on a parameter d1, such that for output symbols of degree less than d1, the field GF(2) is chosen for all input symbols in the neighbor set of the output symbol and for output symbols of degree d1 or greater than the field GF(256) is chosen for at least one, some or all of the members of the neighbor set of the output symbol and the field GF(2) is chosen for the remaining elements of the neighbor set, if any.
The process for selecting the finite field element from the selected field may the simple random process in which an element is chosen uniformly at random from amongst the non-zero elements of the field.
A decoder receiving data encoded by an MFSS encoder as described above might decode the output symbols to regenerate the input symbols by forming a matrix representation of the code according to the method described above, this matrix including no static rows and one dynamic row for each output symbol of the code, and then applying Gaussian Elimination to find the inverse of this matrix, ensuring that at each stage of the Gaussian Elimination process pivot rows of minimal degree are chosen.
As will be clear to those of ordinary skill in the art, many of the well-known properties of the codes described in Luby I and Luby TI are equally applicable to the codes described above and in particular the choice of an appropriate degree distribution can ensure that with high probability the Gaussian Elimination process is able to identify a row of remaining degree one and thus the decoding process operates as a chain reaction process as described in Luby I and Luby II.
This MFSS code has several further advantages over codes known in the art. Firstly, the inclusion of elements from the field GF(256) reduces significantly the probability that any given received output symbol is not information additive with respect to previously received output symbols. As a result, the decoding error probability of this code is much lower than previous codes. For example, in some instances, the failure probability of the codes described in Luby I and Luby II is improved upon.
An advantage of this code over other codes based on large fields is that output symbols of low degree will generally be processed first by the Gaussian Elimination process and as a result the inclusion of elements from GF(256) need not be considered until later in the decoding process. Since operations over GF(256) are relatively expensive compared to those over GF(2), this results in greatly reduced computational complexity compared to codes where many or all of the symbols are constructed using elements from GF(256) or other large finite fields.
A further advantage over other codes based on large fields is that for those output symbols generated using the larger field, only one element of the neighbor set has a coefficient which is taken from the larger field and as a result only one operation between a symbol and a finite field element is required for each such output symbol. This results in low overall computational complexity.
It is known that using inner codes and outer codes to encode input symbols using two (or more) coding procedures leads to a simple code scheme that provides benefits often found in more complex codes. With the use of inner codes and outer codes, source symbols are first encoded using one of the codes and the output of the first encoder is provided to a coder that codes according to the other code and that result is output as the output symbols. Using an MFSS is, of course, different from the use of inner/outer codes. For one, the output symbols are derived from neighbor sets of input codes. In many of the embodiments described herein, each output symbol is a linear combination of input symbols. With multi-stage codes, each output symbol might be a linear combination of input symbols and/or redundant and/or intermediate symbols.
Dense Multi-Field Codes and Encoders/Decoders for Such Codes
In a variation of the teachings described above, the matrix representation of the code is a dense matrix. As is well known, error correction codes can be constructed from dense random matrices over finite fields. For example, a generalized matrix may be constructed in which there are no static rows and each dynamic row comprises elements from GF(2q), with each element chosen randomly. A fixed rate code may then be constructed in which each output symbol corresponds to one of the dynamic rows and is generated as the linear combination of those input symbols for which there is a non-zero element in the corresponding column of this row of the matrix, using these elements as coefficients in the linear combination process.
It is well known to those of skill in the art that the probability that a randomly chosen matrix with K rows and K+A columns with coefficients that are independently and randomly chosen from GF(2q) has a rank that is smaller than K is at most 2−qA. Therefore, the decoding error probability of a code with K input and K/R output symbols in which the output symbols are generated independently and randomly from the input symbols using randomly chosen coefficients from GF(2q) is at most 2−qA, if the number of encoded symbols received is K+A.
In the case of q=1, the code described above has the advantage of reasonable computational complexity, since all operations are within the field GF(2) and thus correspond to conventional XOR operations. However, in this case the lower bound on the failure probability of 2−A once A additional symbols have been received is much higher than desirable.
In the case of q=8, the code described above has the advantage of a lower failure probability (bounded by 2−8A for A additional symbols received). However, in this case all operations are within the field GF(256) and are thus relatively computationally expensive.
A further embodiment allows decoding error probabilities close to those achievable using large values of q to be achieved with computational complexity close to that achievable with small values of q. In this embodiment, output symbols are generated as linear combinations of input symbols with coefficients taken from either GF(2q) or GF(2q) where p<q. In one specific embodiment, exactly (K−2p/q)/R output symbols are generated using coefficients from GF(2q) and the remaining 2p/(qR) output symbols are generated using coefficients from GF(2q).
Data received at a destination can be decoded by determining the linear relationships between received output symbols and the input symbols of the code and solving this set of linear relationships to determine the input symbols.
The decoding error probability of this code is at most that of the code in which all coefficients are chosen from the field GF(2p) and may be significantly lower depending on the number of symbols generated using coefficients from the larger field GF(2q). However, since most of the output symbols are generated using coefficients from GF(2p), the computational complexity of encoding is only slightly greater than that of a code in which all symbols are generated using coefficients from GF(2p). Furthermore, the method of decoding may be so arranged that symbols generated with coefficients form GF(2p) are processed first and thus the majority of the decoding operations are performed with operations exclusively in GF(2p). As a result, the computational complexity of the decoding method is similarly close to that for codes constructed using only GF(2p). In a particular preferred embodiment, p=1 and q=8.
Some Properties of Some Multi-Field Codes
In most of the examples described above, the input and output symbols encode for the same number of bits and each output symbol is placed in one packet (a packet being a unit of transport that is either received in its entirety or lost in its entirety). In some embodiments, the communications system is modified so that each packet contains several output symbols. The size of an output symbol value is then set to a size determined by the size of the input symbol values in the initial splitting of the file or blocks of the stream into input symbols, based on a number of factors. The decoding process remains essentially unchanged, except that output symbols arrive in bunches as each packet is received.
The setting of input symbol and output symbol sizes is usually dictated by the size of the file or block of the stream and the communication system over which the output symbols are to be transmitted. For example, if a communication system groups bits of data into packets of a defined size or groups bits in other ways, the design of symbol sizes begins with the packet or grouping size. From there, a designer would determine how many output symbols will be carried in one packet or group and that determines the output symbol size. For simplicity, the designer would likely set the input symbol size equal to the output symbol size, but if the input data makes a different input symbol size more convenient, it can be used.
The above-described encoding process produces a stream of packets containing output symbols based on the original file or block of the stream. Each output symbol in the stream is generated independently of all other output symbols, and there is no lower or upper bound on the number of output symbols that can be created. A key is associated with each output symbol. That key, and some contents of the input file or block of the stream, determines the value of the output symbol. Consecutively generated output symbols need not have consecutive keys, and in some applications it would be preferable to randomly generate the sequence of keys, or pseudorandomly generate the sequence.
Multi-stage decoding has a property that a block of K equal-sized input symbols can be recovered from K+A output symbols on average, with very high probability, where A is small compared to K. For example, in the preferred embodiment first described above, when K=100,
Since the particular output symbols are generated in a random or pseudorandom order, and the loss of particular output symbols in transit is generally unrelated to the values of the symbols, there is only a small variance in the actual number of output symbols needed to recover the input file or block. In many cases, where a particular collection of K+A output symbols are not enough to decode the a block, the block is still recoverable if the receiver can receive more output symbols from one or more sources.
Because the number of output symbols is only limited by the resolution of I, well more than K+A output symbols can be generated. For example, if I is a 32-bit number, 4 billion different output symbols could be generated, whereas the file or block of the stream could include K=50,000 input symbols. In some applications, only a small number of those 4 billion output symbols may be generated and transmitted and it is a near certainty that an input file or block of a stream can be recovered with a very small fraction of the possible output symbols and an excellent probability that the input file or block can be recovered with slightly more than K output symbols (assuming that the input symbol size is the same as the output symbol size).
In some applications, it may be acceptable to not be able to decode all of the input symbols, or to be able to decode all of input symbols, but with a relatively low probability. In such applications, a receiver can stop attempting to decode all of the input symbols after receiving K+A output symbols. Or, the receiver can stop receiving output symbols after receiving less than K+A output symbols. In some applications, the receiver may even only receive K or less output symbols. Thus, it is to be understood that in some embodiments of the present invention, the desired degree of accuracy need not be complete recovery of all the input symbols.
Further, in some applications where incomplete recovery is acceptable, the data can be encoded such that all of the input symbols cannot be recovered, or such that complete recovery of the input symbols would require reception of many more output symbols than the number of input symbols. Such an encoding would generally require less computational expense, and may thus be an acceptable way to decrease the computational expense of encoding.
It is to be understood that the various functional blocks in the above-described figures may be implemented by a combination of hardware and/or software, and that in specific implementations some or all of the functionality of some of the blocks may be combined. Similarly, it is also to be understood that the various methods described herein may be implemented by a combination of hardware and/or software.
The above description is illustrative and not restrictive. Many variations of the invention will become apparent to those of skill in the art upon review of this disclosure. The scope of the invention should, therefore, be determined not with reference to the above description, but instead should be determined with reference to the appended claims along with their full scope of equivalents.
This application claims priority from and is a non-provisional of U.S. Provisional Patent Application No. 60/775,528 filed Feb. 21, 2006. The following references are include here and are incorporated by reference for all purposes: U.S. Pat. No. 6,307,487 entitled “Information Additive Code Generator and Decoder for Communication Systems” issued to Luby (hereinafter “Luby I”); U.S. Pat. No. 6,320,520 issued to Luby et al. entitled “Information Additive Group Code Generator and Decoder for Communication Systems” (hereinafter “Luby II”); U.S. Pat. No. 7,068,729 issued to Shokrollahi et al. entitled “Multi-Stage Code Generator and Decoder for Communication Systems” (hereinafter “Shokrollahi I”); U.S. Pat. No. 6,909,383 entitled “Systematic Encoding and Decoding of Chain Reaction Codes” issued to Shokrollahi et al. (hereinafter “Shokrollahi II”); U.S. Pat. No. 6,856,263 entitled “Systems and Processes for Decoding Chain Reaction Codes through Inactivation,” issued to Shokrollahi et al. (hereinafter “Shokrollahi III”); and U.S. Patent Publication No. 2005/0219070 A1 entitled “Protection of Data from Erasures Using Subsymbol Based Codes” by Shokrollahi, filed Dec. 1, 2004 (hereinafter “Shokrollahi IV”).
Number | Name | Date | Kind |
---|---|---|---|
3909721 | Bussgang et al. | Sep 1975 | A |
4365338 | McRae et al. | Dec 1982 | A |
4589112 | Karim | May 1986 | A |
4901319 | Ross | Feb 1990 | A |
5136592 | Weng | Aug 1992 | A |
5153591 | Clark | Oct 1992 | A |
5329369 | Willis et al. | Jul 1994 | A |
5331320 | Cideciyan et al. | Jul 1994 | A |
5371532 | Gelman et al. | Dec 1994 | A |
5372532 | Robertson, Jr. | Dec 1994 | A |
5379297 | Glover et al. | Jan 1995 | A |
5421031 | De Bey | May 1995 | A |
5425050 | Schreiber et al. | Jun 1995 | A |
5432787 | Chethik | Jul 1995 | A |
5455823 | Noreen et al. | Oct 1995 | A |
5465318 | Sejnoha | Nov 1995 | A |
5517508 | Scott | May 1996 | A |
5524025 | Lawrence et al. | Jun 1996 | A |
5566208 | Balakrishnan | Oct 1996 | A |
5568614 | Mendelson et al. | Oct 1996 | A |
5583784 | Kapust et al. | Dec 1996 | A |
5608738 | Matsushita | Mar 1997 | A |
5617541 | Albanese et al. | Apr 1997 | A |
5642365 | Murakami et al. | Jun 1997 | A |
5659614 | Bailey, III | Aug 1997 | A |
5699473 | Kim | Dec 1997 | A |
5701582 | Debey | Dec 1997 | A |
5751336 | Aggarwal et al. | May 1998 | A |
5754563 | White | May 1998 | A |
5757415 | Asamizuya et al. | May 1998 | A |
5802394 | Baird et al. | Sep 1998 | A |
5805825 | Danneels et al. | Sep 1998 | A |
5835165 | Keate et al. | Nov 1998 | A |
5844636 | Joseph et al. | Dec 1998 | A |
5852565 | Demos | Dec 1998 | A |
5870412 | Schuster et al. | Feb 1999 | A |
5903775 | Murray | May 1999 | A |
5917852 | Butterfield et al. | Jun 1999 | A |
5926205 | Krause et al. | Jul 1999 | A |
5933056 | Rothenberg | Aug 1999 | A |
5936659 | Viswanathan | Aug 1999 | A |
5936949 | Pasternak et al. | Aug 1999 | A |
5953537 | Balicki et al. | Sep 1999 | A |
5970098 | Herzberg | Oct 1999 | A |
5983383 | Wolf | Nov 1999 | A |
5993056 | Vaman et al. | Nov 1999 | A |
6005477 | Deck et al. | Dec 1999 | A |
6011590 | Saukkonen | Jan 2000 | A |
6012159 | Fischer et al. | Jan 2000 | A |
6014706 | Cannon et al. | Jan 2000 | A |
6018359 | Kermode et al. | Jan 2000 | A |
6041001 | Estakhri | Mar 2000 | A |
6044485 | Dent et al. | Mar 2000 | A |
6061820 | Nakakita et al. | May 2000 | A |
6073250 | Luby et al. | Jun 2000 | A |
6079041 | Kunisa et al. | Jun 2000 | A |
6079042 | Vaman et al. | Jun 2000 | A |
6081907 | Witty et al. | Jun 2000 | A |
6081909 | Luby et al. | Jun 2000 | A |
6081918 | Spielman | Jun 2000 | A |
6088330 | Bruck et al. | Jul 2000 | A |
6097320 | Kuki et al. | Aug 2000 | A |
6134596 | Bolosky et al. | Oct 2000 | A |
6141053 | Saukkonen | Oct 2000 | A |
6141787 | Kunisa et al. | Oct 2000 | A |
6141788 | Rosenberg et al. | Oct 2000 | A |
6154452 | Marko et al. | Nov 2000 | A |
6163870 | Luby et al. | Dec 2000 | A |
6166544 | Debbins et al. | Dec 2000 | A |
6175944 | Urbanke et al. | Jan 2001 | B1 |
6178536 | Sorkin | Jan 2001 | B1 |
6185265 | Campanella | Feb 2001 | B1 |
6195777 | Luby et al. | Feb 2001 | B1 |
6223324 | Sinha et al. | Apr 2001 | B1 |
6226259 | Piret | May 2001 | B1 |
6226301 | Cheng et al. | May 2001 | B1 |
6229824 | Marko | May 2001 | B1 |
6243846 | Schuster et al. | Jun 2001 | B1 |
6272658 | Steele et al. | Aug 2001 | B1 |
6278716 | Rubenstein et al. | Aug 2001 | B1 |
6298462 | Yi | Oct 2001 | B1 |
6307487 | Luby | Oct 2001 | B1 |
6314289 | Eberlein et al. | Nov 2001 | B1 |
6320520 | Luby | Nov 2001 | B1 |
6332163 | Bowman-Amuah | Dec 2001 | B1 |
6333926 | Van Heeswyk et al. | Dec 2001 | B1 |
6373406 | Luby | Apr 2002 | B2 |
6393065 | Piret et al. | May 2002 | B1 |
6411223 | Haken et al. | Jun 2002 | B1 |
6415326 | Gupta et al. | Jul 2002 | B1 |
6420982 | Brown | Jul 2002 | B1 |
6421387 | Rhee | Jul 2002 | B1 |
6430233 | Dillon et al. | Aug 2002 | B1 |
6445717 | Gibson et al. | Sep 2002 | B1 |
6459811 | Hurst, Jr. | Oct 2002 | B1 |
6466698 | Creusere | Oct 2002 | B1 |
6473010 | Vityaev et al. | Oct 2002 | B1 |
6486803 | Luby et al. | Nov 2002 | B1 |
6487692 | Morelos-Zaragoza | Nov 2002 | B1 |
6496980 | Tillman et al. | Dec 2002 | B1 |
6497479 | Stoffel et al. | Dec 2002 | B1 |
6510177 | De et al. | Jan 2003 | B1 |
6523147 | Kroeger et al. | Feb 2003 | B1 |
6535920 | Parry et al. | Mar 2003 | B1 |
6577599 | Gupta et al. | Jun 2003 | B1 |
6584543 | Williams et al. | Jun 2003 | B2 |
6609223 | Wolfgang | Aug 2003 | B1 |
6614366 | Luby | Sep 2003 | B2 |
6618451 | Gonikberg | Sep 2003 | B1 |
6631172 | Shokrollahi et al. | Oct 2003 | B1 |
6633856 | Richardson et al. | Oct 2003 | B2 |
6641366 | Nordhoff | Nov 2003 | B2 |
6643332 | Morelos-Zaragoza et al. | Nov 2003 | B1 |
6677864 | Khayrallah | Jan 2004 | B2 |
6678855 | Gemmell | Jan 2004 | B1 |
6694476 | Sridharan et al. | Feb 2004 | B1 |
6704370 | Chheda et al. | Mar 2004 | B1 |
6732325 | Tash et al. | May 2004 | B1 |
6742154 | Barnard | May 2004 | B1 |
6748441 | Gemmell | Jun 2004 | B1 |
6751772 | Kim et al. | Jun 2004 | B1 |
6765866 | Wyatt | Jul 2004 | B1 |
6804202 | Hwang | Oct 2004 | B1 |
6810499 | Sridharan et al. | Oct 2004 | B2 |
6820221 | Fleming | Nov 2004 | B2 |
6831172 | Barbucci et al. | Dec 2004 | B1 |
6849803 | Gretz | Feb 2005 | B1 |
6850736 | McCune, Jr. | Feb 2005 | B2 |
6856263 | Shokrollahi et al. | Feb 2005 | B2 |
6868083 | Apostolopoulos et al. | Mar 2005 | B2 |
6876623 | Lou et al. | Apr 2005 | B1 |
6882618 | Sakoda et al. | Apr 2005 | B1 |
6895547 | Eleftheriou et al. | May 2005 | B2 |
6909383 | Shokrollahi et al. | Jun 2005 | B2 |
6928603 | Castagna et al. | Aug 2005 | B1 |
6937618 | Noda et al. | Aug 2005 | B1 |
6956875 | Kapadia et al. | Oct 2005 | B2 |
6965636 | DesJardins et al. | Nov 2005 | B1 |
6985459 | Dickson | Jan 2006 | B2 |
6995692 | Yokota et al. | Feb 2006 | B2 |
7010052 | Dill et al. | Mar 2006 | B2 |
7030785 | Shokrollahi et al. | Apr 2006 | B2 |
7031257 | Lu et al. | Apr 2006 | B1 |
7057534 | Luby | Jun 2006 | B2 |
7068681 | Chang et al. | Jun 2006 | B2 |
7068729 | Shokrollahi et al. | Jun 2006 | B2 |
7072971 | Lassen et al. | Jul 2006 | B2 |
7073191 | Srikantan et al. | Jul 2006 | B2 |
7100188 | Hejna, Jr. | Aug 2006 | B2 |
7110412 | Costa et al. | Sep 2006 | B2 |
7139660 | Sarkar et al. | Nov 2006 | B2 |
7139960 | Shokrollahi | Nov 2006 | B2 |
7143433 | Duan et al. | Nov 2006 | B1 |
7151754 | Boyce et al. | Dec 2006 | B1 |
7154951 | Wang | Dec 2006 | B2 |
7164370 | Mishra | Jan 2007 | B1 |
7164882 | Poltorak | Jan 2007 | B2 |
7168030 | Ariyoshi | Jan 2007 | B2 |
7219289 | Dickson | May 2007 | B2 |
7231404 | Paila et al. | Jun 2007 | B2 |
7233264 | Luby | Jun 2007 | B2 |
7240236 | Cutts et al. | Jul 2007 | B2 |
7240358 | Horn et al. | Jul 2007 | B2 |
7243285 | Foisy et al. | Jul 2007 | B2 |
7249291 | Rasmussen et al. | Jul 2007 | B2 |
7254754 | Hetzler et al. | Aug 2007 | B2 |
7257764 | Suzuki et al. | Aug 2007 | B2 |
7265688 | Shokrollahi et al. | Sep 2007 | B2 |
7293222 | Shokrollahi et al. | Nov 2007 | B2 |
7295573 | Yi et al. | Nov 2007 | B2 |
7304990 | Rajwan | Dec 2007 | B2 |
7318180 | Starr | Jan 2008 | B2 |
7320099 | Miura et al. | Jan 2008 | B2 |
7363048 | Cheng et al. | Apr 2008 | B2 |
7391717 | Klemets et al. | Jun 2008 | B2 |
7394407 | Shokrollahi et al. | Jul 2008 | B2 |
7398454 | Cai et al. | Jul 2008 | B2 |
7409626 | Schelstraete | Aug 2008 | B1 |
7412641 | Shokrollahi | Aug 2008 | B2 |
7418651 | Luby et al. | Aug 2008 | B2 |
7451377 | Shokrollahi | Nov 2008 | B2 |
7483447 | Chang et al. | Jan 2009 | B2 |
7483489 | Gentric et al. | Jan 2009 | B2 |
7512697 | Lassen et al. | Mar 2009 | B2 |
7525994 | Scholte | Apr 2009 | B2 |
7529806 | Shteyn | May 2009 | B1 |
7532132 | Shokrollahi et al. | May 2009 | B2 |
7555006 | Wolfe et al. | Jun 2009 | B2 |
7559004 | Chang et al. | Jul 2009 | B1 |
7570665 | Ertel et al. | Aug 2009 | B2 |
7574706 | Meulemans et al. | Aug 2009 | B2 |
7590118 | Giesberts et al. | Sep 2009 | B2 |
7597423 | Silverbrook | Oct 2009 | B2 |
7613183 | Brewer et al. | Nov 2009 | B1 |
7633413 | Shokrollahi et al. | Dec 2009 | B2 |
7633970 | Van Kampen et al. | Dec 2009 | B2 |
7644335 | Luby et al. | Jan 2010 | B2 |
7650036 | Lei et al. | Jan 2010 | B2 |
7668198 | Yi et al. | Feb 2010 | B2 |
7676735 | Luby et al. | Mar 2010 | B2 |
7720096 | Klemets | May 2010 | B2 |
7720174 | Shokrollahi et al. | May 2010 | B2 |
7721184 | Luby et al. | May 2010 | B2 |
7812743 | Luby | Oct 2010 | B2 |
7831896 | Amram et al. | Nov 2010 | B2 |
7924913 | Sullivan et al. | Apr 2011 | B2 |
7956772 | Shokrollahi et al. | Jun 2011 | B2 |
7961700 | Malladi et al. | Jun 2011 | B2 |
7971129 | Watson et al. | Jun 2011 | B2 |
7979769 | Lee et al. | Jul 2011 | B2 |
8027328 | Yang et al. | Sep 2011 | B2 |
8028322 | Riedl et al. | Sep 2011 | B2 |
8081716 | Kang et al. | Dec 2011 | B2 |
8135073 | Shen | Mar 2012 | B2 |
8185794 | Lohmar et al. | May 2012 | B2 |
8185809 | Luby et al. | May 2012 | B2 |
RE43741 | Shokrollahi et al. | Oct 2012 | E |
8301725 | Biderman et al. | Oct 2012 | B2 |
8327403 | Chilvers et al. | Dec 2012 | B1 |
8340133 | Kim et al. | Dec 2012 | B2 |
8422474 | Park et al. | Apr 2013 | B2 |
8462643 | Walton et al. | Jun 2013 | B2 |
8544043 | Parekh et al. | Sep 2013 | B2 |
8572646 | Haberman et al. | Oct 2013 | B2 |
8615023 | Oh et al. | Dec 2013 | B2 |
8638796 | Dan et al. | Jan 2014 | B2 |
8713624 | Harvey et al. | Apr 2014 | B1 |
8737421 | Zhang et al. | May 2014 | B2 |
8812735 | Igarashi | Aug 2014 | B2 |
20010015944 | Takahashi et al. | Aug 2001 | A1 |
20010033586 | Takashimizu et al. | Oct 2001 | A1 |
20020009137 | Nelson et al. | Jan 2002 | A1 |
20020053062 | Szymanski | May 2002 | A1 |
20020083345 | Halliday et al. | Jun 2002 | A1 |
20020085013 | Lippincott | Jul 2002 | A1 |
20020133247 | Smith et al. | Sep 2002 | A1 |
20020141433 | Kwon et al. | Oct 2002 | A1 |
20020143953 | Aiken | Oct 2002 | A1 |
20020191116 | Kessler et al. | Dec 2002 | A1 |
20030005386 | Bhatt et al. | Jan 2003 | A1 |
20030037299 | Smith | Feb 2003 | A1 |
20030086515 | Trans et al. | May 2003 | A1 |
20030101408 | Martinian et al. | May 2003 | A1 |
20030106014 | Dohmen et al. | Jun 2003 | A1 |
20030138043 | Hannuksela | Jul 2003 | A1 |
20030194211 | Abecassis | Oct 2003 | A1 |
20030207696 | Willenegger et al. | Nov 2003 | A1 |
20030224773 | Deeds | Dec 2003 | A1 |
20040015768 | Bordes et al. | Jan 2004 | A1 |
20040031054 | Dankworth et al. | Feb 2004 | A1 |
20040049793 | Chou | Mar 2004 | A1 |
20040066854 | Hannuksela | Apr 2004 | A1 |
20040081106 | Bruhn | Apr 2004 | A1 |
20040096110 | Yogeshwar et al. | May 2004 | A1 |
20040117716 | Shen | Jun 2004 | A1 |
20040151109 | Batra et al. | Aug 2004 | A1 |
20040162071 | Grilli et al. | Aug 2004 | A1 |
20040207548 | Kilbank | Oct 2004 | A1 |
20040231004 | Seo | Nov 2004 | A1 |
20040240382 | Ido et al. | Dec 2004 | A1 |
20040255328 | Baldwin et al. | Dec 2004 | A1 |
20050018635 | Proctor | Jan 2005 | A1 |
20050028067 | Weirauch | Feb 2005 | A1 |
20050041736 | Butler-Smith et al. | Feb 2005 | A1 |
20050071491 | Seo | Mar 2005 | A1 |
20050091697 | Tanaka et al. | Apr 2005 | A1 |
20050097213 | Barrett et al. | May 2005 | A1 |
20050102371 | Aksu | May 2005 | A1 |
20050105371 | Johnson et al. | May 2005 | A1 |
20050123058 | Greenbaum et al. | Jun 2005 | A1 |
20050138286 | Franklin et al. | Jun 2005 | A1 |
20050160272 | Teppler | Jul 2005 | A1 |
20050163468 | Takahashi et al. | Jul 2005 | A1 |
20050169379 | Shin et al. | Aug 2005 | A1 |
20050180415 | Cheung et al. | Aug 2005 | A1 |
20050193309 | Grilli et al. | Sep 2005 | A1 |
20050195752 | Amin et al. | Sep 2005 | A1 |
20050195899 | Han | Sep 2005 | A1 |
20050195900 | Han | Sep 2005 | A1 |
20050207392 | Sivalingham et al. | Sep 2005 | A1 |
20050216472 | Leon et al. | Sep 2005 | A1 |
20050216951 | MacInnis | Sep 2005 | A1 |
20050219070 | Shokrollahi | Oct 2005 | A1 |
20050254575 | Hannuksela et al. | Nov 2005 | A1 |
20060015568 | Walsh et al. | Jan 2006 | A1 |
20060020796 | Aura et al. | Jan 2006 | A1 |
20060031738 | Fay et al. | Feb 2006 | A1 |
20060037057 | Xu | Feb 2006 | A1 |
20060093634 | Lutz et al. | May 2006 | A1 |
20060107174 | Heise | May 2006 | A1 |
20060109805 | Malamal Vadakital et al. | May 2006 | A1 |
20060120464 | Hannuksela | Jun 2006 | A1 |
20060193524 | Tarumoto et al. | Aug 2006 | A1 |
20060212444 | Handman et al. | Sep 2006 | A1 |
20060212782 | Li | Sep 2006 | A1 |
20060229075 | Kim et al. | Oct 2006 | A1 |
20060244824 | Debey | Nov 2006 | A1 |
20060244865 | Simon | Nov 2006 | A1 |
20060248195 | Toumura et al. | Nov 2006 | A1 |
20060256851 | Wang et al. | Nov 2006 | A1 |
20060262856 | Wu et al. | Nov 2006 | A1 |
20060279437 | Luby et al. | Dec 2006 | A1 |
20070002953 | Kusunoki | Jan 2007 | A1 |
20070006274 | Paila et al. | Jan 2007 | A1 |
20070016594 | Visharam et al. | Jan 2007 | A1 |
20070022215 | Singer et al. | Jan 2007 | A1 |
20070028099 | Entin et al. | Feb 2007 | A1 |
20070078876 | Hayashi et al. | Apr 2007 | A1 |
20070081562 | Ma | Apr 2007 | A1 |
20070081586 | Raveendran et al. | Apr 2007 | A1 |
20070110074 | Bradley et al. | May 2007 | A1 |
20070127576 | Henocq et al. | Jun 2007 | A1 |
20070134005 | Myong et al. | Jun 2007 | A1 |
20070140369 | Limberg et al. | Jun 2007 | A1 |
20070157267 | Lopez-Estrada | Jul 2007 | A1 |
20070162568 | Gupta et al. | Jul 2007 | A1 |
20070162611 | Yu et al. | Jul 2007 | A1 |
20070176800 | Rijavec | Aug 2007 | A1 |
20070177811 | Yang et al. | Aug 2007 | A1 |
20070185973 | Wayda et al. | Aug 2007 | A1 |
20070200949 | Walker et al. | Aug 2007 | A1 |
20070201549 | Hannuksela et al. | Aug 2007 | A1 |
20070204196 | Watson et al. | Aug 2007 | A1 |
20070230568 | Eleftheriadis et al. | Oct 2007 | A1 |
20070233784 | O'Rourke et al. | Oct 2007 | A1 |
20070255844 | Shen et al. | Nov 2007 | A1 |
20070277209 | Yousef | Nov 2007 | A1 |
20070300127 | Watson et al. | Dec 2007 | A1 |
20080010153 | Pugh-O'Connor et al. | Jan 2008 | A1 |
20080034273 | Luby | Feb 2008 | A1 |
20080052753 | Huang et al. | Feb 2008 | A1 |
20080058958 | Cheng | Mar 2008 | A1 |
20080059532 | Kazmi et al. | Mar 2008 | A1 |
20080066136 | Dorai et al. | Mar 2008 | A1 |
20080075172 | Koto | Mar 2008 | A1 |
20080086751 | Horn et al. | Apr 2008 | A1 |
20080101478 | Kusunoki | May 2008 | A1 |
20080134005 | Izzat et al. | Jun 2008 | A1 |
20080152241 | Itoi et al. | Jun 2008 | A1 |
20080168133 | Osborne | Jul 2008 | A1 |
20080168516 | Flick et al. | Jul 2008 | A1 |
20080170564 | Shi et al. | Jul 2008 | A1 |
20080170806 | Kim | Jul 2008 | A1 |
20080172430 | Thorstensen | Jul 2008 | A1 |
20080172712 | Munetsugu | Jul 2008 | A1 |
20080181296 | Tian et al. | Jul 2008 | A1 |
20080189419 | Girle et al. | Aug 2008 | A1 |
20080192818 | DiPietro et al. | Aug 2008 | A1 |
20080215317 | Fejzo | Sep 2008 | A1 |
20080232357 | Chen | Sep 2008 | A1 |
20080243918 | Holtman | Oct 2008 | A1 |
20080256418 | Luby et al. | Oct 2008 | A1 |
20080281943 | Shapiro | Nov 2008 | A1 |
20080285556 | Park et al. | Nov 2008 | A1 |
20080303893 | Kim et al. | Dec 2008 | A1 |
20080303896 | Lipton et al. | Dec 2008 | A1 |
20080309525 | Shokrollahi et al. | Dec 2008 | A1 |
20080313191 | Bouazizi | Dec 2008 | A1 |
20090003439 | Wang et al. | Jan 2009 | A1 |
20090019229 | Morrow et al. | Jan 2009 | A1 |
20090031199 | Luby et al. | Jan 2009 | A1 |
20090043906 | Hurst et al. | Feb 2009 | A1 |
20090055705 | Gao | Feb 2009 | A1 |
20090067551 | Chen et al. | Mar 2009 | A1 |
20090083806 | Barrett et al. | Mar 2009 | A1 |
20090089445 | Deshpande | Apr 2009 | A1 |
20090092138 | Joo et al. | Apr 2009 | A1 |
20090100496 | Bechtolsheim et al. | Apr 2009 | A1 |
20090103523 | Katis et al. | Apr 2009 | A1 |
20090106356 | Brase et al. | Apr 2009 | A1 |
20090125636 | Li et al. | May 2009 | A1 |
20090150557 | Wormley et al. | Jun 2009 | A1 |
20090158114 | Shokrollahi | Jun 2009 | A1 |
20090164653 | Mandyam et al. | Jun 2009 | A1 |
20090189792 | Shokrollahi et al. | Jul 2009 | A1 |
20090195640 | Kim et al. | Aug 2009 | A1 |
20090201990 | Leprovost et al. | Aug 2009 | A1 |
20090204877 | Betts | Aug 2009 | A1 |
20090210547 | Lassen et al. | Aug 2009 | A1 |
20090222873 | Einarsson | Sep 2009 | A1 |
20090248697 | Richardson et al. | Oct 2009 | A1 |
20090257508 | Aggarwal et al. | Oct 2009 | A1 |
20090287841 | Chapweske et al. | Nov 2009 | A1 |
20090297123 | Virdi et al. | Dec 2009 | A1 |
20090300203 | Virdi et al. | Dec 2009 | A1 |
20090300204 | Zhang et al. | Dec 2009 | A1 |
20090307565 | Luby et al. | Dec 2009 | A1 |
20090319563 | Schnell | Dec 2009 | A1 |
20090328228 | Schnell | Dec 2009 | A1 |
20100011061 | Hudson et al. | Jan 2010 | A1 |
20100011117 | Hristodorescu et al. | Jan 2010 | A1 |
20100011274 | Stockhammer et al. | Jan 2010 | A1 |
20100020871 | Hannuksela et al. | Jan 2010 | A1 |
20100023525 | Westerlund et al. | Jan 2010 | A1 |
20100046906 | Kanamori et al. | Feb 2010 | A1 |
20100049865 | Hannuksela et al. | Feb 2010 | A1 |
20100061444 | Wilkins et al. | Mar 2010 | A1 |
20100067495 | Lee et al. | Mar 2010 | A1 |
20100103001 | Shokrollahi et al. | Apr 2010 | A1 |
20100131671 | Kohli et al. | May 2010 | A1 |
20100153578 | Van Gassel et al. | Jun 2010 | A1 |
20100165077 | Yin et al. | Jul 2010 | A1 |
20100174823 | Huang | Jul 2010 | A1 |
20100189131 | Branam et al. | Jul 2010 | A1 |
20100198982 | Fernandez | Aug 2010 | A1 |
20100211690 | Pakzad et al. | Aug 2010 | A1 |
20100223533 | Stockhammer et al. | Sep 2010 | A1 |
20100235472 | Sood et al. | Sep 2010 | A1 |
20100235528 | Bocharov et al. | Sep 2010 | A1 |
20100257051 | Fernandez | Oct 2010 | A1 |
20100318632 | Yoo et al. | Dec 2010 | A1 |
20110019769 | Shokrollahi et al. | Jan 2011 | A1 |
20110055881 | Yu et al. | Mar 2011 | A1 |
20110083144 | Bocharov et al. | Apr 2011 | A1 |
20110096828 | Chen et al. | Apr 2011 | A1 |
20110103519 | Shokrollahi et al. | May 2011 | A1 |
20110119394 | Wang et al. | May 2011 | A1 |
20110119396 | Kwon et al. | May 2011 | A1 |
20110216541 | Inoue et al. | Sep 2011 | A1 |
20110231519 | Luby et al. | Sep 2011 | A1 |
20110231569 | Luby et al. | Sep 2011 | A1 |
20110238789 | Luby et al. | Sep 2011 | A1 |
20110239078 | Luby et al. | Sep 2011 | A1 |
20110258510 | Watson et al. | Oct 2011 | A1 |
20110268178 | Park et al. | Nov 2011 | A1 |
20110280311 | Chen et al. | Nov 2011 | A1 |
20110280316 | Chen et al. | Nov 2011 | A1 |
20110299629 | Luby et al. | Dec 2011 | A1 |
20110307545 | Bouazizi | Dec 2011 | A1 |
20110307581 | Furbeck et al. | Dec 2011 | A1 |
20120013746 | Chen et al. | Jan 2012 | A1 |
20120016965 | Chen et al. | Jan 2012 | A1 |
20120020413 | Chen et al. | Jan 2012 | A1 |
20120023249 | Chen et al. | Jan 2012 | A1 |
20120023254 | Park et al. | Jan 2012 | A1 |
20120033730 | Lee | Feb 2012 | A1 |
20120042050 | Chen et al. | Feb 2012 | A1 |
20120042089 | Chen et al. | Feb 2012 | A1 |
20120042090 | Chen et al. | Feb 2012 | A1 |
20120047280 | Park et al. | Feb 2012 | A1 |
20120099593 | Luby | Apr 2012 | A1 |
20120151302 | Luby et al. | Jun 2012 | A1 |
20120185530 | Reza | Jul 2012 | A1 |
20120202535 | Chaddha et al. | Aug 2012 | A1 |
20120207068 | Watson et al. | Aug 2012 | A1 |
20120208580 | Luby et al. | Aug 2012 | A1 |
20120210190 | Luby et al. | Aug 2012 | A1 |
20120317305 | Einarsson et al. | Dec 2012 | A1 |
20130002483 | Rowitch et al. | Jan 2013 | A1 |
20130007223 | Luby et al. | Jan 2013 | A1 |
20130067295 | Luby et al. | Mar 2013 | A1 |
20130091251 | Walker et al. | Apr 2013 | A1 |
20130246643 | Luby et al. | Sep 2013 | A1 |
20130254634 | Luby et al. | Sep 2013 | A1 |
20130287023 | Bims | Oct 2013 | A1 |
20140009578 | Chen et al. | Jan 2014 | A1 |
20140380113 | Luby | Dec 2014 | A1 |
Number | Date | Country |
---|---|---|
1338839 | Mar 2002 | CN |
1425228 | Jun 2003 | CN |
1481643 | Mar 2004 | CN |
1708934 | Dec 2005 | CN |
1714577 | Dec 2005 | CN |
1792056 | Jun 2006 | CN |
1806392 | Jul 2006 | CN |
1819661 | Aug 2006 | CN |
1868157 | Nov 2006 | CN |
101390399 | Mar 2009 | CN |
101729857 | Jun 2010 | CN |
0669587 | Aug 1995 | EP |
0701371 | Mar 1996 | EP |
0784401 | Jul 1997 | EP |
0853433 | Jul 1998 | EP |
0854650 | Jul 1998 | EP |
0903955 | Mar 1999 | EP |
0986908 | Mar 2000 | EP |
1024672 | Aug 2000 | EP |
1051027 | Nov 2000 | EP |
1124344 | Aug 2001 | EP |
1241795 | Sep 2002 | EP |
1298931 | Apr 2003 | EP |
1406452 | Apr 2004 | EP |
1455504 | Sep 2004 | EP |
1468497 | Oct 2004 | EP |
1501318 | Jan 2005 | EP |
1670256 | Jun 2006 | EP |
1755248 | Feb 2007 | EP |
2046044 | Apr 2009 | EP |
2071827 | Jun 2009 | EP |
2096870 | Sep 2009 | EP |
1700410 | Apr 2010 | EP |
2323390 | May 2011 | EP |
H07183873 | Jul 1995 | JP |
08186570 | Jul 1996 | JP |
8289255 | Nov 1996 | JP |
9252253 | Sep 1997 | JP |
11041211 | Feb 1999 | JP |
11112479 | Apr 1999 | JP |
11164270 | Jun 1999 | JP |
2000151426 | May 2000 | JP |
2000216835 | Aug 2000 | JP |
2000513164 | Oct 2000 | JP |
2000307435 | Nov 2000 | JP |
2000353969 | Dec 2000 | JP |
2001036417 | Feb 2001 | JP |
2001094625 | Apr 2001 | JP |
2001189665 | Jul 2001 | JP |
2001223655 | Aug 2001 | JP |
2001251287 | Sep 2001 | JP |
2001274776 | Oct 2001 | JP |
2001274855 | Oct 2001 | JP |
2002073625 | Mar 2002 | JP |
2002204219 | Jul 2002 | JP |
2002543705 | Dec 2002 | JP |
2003018568 | Jan 2003 | JP |
2003507985 | Feb 2003 | JP |
2003092564 | Mar 2003 | JP |
2003510734 | Mar 2003 | JP |
2003174489 | Jun 2003 | JP |
2003256321 | Sep 2003 | JP |
2003318975 | Nov 2003 | JP |
2003319012 | Nov 2003 | JP |
2003333577 | Nov 2003 | JP |
2004048704 | Feb 2004 | JP |
2004070712 | Mar 2004 | JP |
2004135013 | Apr 2004 | JP |
2004165922 | Jun 2004 | JP |
2004516717 | Jun 2004 | JP |
2004192140 | Jul 2004 | JP |
2004193992 | Jul 2004 | JP |
2004529533 | Sep 2004 | JP |
2004289621 | Oct 2004 | JP |
2004343701 | Dec 2004 | JP |
2004348824 | Dec 2004 | JP |
2004362099 | Dec 2004 | JP |
2005094140 | Apr 2005 | JP |
2005136546 | May 2005 | JP |
2005514828 | May 2005 | JP |
2005204170 | Jul 2005 | JP |
2005223433 | Aug 2005 | JP |
2005277950 | Oct 2005 | JP |
2006503463 | Jan 2006 | JP |
2006505177 | Feb 2006 | JP |
2006506926 | Feb 2006 | JP |
2006074335 | Mar 2006 | JP |
2006074421 | Mar 2006 | JP |
2006115104 | Apr 2006 | JP |
3809957 | Jun 2006 | JP |
2006174032 | Jun 2006 | JP |
2006174045 | Jun 2006 | JP |
2006186419 | Jul 2006 | JP |
2006519517 | Aug 2006 | JP |
2006287422 | Oct 2006 | JP |
2006319743 | Nov 2006 | JP |
2007013675 | Jan 2007 | JP |
2007089137 | Apr 2007 | JP |
3976163 | Jun 2007 | JP |
2007158592 | Jun 2007 | JP |
2007174170 | Jul 2007 | JP |
2007520961 | Jul 2007 | JP |
2007228205 | Sep 2007 | JP |
2008011404 | Jan 2008 | JP |
2008016907 | Jan 2008 | JP |
2008502212 | Jan 2008 | JP |
2008508761 | Mar 2008 | JP |
2008508762 | Mar 2008 | JP |
2008283232 | Nov 2008 | JP |
2008283571 | Nov 2008 | JP |
2008543142 | Nov 2008 | JP |
2008546361 | Dec 2008 | JP |
2009027598 | Feb 2009 | JP |
2009522921 | Jun 2009 | JP |
2009522922 | Jun 2009 | JP |
2009171558 | Jul 2009 | JP |
2009527949 | Jul 2009 | JP |
2009277182 | Nov 2009 | JP |
2009544991 | Dec 2009 | JP |
2010539832 | Dec 2010 | JP |
2011087103 | Apr 2011 | JP |
1020030071815 | Sep 2003 | KR |
1020030074386 | Sep 2003 | KR |
20040107152 | Dec 2004 | KR |
20040107401 | Dec 2004 | KR |
20050009376 | Jan 2005 | KR |
100809086 | Mar 2008 | KR |
20080083299 | Sep 2008 | KR |
20090098919 | Sep 2009 | KR |
20100028156 | Mar 2010 | KR |
99117925 | Jul 2001 | RU |
2189629 | Sep 2002 | RU |
2265960 | Dec 2005 | RU |
2290768 | Dec 2006 | RU |
2297663 | Apr 2007 | RU |
2312390 | Dec 2007 | RU |
2357279 | May 2009 | RU |
L246841 | Jan 2006 | TW |
L354908 | Dec 2011 | TW |
L355168 | Dec 2011 | TW |
WO9634463 | Oct 1996 | WO |
WO-9750183 | Dec 1997 | WO |
WO9804973 | Feb 1998 | WO |
WO9832231 | Jul 1998 | WO |
WO-9832256 | Jul 1998 | WO |
WO0014921 | Mar 2000 | WO |
WO0018017 | Mar 2000 | WO |
WO0052600 | Sep 2000 | WO |
WO0120786 | Mar 2001 | WO |
WO0157667 | Aug 2001 | WO |
WO0158130 | Aug 2001 | WO |
WO0158131 | Aug 2001 | WO |
WO0227988 | Apr 2002 | WO |
WO02047391 | Jun 2002 | WO |
02063461 | Aug 2002 | WO |
WO-03046742 | Jun 2003 | WO |
WO03056703 | Jul 2003 | WO |
WO03105350 | Dec 2003 | WO |
WO-03105484 | Dec 2003 | WO |
WO2004008735 | Jan 2004 | WO |
WO2004015948 | Feb 2004 | WO |
WO2004019521 | Mar 2004 | WO |
WO2004030273 | Apr 2004 | WO |
WO2004034589 | Apr 2004 | WO |
WO-2004036824 | Apr 2004 | WO |
WO2004040831 | May 2004 | WO |
WO-2004047019 | Jun 2004 | WO |
WO2004047455 | Jun 2004 | WO |
WO-2004088988 | Oct 2004 | WO |
WO-2004109538 | Dec 2004 | WO |
WO2005036753 | Apr 2005 | WO |
WO2005041421 | May 2005 | WO |
WO2005078982 | Aug 2005 | WO |
WO-2005107123 | Nov 2005 | WO |
WO2005112250 | Nov 2005 | WO |
WO-2006013459 | Feb 2006 | WO |
WO2006020826 | Feb 2006 | WO |
WO-2006036276 | Apr 2006 | WO |
2006060036 | Jun 2006 | WO |
WO-2006057938 | Jun 2006 | WO |
WO2006084503 | Aug 2006 | WO |
WO-2006116102 | Nov 2006 | WO |
WO-2006135878 | Dec 2006 | WO |
WO2007042916 | Apr 2007 | WO |
2007078253 | Jul 2007 | WO |
WO2007090834 | Aug 2007 | WO |
WO-2007098480 | Aug 2007 | WO |
2008011549 | Jan 2008 | WO |
WO-2008023328 | Apr 2008 | WO |
WO2008054100 | May 2008 | WO |
2008086313 | Jul 2008 | WO |
WO2008085013 | Jul 2008 | WO |
WO-2008131023 | Oct 2008 | WO |
2008144004 | Nov 2008 | WO |
WO2008148708 | Dec 2008 | WO |
WO2008156390 | Dec 2008 | WO |
WO-2009065526 | May 2009 | WO |
WO-2009137705 | Nov 2009 | WO |
2009143741 | Dec 2009 | WO |
WO2010085361 | Jul 2010 | WO |
WO2010088420 | Aug 2010 | WO |
WO2010120804 | Oct 2010 | WO |
WO-2011038013 | Mar 2011 | WO |
WO-2011038034 | Mar 2011 | WO |
2011059286 | May 2011 | WO |
2011070552 | Jun 2011 | WO |
2011102792 | Aug 2011 | WO |
WO-2012021540 | Feb 2012 | WO |
WO-2012109614 | Aug 2012 | WO |
Entry |
---|
Davey, M.C. et al.: “Low Density Parity Check Codes over GF(q)” IEEE Communications Letters, vol. 2, No. 6 pp. 165-167 (1998). |
International Search Report, PCT/US2007/062302—International Search Authority—US Dec. 21, 2007. |
Roca, V. et al.: Design, Evaluation and Comparison of Four Large Block FEC Codecs, LDPC, LDGM, LDGM Staircase and LDGM Triangle, plus a Reed-Solomon Small Block FEC Codec, INRIA Research Report RR-5225 (2004). |
Written Opinion, PCT/US2007/062302—International Search Authority—US, Dec. 21, 2007. |
3GPP TS 26.234 V9.1.0 ,“3rd Generation Partnership Project; Technical Specification Group Services and System Aspects; Transparent end-to-end Packet-switched Streaming Service (PSS); Protocols and codecs (Release 9)”, Dec. 2009, 179 pages. |
3GPP TS 26.244 V9.1.0, 3rd Generation Partnership Project; Technical Specification Group Services and System Aspects; Transparent end-to-end packet switched streaming service (PSS); 3GPP file format (3GP), (Release 9), Mar. 2010, 55 pp. |
3GPP TS 26.247, v1.5.0, 3rd Generation Partnership Project; Technical Specification Group Services and System Aspects Transparent end-to-end Packet-switched Streaming Service (PSS); Progressive Download and Dynamic Adaptive Streaming over HTTP (3GP-DASH) (Release 10), 2010, 91 pages. |
3rd Generation Partnership Project, Technical Specification Group Services and System Aspects Transparent end-to-end packet switched streaming service (PSS), 3GPP file format (3GP) (Release 8) , 3GPP Standard, 3GPP TS 26.244, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre , 650, Route Des Lucioles , F-06921 Sophia-Antipolis Cedex , France, No. V8.1.0, Jun. 1, 2009, pp. 1-52, XP050370199. |
3rd Generation Partnership Project; Technical Specification Group Services and System Aspects; Transparent end-to-end packet switched streaming service (PSS); 3GPP file format (3GP) (Release 9) , 3GPP Standard; 3GPP TS 26.244, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre; 650, Route Des Lucioles; F-06921 Sophia-Antipolis Cedex; France, No. V9.2.0, Jun. 9, 2010, pp. 1-55, XP050441544, [retrieved on Jun. 6, 2010]. |
Afzal, et al., “Video Streaming over MBMS: A System Design Approach”, Journal of Multimedia, vol. 1, No. 5, Aug. 2006, pp. 25-35. |
Aggarwal, C. et al.: “A Permutation-Based Pyramid Broadcasting Scheme for Video-on-Demand Systems,” Proc. IEEE Int'l Conf. on Multimedia Systems, Hiroshima, Japan (Jun. 1996). |
Aggarwal, C. et al.: “On Optimal Batching Policies for Video-on-Demand Storage Servers,” Multimedia Systems, vol. 4, No. 4, pp. 253-258 (1996). |
Albanese, A., et al., “Priority Encoding Transmission”, IEEE Transactions on Information Theory, vol. 42, No. 6, pp. 1-22, (Nov. 1996). |
Alex Zambelli,“IIS Smooth Streaming Technical Overview”, Microsoft Mar. 25, 2009, XP002620446, Retrieved from the Internet: URL:http://www.microsoft.com/downloads/en/details.aspx?FamilyID=03d22583-3ed6-44da-8464-blb4b5ca7520, [retrieved on Jan. 21, 2011]. |
Aljoscha Smolic et al., “Development of a New MPEG Standard for Advanced 3D Video Applications”, IEEE International Symposium on Image and Signal Processing and Analysis, Sep. 16, 2009, pp. 400-407, XP031552049, ISBN: 978-953-184-135-1. |
Almeroth, et al., “The use of multicast delivery to provide a scalable and interactive video-on-demand service”, IEEE Journal on Selected Areas in Communication, 14(6): 1110-1122, (1996). |
Alon, et al.: “Linear Time Erasure Codes with Nearly Optimal Recovery,” Proceedings of the Annual Symposium on Foundations of Computer Science, US, Los Alamitos, IEEE Comp. Soc. Press, vol. Symp. 36, pp. 512-516 (Oct. 23, 1995) XP000557871. |
Amin Shokrollahi: “LDPC Codes: An Introduction” Internet Citation 2 Apr. 1 2003, XP002360065 Retrieved from the Internet: URL : http ://www . ipm. ac . ir/IPM/homepage/Amin 2. pdf [retrieved on Dec. 19, 2005]. |
Amon P. et al., “File Format for Scalable Video Coding”, IEEE Transactions on Circuits and Systems for Video Technology, IEEE Service Center, Piscataway, NJ, US, vol. 17, No. 9, Sep. 1, 2007, pp. 1174-1185, XP011193013, ISSN: 1051-8215, DOI:10.1109/TCSVT.2007.905521. |
Anonymous: [Gruneberg, K., Narasimhan, S. and Chen, Y., editors] “Text of ISO/IEC 13818-1:2007/PDAM 6 MVC operation point descriptor”, 90 MPEG Meeting; Oct. 26, 2009-Oct. 30, 2009; Xian; (Motion Picture Expertgroup or ISO/IEC JTC1/SC29/WG111), No. N10942, Nov. 19, 2009, XP030017441. |
Anonymous: “Text of ISO/IEC 14496-12 3rd Edition”, 83 MPEG Meeting; Jan. 14, 2008-Jan. 18, 2008; Antalya; (Motion PictureExpert Group or ISO/IEC JTC1/SC29/WG11), No. N9678, Apr. 22, 2008, XP030016172. |
Anonymous: “Text of ISO/IEC 14496-12:2008/PDAM 2 Sub-track selection & switching”, 91. Mpeg Meeting; Jan. 18, 2010-Jan. 22, 2010; Kyoto; (Motion Picture Expertgroup or ISO/IEC JTC1/SC29/WG11), No. N11137, Jan. 22, 2010, XP030017634, ISSN: 0000-0030. |
Anonymous: “Text of ISO/IEC 14496-15 2nd edition”, 91 MPEG Meeting; Jan. 18, 2010-Jan. 22, 2010; Kyoto; (Motion Picture Expertgroup or ISP/IEC JTC1/SC29/WG11) No. N11139, Jan. 22, 2010, XP030017636. |
Apple Inc., “On the time-stamps in the segment-inbox for httpstreaming (26.244, R9)”, TSG-SA4#58 meeting, Vancouver, Canada, Apr. 2010, p. 5. |
Bar-Noy, et al., “Competitive on-line stream merging algorithms for media-on-demand”, Draft (Jul. 2000), pp. 1-34. |
Bar-Noy et al. “Efficient algorithms for optimal stream merging for media-on-demand,” Draft (Aug. 2000), pp. 1-43. |
Bigloo, A. et al.: “A Robust Rate-Adaptive Hybrid ARQ Scheme and Frequency Hopping for Multiple-Access Communication Systems,” IEEE Journal on Selected Areas in Communications, US, IEEE Inc, New York (Jun. 1, 1994) pp. 917-924, XP000464977. |
Bitner, J.R., et al.: “Efficient Generation of the Binary Reflected Gray code and Its Applications,” Communications of the ACM, pp. 517-521, vol. 19 (9), 1976. |
Blomer, et al., “An XOR-Based Erasure-Resilient Coding Scheme,” ICSI Technical Report No. TR-95-048 (1995) [avail. At ftp://ftp.icsi.berkeley.edu/pub/techreports/1995/tr-95-048.pdf]. |
Byers, J.W. et al.: “A Digital Fountain Approach to Reliable Distribution of Bulk Data,” Computer Communication Review, Association for Computing Machinery. New York, US, vol. 28, No. 4 (Oct. 1998) pp. 56-67 XP000914424 ISSN:0146-4833. |
Byers, J.W. et al.: “Accessing multiple mirror sites in parallel: using Tornado codes to speed up downloads,” 1999, Eighteenth Annual Joint Conference of the IEEE Comupter and Communications Socities, pp. 275-283, Mar. 21, 1999, XP000868811. |
Charles Lee L.H, “Error-Control Block Codes for Communications Engineers”, 2000, Artech House, XP002642221 pp. 39-45. |
Chen, et al., U.S. Patent Application titled “Frame Packing for Asymmetric Stereo Video”, having filed Feb. 25, 2011. |
Chen, et al., U.S. Patent Application titled “One-Stream Coding for Asymmetric Stereo Video”, having filed Feb. 25, 2011. |
Chen Ying et al., “Coding techniques in Multiview Video Coding and Joint Multiview Video Model”, Picture Coding Symposium, 2009, PCS 2009, IEEE, Piscataway, NJ, USA, May 6, 2009, pp. 1-4, XP031491747, ISBN: 978-1-4244-4593-6. |
Choi S: “Temporally enhanced erasure codes for reliable communication protocols” Computer Networks, Elsevier Science Publishers B.V., Amsterdam, NL, vol . 38, No. 6, Apr. 22, 2002, pp. 713-730, XP004345778, ISSN: 1389-1286, DOI:10.1016/S1389-1286(01)00280-8. |
Clark G.C., et al., “Error Correction Coding for Digital Communications, System Applications,” Error Correction Coding for Digital Communications, New York, Plenum Press, US, Jan. 1, 1981, pp. 339-341. |
D. Gozalvez et,al. “AL-FEC for Improved Mobile Reception of MPEG-2 DVB-Transport Streams” Hindawi Publishing Corporation, International Journal of Digital Multimedia Broadcasting vol. 2009, Dec. 31, 2009, pp. 1-10, XP002582035 Retrieved from the Internet: URL:http://www.hindawi.com/journals/ijdmb/2009/614178.html> [retrieved on May 12, 2010]. |
Dan, A. et al.: “Scheduling Policies for an On-Demand Video Server with Batching,” Proc. ACM Multimedia, pp. 391-398 (Oct. 1998). |
David Singer, et al., “ISO/IEC 14496-15/FDIS, International Organization for Standardization Organization Internationale De Normalization ISO/IEC JTC1/SC29/WG11 Coding of Moving Pictures and Audio”, ISO/IEC 2003, Aug. 11, 2003, pp. 1-34. |
Digital Fountain: “Raptor code specification for MBMS file download,” 3GPP SA4 PSM AD-HOC #31 (May 21, 2004) XP002355055 pp. 1-6. |
Digital Fountain: “Specification Text for Raptor Forward Error Correction,” TDOC S4-050249 of 3GPP TSG SA WG 4 Meeting #34 [Online] (Feb. 25, 2005) pp. 1-23, XP002425167, Retrieved from the Internet: URL:http://www.3gpp.org/ftp/tsg—sa/WG4—CODEC/TSGS4—34/Docs. |
“Digital Video Broadcasting (DVB); Guidelines for the implementation of DVB-IP Phase 1 specifications; ETSI TS 102 542” ETSI Standards, LIS, Sophia Antipoliscedex, France, vol. BC, No. V1.2.1, Apr. 1, 2008, XP014041619 ISSN: 0000-0001 p. 43 p. 66 pp. 70, 71. |
DVB-IPI Standard: DVB Blue Book A086r4 (03/07) Transport of MPEG 2 Transport Streatm (TS) Based DVB Services over IP Based Networks, ETSI Technical Specification 102 034 v1.3.1. |
Eager, et al. “Minimizing bandwidth requirements for on-demand data delivery,” Proceedings of the International Workshop on Advances in Multimedia Information Systems,p. 80-87 (Indian Wells, CA Oct. 1999). |
Eager, et al., “Optimal and efficient merging schedules for video-on-demand servers”, Proc. ACM Multimedia, vol. 7, pp. 199-203 (1999). |
Esaki, et al.: “Reliable IP Multicast Communication Over ATM Networks Using Forward Error Correction Policy,” IEICE Transactions on Communications, JP, Institute of Electronics Information and Comm. ENG. Tokyo, vol. E78-V, No. 12, (Dec. 1995), pp. 1622-1637, XP000556183. |
Feng, G., Error Correcting Codes over Z2m for Algorithm-Based Fault-Tolerance, IEEE Transactions on Computers, vol. 43, No. 3, Mar. 1994, pp. 370-374. |
Fernando, et al., “httpstreaming of MPEG Media—Response to CfP”, 93 MPEG Meeting; Jul. 7, 2010-Jul. 30, 2010; Geneva; (Motion Picture Expert Group or ISO/IEC JTC1/SCE29/WG11), No. M17756, Jul. 22, 2010, XP030046346. |
Fielding et al., “RFC 2616: Hypertext Transfer Protocol HTTP/1.1”, Internet Citation, Jun. 1999, pp. 165, XP002196143, Retrieved from the Internet: URL:http://www.rfc-editor-org/ [retrieved on Apr. 15, 2002]. |
Frojdh, et al., “File format sub-track selection and switching,” ISO/IEC JTC1/SC29/WG11 MPEG2009 M16665, London UK., Jul. 2009, 14 pp. |
Gao, L. et al.: “Efficient Schemes for Broadcasting Popular Videos,” Proc. Inter. Workshop on Network and Operating System Support for Digital Audio and Video, pp. 1-13 (1998). |
Gasiba, Tiago et al., “System Design and Advanced Receiver Techniques for MBMS Broadcast Services” PROC. 2006 International Conference on Communications (ICC 2006), Jun. 1, 2006, pp. 5444-5450, XP031025781 ISBN: 978-1-4244-0354-7. |
Gemmell, et al., “A Scalable Multicast Architecture for One-To-Many Telepresentations”, Multimedia Computing and Systems, 1998/Proceedings. IEEE International Conference on Austin, TX, USA Jun. 28-Jul. 1, 1998, Los Alamitos, CA USA, IEEE Comput. Soc, US, Jun. 28, 1998, pp. 128-139, XP010291559. |
Goyal: “Multiple Description Coding: Compression Meets the Network,” In Signal Processing Magazine, IEEE, vol. 18., Issue 5 (Sep. 2001) pp. 74-93 URL:http://www.rle.mit.edu/stir/documents/Goyal—SigProcMag2001—MD.pdf [Apr. 7, 2011]. |
Gozalvez D et, al: “Mobile reception of DVB-T services by means of AL-FEC protection” Proc. IEEE Intern. Symposium on Broadband Multimedia Systems and Broadcasting (BMSB '09), IEEE, Piscataway, NJ, USA, May 13, 2009, pp. 1-5, XP031480155 ISBN: 978-1-4244-2590-7. |
Grineberg, et al., “Deliverable D3.2 MVC/SVC storage format” Jan. 29, 2009, XP002599508 Retrieved from the Internet: URL:http://www.ist-sea.eu/Public/SEA—D3.2—HHI FF—20090129.pdf [retrieved on Sep. 1, 2010] paragraph [02.3]. |
Hagenauer, J. : “Soft is better than hard” Communications, Coding and Cryptology, Kluwer Publication May 1994, XP002606615 Retrieved from the Internet : URL: http://www. Int . ei .turn. de/veroeffentlic hungen/I994/ccc94h. pdf [retrieved on Oct. 25, 2010]. |
He Wenge et al., “Asymmetric Stereoscopic Video Encoding Algorithm Based on Joint Compensation Prediction”, IEEE International Conference on Communications and Mobile Computing, Jan. 6, 2009, pp. 191-194, XP031434775, ISBN: 978-0-7695-3501-2. |
Hershey, et al., “Random Parity Coding (RPC)”, 1996 IEEE International Conference on Communications (ICC). Converging Technologies for Tomorrow'S Applications. Dallas, Jun. 23-27, 1996, IEEE International Conference on Communications (ICC), New York, IEEE, US, vol. 1, Jun. 1996, pp. 122-126, XP000625654. |
Hitachi Ltd. et al., “High-Definition Multimedia Interface,” Specification Version 1.4, Jun. 5, 2009, 425 pp. |
Hua, et al., “Skyscraper broadcasting: A new broadcsting system for metropolitan video-on-demand systems”, Proc. ACM SIGCOMM, pp. 89-100 (Cannes, France, 1997). |
Ian Trow, “Is 3D Event Coverage Using Existing Broadcast Infrastructure Technically Possible?”, International Broadcasting Conference, Sep. 9, 2009-Sep. 13, 2009, XP030081671, pp. 4-5, “3D transmission over broadcast infrastructure” pp. 7-8, “Screen signaling”—Conclusions on 3D systems. |
IETF RFC 2733: Rosenberg, J. et al. “An RTP Payload Format for Generic Forward Error Correction,” Network Working Group, RFC 2733 (Dec. 1999). |
Information Technology -Generic Coding of Moving Pictures and Audio: Systems, Amendment 4: Transport of Multiview Video over ITU-T Rec H.222.0 | ISO/IEC 13818-1 “Text of ISO/IEC 13818-1:2007/FPDAM 4—Transport of Multiview Video over ITU-T Rec H.222.0 | ISO/IEC 13818-1,” Lausanne, Switzerland, 2009, 21 pp. |
International Search Report and Written Opinion—PCT/US2011/044745—ISA/EPO—Dec. 21, 2011 (100875WO). |
International Standard ISO/IEC 13818-1:2000(E), “Information Technology—Generic Coding of Moving Pictures and Associated Audio Information: Systems,” Second edition, Dec. 1, 2000, pp. 1-174. |
International Standard ISO/IEC 14496-12, Information Technology—Coding of audio-visual objects Part 12: ISO base media file format, Third Edition, Oct. 15, 2008, 120 pp. |
International Telecommunication Union, “ITU-T H.264, Series H: Audiovisual and Multimedia Systems, Infrastructure of audiovisual services—Coding of moving video, Advanced video coding for generic audiovisual services,” Mar. 2010, 669 pp. |
ISO/IEC JTC 1/SC 29, ISO/IEC FCD 23001-6, Information technology—MPEG systems technologies—Part 6: Dynamic adaptive streaming over HTTP (DASH), Jan. 28, 2011. |
ISO/IEC JTC1/SC29/WG11: “Requirements on HTTP Streaming of MPEG Media”, 92. MPEG Meeting; Apr. 19, 2010-Apr. 23, 2010; DRESDEN; No. N11340, May 14, 2010, XP030017837, ISSN: 0000-0029. |
Jin Li, “The Efficient Implementation of Reed-Solomon High Rate Erasure Resilient Codes” Proc. 2005 IEEE International Conference on Acoustics, Speech, and Signal Processing, Philadelphia, PA, USA, IEEE, Piscataway, NJ, vol . 3, Mar. 18, 2005, pp. 1097-1100, XP010792442, DOI: 10.1109/ICASSP.2005.1415905 ISBN: 978-0-7803-8874-1. |
“Joint Draft 8.0 on Multiview Video Coding”, 28th JVT meeting, Hannover, Germany, Document: JVT-AB204 (rev.1), Jul. 2008. available from http:// wftp3. itu.int/av-arch/jvt-site/2008—07—Hannover/JVT-AB204. |
Juhn, L. et al.: “Adaptive Fast Data Broadcasting Scheme for Video-on-Demand Service,” IEEE Transactions on Broadcasting, vol. 44, No. 2, pp. 182-185 (Jun. 1998). |
Juhn, L. et al.: “Harmonic Broadcasting for Video-on-Demand Service,” IEEE Transactions on Broadcasting, vol. 43, No. 3, pp. 268-271 (Sep. 1997). |
Kallel, “Complementary Punctured Convolutional (CPC) Codes and Their Applications”, IEEE Transactions on Communications, IEEE Inc., New York, US, vol. 43, No. 6, Jun. 1, 1995, p. 2005-2009. |
Kimata H et al., “Inter-View Prediction With Downsampled Reference Pictures”, ITU Study Group 16—Video Coding Experts Group -ISO/IEC MPEG & ITU-T VCEG(ISO/IEC JTC1/SC29/WG11 and ITU-T SG16 Q6), No. JVT-W079, Apr. 19, 2007, XP030007039. |
Kozamernik F: “Media streaming over the Internet”, Internet Citation, Oct. 2002, XP002266291, Retrieved from the Internet: URL: http://www.ebu.ch/trev—292-kozamerni k. pdf [retrieved on Jan. 8, 2004]section “Video codecs for scalable streaming”. |
Lee L., et al.,“VLSI implementation for low density parity check decoder”, Proceedings of the 8th IEEE International Conference on Elecctronics, Circuits and Systems, 2001. ICECS 2001, Sep. 2, 2001, vol. 3, pp. 1223-1226. |
Lin, S. et al.: “Error Control Coding-Fundamentals and Applications,” 1983, Englewood Cliffs, pp. 288, XP002305226. |
Luby Digital Fountain A Shokrollahi Epfl M Watson Digital Fountain T Stockhammer Nomor Research M: “Raptor Forward Error Correction Scheme for Object Delivery; rfc5053.txt”, IETF Standard, Internet Engineering Task Force, IETF, Ch, Oct. 1, 2007, XP015055125, ISSN: 0000-0003. |
Luby, et al., “Analysis of Low Density Codes and Improved Designs Using Irregular Graphs”, 1998, Proceedings of the 30th Annual ACM Symposium on Theory of Computing, May 23, 1998, pp. 249-258, XP000970907. |
Luby, et al., “FLUTE -File Delivery over Unidirectional Transport”, IETF RFC 3926, pp. 1-35, (Oct. 2004). |
Luby et al., “Improved Low-Density Parity-Check Codes Using Irregular Graphs and Belief Propagation”, Information Theory, 1998. Proceedings. 1998 IEEE International Symposium on Cambridge, MA, USA Aug. 16-21, 1998, New York, NY, USA, IEEE, US Aug. 16, 199. |
Luby et, al. “Layered Coding Transport (LCT) Building Block”, IETF RFC 5651, pp. 1-42, (Oct. 2009). |
Luby, et al.: “Analysis of Low Density Codes and Improved Designs Using Irregular Graphs,” International Computer Science Institute Technical Report TR-97-045 (Nov. 1997) [available at ftp://ftp.icsi.berkeley.edu/pub/techreports/1997/tr-97-045.pdf]. |
Luby, M., et, al. “Forward Error Correction (FEC) Building Block”, IETF RFC 5052, pp. 1-31, (Aug. 2007). |
Luby, M., et al., “Raptor Forward Error Correction Scheme for Object Delivery”, IETF RFC5053, pp. 1-46 (Sep. 2007). |
Luby, M., et al., “RaptorQ Forward Error Correction Scheme for Object Delivery”, IETF draft ietf-rmt-bb-fec-raptorq-04, Reliable Multicast Transport, pp. 1-68, (Aug. 24, 2010). |
Luby, M., et al., “Request for Comments: 3453: The Use of Forward Error Correction (FEC) in Reliable Multicast,” Internet Article, [Online] Dec. 2002, pp. 1-19. |
Luby, M. et al.: “Efficient Erasure Correction Codes,” 2001, IEEE Transactions on Information Theory, Vo. 47, No. 2, pp. 569-584, XP002305225. |
Luby M et al: “IPTV Systems, Standards and Architectures: Part II -Application Layer FEC in IPTV Services” IEEE Communications Magazine, IEEE Service Center, Piscataway, US LNKDDOI: 10.1109/MCOM.2008.4511656, vol. 46, No. 5, May 1, 2008, pp. 94-101, XP011226858 ISSN: 0163-6804. |
Luby, M. et al.: “Pairwise Independence and Derandomization,” Foundations and Trends in Theoretical Computer Science, vol. 1, Issue 4, 2005, Print ISSN 1551-305X, Online ISSN 1551-3068. |
Luby, M. et al.: “Practical Loss-Resilient Codes: Tornado Codes,” 29th Annual ACM Symposium on Theory of Computing, vol. SYMP. 29, May 4, 1997, pp. 1-10, XP002271229. |
Luby, Michael G. “Analysis of Random Processes via And-Or Tree Evaluation,” Proceedings of the 9th Annual ACM-SIAM Symposium on Discrete Algorithms,TR-97-0, 1998, pp. 364-373, (search date: Jan. 25, 2010) URL: <http://portal.acm.prg.citation.cfm?id=314722>. |
Mandelbaum D.M., “An Adaptive-Feedback Coding Scheme Using Incremental Redundancy”, IEEE Trans on Information Theory, vol. May 1974, pp. 388-389, XP002628271, the whole document. |
Marpe, et al., “The H.264/MPEG4 Advanced Video Coding Standard and its Applications,” Standards Report, IEEE Communications Magazine, Aug. 2006, pp. 134-143. |
Matsuoka H., et al., “Low-Density Parity-Check Code Extensions Applied for Broadcast-Communication Integrated Content Delivery”, Research Laboratories, NTT DOCOMO, Inc., 3-6, Hikari-No-Oka, Yokosuka, Kanagawa, 239-8536, Japan, ITC-SS21, 2010 IEICE, pp. 59-63. |
McCanne, et al., “Low-Complexity Video Coding for Receiver-Driven Layered Multicast”, IEEE Journal on Selected Areas in Communication IEEE Service Center, Aug. 1, 1997, vol. 15, No. 6, pp. 983-1001, Piscataway, US, XP011054678, ISSN: 0733-8716. |
Mimnaugh, A et, al. “Enabling Mobile Coverage for DVB-T” Digital Fountain Whitepaper Jan. 29, 2008, pp. 1-9, XP002581808 Retrieved from the Internet: URL:http://www.digitalfountain.com/ufiles/ library/DVB-T-whitepaper.pdf> [retrieved on May 10, 2010]. |
Min-Goo Kim: “On systematic punctured convolutional codes”, IEEE Trans on Communications, vol. 45, No. 2, Feb. 1997, XP002628272, the whole document, pp. 133-139. |
Muller, et al., “A test-bed for the dynamic adaptive streaming over HTTP featuring session mobility” MMSys '11 Proceedings of the second annual ACM conference on Multimedia systems, Feb. 23-25, 2011, San Jose, CA, pp. 271-276. |
Naguib, Ayman, et al., “Applications of Space-Time Block Codes and Interference Suppression for High Capacity and High Data Rate Wireless Systems,” IEEE, 1998, pp. 1803-1810. |
Narayanan, et al., “Physical Layer Design for Packet Data Over IS-136”, Vehicular Technology Conference, 1997, IEEE 47th Phoenix, AZ, USA May 4-7, 1997, New York, NY, USA, IEEE, US May 4, 1997, p. 1029-1033. |
Nokia: “Reed-Solomon Code Specification for. MBMS Download and Streaming Services”, 3GPP Draft; 54-050265—RS—SPEC, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre ; 650, Route Des Lucioles ; F-06921 Sophia-Antipolis Cedex ; France, vol. SA WG4, No. San Diego, USA; 20050415, Apr. 15, 2005, XP050287675, [retrieved on Apr. 15, 2005]. |
Nokia Corp., “Usage of ‘mfra’ box for Random Access and Seeking,” S4-AHI127, 3GPP TSG-SA4 Ad-Hoc Meeting, Dec. 14-16, 2009, Paris, FR, 2 pp. |
Nonnenmacher, et al., “Parity-Based Loss Recovery for Reliable Multicast Transmission”, IEEE / ACM Transactions on Networking, IEEE Inc. New York, US, vol. 6, No. 4, Aug. 1, 1998, p. 349-361. |
Ozden, B. et al.: “A Low-Cost Storage Service for Movie on Demand Databases,” Proceedings of the 20th Very Large DataBases (VLDB) Conference, Santiago, Chile (1994). |
PA. Chou, A. Mohr, A. Wang, S. Mehrotra, “FEC and Pseudo-ARQ for Receiver-Driven Layered Multicast of Audio and Video,” pp. 440-449, IEEE Computer Society, Data Compression Conference (2000). |
Pantos R et al., “HTTP Live Streaming; draft-pantos-http-1ive-streaming-OT.txt ”, HTTP Live Streaming; Draft-Pant0s-HTTP-Live-Streaming-01.Txt, Internet Engineering Task Force, IEFT; Standardworkingdraft, Internet Society (ISOC) 4, Rue Des Falaises Ch- 1205 Geneva, Switzerland, No. 1, Jun. 8, 2009, XP015062692. |
Paris, et al., “A low bandwidth broadcasting protocol for video on demand”, Proc. International Conference on Computer Communications and Networks, vol. 7, pp. 690-697 (Oct. 1998). |
Paris, et al., “Efficient broadcasting protocols for video on demand”, International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication systems (MASCOTS), vol. 6, pp. 127-132 (Jul. 1998). |
Perkins, et al.: “Survey of Packet Loss Recovery Techniques for Streaming Audio,” IEEE Network; Sep./Oct. 1998, pp. 40-48. |
Petition decision for Petition Under 37 C.F.R. § 1.78 to Accept an Unintentionally Delayed Priority Claim under 35 U.S.C. § 120 in U.S. Patent No. 7,532,132, dated Jul. 21, 2011, 2 pages. |
Petition under 37 C.F.R. § 1.78 to Accept an Unintentionally Delayed Priority Claim under 35 U.S.C. § 120 in U.S. Patent No. 7,532,132, dated May 27, 2011, 2 pages. |
Plank J. S., “A Tutorial on Reed-Solomon Coding for Fault-Tolerance I N Raid-Like Systems”, Software Practice & Experience, Wiley & Sons, Bognor Regis, GB, vol. 27, No. 9, Sep. 1, 1997, pp. 995-1012, XP00069594. |
Pless and WC Huffman EDS V S: Algebraic geometry codes, Handbook of Coding Theory, 1998, pp. 871-961, XP002300927. |
Pursley, et al.: “Variable-Rate Coding for Meteor-Burst Communications,” IEEE Transactions on Communications, US, IEEE Inc. New York (1989) vol. 37, No. 11, pp. 1105-1112 XP000074533. |
Pursley, M. et al.: A Correction and an Addendum for “Variable-Rate Coding for Meteor-Burst Communications,” IEEE Transactions on Communications, vol. 43, No. 12 pp. 2866-2867 (Dec. 1995). |
Pyle, et al., “Microsoft http smooth Streaming: Microsoft response to the Call for Proposal on httpstreaming”, 93 MPEG Meeting; Jul. 27, 2010-Jul. 30, 2010; Geneva; (Motion Picture Expert Group or ISO/IEC JTC1/SCE29/WG11), No. M17902, Jul. 22, 2010, XP030046492. |
Qualcomm Europe S A R L: “Baseline Architecture and Definitions for HTTP Streaming”, 3GPP Draft; 54-090603—HTTP—Streaming—Architecture, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre; 650, Route Des Lucioles; F-06921 Sophia-Antipolis Cedex; France, No. Kista; 20090812, 12 Aug. 12, 2009, XP050356889. |
Qualcomm Incorporated: “Use Cases and Examples for Adaptive httpstreaming”, 3GPP Draft; 54-100408-Usecases-HSD, 3rd Generation Partnership Project (JGPP), Mobile Competence Centre; 650, Route Des Lucioles; F-06921 Sophia-Antipolis Cedex; France, vol. SA WG4, No. Prague, Czech Republic; 20100621, Jun. 17, 2010, XP050438085, [retrieved on Jun. 17, 2010]. |
Rangan, et al., “Designing an On-Demand Multimedia Service,” IEEE Communication Magazine, vol. 30, pp. 56-64, (Jul. 1992). |
Realnetworks Inc, et al., “Format for httpstreaming Media Presentation Description”, 3GPP Draft; S4-100020, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre; 650, Route Des Lucioles; F-06921 Sophia- Antipolis Cedex; France, vol. SA WG4, No. S t Julians, Malta; 20100125, Jan. 25, 2010, Jan. 20, 2010, XP050437753, [retrieved on Jan. 1, 2010]. |
Research in Motion UK Limited: “An MPD delta file for httpstreaming”, 3GPP Draft; S4-100453, 3rd Generation Partnership Project (SGPP), Mobile Competence Centre; 650, Route Des Lucioles; F-06921 Sophia-Antipolis Cedex; France, vol. SA WG4, No. Prague, Czech Republic; 20100621, Jun. 21, 2010, Jun. 16, 2010, XP050438066, [retrieved on Jun. 16, 2010]. |
Rhyu, et al., “Response to Call for Proposals on httpstreaming of MPEG Media”, 93 MPEG Meeting; Jul. 26, 2010-Jul. 30, 2010; Geneva; (Motion Picture Expert Group or ISO/IEC JTC1/SCE29/WG11) No. M17779, Jul. 26, 2010, XP030046369. |
Rizzo, L. “Effective Erasure Codes for Reliable Computer Communication Protocols,” Computer Communication Review, 27 (2) pp. 24-36 (aPR. 1, 1997), XP000696916. |
Roca, V., et, al. “Low Density Parity Check (LDPC) Staircase and Triangle Forward Error Correction (FEC) Schemes”, IETF RFC 5170 (Jun. 2008), pp. 1-34. |
Rost, S. et al.: “The Cyclone Server Architecture: streamlining delivery of popular content,” 2002, Computer Communications, vol. 25, No. 4, pp. 403-412. |
Roth, R., et al., “A Construction of Non-Reed-Solomon Type MDS Codes”, IEEE Transactions of Information Theory, vol. 35, No. 3, May 1989, pp. 655-657. |
Roth, R., “On MDS Codes via Cauchy Matrices”, IEEE Transactions on Information Theory, vol. 35, No. 6, Nov. 1989, pp. 1314-1319. |
Schwarz, Heiko et al., “Overview of the Scalable Video Coding Extension of the H.264/AVC Standard”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 17, No. 9, Sep. 2007, pp. 1103-1120. |
Seshan, S. et al.: “Handoffs in Cellular Wireless Networks: The Daedalus Implementation and Experience,” Wireless Personal Communications, NL; Kluwer Academic Publishers, vol. 4, No. 2 (Mar. 1, 1997) pp. 141-162, XP000728589. |
Shacham: “Packet Recovery and Error Correction in High-Speed Wide-Area Networks,” Proceedings of the Military Communications Conference. (Milcom), US, New York, IEEE, vol. 1, pp. 551-557 (1989) XP000131876. |
Shierl T; Gruneberg K; Narasimhan S; Vetro A: “ISO/IEC 13818-1:2007/FPDAM 4—Information Technology Generic Coding of Moving Pictures and Audio Systems amendment 4: Transport of Multiview Video over ITU-T Rec H.222.0 ISO/IEC 13818-1” ITU-T REC. H.222.0(May 2006)FPDAM 4, vol. MPEG2009, No. 10572, May 11, 2009, pp. 1-20, XP002605067 p. 11, last two paragraphs sections 2.6.78 and 2.6.79 table T-1. |
Shokrollahi, A.: “Raptor Codes,” Internet Citation [Online] (Jan. 13, 2004) XP002367883, Retrieved from the Internet: URL:http://www.cs.huji.ac.il/labs/danss/p2p/resources/raptor.pdf. |
Shokrollahi, Amin. “Raptor Codes,” IEEE Transactions on Information Theory, Jun. 2006, vol. 52, No. 6, pp. 2551-2567, (search date: Feb. 1, 2010) URL:<http://portal.acm.org/citation.cfm?id=1148681>. |
Shokrollahi et al., “Design of Efficient Easure Codes with Differential Evolution”, IEEE International Symposium on Information Theory, Jun. 25, 2000, pp. 5-5. |
Sincoskie, W. D., “System Architecture for Large Scale Video on Demand Service,” Computer Network and ISDN Systems, pp. 155-162, (1991). |
Stockhammer, “WD 0.1 of 23001-6 Dynamic Adaptive Streaming over HTTP (DASH)”, MPEG-4 Systems, International Organisation for Standardisation, ISO/IEC JTC1/SC29/WG11, Coding of Moving Pictures and Audio, MPEG 2010 Geneva/m11398, Jan. 6, 2011, 16 pp. |
Sullivan et al., Document: JVT-AA007, “Editors' Draft Revision to ITU-T Rec. H.264|ISO/IEC 14496-10 Advanced Video Coding—In Preparation for ITU-T SG 16 AAP Consent (in integrated form),” Joint Video Team (JVT) of ISO/IEC MPEG & ITU-T VCEG (ISO/IEC JTC1/SC29/WG11 and ITU-T SG16 Q.6), 30th Meeting: Geneva, CH, Jan. 29-Feb. 3, 2009, pp. 1-683, http://wftp3.itu.int/av-arch/jvt-site/2009—01—Geneva/JVT-AD007.zip. |
Sun, et al., “Seamless Switching of Scalable Video Bitstreams for Efficient Streaming,” IEEE Transactions on Multimedia, vol. 6, No. 2, Apr. 2004, pp. 291-303. |
Telefon AB LM Ericsson, et al., “Media Presentation Description in httpstreaming”, 3GPP Draft; S4-100080-MPD, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre; 650, Route Des Lucioles; F-06921 Sophia- Antipolis Cedex; France, vol. SA WG4, No. St Julians, Malta; 20100125, Jan. 20. 2010, XP050437773, [retrieved on Jan. 20, 2010]. |
Thomas Wiegand, et al., “Joint Draft ITU-T Rec. H.264 | ISO/IEC 14496-10 / Amd.3 Scalable video coding”, Joint Video Team (JVT) of ISO/IEC MPEG & ITU-T VCEG (ISO/IEC JTC1/SC29/WG11 and ITU-T SG16 Q.6) 24th Meeting: Geneva, Switzerland, Jun. 29-Jul. 5, 2007, pp. 1-559. |
Tsunoda T., et al., “Reliable Streaming Contents Delivery by Using Multiple Paths,” Technical Report of the Institute of Electronics, Information and Communication Engineers, Japan, Mar. 2004, vol. 103, No. 692, pp. 187-190, NS2003-331, IN2003-286. |
U.S. Appl. No. 12/840,146, by Ying Chen et al., filed Jul. 20, 2010. |
U.S. Appl. No. 12/908,537, by Ying Chen et al., filed Oct. 20, 2010. |
U.S. Appl. No. 12/908,593, by Ying Chen et al., filed Oct. 20, 2010. |
U.S. Appl. No. 13/082,051, by Ying Chen et al., filed Apr. 7, 2011. |
U.S. Appl. No. 13/205,559, by Ying Chen et al., filed Aug. 8 2011. |
U.S. Appl. No. 13/205,565, by Ying Chen et al., filed Aug. 8, 2011. |
U.S. Appl. No. 13/205,574, by Ying Chen et al., filed Aug. 8, 2011. |
Universal Mobile Telecommunications System (UMTS); LTE; Transparent end-to-end Packet-switched Streaming Service (PSS); Protocols and codecs (3GPP TS 26.234 version 9.3.0 Release 9), Technical Specification, European Telecommunications Standards Institute (ETSI), 650, Route Des Lucioles; F-06921 Sophia-Antipolis; France, vol. 3GPP SA, No. V9.3.0, Jun. 1, 2010, XP014047290, paragraphs [5.5.4.2], [5.5.4.3], [5.5.4.4], [5.4.5], [5.5.4.6] paragraphs [10.2.3], [11.2.7], [12.2.3], [12.4.2], [12.6.2] paragraphs [12.6.3], [12.6.3.1], [12.6.4], [12.6.6]. |
Viswanathan, et al., “Metropolitan area video-on-demand services using pyramid broadcasting”, Multimedia Systems, 4(4): 197-208 (1996). |
Viswanathan, et al., “Pyramid Broadcasting for Video-on-Demand Service”, Proceedings of the SPIE Multimedia Computing and Networking Conference, vol. 2417, pp. 66-77 (San Jose, CA, Feb. 1995). |
Viswanathan,Subramaniyam R., “Publishing in Wireless and Wireline Environments,” Ph. D Thesis, Rutgers, The State University of New Jersey (Nov. 1994), 180pages. |
Wang,“On Random Access”, Joint Video Team (JVT) of ISO/IEC MPEG & ITU-T VCEG (ISO/IEC JTC1ISC29/WG11 and ITU-T SG16 Q.6), 4th Meeting: Klagenfurt, Austria, Jul. 22-26, 2002, p. 13. |
Watson, M., et, al. “Asynchronous Layered Coding (ALC) Protocol Instantiation”, IETF RFC 5775, pp. 1-23, (Apr. 2010). |
Wenger, et al., RFC 3984, “RTP Payload Format for H.264 Video,” Feb. 2005, 84 pp. |
Wong, J.W., “Broadcast delivery”, Proceedings of the IEEE, 76(12): 1566-1577, (1988). |
Yamauchi, Nagamasa. “Application of Lost Packet Recovery by Front Error Correction to Internet Multimedia Transfer” Proceedings of Workshop for Multimedia Communication and Distributed Processing, Japan, Information Processing Society of Japan (IPS), Dec. 6, 2000, vol. 2000, No. 15, pp. 145-150. |
Yin et al., “Modified Belief-Propogation algorithm for Decoding of Irregular Low-Density Parity-Check Codes”, Electronics Letters, IEE Stevenage, GB, vol. 38, No. 24, Nov. 21, 2002, pp. 1551-1553. |
Ying Chen et al: “Response to the CfP on HTTP Streaming: Adaptive Video Streaming based on AVC”, 93 MPEG Meeting; Jul. 26, 2010-Jul. 30, 2010; Geneva; (Motion Picture Expert Group or ISO/IEC JTC1/SC29/WG11), No. M17909, Jul. 26, 2010, XP030046499. |
Zorzi, et al.: “On the Statistics of Block Errors in Bursty Channels,” IEEE Transactions on Communications, vol. 45, No. 6, Jun. 1997, pp. 660-667. |
Cataldi et al., “ Sliding-Window Raptor Codes for Efficient Scalable Wireless Video Broadcasting With Unequal Loss Protection”, IEEE Transactions on Image Processing, Jun. 1, 2010, pp. 1491-1503, vol. 19, No. 6, IEEE Service Center, XP011328559, ISSN: 1057-7149, DOI: 10.1109/TIP.2010.2042985. |
Gracie et al., “ Turbo and Turbo-Like Codes: Principles and Applications in Telecommunications”, Proceedings of the IEEE, Jun. 1, 2007, pp. 1228-1254, vol. 95, No. 6, IEEE, XP011189323, ISSN: 0018-9219, DOI: 10.1109/JPR0C.2007.895197. |
Huawei et al., “Implict mapping between CCE and PUCCH for ACK/NACK TDD”, 3GPP Draft; R1-082359, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre; 650, Route Des Lucioles ; F-06921 Sophia-Antipolis Cedex ; France, vol. Ran WG1, No. Warsaw, Poland, Jun. 24, 2008, XP050110650, [retrieved on Jun. 24, 2008] . |
International Search Report and Written Opinion—PCT/US2012/024737—ISA/EPO—May 11, 2012 (092888U1WO). |
Kimura et al., “A Highly Mobile SDM-0FDM System Using Reduced-Complexity-and-Latency Processing”, IEEE International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC), Sep. 1, 2007, pp. 1-5, IEEE, XP031168836, ISBN: 978-1-4244-1143-6, DOI: 10.1109/PIMRC.2007.4394758. |
Luby Qualcomm Incorporated, “Universal Object Delivery using RaptorQ; draft-luby-uod-raptorq-OO.txt”, Internet Engineering Task Force (IETF), Standardworkingdraft, Internet Society (ISOC), Mar. 7, 2011, pp. 1-10, XP015074424, [retrieved on Mar. 7, 2011]. |
Mackay, “Fountain codes Capacity approaching codes design and implementation”, IEE Proceedings: Communications, Dec. 9, 2005, pp. 1062-1068, vol. 152, No. 6, Institution of Electrical Engineers, XP006025749, ISSN: 1350-2425, DOI: 10.1049/IP-C0M:20050237 . |
Todd, “Error Correction Coding: Mathematical Methods and Algorithms”, Mathematical Methods and Algorithms, Jan. 1, 2005, pp. 451-534, Wiley, XP002618913. |
Bross, et al., “High efficiency video coding (HEVC) text specification draft 6,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 JCTVC-H1003, 7th Meeting: Geneva, CH, Nov. 21-30, 2011, pp. 259. |
Bross, et al., “High efficiency video coding (HEVC) text specification draft 7,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 9th Meeting: Geneva, CH, Apr. 27-May 7, 2012, JCTVC-I1003—d21, pp. 290. |
Bross, et al., “High efficiency video coding (HEVC) text specification draft 8,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 10th Meeting: Stockholm, SE, Jul. 11-20, 2012, JCTVC-J1003—d7, pp. 261. |
Bross et al., “WD4: Working Draft 4 of High-Efficiency Video Coding,” JCTVC-F803—d2, (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 Joint Collaborative Team on Video Coding, 6th Meeting, Torino, IT, Jul. 1-22, 2011, 226 pages. |
Bross et al., “WD5: Working Draft 5 of High-Efficiency Video Coding,” JCTVC-G1103—d2, (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 Joint Collaborative Team on Video Coding, 7th Meeting, Geneva, Switzerland (Nov. 2011), 214 pages. |
European Search Report—EP10013235—Search Authority—The Hague—Aug. 20, 2012. |
ITU-T H.264, Series H: Audiovisual and Multimedia Systems, Infrastructure of audiovisual services—Coding of moving video, Advanced video coding for generic audiovisual services, The International Telecommunication Union. Jun. 2011, 674 pp. |
Jiang., File Format for Scalable Video Coding, PowerPoint Presentation for CMPT 820, Summer 2008. |
Pantos, “HTTP Live Streaming draft-pantos-http-live-streaming-02”, Informational, Internet-Draft, Intended status: Informational, Expires: Apr. 8, 2010, http://tools.ietf.org/html/draft-pantos-http-live-streaming-02, pp. 1-20, Oct. 5, 2009. |
Thomas Wiegand et al.,“ WD1: Working Draft 1 of High-Efficiency Video Coding”, JCTVC-C403, Joint Collaborative Team on Video Coding (JCT-VC), of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, 3rd Meeting: Guangzhou, CN, Oct. 7-15, 2010. |
Wiegand et al., “WD3: Working Draft 3 of High-Efficiency Video Coding,” Document JCTVC-E603, 5th Meeting: Geneva, CH, Mar. 16-23, 2011,193 pp. |
Wiegand T. et al., “WD2: Working Draft 2 of High-Efficiency Video Coding”, Jan. 28, 2011, No. JCTVC-D503, Jan. 28, 2011, XP002679642, Retrieved from the Internet: URL: http://wftp3.itu.int/av-arch/jctvc-site/2011—01—D—Daegu/ [retrieved on Jul. 11, 2012]. |
Yamanouchi N., et al., “Internet Multimedia Transmission with Packet by Using Forward Error Correction,” Proceedings of DPS Workshop, The Information Processing Society of Japan, Dec. 6, 2000, vol. 2000, No. 15, pp. 145-150. |
3rd Generation Partnership Project; Technical Specification Group Services and System Aspects; Transparent end-to-end Packet-switched Streaming Service (PSS); Progressive Download and Dynamic Adaptive Streaming over HTTP (3GP-DASH) (Release 10), 3GPP Standard; 3GPP TS 26.247, 3RD Generation Partnership Project (3GPP), Mobile Competence Centre; 650, Route Des Lucioles; F-06921 Sophia-Antipolis Cedex; France, vol. SA WG4, No. V10.0.0, Jun. 17, 2011, pp. 1-94, XP050553206, [retrieved on Jun. 17, 2011]. |
Anonymous: “Technologies under Consideration”, 100. MPEG Meeting;30-4-2012—Apr. 5, 2012; Geneva;(Motion Picture Expert Group or ISO/IEC JTC1/SC29/WG11) No. N12682, Jun. 7, 2012, XP030019156. |
Atis: “PTV Content on Demand Service”, IIF-WT-063R44, Nov. 11, 2010, pp. 1-124, XP055045168, Retrieved from the Internet: URL:ftp://vqeg.its.bldrdoc.gov/DocumentsNOEG—Atlanta—Nov10/MeetingFiles/Liaison/IIF-WT-063R44—Content—on—Demand.pdf [retrieved on Nov. 22, 2012]. |
Bouazizi I., et al., “Proposals for ALC/FLUTE server file format (14496-12Amd.2)”, 77. MPEG Meeting; 17-07-2006—Dec. 7, 2006; Klagenfurt; (Motion Pictureexpert Group or ISO/IEC JTC1/SC29/WG11), No. M13675, Jul. 12, 2006, XP030042344, ISSN: 0000-0236. |
Frojdh P., et al., “Study on 14496-12:2005/PDAM2 ALU/ FLUTE Server File Format”, 78.MPEG Meeting; Oct. 23, 2006-Oct. 27, 2006; Hangzhou: (Motion Picturexpert Group or ISO/IEC JTC1/SC29/WG11),, No. M13855, Oct. 13, 2006, XP030042523, ISSN: 0000-0233. |
Gil A., et al., “Personalized Multimedia Touristic Services for Hybrid Broadcast/Broadband Mobile Receivers,” IEEE Transactions on Consumer Electronics, 2010, vol. 56 (1), pp. 211-219. |
Hannuksela M. M., et al., “DASH: Indication of Subsegments Starting with SAP”, 97. MPEG Meeting; Jul. 18, 2011-Jul. 22, 2011; Torino; (Motion Picture Expert Group or ISO/IEC JTC1/SC29/WG11) No. m21096, Jul. 21, 2011, XP030049659. |
Hannuksela M. M., et al., “ISOBMFF: SAP definitions and 'sidx' box”, 97. MPEG Meeting; Jul. 18, 2011-Jul. 22, 2011; Torino; (Motion Picture Expert Group or ISO/IEC JTC1/SC29/WG11) No. m21435, Jul. 22, 2011, XP030049998. |
International Search Report and Written Opinion—PCT/US2012/053394—ISA/EPO—Feb. 6, 2013. |
Li, M., et al., “Playout Buffer and Rate Optimization for Streaming over IEEE 802.11 Wireless Networks”, Aug. 2009, Worcester Polytechnic Institute, USA. |
Luby et al., RaptorQ Forward Error Correction Scheme for Object Delivery draft-ietf-rmt-bb-fec-raptorq-00, Qualcomm, Inc. Jan. 28, 2010. |
Michael G et al., “Improved low-density parity-check codes using irregular graphs”, Information Theory, IEEE Transactions on, Feb. 2001, vol. 47, No. 2, pp. 585-598. |
Motorola et al: “An Analysis of DCD Channel Mapping to BCAST File Delivery Sessions; OMA-CD-DCD-2007-0112-INP—DCD—Channel—Mapping—to—BCAST—Fi1e—Delivery”, OMA-CD-DCD-2007-0112-INP—DCD—Channel—Mapping—to—BCAST—File—Delivery, Open Mobile Alliance (OMA), 4330 La Jolla Village Dr., Suite 110 Dr., Suite 110 San Diego, CA 92122; USA Oct. 2, 2007, pp. 1-13, XP064036903. |
Ohashi A et al., “Low-Density Parity-Check (LDPC) Decoding of Quantized Data,” Technical Report of the Institute of Electronics, Information and Communication Engineers, Aug. 23, 2002, vol. 102, No. 282, pp. 47-52, RCS2002-154. |
Roumy A., et al., “Unequal Erasure Protection and Object Bundle Protection with the Generalized Object Encoding Approach”, Inria-00612583, Version 1, Jul. 29, 2011, 25 pages. |
Schulzrinne, et al., “Real Time Streaming Protocol (RTSP)” Network Working Group, Request for Comments: 2326, Apr. 1998, pp. 1-92. |
Stockhammer T., et al., “DASH: Improvements on Representation Access Points and related flags”, 97. MPEG Meeting; Jul. 18, 2011-Jul. 22, 2011; Torino; (Motion Picture Expert Group or ISO/IEC JTC1/SC29/WG11) No. m20339, Jul. 24, 2011, XP030048903. |
Supplementary European Search Report—EP07757111—Search Authority—The Hague—Jan. 22, 2013. |
Wadayama T, “Introduction to Low Density Parity Check Codes and Sum-Product Algorithm,” Technical Report of the Institute of Electronics, Information and Communication Engineers, Dec. 6, 2001, vol. 101, No. 498, pp. 39-46, MR2001-83. |
Yamazaki M., et al., “Multilevel Block Modulation Codes Construction of Generalized DFT,” Technical Report of the Institute of Electronics, Information and Communication Engineers, Jan. 24, 1997, vol. 96, No. 494, pp. 19-24, IT96-50. |
3GPP: “3rd Generation Partnership Project; Technical Specification Group Services and system Aspects; Multimedia Broadcast/Multicast Service (MBMS); Protocols and codecs (Release 6)”, Sophia Antipolis, France, Jun. 1, 2005, XP002695256, Retrieved from the Internet: URL:http://www.etsi.org/deliver/etsi—ts/126300—126399/126346/06.01.00—60/ts—126346v060100p.pdf. |
Anonymous: “Technologies under Consideration”, 98. MPEG Meeting; Nov. 28, 2011-Dec. 2, 2011; Geneva; (Motion Picture Expert Group or ISO/IEC JTC1/SC29/WG11),, No. N12330, Dec. 3, 2011, XP030018825. |
Anonymous: “Text of ISO/IEC IS 23009-1 Media Presentation Description and Segment Formats”, 98. MPEG Meeting; Nov. 28, 2011-Dec. 2, 2012; Geneva; (Motion Picture Expert Group or ISO/IEC JTC1/SC29/WG11) No. N12329, Jan. 6, 2012, XP030018824. |
“Digital Video Broadcasting (DVB); Framing structure, channel coding and modulation for digital terrestrial television; ETSI EN 300 744” ETSI Standards, LIS, Sophia Antipolis Cedex, France, V1.6.1, p. 9, Jan. 10, 2009. |
Kim J., et al., “Enhanced Adaptive Modulation and Coding Schemes Based on Multiple Channel Reportings for Wireless Multicast Systems”, 62nd IEEE Vehicular Technology Conference, VTC-2005-FALL, Sep. 25-28, 2005, vol. 2, pp. 725-729, XP010878578, DOI: 1 0.11 09/VETECF.2005.1558019, ISBN: 978-0/7803-9152-9. |
Moriyama, S., “5. Present Situation of Terrestrial Digital Broadcasting in Europe and USA”, Journal of The Institute of Image Information and Television Engineers, Nov. 20, 1999, vol. 53, No. 11, pp. 1476-1478. |
Qualcomm Incorporated: “RaptorQ Technical Overview”, pp. 1-12, Oct. 1, 2010. |
Bross, et al., “High efficiency video coding (HEVC) text specification draft 6,” JCTVC-H1003, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, 8th Meeting: San José, CA, USA, Feb. 1-10, 2012, 259 pp. |
Makoto N., et al., “On Tuning of Blocking LU decomposition for VP2000 series” The 42th Information Processing Society of Japan Conference (1st term in 1991), Feb. 25, 1991, pp. 71-72, 4B-8. |
Miller G., et al., “Bounds on the maximum likelihood decoding error probability of low density parity check codes”, Information Theory, 2000. Proceedings. IEEE International Symposium on, 2000, p. 290. |
Muramatsu J., et al., “Low density parity check matrices for coding of multiple access networks”, Information Theory Workshop, 2003. Proceedings. 2003 IEEE, Apr. 4, 2003, pp. 304-307. |
Samukawa, H. “Blocked Algorithm for LU Decomposition” Journal of the Information Processing Society of Japan, Mar. 15, 1993, vol. 34, No. 3, pp. 398-408. |
3GPP TSG-SA4 #57 S4-100015, IMS based PSS and MBMS User Service extensions, Jan 19, 2010, URL: http://www.3gpp.org/ftp/tsg—sa/WG4—CODEC/TSGS4—57/docs/S4-100015.zip. |
3rd Generation Partnership Project; Technical Specification Group Services and System Aspects; Transparent end-to-end Packet-switched Streaming Service (PSS); Protocols and codecs (Release 9) 3GPP TS 26.234 V9.3.0, Jun. 23, 2010 pp. 85-102, URL:http://www.3gpp.org/ftp/TSG—SA/WG4—CODEC/TSGS4—59/Docs/S4-100511.zip, 26234-930.zip. |
Lee, J. Y., “Description of Evaluation Experiments on ISO/IEC 23001-6, Dynamic Adaptive Streaming over HTTP”, ISO/IEC JTC1/SC29/WG11MPEG2010/N11450, Jul. 31, 2010, 16 pp. |
Luby M., “Simple Forward Error Correction (FEC) Schemes,” draft-luby-rmt-bb-fec-supp-simple-00.txt, pp. 1-14, Jun. 2004. |
Luby M., “LT Codes”, Foundations of Computer Science, 2002, Proceedings, The 43rd Annual IEEE Symposium on, 2002. |
Morioka S., “A Verification Methodology for Error Correction Circuits over Galois Fields”, Tokyo Research Laboratory, IBM Japan Ltd, pp. 275- 280, Apr. 22-23, 2002. |
Qualcomm Incorporated: “Adaptive HTTP Streaming: Complete Proposal”, 3GPP TSG-SA4 AHI Meeting S4-AHl170, Mar. 2, 2010, URL, http://www.3gpp.org/FTP/tsg—sa/WG4—CODEC/Ad-hoc—MBS/Docs—AHI/S4-AHl170.zip, S4-AH170—CR—AdaptiveHTTPStreaming-Full.doc. |
Qualcomm Incorporated: “Corrections to 3GPP Adaptive HTTP Streaming”, 3GPP TSG-SA4 #59 Change Request 26.234 CR0172 S4-100403, Jun. 16, 2010, URL, http://www.3gpp.org/FTP/tsg—sa/WG4 CODEC/TSGS4—59/Docs/S4-100403.zip, S4-100403—CR—26234-0172-AdaptiveHTTPStreaming-Rel-9.doc. |
Gerard F., et al., “HTTP Streaming MPEG media—Response to CFP”, 93. MPEG Meeting, Geneva Jul. 26, 2010 to Jul. 30, 2010. |
Chikara S., et al., “ Add-on Download Scheme for Multicast Content Distribution Using LT Codes”, IEICE. B, Communications, Aug. 1, 2006, J89-B (8), pp. 1379-1389. |
Hasan M A., et al., “Architecture for a Low Complexity Rate-Adaptive Reed-Solomon Encoder”, IEEE Transactions on Computers, IEEE Service Center, Los Alamitos, CA, US, vol. 44, No. 7, Jul. 1, 1995, pp. 938-942, XP000525729, ISSN: 0018-9340, DOI: 10.1109/12.392853. |
Tetsuo M., et al., “ Comparison of Loss Resilient Ability between Multi-Stage and Reed-Solomon Coding”, Technical report of IEICE. CQ, Communication Quality, vol. 103 (178), Jul. 4, 2003, pp. 19-24. |
Watson M., et al., “Forward Error Correction (FEC) Framework draft-ietf-fecframe-framework-11,” 2011, pp. 1-38, URL, http://tools.ietf.org/pdf/draft-ietf-fecframe-framework-11.pdf. |
Watson M., et al., “Raptor FEC Schemes for FECFRAME draft-ietf-fecframe-raptor-04,” 2010, pp. 1-21, URL, http://tools.ietf.org/pdf/draft-ietf-fecframe-raptor-04.pdf. |
Qualcomm Incorporated: “RatorQ Forward Error Correction Scheme for Object Delivery draft-ietf-rmt-bb-fec-raptorq-04”, Internet Engineering Task Force, IETF, pp. 1-68, Aug. 24, 2010. |
Ramsey B, “HTTP Status: 206 Partial Content and Range Requests,” May 5, 2008 obtained at http://benramsey.com/blog/2008/05/206-partial-content-and-randge-requests/. |
Number | Date | Country | |
---|---|---|---|
20070195894 A1 | Aug 2007 | US |
Number | Date | Country | |
---|---|---|---|
60775528 | Feb 2006 | US |