The present document relates to wireless communication, and more specifically, to forward error corrections codes used therein.
Forward error correction is used in a wide variety of devices including wireless communications systems, wired communications systems, optical communication systems, disk drives, as well as many others. Forward error correction adds redundancy to transmitted information that allows for correction of errors at a receiver. For example, errors caused by noise in a transmission channel between a transmitter and receiver may be corrected at the receiver. Low density parity check codes provide excellent forward error correction performance but new techniques are needed to improve their performance even more.
This document relates to methods, systems, and devices for forward error correction in orthogonal time frequency space (OTFS) communication systems using non-binary low-density parity-check (NB-LDPC) codes. Embodiments of the disclosed technology formulate a parity-check matrix that includes non-binary entries, which provides better permutations, thereby lessening error triggering events and producing extremely low error floors such as 10−11 or 10−12.
In one exemplary aspect, a method for forward error correction is disclosed. The method includes receiving information bits, encoding the information bits via a non-binary low density parity check (NB-LDPC) code, wherein the NB-LDPC code is formulated as a matrix with binary and non-binary entries, modulating the encoded information bits to generate a signal, and transmitting the signal. The signal is modulated using an OTFS or OFDM modulation scheme.
In another exemplary aspect, a method for forward error correction is disclosed. The method includes receiving a signal, demodulating the received signal to produce data, decoding the data via a NB-LDPC code, wherein the NB-LDPC code is formulated as a matrix with binary and non-binary entries, and providing the decoded data to a data sink.
In yet another exemplary aspect, the above-described method(s) are embodied in the form of processor-executable code and stored in a computer-readable program medium.
In yet another exemplary aspect, a device that is configured or operable to perform the above-described methods is disclosed.
The above and other aspects and their implementations are described in greater detail in the drawings, the descriptions, and the claims.
Section headings are used in the present document to improve readability of the description and do not, in any way, limit the discussion to the respective sections only. Section headings are used only to facilitate readability and are not intended to limit the embodiments and technology described in each section only to that section. Furthermore, for ease of explanation, a number of simplifying assumptions have been made. Although these simplifying assumptions are intended to help convey ideas, they are not intended to be limiting.
Next generation of applications and services that use wireless transmission will demand high reliability of data transmission. For example, it is expected that applications such as autonomous vehicle driving and medical patient care may rely on wireless transmission. Therefore, it is becoming important to provide very high reliability and low error rate of data transmission in wireless networks.
Traditional error correction codes often are not able to meet the stringent bit error rate (BER) requirements of the next generation wireless networks. In some cases, the codes may mathematically allow for very low BER operations, but the computational complexity for such implementations may be excessive for next generation wireless devices. For example, many devices may be powered from a battery that is not easily replaceable (e.g., machine to machine communication devices) or may have minimal power storage capability (e.g., IoT devices, wireless devices with small, thin form factors).
The techniques provided in the present document overcome these challenges, and others. In one example aspect, a mathematically powerful error correction code is disclosed, along with its encoding and decoding implementations. The code uses non-binary symbols and achieves very high degree of reliability without causing significant increase in computational complexity. The use of such codes, along with recently developed OTFS modulation technique, is expected to meet the stringent reliability and other operational needs of next generation wireless networks.
To explain the disclosed subject matter, a brief introduction to OTFS modulation is first provided, followed by various embodiments of the error coding technology disclosed herein.
OTFS modulation has numerous benefits that tie into the challenges that 5G systems are trying to overcome. Arguably, the biggest benefit and the main reason to study this modulation is its ability to communicate over a channel that randomly fades within the time-frequency frame and still provide a stationary, deterministic and non-fading channel interaction between the transmitter and the receiver. As will be seen, in the OTFS domain all information symbols experience the same channel and the same SNR, e.g., there is no concept of time-selective or frequency-selective fading.
Further, OTFS best utilizes the fades and power fluctuations in the received signal to maximize capacity. To illustrate this point, assume that the channel consists of two reflectors which introduce peaks and valleys in the channel response either across time or across frequency or both. An OFDM system can theoretically address this problem by allocating power resources according to the waterfilling principle. However, due to practical difficulties such approaches are not pursued in wireless OFDM systems, leading to wasteful parts of the time-frequency frame having excess received energy, followed by other parts with too low received energy. An OTFS system resolves the two reflectors and the receiver equalizer coherently combines the energy of the two reflectors, providing a non-fading channel with the same SNR for each symbol. It therefore provides a channel interaction that is designed to maximize capacity under the transmit assumption of equal power allocation across symbols (which is common in existing wireless systems), using only standard AWGN codes.
In addition, OTFS operates in a domain in which the channel can be characterized in a very compact form. This has significant implications for addressing the channel estimation bottlenecks that plague current multi-antenna systems and can be a key enabling technology for addressing similar problems in future massive MIMO systems. One key benefit of OTFS is its ability to easily handle extreme Doppler channels. This is not only useful in vehicle-to-vehicle, high speed train and other 5G applications that are Doppler intensive, but can also be an enabling technology for mm-Wave systems where Doppler effects will be significantly amplified.
Further, as will be seen, OTFS provides a natural way to deliver massive processing gain, and two dimensional CDMA random access to multicarrier systems. The processing gain can address the challenge of deep building penetration needed for IoT applications, while the CDMA multiple access scheme can address the battery life challenges and short burst efficiency needed for IOT deployments.
Last but not least, the compact channel estimation process that OTFS provides can be essential to the successful deployment of advanced technologies like Cooperative Multipoint (Co-MP) and distributed interference mitigation or network MIMO.
OTFS works in the Delay-Doppler coordinate system using a set of basis functions orthogonal to both time and frequency shifts.
Through this transform, each QAM symbol is spread throughout the Time-Frequency plane (e.g., across the full signal bandwidth and symbol time) utilizing a different basis function. As a result, all symbols have the same SNR and experience exactly the same channel. The implication is that there is no frequency or time selective fading of QAM symbols, in contrast with existing modulations such as OFDM or TDMA. Since FEC is not required to counter this frequency or time selective fading impact, the full power of the code can be applied to the non-faded signal.
After the 2D OTFS transform, the signal is now spread throughout the Time-Frequency plane. In fact, the transform results in a doubly periodic extension that extends throughout frequency and time. This signal is be windowed, as will be explained below. The resulting transformed and windowed signal lies in the same domain as the familiar OFDM symbols. The OTFS signal then follows the same data path as an OFDM symbol, namely through a transmit filter bank. At the receive side, the inverse processing is performed. Note that in OFDM the transmit and receive symbols in the Time-Frequency domain are related through the multiplicative channel H(f, t), whereas in OTFS the transmit and receive symbols in the Delay-Doppler domain are related through the convolutive channel h(τ, ν).
An alternate illustration of this construction is shown in
Another observation worth noting in
To summarize, in OTFS information symbols are indexed by points on a lattice or grid in the Delay-Doppler domain. Through the OTFS Transform each QAM symbol weights a 2D basis function defined in the Time-Frequency domain. The frequency domain samples at each time are transformed into time domain waveforms using filter banks.
In OTFS, the information QAM symbols are arranged over an N×M grid on the Delay-Doppler plane, as shown in
After transmission and demodulation, the received OTFS symbols are given by the two dimensional convolution of the transmitted QAM symbols above with the Delay-Doppler channel. The result is shown in
The various components of OTFS modulation include the 2D OTFS transform, as shown in
The first step in the modulation of the QAM symbols is the 2D OTFS transform. This is given by a variant of the 2D FFT called the Symplectic Finite Fourier Transform (SFFT), defined as
Where x(m, n) are the QAM symbols in the Delay-Doppler domain, bm,n(k, l) are the basis functions associated with the [m, n]th QAM symbol in the Time-Frequency domain (with time and frequency indexed by k and l, respectively), and M and N are the number of points in the Delay and Doppler dimensions, respectively. Alternatively, M is equivalent to the number of subcarriers and N to the number of multi-carrier symbols. Notice that the Symplectic Fourier Transform differs from the more well-known Cartesian Fourier Transform in that the exponential functions across each of the two dimensions have opposing signs and the coordinates are flipped in the two domains. This is necessary as it matches the behavior of the Delay-Doppler channel representation relative to the time-varying frequency response representation of the channel.
To visualize the 2D basis functions, consider the continuous time representations of the Delay-Doppler and Time-Frequency domains. In
To summarize the initial step in the modulation process:
At the receiver, the corresponding final demodulation step is the Inverse Symplectic Finite Fourier Transform, given by
In some embodiments, the basis functions, X(k, l) are doubly periodic with period [N, M], or equivalently, as seen in
The signal in the Time-Frequency domain is thus given by
Θ(k,l)=Wtr[k,l]·SFFT{x(m,n)}
The window in general could extend beyond the period of the information symbols [N, M] and could have a shape different from a rectangular pulse. This would be akin to adding cyclic prefix/suffix in the dimensions of both time and frequency with or without shaping. The choice of window has implications on the shape and resolution of the channel response in the information domain. In some embodiments, the OTFS window also enables the multiplexing of traffic to or from multiple users.
In some digital communication systems, low density parity check (LDPC) error correction is used. While “textbook” LDPC can provide good performance that is close to the well-known Shannon limit, for practical reasons, real implementations often use sub-optimal LDPC such as repeat-accumulate techniques for implementing LDPC. As a result, the error floor seen in these implementations is of the order of 10−5 to 10′. In applications that require better performance, an outer code such as ReeD Solomon code or BCH code is used. The use of outer code increases the complexity and latency of an implementation.
The non-binary LDPC codes described in the present document can be used in embodiments in which an extremely low error floor (10−11 or 10−12) can be achieved. Due to the formulation of an H matrix that includes non-binary entries, as further described herein, the H matrix provides better permutations, thereby lessening error triggering events, thus resulting in superior performance.
In addition, the disclosed non-binary LDPC codes are especially useful in orthogonal time frequency space (OTFS) modulation based communication systems. Certain error correction codes such as LDPC are known to provide better results on AWGN channels (as compared to fading channels). Due to the property that a symbol in OTFS is spread over the entire channel spectrum, the impairment due to channel is averaged and the resulting error statistics looks more like AWGN, thus making LDPC implementations particularly suitable for OTFS modulation.
Exemplary embodiments of NB-LDPC codes in OTFS systems are described in the following subsections.
Definition: Let be a set of objects on which two operations ‘+’ and x′ are defined. is said to be a field if and only if (iff):
1. forms a commutative group under ‘+’. The additive identity element is ‘0’.
2. \ {0}−(read as: take away 0) (the set with the additive identity removed) forms a commutative group under ‘×’. The multiplicative identity element is ‘1’.
3. The operations ‘+’ and ‘×’ distribute: a×(b+c)=(a×b)+(a×c).
A field can also be defined as a commutative ring with identity in which every element has a multiplicative inverse.
Example: Rational numbers form one of the ‘infintite fields’.
Example: The real numbers form another infinite field, and so do the complex numbers.
Example: The integers do not form a field, for most of the integers do not have an integer multiplicative inverse.
Definition: Fields of finite order are known as Galois fields. A Galois field of order q is denoted as GF(q). Examples of GF(2) and GF(7) are given in Tables 2.1, 2.2, 2.3, and 2.4.
(2)-Addition
(2)-Multiplication
(7)-Addition
(7)-Multiplication
Example Tables 2.5 and 2.6 describe the ‘+’ and ‘×’ operations of (22). Since the rules of addition and multiplication are not straightforward, it is best to implement it in the hardware using two dimensional look up tables. In a simple implementations, two tables each of size q2 would be required for this for this purpose. However, this memory requirement can be brought down to q.
Linear Block Codes. A block error control code C consists of a set of M code words {c0, c1, c2 . . . , cM-1}. Each code word is of the form ci=(c0, c1 . . . cn-1); if the individual coordinates take on values from the Galois Field GF(q), then the code C is said to be q-ary. C is a q-ary linear code of length n iff C forms a vector sub space over the vector space V consisting of all n-tuples of GF(q).
Let {g0, g1, . . . , gk-1} be a basis of the code words for the (n,k) q-ary code C. There exists a unique representation c=a0g0+a1g1+ . . . +ak-1gk-1 for every code word ci∈C. Since every linear combination of the basis elements must also be a code word, there is a one-to-one mapping between the sets of k-symbols (a0, a1, . . . , ak-1) over GF(q) and the code words in C. A generator matrix G can be constructed by taking as its rows the vectors in the basis.
G can be used to directly encode k-symbol data blocks in the following manner. Let m=(m0, m1, . . . , mk-1) be a q-ary block of un-coded data.
As discussed, a q-ary code C of length n forms a vector sub space of dimension k, within the vector space V of all n-tuples over GF(q). Given this, we can talk about the dual space of C within V. Two sub spaces, S and S⊥ of the vector space V are dual to each other if for all a e S and b∈S⊥, a·b=0. The dual space of a linear code is called the dual code of C. It is denoted by C⊥. It is a vector space of dimension n−k. It follows that a basis {h0, h1, . . . hn-k-1} for C⊥ can be found. This is used to construct a parity-check matrix H.
For every c∈C, cHT=0. Similarly if cHT=0, c is a code word. It follows that GHT=0.
Using Gaussian elimination and column reordering on a linear code C with a generator matrix G, it is always possible to obtain a generator matrix of the form in Eq. 2.4.
When a data block is encoded using G, the data block is embedded without modification in the first k coordinates of the resulting code word. This is called systematic encoding.
After decoding, the first k symbols are removed from the selected code word and passed along to the data sink.
Given a systematic generator matrix of the form Eq. 2.4, a corresponding parity check matrix can be obtained as:
Circulant Matrix: An n×n circulant matrix C takes the form:
A circulant matrix is fully specified by one vector, c, which appears as the first column of C. The remaining columns of C are each cyclic permutations of the vector c with offset equal to the column index. The last row of C is the vector c in reverse order, and the remaining rows are each cyclic permutations of the last row. In the context of non-binary LDPC codes, a typical circulant matrix will have the following form:
Permutation Matrix: The permutation matrix Pπ corresponding to the permutation:
In the context of non-binary LDPC codes, the non-zero elements can, in general, be different.
Let b, k, and t be positive integers such that k<tb. A (tb, k) linear code Cqc over GF (2m) is called a quasi-cyclic code when: 1) Each code word in Cqc consists of t sections of b bits each; and 2) every t-sectioned cyclic-shift of a codeword in Cqc is also a codeword in Cqc. Such a QC code is called a t-section QC code.
Regular Codes: If the number of non-zero elements in each column of the H matrix is same, we call it a regular code. On the other hand, if the weights of the columns differ, then its called irregular code.
There are several types of probabilities that can be associated with a random variable, x. For the event {x=ak}, suppose that E is an event whose effect on the variable x is under question. The prior probability refers to the probability P (x=ak) that the variable x takes the value ak. It is also known as the intrinsic probability. The prior/intrinsic probability for the variable x with respect to the event E is denoted as PEint (x=a)=P (x=a). The posterior (a posteriori) probability is the conditional probability for random variable x with respect to event E is denoted as PEpost (x=a)=P (x=a|E). The intrinsic and posterior probabilities represent the probability before and after taking into account the event E. Using Bayes' rule, the posterior probability can be written as:
The term P (x=a) is the intrinsic probability; the term P (E|x=a) is proportional to the ‘extrinsic’ probability, which is the probability that describes the new information for x that has been obtained from the event E. The extrinsic probability for x with respect to E is defined by, PEext (x=a)=cx′P (E|x=a) where c′ is the normalization constant.
These terms will come up again when we discuss the messages passed in the sum-product algorithm.
Block codes, such as LDPC codes, can be represented using bipartite graphs known as Tanner Graphs. This is best illustrated with an example. The Tanner graph for the parity check matrix H in (Equation 2.12) is shown in
The reliability of each variable node (the variable node's belief that it takes on each of its possible values, which is a conditional probability distribution or LLR) is initialized from the received symbols (after equalization, so it is assumed that soft information about each variable is an independent piece of information). In the sum-product algorithm (SPA), or message passing algorithm, this reliability information is refined by being passed back and forth between variable nodes and check (constraint) nodes iteratively. In each iteration, the information (hopefully) becomes more reliable.
The a priori, or intrinsic, information associated with a node is the information known about it before the decoding starts. For a variable node, this is the information from the received symbols from the channel. The extrinsic information associated with a node a is the information provided by the decoder except for the information provided by node a. The a posteriori information for node a is all of the information provided by the decoder without removing the information provided by node a. This is the information that a variable node will use to make a final decision on its value.
Building a decoder for a generic LDPC code is not viable in hardware. It would require an enormous amount of wiring and silicon. For this reason, commercial LDPC implementations use a special class of codes known as Quasi Cyclic (QC) LDPC codes. Decoder architecture for QC binary and non-binary (NB) codes are simple to implement. QC-binary systematic encoders can be implemented using simple shift-registers with linear complexity owing to a QC-Generator matrix. However, a QC-non-binary encoder does not entail, in general, a QC-G matrix and so would require quadratic complexity. However, for short/medium length codes, such as some of the example embodiments disclosed here, encoder complexity is not of prime concern currently.
Two examples of the disclosed subject matter include LDPC codes a) structured regular systematic repeat accumulate codes (RA codes) over the non-binary, structures (qn) and b) codes that are created out of just permutation matrices. Code/H-matrix designs for a) are based on computer search whereas b) are based on analytic (Euclidean geometry) techniques. Both these methods are described in the sequel.
Construction of a parity check matrix, H, for a non-binary RRA (regular repeat accumulate) code consists of two parts; 1) construct the binary-H matrix and 2) convert it to non-binary.
Parity check matrix for RRA code is given by, H=[H1, H2], where H1 is the parity part and H2 is the systematic part. Both H1 and H2 are composed of blocks of circulant permutation matrices and zero matrices of block size Q. There are Q different possibilities for each block/sub-matrix. Each block can be fully specified by an exponent that denotes the number of right shifts required for the non-zero element of the first row. For example, an exponent of 0 denotes an identity sub-matrix of size Q. The distinct exponents can thus be 0, 1 . . . (Q−1). An all-zero block is denoted with an exponent of −1 (or 16, at times). The parity part, H1, includes circulant permutation blocks while the systematic part, H2 is square and double diagonal on the block level, meaning that each Q-sized block in H2 consists of identity elements on its main and lower diagonal, and zero blocks elsewhere.
Henceforth the methodology by which H1 can be constructed to contain no four-cycles in its Tanner graph is discussed. Four cycles can be formed either internal to H1 or between H1 and H2. The method for reducing the amount of six-cycles is mentioned as well, although our application total removal of 6-cycles may not be feasible.
Let L be the number of Q-sized block rows and J be the number of block columns in H1. From the definition of RRA, the number of block columns in H2 is J. N, the number of columns in [H1, H2], is N=(L+J)Q. A necessary and sufficient condition for removing 4-cycles for any configuration of rectangle abcd below (note that a, b, c, d are block circulant matrices as described previously) is given by,
Now offsetting column (a,c) by X and column (b,d) by Y (offsets are mod Q), as well as row (a,b) by D and row (c,d) by E. The transformation results in
aa+X+D
cc+X+E
bb+Y+D
dd+Y+E (Equation 3.2)
and the new equation is
(a+X+D)−(c+X+E)≠(b+Y+D)−(d+Y+E)⇒(a+D)—(c+E)≠(b+D)−(d+E)⇒a−≠b−d (Equation 3.3)
This shows that the rectangle rule is invariant under constant offsets to entire rows or columns. Therefore, without loss of generality, we can add offsets such that the first row and the first column consist of identity sub-matrices (‘0’ exponent blocks). By exploiting this property the search space for solutions is decreased. Having the first row and first column exponents as 0 implies that each row and each column must contain unique values lest a 4-cycle be formed. Note that, a 6-cycle will be formed with the first row and column if a≡d (mod Q).
The algorithm used to design the H matrix works as follows: Let W be an array containing all ones, [1,1 . . . 1], of length Q. Let it denote the inclusion of numbers {0 . . . (Q−1)} in a given set, say A. That is W=[1,1, . . . 1] implies A={0,1, . . . 15}. As another example, let W=[1,0,1,1,0,1,1,1,1,1,0,0,0,1,1,1]. This implies that A={0,2,3,5,6,7,8,9,13,14,15}. Similarly for other values of W.
Recall that the first row and column of H1 are preselected to have 0 exponents. Starting with index (2,2) of H1, and working row-wise (or column-wise, since the rules are invariant under transposition), solve for the set of possible exponent values for each index of H1 that will form 4-cycles. This is done as follows: For each index, compare it with all slots with indexes less than its own, such that the current index is d and the one it is compared with is a, with b and c found to form the rectangle. Solve Equation 3.1 (recall a, b, c are known values) to find the value of d that will form a 4-cycle. Remove this value from A by suitably modifying W.
After this process, two situations may arise. Either there is at least one option for d, in which case pick one at random, or there are no options, in which case the run is aborted and the process must be restarted. If a matrix is successfully generated, it is passed into a function that runs a quick check to make sure that it indeed does pass the conditions for no 4-cycles, as well as counts the number of instances of a a≡d (mod Q) to get a metric for the amount of 6-cycles.
In some example embodiments, the following choices may be made: N=256, Q=16, and L=1 . . . 8, J=15 . . . 8. High rate codes,
such as 15/16, 14/16 will have low number of layers (1 and 2). This is undesirable due to the low degree of H1. This is especially so for 15/16 where the degree is equal to 1, rendering the code ineffective. Simulation results of these codes and ensuing discussions are detailed below.
Good high rate codes may be designed by increasing the number of layers. Since J must be increased by the same proportion as L to keep the rate fixed, N=(L+J)Q has to increase for a given constant value of Q. However, increasing N will cause greater latency and hardware resources. Since each row and column must contain unique values (prior to adding an offset of one to even rows), Q≥J is necessary to eliminate 4-cycles.
Relation between Code Rate (R), Number of Layers (L), sub-matrix Size (Q), Code Size (N) are depicted in Table 3.1.
Wherever there are no entries in Table 3.1, it implies that there is no solution existing using this algorithm. For example, for a 9/16 code, the only possible value of N is 256 and so on. Entries in the table with * indicates example embodiments. Entries with ˜ indicates there are better alternative parameters possible for the given value of N. Entries with {circumflex over ( )} denote 256≤N≤640.
As can be seen in Table 3.1, there is no solution for R=15/16, L>1, N≤640. Only one possible other option exists for R=14/16, none other exist for R=13/16 and several possibilities exist for R=12/16.
Currently the elements in a sub-matrix are chosen at random. However, this has to be changed to maximize the entropy. This is done as follows: Say we have a code that has M=N−K=16. We have 16 entries in the syndrome vector. Consider the 1st entry; by maximizing the entropy of the bit/symbol at that location.
The LDPC encoder belongs to the transmit chain, and it sits between the scrambler and the data/symbol mapper, as shown in
The LDPC encoder encodes the output of the scrambler whether or not the RS code is not being used. It takes a k-symbol message vector and encodes it into an n-symbol codeword vector (n>m). This adds structured redundancy that the LDPC decoder can exploit to correct errors in the codeword caused by the channel. For this reason, the encoder comes before the data is mapped to constellation symbols and modulated
The architecture must be able to implement any of the codes from the set {8/16, 9/16, 10/16, 11/16, 12/16, 13/16, 14/16, 15/16}. Switching between code rates should be able to happen from frame to frame. The encoder must support shortened codewords in the form of “zero stuffing” and must support zero padding.
At the top level, the LDPC encoder has four components: (1) the top controller, (2) the input interface to the scrambler, (3) the core, and (4) the Tx interface at the output. The signals and connections between the blocks are shown in
The top controller controls switching between normal encoding operation, zero stuffing operation, and zero padding operation based on configuration registers that set the number of information bytes per frame, the number of zero to stuff/pad, and whether to stuff or pad if the number of zeros is larger than 0. It also controls the ready signal that goes to frame buffer FSM and times it so that the information symbols coming into the encoder pause at the correct time
The input interface converts the 8-bit output from the scrambler to a 32-bit input for the LDPC encoder and pulses a valid signal when the 32-bits are ready. During the last codeword of the frame and if zero stuffing is enabled, zeros may be appended to the input in order to make up a full codeword. These zeros are not part of the information, so they must be removed before the codeword is written to the Tx FIFO in the output interface.
The core stores the information symbols from the input interface in a memory, encodes the information when enough information symbols have been stored (the number depends on the code rate), and outputs a 512-bit codeword 32-bits (16 symbols) at a time. It also raises a valid signal concurrently with each 32-bit codeword segment.
The Tx interface is an asynchronous FIFO that accepts a 32-bit input and a concurrent data valid signal and reads out between 2 and 10 valid bits depending on the current QAM constellation used by the modulator. During the last codeword and if zero stuffing is enabled, zeros added in the input interface must be removed.
As shown in
The I memory is a ping-pong memory where the scrambler writes into one memory and the LDPC encoder reads from the other memory. The memory to write to and read from is controlled by the LDPC encoder's controller. The scrambler supplies the write data and data valid signals, and the LDPC encoder provides the write address, read address, and the memory select signals. The memory width is 32 because an entire submatrix's worth of information symbols is read out every cycle (z=16, and log 2(q)=2, so the total width is 32). The memory depth is 15 since the highest rate code has 15 block rows in its generator matrix. The other code rates will not use the full depth of the memory.
The G memory is a ROM that stores the submatrices for all G matrices (all code rates) so that it can switch between matrices quickly by just changing an offset to the read address that puts its first read at the desired matrix's first address (
The width of the ROM is calculated as nSPU·z2·log2 q, where nSPU is the number of SPUs in the architecture (here nSPU=2, so the width is 1024). The depth is set as the sum of the depth required by each matrix and the depth of each matrix is the number of block rows in the matrix multiplied by the number of groups of block columns, which is the number of block columns in P divided by nSPU rounded up to the nearest integer (
A single SPU computes one submatrix multiply operation (one miPi term in the sum in Equation 2.2) every clock cycle.
The controller generates all read addresses for the I and G memories, the memory select signal for the I memory, the signal that resets the running sum in the SPUs when a new set of block columns is processed, and the ready signal that controls the frame buffer FSM that feeds information symbols to the RS encoder and scrambler. The controller keeps track of the number of writes to the I memory and the status of the encoder (based on the number of block column sets supplied by configuration registers and the cycle count) to indicate whether the encoder is ready for more data to be written to the I memory and which memory to should be written.
Operation of the encoder includes the following:
The message passing equations detailing how to process the reliability information at the variable nodes and check nodes is detailed below.
Linear Combinations of RVs over Finite Algebraic Structures: Consider the discrete random variable x∈GF (q), which can take on values in the set {0,1, . . . , q−1}. It has a probability mass function (PMF) P (x) given by
P(x)={(P(x=0)=p0),(P(x=1)=p1), . . . ,(P(x=q−1)=pq−1)} (Equation 4.1)
where p0, p1, . . . , pq-1 sum to 1. The question is: what is P (h×x) where h∈GF (q)?
In other words, we are asking for each of the probabilities P(h×x=0), P(h×x=1), . . . , P (h×x=q−1). Rearranging these terms:
In general P(h×x=r)⇒P (x=h−1×r). Note that h−1×{0,1, . . . , q−1} results just in reordering (shuffle) of the set {0,1, 2, . . . , q−1} by virtue of the closure property of the field multiplication. Thus we see that P (h×x) is a shuffle of the elements of the vector/PMF P (x) by h−1. The exact shuffle depends on the primitive polynomial that was used to build the field GF (q), which sets the multiplication table for the field.
Now, consider the following check equation, where all elements are from GF (q):
h1×x1+h2×x2+h3×x3=0 (Equation 4.3)
Each xi has vector of probabilities, one for each of its possible values, where all elements sum to 1. Now let us ask the question: Given the above relation/constraint between x1, x2 and x3, what is the distribution of x1, given the distributions of x2 and x3?
From a Galois field Fourier transform, for y, s, t, ∈GF (q), if y=s+t, the distribution of y will be the finite field convolution of the distributions of s and t. In this way, the distribution of x1 can be found as detailed below. From Equation 4.3:
−h1×x1=h2x x2+h3×x3 (Equation 4.4)
The distribution of −h1×x1 in the above equation, thus, will be the finite field convolution of the distributions of h2×x2 (a shuffle of the distribution of x2 by h2−1) and that h3×x3 (a shuffle of the distribution of x3 by h3−1). Note that −h1 (viz. the additive inverse of h1) is the same as h1 in finite fields of the form GF (2n). So the distribution of x1 is obtained by shuffling the result of the above convolution by (h1−1)−1=h1.
There is another way of looking at the above equation. This interpretation will be appreciated later when we deal with the mixed-mode sum-product algorithm (SPA). Equation 4.4 can be re-written for finite fields (note that the negative sign is dropped, which presumes we are in a field of the form GF (2m)) as
x1=h1−1×h2×x2+h1−1×h3×x3 (Equation 4.5)
From Equation 4.5, the distribution of x1 is obtained in two steps: (a) shuffle the distributions of x2 by h1×h2−1 and x3 by h1×h1−1, and (b) take the (finite field) convolution of the shuffled distributions.
Moving from the Linear to the Log Domain based on a Linear Combination of Two Random Variables: Consider the discrete random variable x∈GF (2). The log likelihood ration (LLR) of x, LLR(x), is a scalar defined as:
Now consider the discrete random variable y∈GF (22). The LLR of y, (LLR(y)), is a vector of length 22−1=3. It is defined as:
The LLR for elements from higher order fields can be defined similarly with q−1 elements in the LLR vector.
Let us now address the question: Given the LLR vectors of x1 and x2, how can we compute the LLR vector of h1×x1+h2×x2 where (h1, h2∈GF(4))? To answer this, consider an element, αi∈GF(q) (i.e. αi∈GF (4) can take values {0,1,2,3}). Then,
LLR1(v) corresponds to the LLR value of x1 at x1=v, and similarly for LLR2 (v). The notation v∈GF (q)\{0, h1−1×αi} denotes that v takes all the values in GF (q) except {0,h1−1×αi}. The second equality is obtained by dividing the top and bottom terms by P (x1=0)·P (x2=0) and by expanding the sums inside the P(·)s to account for all events in which the sum is true. The third equality comes from expanding the numerator and denominator and separating the log of the division into a difference of logs. Equation 4.8 can be used for each value of αi (for a total of q−1 times) to calculate each element of the LLR vector LLR (h1×x1+h2×x2) and thus obtain the entire LLR vector of h1×x1+h2×x2.
As has been detailed, in (·), (αi−h1×v) is (αi−h1×v)(·) and (h2−1×(−h1×v)) is (h2−1×(h1×v))(·). In the following, the subscripts indicating the algebraic structure will be omitted for clarity.
Consider α∈GF (4). α takes value 0,1,2,3. Let h1=2 and h2=3. From Table 2.6, h1−1=3 and h2−1=2. Therefore,
Similarly, we can compute, LLR(h1×x1+h2×x2=2) and LLR(h1×x1+h2×x2=3).
Linear Combination of n Random Variables: Once again, consider the following check equation,
h1x1+h2x2+h3x3+h4x4=0 (Equation 4.11)
Where h1, . . . , h4, x1, . . . , x4 are elements from GF (q). Next, define σ and ρ as:
σ1=h1x1,ρ1=h1x1+h2x2+h3x3+h4x4
σ2=h1x1+h2x2,ρ2=h2x2+h3x3+h4x4
σ3=h1x1+h2x2+h3x3,ρ3=h3x3+h4x4
σ4=h1x1+h2x2+h3x3+h4x4,ρ4=h4x4 (Equation 4.12)
which can be re-written as,
σ1=h1x1,ρ1=h1x1+ρ2
σ2=σ1+h2x2,ρ2=h2x2+ρ3
σ3=σ2+h3x3,ρ3=h3x3+ρ4
σ4=σ3+h4x4,ρ4=h4x4 (Equation 4.13)
In accordance with the foregoing description, the probability distributions of each b and each p can be computed recursively using convolutions in the linear domain, and LLR vectors of σ1, σ2 . . . σn-1 and ρ1, ρ2 . . . ρn-1 can be computed recursively using Equation 4.10 in the log domain. The σ computations represent a forward recursion, and the ρ computations represent a backward recursion. This is a very efficient method to compute these values. However, the band p quantities are not the LLR vectors of the x1s. Therefore, the question is how can we compute the LLR vectors LLR(x1), LLR(x2), LLR(x3), LLR(x4) from the LLR vectors of σs . . . and ρs?
As discussed earlier, LLR(x1), is the shuffle of LLR(σ1) by (h1−1)−1=h1. Note that this is true because we are shuffling by the inverse h1−1 and the value to shuffle by is the inverse of the shuffle value. LLR(x4) can similarly be obtained from LLR(σ4). (4.11) can be written in terms of σs' and ρs' as below.
σ1+h2x2+ρ3 (Equation 4.14)
ρ2+h3x3+σ4 (Equation 4.15)
Therefore,
σ1+ρ3=−h2x2 (Equation 4.16)
σ2+ρ4=−h3x3 (Equation 4.17)
Using Equation 4.10 we can obtain LLR(σ1+ρ3), LLR(x2) is obtained by shuffling this distribution by (−h2−1)−1=h2 (in GF(2m), but differs in GF(pm) where ρ is a prime not equal to 2 and ∈(·)). Similarly proceeding we can obtain LLR(x3).
This idea can be generalized for obtaining LLR(xi) for i=1 . . . N.
Methods to Calculate max*: Computation of max*(x,y)=ln(ex+ey) is an important component finding the linear combination of RVs in the log domain. To see options for its calculation, consider the Jacobian logarithm:
The second equality can be found by assuming x>y and factoring out ex and then doing the same assuming y>x.
An implementation of the same is also considered as shown in
Now consider a 3-input max* (max3*) as given by Equation 4.19. This enables us a low resource high delay implementation of max3* as given in
An approximate implementation of the same can be achieved similarly by concatenating
In general,
max*(x1,x2,x3, . . . xn)=max*[x1 max*(x2 max*(x3,max*(x4. . . xn−2,max*(xn−1,xn)] (Equation 4.20)
Following the architecture given in
Now consider the four input max* as below,
Equation 4.21 points to a parallel architecture (high resource but low delay) for the implementation of max4* as given in
A parallel implementation for max* with N (power of 2) elements (maxN*), will use N−1 max2* elements and would use log N2 units of clocks.
Message passing equations from variable to check messages: There are at least two types of messages passed in the SPA: from variable nodes to check nodes (V2C) and check nodes to variable nodes (C2V). First, let us look at the V2C messages. The variable node basically enforces a constraint that all incoming messages to it should agree on its probability of taking on each value.
Suppose that the variable node N is connected to K+1 edges with associated nodes x0, x1, . . . , xk. The constraint set for this node is: SN={(x0, x1, . . . , xk)|(x0=x1=x2 . . . =xk=v1)}. Let P(x1=v1), P(x2=v1)·P(xk=v1) represent the probabilities of x1, x2, . . . , xk to take the value v1. Now, the probability of the output variable of the node x0 to take value v1 is given by
where c′ is a normalization constant that ensures Σv∈GF(q)P(x0=v)=1.
In the log domain, Equation 4.22 becomes
Note that no normalization constant is needed here. In summary, for an equality node, the outgoing LLR along one edge is the sum of the incoming LLRs along all of the other edges.
Check to Variable Messages: Similar to variable nodes enforcing an equality constraint on incoming messages, check nodes enforce a parity check constraint defined by a row of the parity check matrix H. A parity constant states that a linear combination of variables must equal a particular value (usually 0), such as in the following:
hi,1×vi,1+hi,2+ . . . +hi,k×vi,k=0 (Equation 4.24)
In the context of the SPA, the k edges entering a check node represent probabilities that the variable node on the other side of the edge takes on each of its possible values. When viewed as k−1 incoming edges and 1 outgoing edge, the k−1 incoming edges and the parity check equation that must be satisfied are used to calculate the probability/LLR vector of the variable node attached to the outgoing edge (note that the outgoing variable node's message is not used). Intuitively, the probability that the node attached to the outgoing edge takes on a value αi is equal to the probability that nodes attached to the incoming edge sum to the additive inverse of αi, −αi (in the case of GF(2m), αi=−αi). To actually calculate the probability/LLR vector of the kth variable node, the equations derived in earlier sections are used.
In particular, the forward/backward recursion is done using all incoming messages to the check node to calculate each a and p, and then the appropriate σ, ρ LLR vectors are “added” (where added means the operation defined in (4.6) or (4.10)) to form each variable's C2V message.
Tanner Graph and Message Passing: A pictorial representation of the above using Tanner graphs is shown in
Layered SPA: The LDPC decoder belongs to the receive chain, and it sits between the Slicer/LLR Computer and the Reed-Solomon (RS) Decoder (
There are two different options when it comes to decoding algorithms. A flooding type algorithm where each variable node and check node pass information iteratively until convergence and a layered algorithm where only a subset of variable and check nodes pass information iteratively until convergence. In this example implementation the layered algorithm is implemented. It is more easily implemented in hardware and converges faster.
Decoding operation begins with the LLR values coming from the LLR computer (described in section 5). VNs receive this a priori LLR values for each of the VNs and initialize the Q memory with these values. V2C messages are calculated. In the first iteration they are just the initial a priori LLR values). In general, V2C messages are calculated as follows: (Qnew=Qold+Rnew−Rold). C2V messages are responsible for the R values both old and new are used to calculate the extrinsic Q values. After all of the iterations and layers have completed, these Q values (without subtracting the old R values) are sent to the LLR Slicer to be transformed into bits.
A Variable Node (VN), also referred to as the bit node, initially receives the a priori LLR values from the LLR computer. These values initialize the Q memory before any extrinsic or a posterior LLR values have been computed. The VN provides the V2C messages that allow for the CN to calculate the calculate the σ, ρ, and R values which are required in order to calculate the a posterior LLR values. The final LLR values are stored in the Q Memory and ultimately end up in the LLR slicer for conversion from LLRs to bits.
The Quasi-Cyclic Router (QC_ROUTER) takes in as inputs the current submatrix to provide the routing requirements for the given submatrix that is being processed. The QC Router is implemented using a series of mux stages in order to route the inputs to the correct outputs.
The LLR Permuter (LLR_SHIFT) takes in as inputs the current set of 3 LLRs and permutes them based on the input hi,j element. For example, the hi,j element (which is an element of GF(4) {1,2,3} only nonzero values are considered) could be 3. The GF(4) inverse of 3 is 2, so this would permute the incoming LLRs by 2 shifts such that {LLR1, LLR2, LLR3} would become {LLR3, LLR1, LLR2}. The architecture of the LLR Permuter is also known as LLR_Shift.
The Check Node (CN) is used for the computation of σ and ρ, the forward and backward recursion respectively, and ultimately, the computation of R. R is essentially the set of LLR a posterior values for a given code symbol and its value is iterated on until the decoder is complete. The final R values will ultimately end up at the VN where the final Q values will be calculated. The input to the CN is the extrinsic LLR values that are computed by subtracting the new Q values with the previous R values as explained earlier. Internally, the CN will store the previous extrinsic information and that will be used to calculate the new R values. This process will continue until all layers and iterations have completed. The value of the row weight will determine the number of variable nodes that each check node sends a message to. The Check Node is comprised of two elementary check nodes (ECN) to parallelize the σ, ρ, and R calculations.
The Elementary Check Node (ECN) is the building block for the Check Node. Its purpose is to calculate the following equation:
Herein. the above equation may be interpreted as three separate equations, one for each value of α.
Due to the LLR permuters before and after the CN, all values of h and h−1 are equal to 1 since 1 is the inverse of 1 in GF(4). This simplification makes the computation of all the different combinations much less hardware intensive. The Elementary Check Node architecture is shown in
In this architecture, all combinations of LLR additions are created and sent to the correct max* operation depending on α. The max* operation is in its own module due to the need to approximate it as simply a max operation. Currently it is implemented as purely a max operation but can be quickly adapted to any variety of approximations.
In some embodiments, σ is computed as (given by Eq. 4.12) in the forward recursion. while ρ is computed in the backward (Eq. 4.12) recursion, and R (Eq. 4.8) is computed using σ and ρ, as shown below and with a pictorial representation of it given in
Received symbols from the equalizer prior to slicing can be represented as,
y(k)=γkx(k)+n(k) (Equation 5.1)
where y(k) is the received QAM symbol, x(k) is the transmitted QAM symbol, γk is a channel quality measure (relative attenuation/fading the transmitted symbol have experienced) and n(k) is Additive White Gaussian Noise (AWGN). In OFDM systems, for example, due to fading, certain sub-carriers undergo erasure and γk can be thought of as a quantity inversely related to the intensity of erasure. In OTFS systems, the fading is absent and here we assume γk to be always 1.
We use LDPC codes in GF(4). Thus we have 2 bits per each LDPC symbol. LDPC decoder requires a measure of reliability for each symbol, viz. LLR of all the 22−1 bit sequences to perform the soft decoding (equation 4.7). This can be calculated in two ways, a) symbol wise and b) bit wise. Symbol-wise LLR computation is applicable to even order constellations (QAM4, QAM16, QAM64, QAM256 etc.) whereas bit-wise LLR computer may be used for both even and odd order constellations (QAM4, QAM8, QAM16, QAM32, . . . )
LDPC symbol-wise LLR computer: Consider a QAM constellation. Each constellation point consists of say dR number of bits. For e.g., dR=6 for QAM64 and so on. Now, consider a set with c number of consecutive bits in this constellation. Let's say c=2. then we have 3 such sets associated with each QAM64 symbol (first, middle and last). Consider one such set. It can take 2c(22=4 in our example) distinct values in general. Let us assume that one of the constellation points, say, x was transmitted. Let us say y was received. y can fall anywhere on the constellation plane. Out of the 2dR constellation points, let there be a number of constellation points where these c bit positions of x are identical (s(c)=x1x2 . . . xc) where s is the LDPC symbol). Let there be β number of constellation points where these c bit positions are a 0 (s(c)=0102 . . . (·)0c).
Since we assume AWGN in Equation 5.1, we get,
P(x(k)=(k))=a/y(k))=(K/σ)×e−(y(k)−a)
where K is a constant and σ2 is the noise variance.
From above, LLRs can be written as,
An LLR vector is computed by computing Equation 5.3 for all combinations of x1x2 . . . xc. (Thus there are a total of 2c−1 number of LLRs in a LLR vector.)
From Equation 5.2, this can be seen to be of the form,
For higher order modulations, α and β in the above expression are prohibitively expensive. So, for computational tractability, we limit the search space to a few salient constellation points determined by the position of the current received symbol y(k). This approximation translates to only using the closest constellation points that has s(c)=x1x2 . . . xc and s(c)=0102 . . . 0c in the computation of LLR(s(c)=x1x2 . . . xc), and has yielded good results in a gray coded constellation.
Bit-wise LLR computer: To find LLR(s(c)=x1x2 . . . xc), we can, instead, look at the LLRs of individual bits and combine them to get the required LLR of a certain group of bits.
For example, the LLR for the previous example is recalculated bit-wise, as shown in
From the definition of LLRs, we know that P(s(1)=1)=P(s(1)=0)×eLLR(s
Computing similarly we can see that, LLR(s(c)=10)=LLR(s(2))−0, LLR(s(c)=01)=0+LLR(s(1)) and LLR(s(c)=00)=0+0.
LLR computation described thus, works for both odd and even order constellation whereas the symbol-wise LLR computer works only for even order constellations.
Conversion between symbol-wise and bit-wise LLRs: Given a finite sequence of bits:
{b1,b2, . . . ,bn}≡{bi} (Equation 5.6)
Each bit has an associated probability, P(bi=1),P(bi=0) where P(bi=1)+P(bi=0)=1. Since the bits are assumed to be independent, then
Every sequence of bits that bi can be represented with an unsigned integer (e.g., 3=‘10’). Let P(k) denote the probability that bi takes the form of the sequence of bits which is the binary representation of the unsigned integer k (i.e., k=b(1)×20+b(2)×21+ . . . b(n)×2n−1). Using this notation, LLR vector of size n sequence (of bits) is:
As expected, this LLR vector is of size 2n−1. Another way to represent this LLR vector is by individually describing each element.
Where k[i] denotes the ith bit in the binary representation of the unsigned integer k. Since k|i| can either be 0 or 1, in the case where it is 0 it cancels out with the denominator. We now see the resemblance of an LLR of an individual bit, represented by LLR(bi) as the LLR of bit bi:
In summary, it has been found that
if and only if the sequence of bits bi is independent. Thus a length 2n−1 LLR vector can be represented with n bit-wise LLRs. In particular, if (LDPC) symbols consist of 2 bits, then a symbol-wise LLR will have three elements: (a, b, c). This result shows that the bits are independent if and only if a+b=c.
LLR calculation: Given a constellation with an associated bit pattern, it can be converted into two tables containing the individual bits (see below, for example):
For each such table of bits, given a soft decision point (output of the equalizer), a single bit LLR can be calculated using:
1) Find the distance to the nearest point containing a 0, let this distance be d0.
2) Find the distance to the nearest point containing a 1, let this distance be d1.
LLR=d02−d12 (Equation 5.18).
3) Check to see if this makes sense—if very close to 1, then much farther from 0, and d02>d12, meaning that the LLR will be positive. If we are very close to 0 and far from 1, then the LLR will be negative, as expected with the convention that
In some embodiments, if the constellation bit pattern obeys a regular symmetry (as do the even constellations) then simple computation can be done to deduce the closest 0 point and closest 1 point. If, on the other hand, the constellation bit-pattern does not have such a symmetry, a look-up table based technique is be used to find these points. The look-up table, with finite number of elements, in some cases will work as an approximation to the direct computation (and thus would be not as exact as the direct computation)
In some embodiments, hardware supports symbol-wise LLR computer.
LLR computer Architecture: As shown in
LLR computer: Even order constellations has even number of bits in each symbol. Each LDPC symbol in GF(4) consists of 2 bits. Each such symbol will have 4 neighbors. Thus in the case of 4-QAM there's only one set of neighbors, in the case of 16-QAM there are 2 sets of neighbors and so on up to 1024-QAM in which case there are 5 sets of neighbors. Each set of 4 neighbors is used to calculate 4 Euclidean distances between the neighbors and the input soft estimate, resulting in sets of 4 such distances which in fact represent LLRs. One such LLR corresponds to the LDPC symbol ‘00’. This LLR will be subtracted from the other 3, (the normalization of LLR). These 3 normalized LLRs will then be sent to the LDPC Decoder to be processed.
The Euclidean distances between the signal y=y_I+jy_Q and a point of the constellation n=n_I+jn_Q are calculated as follows:
∥y−n∥2=(yI−nI)2+(yQ−Q)2 (Equation 5.20).
A block diagram of the algorithm described above shown in
The hard decisions from the Slicer (signal s), together with the respective QAM order are used to derive the addresses to a ROM that contains all the “neighbors” associated with every point in all the square QAM constellations.
To calculate the ROM allocated for each constellation, the following may be used:
The method 2400 includes, at step 2420, encoding the information bits via a non-binary low density parity check code, which can be formulated as a matrix with binary and non-binary entries. Section 2 of the present document provides examples of the matrix representation of the NB-LDPC code.
The method 2400 includes, at step 2430, modulating the encoded information bits to generate a signal. In some embodiments, the modulating is performed in accordance with orthogonal time frequency space (OTFS) modulation. In other embodiments, the modulating is performed in accordance with orthogonal frequency division multiplexing (OFDM).
The method 2400 includes, at step 2440, transmitting the signal over a channel. As previously discussed, one advantageous aspect of the use of NB-LDPC is to both use the AWGN statistical properties of noise ingress and also use the low complexity implementation of NB-LDPC to achieve very low BER performance. In particular, because OTFS modulation spreads a symbols over the entire bandwidth of the channel, frequency localized noise ingress still does not degrade performance of NB-LDPC, due to “averaging” or “whitening” effect over the entire bandwidth of the symbol.
In some embodiments, transmitting the signal includes transmitting a radio frequency signal over the air or a wire. In other embodiments, transmitting the signal includes transmitting an optical signal over a fiber optic cable.
The method 2500 includes, at step 2520, demodulating the received signal to produce data. In some embodiments, the demodulating is performed in accordance with orthogonal time frequency space (OTFS) modulation. In other embodiments, the demodulating is performed in accordance with orthogonal frequency division multiplexing (OFDM).
The method 2500 includes, at step 2530, decoding the data via a non-binary low density parity check decoder, which is characterized by a matrix with binary and non-binary entries.
The method 2500 includes, at step 2540, providing the decoded data to a data sink. The data sink may be, for example, applications and services running on the wired or wireless device that receives the signal.
Aspects disclosed in this patent document include a forward error correction method, including encoding and decoding data which using a non-binary LDPC error correction code. Non-binary LDPC error correction code may be transmitted and received over a wireless or wireline medium, or used in devices such as disk drives. Encoded data may be further modulated using OTFS or OFDM or other modulation. An encoder H matrix is described. The H matrix may be found via a search methodology including the removing of 4 cycles and 6 cycles, followed by interleaving and choosing non-binary values for the circulant matrix. The error floor of a non-binary LDPC code found using this methodology to determine an H matrix has very low error flow far below other codes. A single H matrix may be used to generate punctured codes with a variety of rates such 8/16, 9/16, . . . , 15/16, and others where each code rate is derived from the same H matrix An architecture for the above is disclosed including a partially parallel architecture for the decoder. Compared to binary LDPC, a modified permuter may be used. An architecture for decoding the NB LDPC layered SPA (sum product algorithm) modified for non-binary use and architecture. Two different forms of max* computation are disclosed (the serial and parallel) in the decoder and their architecture. Max* is a modified version for a NB-Code. Normalization of the log likelihood ratio is disclosed. LDPC symbol-wise normalized LLR computation methodology is disclosed. LDPC bit-wise normalized LLR computation methodology is disclosed and an architecture is disclosed.
Various example embodiments of the above-described LDPC encoder and decoder operation may be described using a clause-based description format as follows:
1. A forward error correction method performed at a receiver, comprising: receiving, over a channel, a signal that is modulated using an orthogonal time frequency space (OTFS) or an orthogonal frequency division multiplexing (OFDM) modulation scheme, demodulating the received signal to produce data; decoding the data via a non-binary low density parity check (NB-LDPC) code, by formulating the NB-LDPC code as a matrix with binary and non-binary entries; and providing the decoded data to a data sink.
2. The method of clause 1, wherein the signal received by the receiver undergoes symbol spreading over entire bandwidth of the channel, thereby having error statistics that are mathematically represented as an additive white gaussian noise.
3. The method of clause 1, wherein the NB-LDPC code comprises a structured regular systematic repeat accumulate code over a non-binary field.
4. The method of clause 3, wherein a parity matrix H for the NB-LDPC code comprises a non-binary matrix generated from a binary H matrix. Various examples of the H matrix are described in Section 2.
5. The method of clause 4, wherein the binary H matrix is based on a computer search algorithm.
6. The method of clause 5, wherein the computer search algorithm terminates when no N-cycles are present in a Tanner graph representation of the binary H matrix, and wherein N=4 or 6. Additional description is provided with reference to
7. The method of clause 4, wherein the parity check matrix H is represented as H=[H_1,H_2], where is the parity part and is the systematic part, wherein both and are composed of blocks of circulant permutation matrices and zero matrices of block size Q, where Q is an integer.
8. The method of clause 1, wherein the demodulating the received signal to produce data includes operating a slicer and a log likelihood ratio (LLR) computer stage on intermediate data generated from the received signal.
9. The method of clause 1, wherein the decoding data includes performing decoding using one of a flooding type algorithm or a layered algorithm.
10. The method of clause 8, wherein the LLR computer is either a symbol-wise LLR computer or a bit-wise LLR computer.
11. The method of clause 10, wherein the symbol-wise LLR computer calculates
LLR for a LDPC symbol using:
where P is a probability function, i is an integer, c is an integer number of consecutive bits of a constellation in the OTFS or OFDM modulated signal, and x and y represent transmitted and received constellation points, respectively.
12. The forward error correction method of clause 1, wherein the receiving the signal includes receiving a radio frequency signal over the air.
13. The forward error correction method of clause 1, wherein the receiving the signal includes receiving a radio frequency signal over a wire.
14. The forward error correction method of clause 1, wherein the receiving the signal includes receiving an optical signal over a fiber optic cable.
15. A forward error correction method performed at a transmitter, comprising: encoding information bits via a non-binary low density parity check (NB-LDPC) code, wherein the NB-LDPC code is formulated as a matrix with binary and non-binary entries, modulating, using orthogonal time frequency space (OTFS) or orthogonal frequency division multiplexing (OFDM), the encoded information bits to generate a signal; and transmitting the signal over a channel.
16. The method of Clause 15, wherein the signal received by the receiver undergoes symbol spreading over entire bandwidth of the channel, thereby having error statistics that are mathematically represented as an additive white gaussian noise.
17. The method of clause 15, wherein the NB-LDPC code comprises a structured regular systematic repeat accumulate code over a non-binary field.
18. The method of clause 17, wherein a parity matrix H for the NB-LDPC code comprises a non-binary matrix generated from a binary H matrix.
19. The method of clause 15, wherein the data is scrambled prior to encoding with the NB-LDPC code.
20. The method of clause 15, wherein the encoding with the NB-LDPC code comprises processing through a four-stage architecture comprising a top controller, an input interface, a core and a transmit interface for outputting the encoded data.
21. The method of clause 20, further comprising storing information symbols from the data in the core.
22. The method of clause 21, wherein the core comprises an I memory, a G memory, a submatrix processing unit, a readout register and a core controller, wherein the method further includes: holding the information symbols in the I memory during computations; holding a submatrix of calculations in the G memory; computing the parity submatrix and corresponding portion of a message vector by the submatrix processing unit; reading out results of calculations in the core using the readout register; and controlling operation of the core using the core controller.
23. The forward error correction method of clause 15, wherein the transmitting the signal includes transmitting a radio frequency signal over the air.
24. The forward error correction method of clause 15, wherein the transmitting the signal includes transmitting a radio frequency signal over a wire.
25. The forward error correction method of clause 15, wherein the transmitting the signal includes transmitting an optical signal over a fiber optic cable.
26. An apparatus comprising a memory storing instructions and a processor, wherein the instructions, when executed by the processor, cause the processor to perform a method of any of clauses 1 to 25.
27. A computer-readable medium having code stored thereon, the code comprising instructions, when executed, causing a processor to implement a method recited in any one or more of clauses 1 to 25.
Various additional implementation details are provided in the encoder implementation details sections 2.5 and 2.6 and decoder implementation details section 2.8. From the foregoing, it will be appreciated that specific embodiments of the invention have been described herein for purposes of illustration, but that various modifications may be made without deviating from the scope of the invention. Accordingly, the invention is not limited except as by the appended claims.
It will be appreciated that the disclosed techniques can be used to improve reception performance of wireless apparatus and/or reduce complexity of implementation.
The disclosed and other embodiments, modules and the functional operations described in this document can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structures disclosed in this document and their structural equivalents, or in combinations of one or more of them. For example, the various LDPC encoder and decoder configurations may be implemented partially in software, and partly in hardware circuits. Matrix operations and iterative calculations, for example, may be implemented in processor-executed code, while division and accumulation operations may be performed in hardware circuits.
The disclosed and other embodiments can be implemented as one or more computer program products, i.e., one or more modules of computer program instructions encoded on a computer readable medium for execution by, or to control the operation of, data processing apparatus. The computer readable medium can be a machine-readable storage device, a machine-readable storage substrate, a memory device, a composition of matter effecting a machine-readable propagated signal, or a combination of one or more them. The term “data processing apparatus” encompasses all apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them. A propagated signal is an artificially generated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus.
A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a standalone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program does not necessarily correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
The processes and logic flows described in this document can be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
Processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read only memory or a random access memory or both. The essential elements of a computer are a processor for performing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Computer readable media suitable for storing computer program instructions and data include all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto optical disks; and CD ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
While this patent document contains many specifics, these should not be construed as limitations on the scope of an invention that is claimed or of what may be claimed, but rather as descriptions of features specific to particular embodiments. Certain features that are described in this document in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a subcombination or a variation of a sub-combination. Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results.
Only a few examples and implementations are disclosed. Variations, modifications, and enhancements to the described examples and implementations and other implementations can be made based on what is disclosed.
This patent document is a continuation of U.S. patent application Ser. No. 16/651,020, filed Mar. 26, 2020, which is a 371 National Phase Application of PCT Application No. PCT/US2018/053655 entitled “FORWARD ERROR CORRECTION USING NON-BINARY LOW DENSITY PARITY CHECK CODES” filed on Sep. 29, 2018, which claims priority to and benefits of U.S. Provisional Patent Application No. 62/566,190 entitled “FORWARD ERROR CORRECTION USING NON-BINARY LOW DENSITY PARITY CHECK CODES” filed on Sep. 29, 2017. The entire content of the aforementioned patent application is incorporated by reference as part of the disclosure of this patent document.
Number | Name | Date | Kind |
---|---|---|---|
4754493 | Coates | Jun 1988 | A |
5083135 | Nagy et al. | Jan 1992 | A |
5182642 | Gersdorff et al. | Jan 1993 | A |
5623511 | Bar-David et al. | Apr 1997 | A |
5831977 | Dent | Nov 1998 | A |
5872542 | Simons et al. | Feb 1999 | A |
5956624 | Hunsinger et al. | Sep 1999 | A |
6212246 | Hendrickson | Apr 2001 | B1 |
6289063 | Duxbury | Sep 2001 | B1 |
6356555 | Rakib et al. | Mar 2002 | B1 |
6388621 | Lynch | May 2002 | B1 |
6426983 | Rakib et al. | Jul 2002 | B1 |
6608864 | Strait | Aug 2003 | B1 |
6631168 | Izumi | Oct 2003 | B2 |
6704366 | Combes et al. | Mar 2004 | B1 |
6956814 | Campanella | Oct 2005 | B1 |
7327812 | Auer | Feb 2008 | B2 |
7392018 | Ebert et al. | Jun 2008 | B1 |
7773685 | Tirkkonen et al. | Aug 2010 | B2 |
7864877 | Hottinen | Jan 2011 | B2 |
8229017 | Lee et al. | Jul 2012 | B1 |
8259845 | Dent | Sep 2012 | B2 |
8401131 | Fety et al. | Mar 2013 | B2 |
8547988 | Hadani et al. | Oct 2013 | B2 |
8619892 | Vetter et al. | Dec 2013 | B2 |
8879378 | Rakib et al. | Nov 2014 | B2 |
8892048 | Turner | Nov 2014 | B1 |
8976851 | Hadani et al. | Mar 2015 | B2 |
9031141 | Hadani et al. | May 2015 | B2 |
9071285 | Hadani et al. | Jun 2015 | B2 |
9071286 | Hadani et al. | Jun 2015 | B2 |
9083483 | Rakib et al. | Jul 2015 | B1 |
9083595 | Rakib et al. | Jul 2015 | B2 |
9130638 | Hadani et al. | Sep 2015 | B2 |
9166728 | Djordjevic | Oct 2015 | B2 |
9282528 | Hashimoto | Mar 2016 | B2 |
9294315 | Hadani et al. | Mar 2016 | B2 |
9444514 | Hadani et al. | Sep 2016 | B2 |
9548840 | Hadani et al. | Jan 2017 | B2 |
9553984 | Krause et al. | Jan 2017 | B2 |
9590779 | Hadani et al. | Mar 2017 | B2 |
9634719 | Rakib et al. | Apr 2017 | B2 |
9660851 | Hadani et al. | May 2017 | B2 |
9667307 | Hadani et al. | May 2017 | B1 |
9668148 | Hadani et al. | May 2017 | B2 |
9712354 | Hadani et al. | Jul 2017 | B2 |
9722741 | Rakib et al. | Aug 2017 | B1 |
9729281 | Hadani et al. | Aug 2017 | B2 |
10651912 | Wang et al. | May 2020 | B2 |
10667148 | Hadani et al. | May 2020 | B1 |
10681568 | Hadani et al. | Jun 2020 | B1 |
10693581 | Rakib et al. | Jun 2020 | B2 |
10693692 | Hadani et al. | Jun 2020 | B2 |
10716095 | Rakib et al. | Jul 2020 | B2 |
10749651 | Hebron et al. | Aug 2020 | B2 |
10826728 | Tsatsanis et al. | Nov 2020 | B2 |
10855425 | Kons et al. | Dec 2020 | B2 |
10873418 | Kons et al. | Dec 2020 | B2 |
10886991 | Akoum et al. | Jan 2021 | B2 |
11050530 | Wang et al. | Jun 2021 | B2 |
20010031022 | Petrus et al. | Oct 2001 | A1 |
20010033614 | Hudson | Oct 2001 | A1 |
20010046205 | Easton et al. | Nov 2001 | A1 |
20020001308 | Heuer | Jan 2002 | A1 |
20020034191 | Shattil | Mar 2002 | A1 |
20020181388 | Jain et al. | Dec 2002 | A1 |
20020181390 | Mody et al. | Dec 2002 | A1 |
20020181607 | Izumi | Dec 2002 | A1 |
20030073464 | Giannakis et al. | Apr 2003 | A1 |
20030185295 | Yousef | Oct 2003 | A1 |
20030235147 | Walton et al. | Dec 2003 | A1 |
20040044715 | Aldroubi et al. | Mar 2004 | A1 |
20040174812 | Murakami et al. | Sep 2004 | A1 |
20040189581 | Sako et al. | Sep 2004 | A1 |
20040218523 | Varshney et al. | Nov 2004 | A1 |
20050157778 | Trachewsky et al. | Jul 2005 | A1 |
20050157820 | Wongwirawat et al. | Jul 2005 | A1 |
20050180517 | Abe | Aug 2005 | A1 |
20050207334 | Hadad | Sep 2005 | A1 |
20050251844 | Martone et al. | Nov 2005 | A1 |
20060008021 | Bonnet | Jan 2006 | A1 |
20060039270 | Strohmer et al. | Feb 2006 | A1 |
20070014272 | Palanki et al. | Jan 2007 | A1 |
20070038691 | Candes et al. | Feb 2007 | A1 |
20070078661 | Sriram et al. | Apr 2007 | A1 |
20070104283 | Han et al. | May 2007 | A1 |
20070110131 | Guess et al. | May 2007 | A1 |
20070136635 | Niu | Jun 2007 | A1 |
20070211952 | Faber et al. | Sep 2007 | A1 |
20070237181 | Cho et al. | Oct 2007 | A1 |
20070253465 | Muharemovic et al. | Nov 2007 | A1 |
20070253504 | Hasegawa | Nov 2007 | A1 |
20070263752 | Guey et al. | Nov 2007 | A1 |
20080043857 | Dias et al. | Feb 2008 | A1 |
20080117999 | Kadous et al. | May 2008 | A1 |
20080186843 | Ma et al. | Aug 2008 | A1 |
20080187062 | Pan et al. | Aug 2008 | A1 |
20080232504 | Ma et al. | Sep 2008 | A1 |
20080310383 | Kowalski | Dec 2008 | A1 |
20090080403 | Hamdi | Mar 2009 | A1 |
20090092259 | Jot et al. | Apr 2009 | A1 |
20090103593 | Bergamo | Apr 2009 | A1 |
20090122854 | Zhu et al. | May 2009 | A1 |
20090161804 | Chrabieh et al. | Jun 2009 | A1 |
20090204627 | Hadani | Aug 2009 | A1 |
20090222226 | Baraniuk et al. | Sep 2009 | A1 |
20090303961 | Popovic et al. | Dec 2009 | A1 |
20100001901 | Baraniuk et al. | Jan 2010 | A1 |
20100008432 | Kim et al. | Jan 2010 | A1 |
20100027608 | Priotti | Feb 2010 | A1 |
20100111138 | Hosur et al. | May 2010 | A1 |
20100142476 | Jiang et al. | Jun 2010 | A1 |
20100187914 | Rada et al. | Jul 2010 | A1 |
20100238787 | Guey | Sep 2010 | A1 |
20100277308 | Potkonjak | Nov 2010 | A1 |
20100303136 | Ashikhmin et al. | Dec 2010 | A1 |
20100322349 | Lee et al. | Dec 2010 | A1 |
20110007789 | Garmany | Jan 2011 | A1 |
20110110532 | Svendsen | May 2011 | A1 |
20110116489 | Grandhi | May 2011 | A1 |
20110116516 | Hwang et al. | May 2011 | A1 |
20110126071 | Han et al. | May 2011 | A1 |
20110131463 | Gunnam | Jun 2011 | A1 |
20110216808 | Tong et al. | Sep 2011 | A1 |
20110286502 | Adachi et al. | Nov 2011 | A1 |
20110287778 | Levin et al. | Nov 2011 | A1 |
20110292971 | Hadani et al. | Dec 2011 | A1 |
20110293030 | Rakib et al. | Dec 2011 | A1 |
20110299379 | Sesia et al. | Dec 2011 | A1 |
20110305267 | Riu et al. | Dec 2011 | A1 |
20120021769 | Lindoff et al. | Jan 2012 | A1 |
20120051457 | Ma et al. | Mar 2012 | A1 |
20120140716 | Baldemair et al. | Jun 2012 | A1 |
20120170684 | Yim et al. | Jul 2012 | A1 |
20120201322 | Rakib et al. | Aug 2012 | A1 |
20120213098 | Sun | Aug 2012 | A1 |
20120235795 | Liao et al. | Sep 2012 | A1 |
20120269201 | Atungsiri et al. | Oct 2012 | A1 |
20120272117 | Stadelmeier et al. | Oct 2012 | A1 |
20120320994 | Loghin et al. | Dec 2012 | A1 |
20130021977 | Yang et al. | Jan 2013 | A1 |
20130058390 | Haas et al. | Mar 2013 | A1 |
20130077579 | Cho et al. | Mar 2013 | A1 |
20130083661 | Gupta et al. | Apr 2013 | A1 |
20130121497 | Smaragdis et al. | May 2013 | A1 |
20130230010 | Kim et al. | Sep 2013 | A1 |
20130246894 | Declercq | Sep 2013 | A1 |
20130260787 | Hashimoto | Oct 2013 | A1 |
20130279627 | Wu et al. | Oct 2013 | A1 |
20130315133 | Wang et al. | Nov 2013 | A1 |
20140143639 | Loghin et al. | May 2014 | A1 |
20140161154 | Hadani et al. | Jun 2014 | A1 |
20140169385 | Hadani et al. | Jun 2014 | A1 |
20140169406 | Hadani et al. | Jun 2014 | A1 |
20140169433 | Hadani et al. | Jun 2014 | A1 |
20140169436 | Hadani et al. | Jun 2014 | A1 |
20140169437 | Hadani et al. | Jun 2014 | A1 |
20140169441 | Hadani et al. | Jun 2014 | A1 |
20140247803 | Arambepola et al. | Sep 2014 | A1 |
20140348252 | Siohan et al. | Nov 2014 | A1 |
20140364128 | Lee et al. | Dec 2014 | A1 |
20150003499 | Boutillon | Jan 2015 | A1 |
20150104168 | Djordjevic | Apr 2015 | A1 |
20150117395 | Hadani et al. | Apr 2015 | A1 |
20150326273 | Rakib | Nov 2015 | A1 |
20150327085 | Hadani et al. | Nov 2015 | A1 |
20150382231 | Jabbar et al. | Dec 2015 | A1 |
20160043835 | Hadani et al. | Feb 2016 | A1 |
20160135132 | Donepudi et al. | May 2016 | A1 |
20160182269 | Hadani et al. | Jun 2016 | A1 |
20160191217 | Hadani et al. | Jun 2016 | A1 |
20160191280 | Hadani et al. | Jun 2016 | A1 |
20160254889 | Shattil | Sep 2016 | A1 |
20160277225 | Frenne et al. | Sep 2016 | A1 |
20160309345 | Tehrani et al. | Oct 2016 | A1 |
20160380743 | Rakib | Dec 2016 | A1 |
20160381576 | Hadani et al. | Dec 2016 | A1 |
20170012749 | Rakib et al. | Jan 2017 | A1 |
20170012810 | Rakib et al. | Jan 2017 | A1 |
20170019297 | Rakib | Jan 2017 | A1 |
20170033899 | Rakib et al. | Feb 2017 | A1 |
20170040711 | Rakib et al. | Feb 2017 | A1 |
20170078054 | Hadani et al. | Mar 2017 | A1 |
20170099122 | Hadani et al. | Apr 2017 | A1 |
20170099607 | Hadani et al. | Apr 2017 | A1 |
20170149594 | Rakib et al. | May 2017 | A1 |
20170149595 | Rakib et al. | May 2017 | A1 |
20170201354 | Hadani et al. | Jul 2017 | A1 |
20170207817 | Hadani et al. | Jul 2017 | A1 |
20170222700 | Hadani et al. | Aug 2017 | A1 |
20170230215 | Rakib et al. | Aug 2017 | A1 |
20170244524 | Hadani et al. | Aug 2017 | A1 |
20170288913 | Rakib et al. | Oct 2017 | A1 |
20170289961 | Rakib et al. | Oct 2017 | A1 |
20180109284 | Hadani et al. | Apr 2018 | A1 |
20180167165 | Kons et al. | Jun 2018 | A1 |
20180205481 | Rakib et al. | Jul 2018 | A1 |
20180227159 | Rakib et al. | Aug 2018 | A1 |
20180242170 | Hadani et al. | Aug 2018 | A1 |
20180262306 | Hadani et al. | Sep 2018 | A1 |
20180288809 | Delfeld et al. | Oct 2018 | A1 |
20180302802 | Fanfelle | Oct 2018 | A1 |
20190036577 | Delfeld et al. | Jan 2019 | A1 |
20190036741 | Hadani et al. | Jan 2019 | A1 |
20190044682 | Hebron et al. | Feb 2019 | A1 |
20190075551 | Hadani et al. | Mar 2019 | A1 |
20190081836 | Hadani et al. | Mar 2019 | A1 |
20190159177 | Rakib et al. | May 2019 | A1 |
20190173617 | Kons et al. | Jun 2019 | A1 |
20190173630 | Kons et al. | Jun 2019 | A1 |
20190215109 | Hadani et al. | Jul 2019 | A1 |
20190238189 | Delfeld et al. | Aug 2019 | A1 |
20190327054 | Kons et al. | Oct 2019 | A1 |
20190342126 | Hadani et al. | Nov 2019 | A1 |
20190342136 | Hadani et al. | Nov 2019 | A1 |
20190379422 | Hadani et al. | Dec 2019 | A1 |
20200045562 | Hadani et al. | Feb 2020 | A1 |
20200119868 | Rakib et al. | Apr 2020 | A1 |
20200137774 | Molisch et al. | Apr 2020 | A1 |
20200145273 | Rakib et al. | May 2020 | A1 |
20200153107 | Rakib | May 2020 | A1 |
20200186397 | Tsatsanis et al. | Jun 2020 | A1 |
20200204309 | Namboodiri | Jun 2020 | A1 |
20200204410 | Hadani | Jun 2020 | A1 |
20200228170 | Delfeld et al. | Jul 2020 | A1 |
20200259604 | Hadani et al. | Aug 2020 | A1 |
20200259692 | Hadani et al. | Aug 2020 | A1 |
20200259697 | Delfeld | Aug 2020 | A1 |
20200280138 | Fanfelle et al. | Sep 2020 | A1 |
20200287672 | Namboodiri et al. | Sep 2020 | A1 |
20200288333 | Rakib et al. | Sep 2020 | A1 |
20200305010 | Hadani et al. | Sep 2020 | A1 |
20200313695 | Namboodiri et al. | Oct 2020 | A1 |
20200313949 | Hadani | Oct 2020 | A1 |
20200322185 | Kons et al. | Oct 2020 | A1 |
20200322202 | Hadani et al. | Oct 2020 | A1 |
20200351836 | Rakib et al. | Nov 2020 | A1 |
20200367252 | Hebron et al. | Nov 2020 | A1 |
20200389268 | Sathyanarayan et al. | Dec 2020 | A1 |
20200403829 | Namboodiri et al. | Dec 2020 | A1 |
20210028877 | Rakib et al. | Jan 2021 | A1 |
20210036823 | Hebron et al. | Feb 2021 | A1 |
20210058114 | Molisch et al. | Feb 2021 | A1 |
20210077625 | Kons et al. | Mar 2021 | A1 |
20210105155 | Kons et al. | Apr 2021 | A1 |
20210126750 | Kons et al. | Apr 2021 | A1 |
20210135905 | Kons et al. | May 2021 | A1 |
20210194732 | Kons et al. | Jun 2021 | A1 |
Number | Date | Country |
---|---|---|
2011150315 | Dec 2011 | WO |
2013148546 | Oct 2013 | WO |
2014004585 | Jan 2014 | WO |
2016014596 | Jan 2016 | WO |
2016014598 | Jan 2016 | WO |
2016176642 | Nov 2016 | WO |
2016183230 | Nov 2016 | WO |
2016183240 | Nov 2016 | WO |
2016209848 | Dec 2016 | WO |
2017003952 | Jan 2017 | WO |
2017011455 | Jan 2017 | WO |
2017011478 | Jan 2017 | WO |
2017044501 | Mar 2017 | WO |
2017087706 | May 2017 | WO |
2017100666 | Jun 2017 | WO |
2017147439 | Aug 2017 | WO |
2017165697 | Sep 2017 | WO |
2017173160 | Oct 2017 | WO |
2017173389 | Oct 2017 | WO |
2017201467 | Nov 2017 | WO |
2018031938 | Feb 2018 | WO |
2018031952 | Feb 2018 | WO |
2018032016 | Feb 2018 | WO |
2018064587 | Apr 2018 | WO |
2018064605 | Apr 2018 | WO |
2018106731 | Jun 2018 | WO |
2018129554 | Jul 2018 | WO |
2018140837 | Aug 2018 | WO |
2018191309 | Oct 2018 | WO |
2018195548 | Oct 2018 | WO |
2018200567 | Nov 2018 | WO |
2018200577 | Nov 2018 | WO |
2019014332 | Jan 2019 | WO |
2019032142 | Feb 2019 | WO |
2019032605 | Feb 2019 | WO |
2019036492 | Feb 2019 | WO |
2019051093 | Mar 2019 | WO |
2019051427 | Mar 2019 | WO |
2019055861 | Mar 2019 | WO |
2019068053 | Apr 2019 | WO |
2019060596 | May 2019 | WO |
2019089986 | May 2019 | WO |
2019113046 | Jun 2019 | WO |
2019157230 | Aug 2019 | WO |
2019173775 | Sep 2019 | WO |
2019241436 | Dec 2019 | WO |
2019241589 | Dec 2019 | WO |
2020142520 | Jul 2020 | WO |
2020206304 | Oct 2020 | WO |
2020227619 | Nov 2020 | WO |
2020247768 | Dec 2020 | WO |
2021026212 | Feb 2021 | WO |
2021062354 | Apr 2021 | WO |
Entry |
---|
Banelli, P. et al., “Modulation Formats and Waveforms for 5G Networks: Who Will Be the Heir of OFDM?,” IEEE Signal Processing Magazine, vol. 81, pp. 80-93, Nov. 2014. |
El Hattachi, R. et al., “Ngmn 5G Initiative White Paper,” NGMN Alliance, Feb. 17, 2015. [Online]. Available: https://www.ngmn.org/uploads/media/NGMN_5G_White_Paper_V1_0.pdf, 125 pages. |
Rusek, F. et al., “Scaling Up MIMO, Opportunities and Challenges with Very Large Arrays,” IEEE Signal Processing Magazine, pp. 40-60 (2013). |
Vodafone, “Cellular Internet of Things: Architectural Aspects,” RP-150869, 3GPP RAN#68, Malmo, Sweden (Jun. 9, 2015), 19 pages. |
Gurevich, S. et al. “Group Representation Design of Digital Signals and Sequences,” S.W. Golomb et al. (eds.), SETA 2008, LNCS 5203, pp. 153-166, Springer-Verlag Berlin Heidelberg (2008). |
“AT&T Annual Report 2014,” Opening Our Network [Online]. Retrieved from the Internet: Sep. 22, 2016. <URL: http://www.att.com/Investor/ATT_Annual/2014/att_introduces_new_concepts_for_telecom_network.html>, 5 pages. |
Catt, “UL ACK/NACK transmission methods for LTE-A,” 3GPP TSG RAN WG1 Meeting #60bis, R1-102453, Beijing, China, Apr. 12-16, 2010, 8 pages. |
Toskala, A. et al., “Physical Layer,” Chapter 5 In: “LTE for UMTS: OFDMA and SC-FDMA Based Radio Access,” Holma, H. et al. (eds.), John Wiley & Sons, Ltd., United Kingdom, 2009, pp. 83-135. |
Mecklenbrauker, W., “A Tutorial on Non-Parametric Bilinear Time-Frequency Signal Representations,” In: Time and Frequency Representation of Signals and Systems, Longo, G. et al. (eds.), Springer-Verlag Wien, vol. 309, pp. 11-68 (1989). |
Nehorai, A. et al., “MURI: Adaptive waveform design for full spectral dominance (2005-2010),” AFOSR FA9550-05-1-0443, Final Report, [online], Mar. 11, 2011 Retrieved on May 11, 2013, Retrieved from the Internet <URL: http://oai.dtic.mil/oai/oai?verb=getRecord&metadataPrefix=html&identifier=ADA565420>, 103 pages. |
Number | Date | Country | |
---|---|---|---|
20220069844 A1 | Mar 2022 | US |
Number | Date | Country | |
---|---|---|---|
62566190 | Sep 2017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16651020 | US | |
Child | 17451251 | US |