The present disclosure relates to signal processing for communication systems, and more specifically, to lattice-based data modulation methods.
A lattice is a periodic arrangement of points in an n-dimensional space. Communications engineers and information theorists use lattices for quantization and modulation, for example to perform lossy compression (“source coding”) and/or noise immunity (“channel coding”).
In some embodiments, a data modulation method includes receiving a bit string at a processor, and identifying a set of binary strings based on the bit string. Each binary string from the set of binary strings is mapped to an element from a set of elements of a lattice-based signal constellation, without using a lookup table. Real-valued points from the set of elements are identified based on the mapping. The method also includes causing transmission of a signal having a modulation based on the real-valued points.
In some embodiments, a data demodulation method includes receiving at a processor and from a transmitter, a signal encoding a bit string. A point of an n-dimensional lattice associated with the signal is identified, via the processor, based on the signal and a closest vector algorithm. The point of the n-dimensional lattice is multiplied, via the processor, by an inverse of a basis of the n-dimensional lattice, to identify an element from a plurality of elements of an n-dimensional matrix ℤn associated with the signal. Bits are recovered based on the signal, via the processor, by reducing components of the signal modulo m, to produce an element of a quotient ℤn/mℤn.
Vast literature exists on modulation techniques for wireless communications. One related area of research is in using various types of dense packings and/or (generally regular) lattices to modulate data into higher dimensional spaces. These ideas go back to the original work of Claude Shannon (see, e.g., Shannon, Claude E., “Probability of Error for Optimal Codes in a Gaussian Channel,” Bell System Tech. J. 38, 611-656) and the Shannon-Hartley theorem, and research continues through today.
One known idea within the wireless communications community is to use points in higher dimensional spaces that are more “dense” than the usual quadrature amplitude modulation / amplitude and phase-shift keying (QAM/APSK) constellations. This allows greater distance (e.g., Hamming distance or Euclidean distance) between the constellation points, which decreases the probability that channel noise or other channel distortions will cause an error. There have also been many attempts to combine coding theory with modulation, such as trellis coded modulation and multilevel codes
Most of the foregoing approaches use either a coding theoretic construction (e.g., one of the approaches discussed in Conway, J and Sloane, N, “Sphere Packings, Lattices, and Groups, Springer, 1993) or a lattice constellation without underlying coding.
Known coding theoretic construction approaches typically include breaking a lattice, or a set of lattices, into a set of lattice cosets. A set of coded bits is then used to select a coset (as a subset of the lattice or set of lattices), and a set of uncoded bits is used to select a point within the subset. For example, a set of message bits may be run through a standard (binary) error correcting code (e.g., a convolutional encoder), and the output (encoded) bits are used to select a coset. The remaining message bits are then used to select a point within that coset.
Another known approach is to use a lattice constellation without underlying coding, but instead to use an elaborate series of lookup tables (which becomes intractable at higher throughputs/larger constellations), or to use more geometrically convenient shaping regions such as rectangles (which reduce the efficiency of the constellations).
In contrast to the foregoing known techniques, some embodiments of the present disclosure facilitate bit-to-symbol mapping and symbol-to-bit mapping for a lattice based constellation, for example in a modem / baseband processor and for mapping bits into complex baseband I/Q points, irrespective of any underlying coding theoretic schemes, and without singling out closest elements of the lattices. Moreover, embodiments of the present disclosure do not use lookup-tables to map bit strings to lattice points, or to subsets of lattice points. Still further, in some implementations, embodiments of the present disclosure do not use rectangular shaping regions.
As used herein, a lattice can refer to a set of points in an n-dimensional space given by all linear combinations with integer coefficients of a basis set of up to n linearly independent vectors. One example of a lattice is a Leech lattice, discussed further below. As used herein, a Voronoi region of a lattice point can refer to the region of the n-dimensional space closer to that lattice point than to all other lattice points. A code is a finite set of codewords having a specified length, a codeword being a sequence of symbols encoding a message to be transmitted within a communication system. Codewords are translated into signals (coded signals) via modulation to real and/or complex values. Coded signals can be represented as points within a signal space. A lattice code is defined by a finite set of lattice points within a predefined region of a given lattice, the predefined region referred to herein as a “shaping region.”
Consider an ordered basis B for an n-dimensional lattice Λn. The basis is an isomorphism from Zn (an n-dimensional vector of integers) to points of Λn. Furthermore, note that B restricts to an isomorphism of subgroups λℤn→ λΛn for λ ∈ ℤ. Illustrating this below, notice that since the square on the left side of the diagram commutes, the composition of the basis with the quotient map Q has kernel (“ker”) λℤn (i.e., ker(Q ∘ B) = λℤn). Now, by the universal property of cokernels, there exists a unique homomorphism, σ:ℤn/λℤn → Λn/λΛn that makes the square on the right commute.
By employing the same strategy using B-1, a map τ:Λn/λΛn→ℤn/λℤn is obtained, which makes the right side commute. Therefore, τ = σ-1, making σ an isomorphism. So, the quotient λn/λΛn is isomorphic as an abelian group to the quotient ℤn/λℤn. The space ℤn/λℤn is the set of n copies of integers modulo λ (i.e., of the form (a1, a2, a3, ..., an), where each ai ∈ (0...,λ-1). This size of this set is λn. And, because of the isomorphism implied by the above commutative diagram, the size of Λn/λΛn is also λn.
If the value of λ is taken to be a power of 2 (such that λ = 2k), then the number of points in the quotient Λn/λΛn is 2nk. This means that there is a bijection from binary strings (or sub-strings) of length nk to the elements of the quotient Λn/λΛn.
As an explicit construction of this bijection, consider the set of 2nk strings of nk bits. If any given bit string is broken into blocks or sub-strings of k bits, each k-bit sub-string can be naturally mapped to the integers mod λ by treating that k-bit sub-string as the binary expansion of the integer. In this way, each of the length-nk bit strings is mapped to ℤn/λℤn - n values taken from the integers mod λ = 2k. So, given a length-nk bit string, mapped to an element α ∈ ℤn/λℤn, a point of the lattice Λ can be produced using the basis. Specifically, the point of the lattice will be p= Bα ∈ Λn.
The point p can be reduced to an enlarged Voronoi cell at the origin by subtracting the point of λΛn that is closest to p, resulting in a point q ∈ Λn/λΛn.
One subtlety that may arise is that the fundamental/Voronoi region of λΛn may pass through some of the points of Λn. Suppose that a point x ∈ Λn intersects the fundamental region of λΛn. By the group property of lattices, this means that -x also intersects the fundamental region of λΛn. As used herein, a “fundamental region” refers to a Voronoi cell of the origin, or a lattice region in which every point is closer to the origin than to any other lattice point. If x = Bα, then by linearity, the point that produces -x is -α. But both α and -α are both elements of ℤn/λℤn.
As an illustration of this, let n = 2 and let k = 1, and let the lattice basis B have basis vectors (1,1) and (1,-1), so the matrix is B =
The set of valid α vectors is (0,0), (0,1), (1,0), and (1,1). B maps these, respectively, to (0,0), (1,-1), (1,1), and (2,0). However, consider the additive inverse of the three non-zero α vectors - (0,-1), (-1,0), and (-1,-1). B maps these, respectively, to (-1,1), (-1,-1), and (-2,0). However, the points (0,-1) and (0,1) are equivalent modulo 2, as are the points (-1,0) and (1,0), and the points (-1,-1) and (1,1). So, for the points that lie on the surface of the fundamental region of λΛn, there is an ambiguity.
The ambiguity can be resolved by creating a convention for the selection of points to be used. This convention can include, for example, selecting a face of a fundamental region to be used, and then only using the point on that (selected) face. Alternatively, the convention can include slightly shifting the fundamental region away from the origin, to “break” the symmetry that resulted in points on the fundamental region surface. For example, slightly shifting the fundamental region (i.e., resulting in a “shifted fundamental region”) can refer to shifting the fundamental region by a smallest amount (and in an appropriate direction) that prevents any points from intersecting the enlarged fundamental region.
Turning now to the drawings,
To illustrate the procedures discussed above, consider the constellation diagrams shown in
By taking k=1, the fundamental region of 2A2 can be included, as shown in
Similarly, for k=2, the fundamental region of 4Λ2 can be drawn, and a convention similar to the convention applied in
In other embodiments, the fundamental region may be shifted, as shown in
Similarly, as shown in the embodiment of
In some implementations, when using a small offset from the origin to produce the 2nk points, it can be advantageous to shift each point of a fundamental region to make the dot labelled “A” the actual center of the set of points, so that the set of all data points has zero mean. The offset is a single vector in Euclidean space (ℝn), and is not made of lattice points.
In some embodiments, a data modulation method includes performing a bit-to-symbol mapping that uses a lattice constellation.
In some embodiments, a data demodulation method includes performing a symbol-to-bit mapping that uses a lattice constellation.
In some embodiments, a lattice-based data modulation method or demodulation method does not include lattice coding, and does not include the use of redundant lattice points.
In some embodiments, a data modulation method or demodulation method includes using a quotient of a lattice with a copy of that lattice scaled by a positive integer.
In some embodiments, a data modulation method includes mapping 2nk binary strings (of length nk) to elements of a quotient Λn/2kΛn, or to a shifted copy (e.g., as discussed and shown with reference to
The real-valued points can then be transmitted (optionally with any of: upsampling, filtering, driving to carrier, sending through a digital-to-analog converter (DAC), etc.), and received by a recipient. The receiver may then perform synchronization and/or equalization. To demodulate / recover the bits, the receiver will then use the pre-selected closest vector algorithm to recover the “closest” point of Λn, and this recovered closest point is either multiplied by the inverse B-1, thereby producing an element of ℤn, or it is used to recover soft information about the received value which can then be fed into an error correction scheme. The individual components (i.e., the elements of ℤn) are then reduced mod 2k (i.e., modular arithmetic is performed on the individual components), to produce an element of ℤn/2kℤn (or, stated more generally, ℤn/mℤn, where m is a positive integer). If the quotient involves an ambiguity resulting from points of Λn that lie on the surface of the fundamental region of 2kΛn, the ambiguity may be resolved (e.g., as discussed above), for example, by enforcing a convention that selects a set of “faces” of the fundamental region to use. Alternatively, the ambiguity can be resolved by enforcing a convention that includes centering the fundamental region of 2kΛn around a point located away from the origin that does not result in elements of Λn that lie on a surface of its fundamental region (as discussed above in reference to
In some implementations, the real-valued points represent in-phase/quadrature (I/Q) points or components.
In some implementations, the set of elements is associated with a quotient of a first lattice and a second lattice, the second lattice being a copy of the first lattice scaled by a positive integer.
In some implementations, the method also includes performing at least one of synchronization or equalization of the signal prior to identifying the point of an n-dimensional lattice associated with the signal.
In some implementations, the point is a first point, and the method also includes identifying, via the processor, an ambiguity associated with the quotient ℤn/mℤn. In response to identifying the ambiguity, a set of faces of a fundamental region of the n-dimensional lattice can be selected via the processor, and a second point of the n-dimensional lattice associated with the signal can be identified via the processor based on the set of faces of a fundamental region of the n-dimensional lattice, the instructions to recover the bits including instructions to recover the bits further based on the second point.
In some implementations, the point is a first point, and the method also includes identifying, via the processor, an ambiguity associated with the quotient ℤn/mℤn. In response to identifying the ambiguity, a fundamental region of the n-dimensional lattice is centered around a second point of the n-dimensional lattice that is distanced from an origin of the n-dimensional lattice, such that the plurality of elements of the n-dimensional matrix ℤn do not lie on a surface of the fundamental region.
Some embodiments of the present disclosure improve upon known communications techniques using Leech lattices as the basis for a lattice-based signal constellation. The Leech lattice is an even unimodular lattice A24 in 24-dimensional Euclidean space, and is an attractive candidate. The following discussion quantifies the gain achievable by Leech lattices, as contrasted with some known constellations, assuming spherical shaping regions, which are approximated, in some instances, using a scaled Voronoi as described herein.
The Leech lattice is expected to have some interesting communication capabilities. Consider the set of points (also referred to herein as “centers”) of Λ contained within and on a sphere of radius 4√n for some fixed positive integer n, and denote this set by Λn i.e., Λn =
Each point can serve as a signal vector. Those signals in Λn are referred to herein as surface vectors, and the others are referred to herein as internal vectors of Λn. The decision region for an internal vector is bounded by 196,560 hyper-planes, which makes it unlikely that an exact expression for the probability of error will be found. An upper bound for the probability of error and/or a lower bound for the probability of error, however, can be found.
Assume that all signals are equally likely, and consider the probability of error given that an internal vector was sent. Further, assume that the channel has the effect of additively perturbing each signal component independently by a gaussian random variable of mean zero and variance N0/2. The probability of a correct decision P(C) is greater than the probability that the perturbed vector remains within a sphere of radius 2√2 about the sent signal vector. To calculate this probability, denote the twenty-four dimensional gaussian distribution of independent variates, each with mean zero and variance N0/2, by:
A sphere of radius r in E24 has a surface area:
Thus, the probability of the signal vector remaining inside a sphere of radius 2√2 is
By the change of variable y = r2/N0 this last expression may be written as:
where P(a, x) is the incomplete gamma function. The probability of a correct decision for a surface vector will be slightly greater than this quantity. Here, the region of integration is over the same sphere as above, plus a conical region exterior to the sphere of radius 2√n. The integration over the conical region may be expressed in terms of a double integral that may be complex to compute, and that may be small except at very high noise levels. For simplicity, this term may be ignored. Hence, a bound on the probability of error PL(E) for the subset of the Leech Lattice Λn is given by:
which may be written as:
where
The signal vectors of Λn are energy constrained by:
and so EN = (16/24)n = 2n/3, where EN is the energy per dimension and the subscript N is 24. Thus, in the upper bound (1),
Compare the bound (1) to the bound PG(E), for signals constrained in energy by 24EN:
where
and
The probabilities for the two error bounds (1) and (2) are plotted in
M being the number of signals. The curves are labelled by their equation numbers. Notice that only certain rates are possible with the Leech Lattice. As can be seen in
For large values of
and Gallagers bound behaves as:
where K is a constant. The bound for Λn in (1) behaves as
which gives exponential decrease rather than n-th power. This is reflected in the behaviour of the curves of
The two bounds behave in approximately the same manner for other values of n. Consider the effect of the change from n = 10 to n = 100. Using the asymptotic value for the number of points in Λ100, one obtains:
The effect on the bound (1) is merely to shift the curve to the right by 10 db. The effect on bound (2) is to multiply values for n = 10 by eNΔR = e24x1.1290 = e27.096 = 5.852 × 1011. For values of EN/N0 such that 10 log10(EN/N0) > 10, i.e., EN/N0 > 10, the behaviour indicated by Eq. (3) becomes a very good approximation. On the logarithmic scales used in the figure, this implies that the bound (2) will be very close to a straight line. Using this straight line approximation to the bound, it may be seen that multiplication by 5.852 × 1011 is equivalent to a right shift of the bound by an amount slightly greater than 10 db.
Note that the bound (2) is valid only for
which implies, for n = 10 and R = 1.2597, that the bound is valid only for:
Attempts were made, without success, to find a reliability or exponent function for the bound of (1).
A Euclidean code is a finite set of points x1, ..., xM in n-dimensional real Euclidean space Rn. Euclidean codes can be used as signal sets for a Gaussian channel, as representative points (or output vectors) in a vector quantizer, and in many other applications. There are a number of desirable properties that a Euclidean code should have:
The foregoing properties (also referred to herein as “problems” to solve) are not necessarily independent of one another, and there may be tradeoffs between / among them. Lattice codes include a subset of points of some lattice in Rn, which is a set of points of a lattice packing of equal n-dimensional spheres. The code is defined by specifying a lattice A in Rn and a certain region of the space Rn, and consists of all lattice points inside this region.
For these codes, the minimum distance is the minimum distance between lattice points, and the number of code vectors is determined by the density of the lattice. Thus, properties 1) and 3) amount to saying that the lattice should have a high density, and property 2) states that the region of space defining the code should be as nearly spherical as possible.
Some known very fast algorithms have been used to satisfy property 6) for a large class of lattice codes. According to some embodiments of the present disclosure, similar lattices are used to satisfy properties 4) and 5). Since if the codes are used for a Gaussian channel 6) is the decoding problem while 4) and 5) are encoding problems, properties 4) and 5) might be expected to be more readily solved than property 6). For these codes, however, properties 4), 5), and 6) appear to be of comparable difficulty.
In accordance with some embodiments, to solve for properties 4) and 5), only lattice codes that are defined by certain very special regions of space, regions that can be called Voronoi-shaped, are considered, and such regions are referred to herein as Voronoi codes.
Voronoi codes can have the following two drawbacks: only certain rates can be attained, and because the Voronoi-shaped regions defining them are not spherical, they do not in general have the lowest possible total energy (although the difference may be small). Under the notation set forth herein, the norm ||x||2 of a vector x is its squared length x • x.
Given an index vector (k1, ..., kn) with
suppose one desires to find that vector x in the Voronoi code CΛ(r, α) for which index (x) = (k1, ..., kn). To do so, first form x′ = ∑kivi, which has index equal to (k1, ..., kn) but need not be in the code. The desired lattice point x is then the unique solution to:
(by the definition of CΛ(r, α)). Equation (4) is readily solved if there is an algorithm available for solving for property 6), i.e., for finding the closest point of A to an arbitrary point of the space. In fact, if one sets z = r-1(x′ - α), and λ is the closest point of A to z, then x = x′ - rλ is the desired lattice point. Thus, property 4) may be solved as follows.
Given the index vector (k1, ..., kn), calculate x′ = ∑kivi and z = r-1(x′ - a). Find the closest point λ ∈ Λ to z (using, for example, the algorithms in [3]). Then x = x′ - rλ - a is in the Voronoi code CΛ(r, a) and has index (k1, ..., kn).
For example, consider a four-dimensional Voronoi code CD
In accordance with some embodiments, an algorithm for finding the closest point of the integer lattice ℤn to an arbitrary point x ∈ ℝn is used. For a real number x, let:
In case of a tie, choose the integer with the smallest absolute value. For x = (x1, · · ·, xn ) ∈ ℝn, let
Also define g(x), which is the same as ƒ(x) except that the worst component of x-that furthest from an integer-is rounded the wrong way. In case of a tie, the component with the lowest subscript is rounded the wrong way. More formally, for x ∈ ℝ,ƒ(x) and the function w(x), which rounds the wrong way, are defined as follows. (Here, m is an integer.)
(Ties are handled so as to give preference to points of smaller norm.) Also:
so that * δ( x ) * ≤ ½ is the distance from x to the nearest integer.
Given x= (x1, · · ·, Xn) ∈ ℝn, let k (1 ≤ k ≤ n) be such that:
and
Then, g(x) is defined by:
Algorithm 1 — To Find the Closest Point of ℤn to x: Given x ∈ ℝn, the closest point of ℤn is ƒ(x). (If x is equidistant from two or more points of ℤn, this procedure finds the one with the smallest norm.)
To see that the procedure works, let u = (u1, · · ·, un) be any point of ℤn. Then,
which is minimized by choosing ui = ƒ(xi) for i = 1, · · ·, n. Because of (5), ties are broken correctly, favoring the point with the smallest norm.
Algorithm 2—To Find the Closest Point of Dn to x: Given x ∈ ℝn, the closest point of Dn is whichever of f(x) and g(x) has an even sum of components (one will have an even sum, the other an odd sum). If x is equidistant from two or more points of Dn, this procedure produces a nearest point having the smallest norm.
This procedure works because ƒ(x) is the closest point of ℤn to x and g(x) is the next closest. ƒ(x) and g(x) differ by one in exactly one coordinate, and so precisely one of ∑ƒ(xi) and ∑g(xi) is even and the other is odd. Again, (5) implies that ties are broken correctly.
Example: Find the closest point of D4 to x (0.6, - 1.1, 1.7, 0.1). Compute:
and
since the first component of x is the furthest from an integer. The sum of the components of ƒ(x) is 1- 1 + 2 + 0 = 2, which is even, while that of g(x) is 0 - 1 + 2 + 0 = 1, which is odd. Therefore, ƒ(x) is the point of D4 closest to x.
To illustrate how ties are handled, suppose:
In fact, x is now equidistant from eight points of D4, namely (0, 0, 0, 0), any permutation of (1, 1, 0, 0), and (1, 1, 1, 1). The algorithm computes:
and selects ƒ(x). Indeed, ƒ(x) does have the smallest norm of the eight neighboring points.
Since E8 is the union of two cosets of D8, the discussion in the previous section leads to the following procedure.
Algorithm 3—To Find the Closest Point of E8 to x: Given x = (x1, · · ·, x8) ∈ ℝ8.
Compute ƒ(x) and g(x), and select whichever has an even sum of components; call it y0. Compute
where:
and select whichever has an even sum of components; add
and call the result y1.
Compare y0 and y1 with x, and choose the closest. For example, to find the closest point of E8 to
compute
and take y0 = g(x). Also,
and so
Finally,
and it can be concluded that y0 = g(x) is the closest point to x.
Finding the Closest Point of
Algorithm 4—To Find the Closest Point of An to x:
Step 1: Given x ∈ ℝn+1, compute s = ∑ xi and replace x by
Step 2: Calculate
and the deficiency
Step 3: Sort the
in order of increasing value of
(defined in Section III). We obtain a rearrangement of numbers 0, 1, ...., n, say i0, i1,..., in, such that:
Step 4: If Δ=0,ƒ(x′) is the closest point of An to x.
If Δ>0, the closest point is obtained by subtracting 1 from the components
If Δ<0, the closest point is obtained by adding 1 to the components
Remarks: As discussed above, ƒ(x) is the closest point of ℤn+1 to x. The procedure described here finds the closest point of An because it makes the smallest changes to the norm of ƒ(x) needed to make
vanish.
Step 1 projects x onto x′, the closest point of the hyperplane ∑xi = 0. Since An is by definition contained in this hyperplane, it may be possible to assume that x already lies there, in which case Step 1 can be omitted.
The only substantial amount of computation needed is for the sort in Step 3, which takes O(n log n) steps. Step 3 can be omitted, however, if x is expected to be close to An. In this case, Δ will be small, and Steps 3 and 4 can be replaced by the following:
Step 3 : If Δ=0, ƒ(x′) is the closest pint of An to x. If Δ>0, find the Δ components of x′, say
for which
is as small (i.e., as close to
) as possible. The closest point of An is obtained by subtracting one from the components
of ƒ(x).
If Δ < 0, find the |Δ| components of x′, say
for which
is as large (i.e., as close to
) as possible. The closest point of An is obtained by adding 1 to the components of
In any case, |Δ| cannot exceed n/2. If Δ is expected to be large, however, the first version of the algorithm is preferable. The closest point of
can be found using the fact that
is the union of n + 1 cosets of An. For example, the hexagonal lattice A2 is shown in
Conversely, the u-coordinates may be obtained from the x-coordinates by:
For example, the points (0, 0), (1, 0),
have x-coordinates (0, 0, 0), (1, 0, -1), (1, -1, 0), (0, -1, 1), respectively. To find the closest point of An to the point P with coordinates:
first find the x-coordinates of P, which are x = (x0, x1, X2) = (0.169, 0.462, -0.631).
Step 1 of the algorithm case be omitted, since x0 + x1 + x2 = 0 holds automatically. Step 2 produces:
with deficiency Δ = -1. At Step 3, one obtains:
At Step 4, 1 is added to ƒ(x1), obtaining:
which is the closest point of A2. The u-coordinates for this point are:
(see
Since A3 ≅ D3, Algorithm 2 is preferable to Algorithm 4 for finding the closest point of the face-centered cubic lattice. Finally, E7 and
can be handled via the algorithm for A7.
In one or more embodiments of the present disclosure, certain lattices can be defined by ‘code formulas’, as follows. Let the standard binary representation of an integer x be:
where xk is called 2k’s-coefficient, k ≥ 0, and the congruence relations
may be used to determine the coefficients recursively (particularly if x is negative). Let C0, C1 C2 ... be a set of binary block codes of length N, which must be nested and satisfy certain other conditions to produce a lattice. Then, define A as the set of all integer N-tuples x whose 2k′s-coefficient N-tuple is a code-word in the code Ck, for all k ≥ 0. The foregoing can be written symbolically by the ‘code formula’:
When, as is always the case, the 2k′s-coefficients are allowed to take on any value for k ≥ K —i.e., when Ck is the universe (N, N) code for k > K, one may write:
where ZN is the N-dimensional integer lattice.
One standard definition for the Leech lattice Λ24 expresses it as the union of a sublattice H24 and a coset H24 + a of H24. Here, H24 is a lattice sometimes called the ‘Leech half-lattice,’ which may be defined by the code formula:
where C0 = (24,12) is the binary Golay code and C1 = (24,23) is a single-parity-check code. The translation 24-tuple a may be taken as a = (-3, 123)/2.
The Golay code has minimum distance 8 and 759 weight-8 codewords. The minimum squared distance between points in Λ24 or H24 in this representation is
H24 has 98,256 lattice points of Euclidean norm 8, namely (24·23/2)·4 = 1,104 points with two coordinates of magnitude 2 and 22 of magnitude 0, plus 759·128 = 97,152 points with eight coordinates of magnitude 1 and 16 of zero. Λ24 has 196,560 points of norm 8, namely those of H24 plus 24·4096 = 98,304 points in H24 + α with one coordinate of magnitude 3/2 and 23 coordinates of magnitude ½.
The Voronoi region Rν (0) of a lattice A is the set of points that are at least as close to 0 as to any other point in A; i.e., the Voronoi region is essentially the decision region of a maximum-likelihood decoding algorithm for A (up to the ambiguity involved in resolving ties on the boundary). The packing radius rmin(Λ), or error-correction radius, is the radius of the largest sphere that can be inscribed in Rν (0), and is equal to dmin(Λ)/2, where
is the minimum squared distance between points in A. The kissing number N0(Λ), or error coefficient, is the number of points on the boundary of Rν (0) with norm
and is equal to the number of points in Λ of norm
For H24,
and N0(H24) = 98256 [2]. For Λ24,
and No(Λ24) =196,560.
Let G24 be the ‘Construction Λ’ lattice consisting of all integer 24-tuples whose ones-coefficient N-tuple is a codeword in the (24,12) binary Golay code; i.e., G24 has the code formula G24 = (24,12) + 2Z24. Then, H24 is a sublattice of G24, and in fact G24 is the union of H24 and a coset H24 + b of H24, where one may take b = (2,023). H24 is thus the subset of G24 in which the twos-coefficient 24-tuple has even parity, and H24 + b is the subset with odd parity.
Any soft-decision decoding algorithm for the Golay code may be used as a decoding algorithm for G24 as follows. Given any real 24-tuple r, first find the closest even and odd integers kjo and kj1 to each coordinate rj of r. The differences in squared distances, + [(r1 -kj0)2 - (r1 - kj1)2 ] /2 = ± [(rj - (kj0 + kj1) /2], may be taken as the ‘metrics’ for 0 and 1, respectively, for that coordinate in any soft-decision decoding algorithm for the Golay code. The decoded Golay codeword is then mapped back to kj0 or kj1, at the jth coordinate, depending on whether the decoded codeword is 0 or 1 in that coordinate. A maximum-likelihood decoding algorithm for the Golay code that uses between about 700 and 800 operations per 24-tuple on average can be used as a maximum-likelihood decoder for G24; i.e., this algorithm can be used to find the closest point in G24 to any given real 24-tuple r.
A decoding algorithm for H24 can then be specified as follows. Decoding Algorithm 1 (H24): Given any real 24-tuple r representing a received word, first find the closest point x0 in G24 to r. Check the parity of the twos-coefficient 24-tuple of x0; if it is even, then it is in H24, so accept it. If it is odd, then change one coordinate of x0 by ±2 in the coordinate xj0 where such a change will increase the squared distance (rj - xj0)2 by the least possible amount—i.e., where |rj - xj0 | is greatest. The resulting point x′0 has even twos-coefficient parity and is thus in H24.
Additional complexity in Decoding Algorithm 1 beyond decoding G24 can include a parity check and/or a computation and comparison of 24 magnitudes |rj - xj0|, however this complexity may be negligible.
Decoding Algorithm 1 always maps r into a point in H24 by construction, but not necessarily into the closest point in H24. For example, the 24-tuple x = (- 1, 17, 016) is in G24 but not in H24, and the point r = x/2 is at squared distance 2 from both x and the origin 0, which is in all lattices. If the G24 decoder resolves this tie by choosing x, then the parity check will fail. Changing one coordinate by ±2, however, cannot result in the origin 0, which is the closest point in H24, but rather must result in some other point in H24 of norm 8 that is at squared distance 4 from r.
Decoding Algorithm 1 does, however, always map r into the closest point x in H24 when r - x is within the error-correction radius of H24, as shall now be shown. In other words, Decoding Algorithm 1 is a bounded-distance decoding algorithm with the same error exponent as a maximum-likelihood decoder for H24.
Lemma 1: Given a 24-tuple r, if a point x exists in H24 such that ||r - x||2 < 2, then Decoding Algorithm 1 decodes r to x.
Proof: Without loss of generality, let x = 0 since if Decoding Algorithm 1 maps r to x, then it maps r + x′ to x + x′; i.e., suppose that ||r||2 < 2. Then, since the first step finds the closest point x0 in G24, x0 must either be 0 or a point x0 in G24 of norm ||x0||2 < 8. The only points in G24 with norm less than 8 are 0 and the points with a single nonzero coordinate of magnitude 2. If x0 is any of the latter points, however, then parity will fail, and 0 will be one of the candidates for the modified point x0′. In fact, 0 must then be chosen because any other candidate point is in H24, therefore has norm at least 8, and thus cannot be closer to r than 0. Hence all points r with ||r||2 < 2 map to 0.
The set of all points r that map to 0 is called the decision region R1(0) of Decoding Algorithm 1. By the translation property, the set of all points that map to any lattice point x is R1 (x) = R1(0) + x. Lemma 1 shows that R1(0) contains all points of norm less than 2. Since spheres of radius 2 drawn around the points of H24 must touch, there must be points of norm 2 on the boundary of R1(0). The number of points on the boundary of R1(0) with norm ||r||2 = 2 is the effective error coefficient N0,eff of Decoding Algorithm 1. Lemma 2 shows that Decoding Algorithm 1 approximately doubles the effective error coefficient of H24.
Lemma 2: The effective error coefficient of Decoding Algorithm 1 is N0,eff = 98,256 + 97,152 = 195,408.
Proof: In addition to the 98,256 points in H24 of norm 8, there are 759.128 = 97,152 points of norm 8 with eight coordinates of magnitude 1 and 16 of 0 that are in G24 but not in H24 (those with odd twos-coefficient parity). If the G24 decoder decodes r to any of these points in the first step of Decoding Algorithm 1, the second step cannot yield x′0 = 0.
To decode the Leech lattice Λ24, Decoding Algorithm 1 may be applied twice to the two cosets of H24 of which Λ24 is the union.
Decoding Algorithm 2 (Λ24): Given any real 24-tuple r, apply Decoding Algorithm 1 to r to find a point x0 in H24; also, Decoding Algorithm 1 to r - a to find a point x1 in H24 whose translate x1 + a is in the coset H24 + a. Compute the squared distances ||r - x0||2 and ||r - x1 + a)||2, and choose x1 or x1 + a according to which distance is smaller.
The complexity of Decoding Algorithm 2 is not much more than twice that of Decoding Algorithm 1 since the complexity of the translations of r and x1 by a and the computation and comparison of the two squared distances is small compared to the complexity of Golay decoding.
It can be shown that Decoding Algorithm 2 is a bounded-distance decoding algorithm that achieves the error-correction radius of Λ24 and increases the effective error coefficient by a factor of only about 1.5. Thus, the effective signal-to-noise ratio required by Decoding Algorithm 2 is only about 0.1 dB worse than that of maximum-likelihood decoding if the noise is Gaussian and the desired error rate is of the order of 10-6.
Theorem 1: Given a 24-tuple r, if there is a point x in Λ24 such that ||r - x||2, then Decoding Algorithm 2 decodes r to x.
Proof: If x is in H24, then by Lemma 1, Decoding Algorithm 1 applied to r yields x; if x is in H24 + a, then x - a is in H24 and thus by Lemma 1, Decoding Algorithm 1 applied to r -a yields x - a. Since
there can be only one point x in Λ24 such that ||r - x||2 < 2. If either of the two trial decodings finds such a point, it must be chosen as the closest point.
Theorem 2: The effective error coefficient of Decoding Algorithm 2 is N0,eff = 196,560 + 97,152 = 293,712.
Proof: The effective error coefficient is the number of points on the boundary of the decision region R2(0) of norm 2, which is the same as the number of points x in G24 or G24 + a of norm 8 that are in Λ24 or cannot be modified to 0 by a change of ±2 in one coordinate. This includes the 196,560 points in Λ24 of norm 8 and also the 97,152 points of norm 8 that are in G24 but not H24 that were mentioned in the proof of Lemma 2. Any point in G24 + a = G24 + (½)24 can be modified to a point in H24 + a by a change of ±2 in the first coordinate, so that there are no further points of this type.
In some embodiments, when selecting the boundary of a signal constellation used for data transmission, one tries to minimize the average energy of the constellation for a given number of points from a given packing. The reduction in the average energy per two dimensions due to the use of a region
as the boundary instead of a hypercube, is called the shaping gain γs of
The price to be paid for shaping involves an increase in the factor constellation-expansion ratio (CERs), an increase in the factor peak-to-average-power ratio (PAR), and an increase in the addressing complexity. There exists a tradeoff between γs and CERs, PAR, however as discussed herein, an N-dimensional shaping region may be selected having a structure that simultaneously optimizes the tradeoffs, as discussed below.
In one or more embodiments of the present disclosure, the integral of a function of the general form
over a final region
is calculated as:
This integral is used to calculate the volume and the second moment of the
region. The results, together with
where is the volume of shaping region
and R2 is the radius, can be used to compute γs, CERs, and PAR.
The expression ψ = β/n(n = N/2) can be used as the normalized parameter for the
region. The complete notation for the region is
For ψ = 1/n(β = 1, RN = R2), the spherical region ℓ2 (RN) is obtained. This case corresponds to the final point on the tradeoff curves. For
by increasing ψ, one moves along the curves towards their initial parts. Finally, for
one has
This results in the starting point on the tradeoff curves. The two cases of 0 < ψ < 1/n and ψ > 1 results in the regions
and
respectively.
Referring to
The S-points correspond to a spherical region and achieve the maximum shaping gain in a given dimensionality. The K-points correspond to rs = N/4 (CERs = (2)½ = 1.41). They achieve almost all of the shaping gain of the S-points, but with a much lower value of CERs, PAR. The L-points correspond to rs = N/8 (CERs = (2)½ = 1.19). They achieve a significant γs with a very low CERs, PAR. The B-points correspond to rs = 3 (CERs = (8)2/N). The A-points correspond to the addressing scheme based on the lattice
and result in rs = 1 (CERs = (2)2/N). For N = 4, this point corresponds to a spherical region.
From
to shape the N′ - D subspaces and then achieve another level of shaping on the n′ = N/N′-fold Cartesian product of these subspaces. This is one example of the application of a multilevel shaping/addressing scheme.
More generally, consider an
region. This region has an
region along each of its constituent N′ — D (N′ even) subspaces. The basic idea is that the
subregions can be modified such that the complexity of the addressing in the N-D space is decreased while the overall suboptimality is small. Specifically, in some of our schemes, 1) the
region is replaced by the region
and/or 2) this region is partitioned into a finite number of energy shells, and then the Cartesian product of the N′ -D subspaces can be shaped.
Additional details regarding multi-dimensional signal constellations, signal modulation, and other methods / techniques compatible with embodiments of the present disclosure can be found in the following references, the entire contents of each of which is incorporated by reference herein for all purposes and appended hereto as Appendices A-D, respectively: G. R. Lang and F. M. Longstaff, “A Leech lattice modem,” in IEEE Journal on Selected Areas in Communications, vol. 7, no. 6, pp. 968-973, Aug. 1989, doi: 10.1109/49.29618 (Appendix A); European Patent Number EP1608081, issued May 6, 2009 and titled “Apparatus and Method for Space-Frequency Block Coding/Decoding in a Communication System” (Appendix B); Zamir, Ram, “Lattice Coding for Signals and Networks,” Cambridge University Press, 2014 (Appendix C); and Hao, W. & Zhang, J.-Q & Song, H.-B (2014), A lattice based approach to the construction of multi-dimensional signal constellations, Tien Tzu Hsueh Pao/Acta Electronica Sinica, 42, 1672-1679. 10.3969/j.issn.0372-2112.2014.09.002 (Appendix D).
Additional details regarding lattice constructions and lattice encoding compatible with embodiments of the present disclosure can be found in Blake, Ian, “The Leech Lattice as a Code for the Gaussian Channel,” Information and Control 19, 66-74 (1971); J. Choi, Y. Nam and N. Lee, “Spatial Lattice Modulation for MIMO Systems,” in IEEE Transactions on Signal Processing, vol. 66, no. 12, pp. 3185-3198, 15 Jun. 15, 2018, doi: 10.1109/TSP.2018.2827325; G. D. Forney, “Coset codes. I. Introduction and geometrical classification,” in IEEE Transactions on Information Theory, vol. 34, no. 5, pp. 1123-1151, Sept. 1988, doi: 10.1109/18.21245; G.D. Forney, “Coset codes. II. Binary lattices and related codes,” in IEEE Transactions on Information Theory, vol. 34, no. 5, pp. 1152-1187, Sept. 1988, doi: 10.1109/18.21246; J. Conway and N. Sloane, “A fast encoding method for lattice codes and quantizers,” in IEEE Transactions on Information Theory, vol. 29, no. 6, pp. 820-824, November 1983, doi: 10.1109/TIT.1983.1056761; U.S. Patent Number 9,692,456, issued Jun. 27, 2017 and titled “Product Coded Modulation Scheme Based on E8 Lattice and Binary and Nonbinary Codes”; U.S. Patent Number 7,036,071, issued Apr. 25, 2006 and titled “Practical Coding and Metric Calculation for the Lattice Interfered Channel”; G. Forney, R. Gallager, G. Lang, F. Longstaff and S. Qureshi, “Efficient Modulation for Band-Limited Channels,” in IEEE Journal on Selected Areas in Communications, vol. 2, no. 5, pp. 632-647, September 1984, doi: 10.1109/JSAC.1984.1146101; A. K. Khandani and P. Kabal, “An efficient block-based addressing scheme for the nearly optimum shaping of multidimensional signal spaces,” in IEEE Transactions on Information Theory, vol. 41, no. 6, pp. 2026-2031, Nov. 1995, doi: 10.1109/18.476330; S. Stem and R. F. H. Fischer, “Lattice-Reduction-Aided Precoding for Coded Modulation over Algebraic Signal Constellations,” WSA 2016; 20th International ITG Workshop on Smart Antennas, 2016, pp. 1-8; and G. D. Forney and G. Ungerboeck, “Modulation and coding for linear Gaussian channels,” in IEEE Transactions on Information Theory, vol. 44, no. 6, pp. 2384-2415, Oct. 1998, doi: 10.1109/18.720542, the entire contents of each of which is incorporated by reference herein for all purposes.
Additional details regarding lattice decoding algorithms compatible with embodiments of the present disclosure can be found in: D. J. Costello and G. D. Forney, “Channel coding: The road to channel capacity,” in Proceedings of the IEEE, vol. 95, no. 6, pp. 1150-1177, June 2007, doi: 10.1109/JPROC.2007.895188; J. Conway and N. Sloane, “Fast quantizing and decoding and algorithms for lattice quantizers and codes,” in IEEE Transactions on Information Theory, vol. 28, no. 2, pp. 227-232, March 1982, doi: 10.1109/TIT.1982.1056484; and G. D. Forney, “A bounded-distance decoding algorithm for the Leech lattice, with generalizations,” in IEEE Transactions on Information Theory, vol. 35, no. 4, pp. 906-909, July 1989, doi: 10.1109/18.32173, the entire contents of each of which is incorporated by reference herein for all purposes.
Additional details regarding boundary selection for signal constellations in data transmission compatible with embodiments of the present disclosure can be found in: A. K. Khandani and P. Kabal, “Shaping multidimensional signal spaces. I. Optimum shaping, shell mapping,” in IEEE Transactions on Information Theory, vol. 39, no. 6, pp. 1799-1808, Nov. 1993, doi: 10.1109/18.265491; and A. K. Khandani and P. Kabal, “Shaping multidimensional signal spaces. II. Shell-addressed constellations,” in IEEE Transactions on Information Theory, vol. 39, no. 6, pp. 1809-1819, Nov. 1993, doi: 10.1109/18.265493, the entire contents of each of which is incorporated by reference herein for all purposes.
Example applications of lattice constellation compatible with embodiments of the present disclosure can be found in: U.S. Pat. No. 9,172,578, issued Oct. 27, 2015 and titled “High Speed Transceiver Based on Embedded Leech Lattice Constellation”; U.S. Pat. No. 8,989,283, issued May 24, 2015 and titled “High Speed Transceiver Based on Concatenates of a Leech Lattice with Binary and Nonbinary Codes”; G. Ungerboeck, “Channel coding with multilevel/phase signals,” in IEEE Transactions on Information Theory, vol. 28, no. 1, pp. 55-67, January 1982, doi: 10.1109/TIT.1982.1056454; and U. Wachsmann, R. F. H. Fischer and J. B. Huber, “Multilevel codes: theoretical concepts and practical design rules,” in IEEE Transactions on Information Theory, vol. 45, no. 5, pp. 1361-1391, July 1999, doi: 10.1109/18.771140, the entire contents of each of which is incorporated by reference herein for all purposes.
Implementations of the various techniques described herein may be implemented in digital electronic circuitry, or in computer hardware, firmware, software (executed or stored in hardware), or in combinations of them. Implementations may be implemented as a computer program product, i.e., a computer program tangibly embodied, e.g., in a machine-readable storage device (computer-readable medium, a non-transitory computer-readable storage medium, a tangible computer-readable storage medium, etc.), for processing by, or to control the operation of, data processing apparatus, e.g., a programmable processor, a computer, or multiple computers. A computer program, such as the computer program(s) described above, can be written in any form of programming language, including compiled or interpreted languages, and can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program can be deployed to be processed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a communication network.
Method steps may be performed by one or more programmable processors executing a computer program to perform functions by operating on input data and generating output. Method steps also may be performed by, and an apparatus may be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit).
Processors suitable for the processing of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read-only memory or a random access memory or both. Elements of a computer may include at least one processor for executing instructions and one or more memory devices for storing instructions and data. Generally, a computer also may include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks. Information carriers suitable for embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks. The processor and the memory may be supplemented by, or incorporated in, special purpose logic circuitry.
To provide for interaction with a user, implementations may be implemented on a computer having a display device, e.g., a liquid crystal display (LCD or LED) monitor, a touchscreen display, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input.
Implementations may be implemented in a computing system that includes a back-end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front-end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation, or any combination of such back-end, middleware, or front-end components. Components may be interconnected by any form or medium of digital data communication, e.g., a communication network. Examples of communication networks include a local area network (LAN) and a wide area network (WAN), e.g., the Internet.
While certain features of the described implementations have been illustrated as described herein, many modifications, substitutions, changes and equivalents will now occur to those skilled in the art. It is, therefore, to be understood that the appended claims are intended to cover all such modifications and changes as fall within the scope of the implementations. It should be understood that they have been presented by way of example only, not limitation, and various changes in form and details may be made. Any portion of the apparatus and/or methods described herein may be combined in any combination, except mutually exclusive combinations. The implementations described herein can include various combinations and/or sub-combinations of the functions, components and/or features of the different implementations described.
This application claims priority to and the benefit of U.S. Provisional Pat. Application No. 63/312,496, filed Feb. 22, 2022 and titled “METHODS AND APPARATUS FOR SIGNAL MODULATION USING LATTICE-BASED SIGNAL CONSTELLATIONS,” the contents of which are incorporated by reference herein in their entirety.
Number | Date | Country | |
---|---|---|---|
63312496 | Feb 2022 | US |