The present document relates to mobile wireless communication, and more particularly, to massive cooperative multipoint network operation.
Due to an explosive growth in the number of wireless user devices and the amount of wireless data that these devices can generate or consume, current wireless communication networks are fast running out of bandwidth to accommodate such a high growth in data traffic and provide high quality of service to users.
Various efforts are underway in the telecommunication industry to come up with next generation of wireless technologies that can keep up with the demand on performance of wireless devices and networks. Many of those activities involve situations in which a large number of user devices may be served by a network.
This document discloses devices and techniques useful for embodiments of wireless technologies in which cooperative multipoint operation of wireless networks is achieved.
In one example aspect, a method of wireless communication is disclosed. The method includes determining, by a network device, a cooperative multipoint (COMP) management status of wireless devices served by the network device, and providing, by the network device, wireless connectivity to the one or more wireless devices, wherein the network device jointly manages transmission resources for a first wireless device due to the COMP management status being a joint COMP status and the network device locally manages transmission resources for a second wireless device due to the COMP management status being a local COMP status.
In yet another example aspect, a wireless communication system is disclosed. The system includes an arrangement of a plurality of network nodes in which each network node is configured to provide wireless connectivity to wireless devices using a mode that includes a joint cooperative multipoint (COMP) mode and a local COMP mode, wherein, in the joint COMP mode, transmission resources for wireless devices are managed cooperatively with other network nodes, and wherein, in the local COMP mode, transmission resources for wireless devices are managed locally, without explicit coordination with other network nodes.
In yet another example aspect, a wireless communication apparatus that implements the above-described methods is disclosed.
In yet another example aspect, the methods may be embodied as processor-executable code and may be stored on a computer-readable program medium.
These, and other, features are described in this document.
Drawings described herein are used to provide a further understanding and constitute a part of this application. Example embodiments and illustrations thereof are used to explain the technology rather than limiting its scope.
To make the purposes, technical solutions and advantages of this disclosure more apparent, various embodiments are described in detail below with reference to the drawings. Unless otherwise noted, embodiments and features in embodiments of the present document may be combined with each other.
Cellular wireless service providers have begun planning and deployment of next generation networks to support deployment of denser deployments of higher bandwidth user devices. Furthermore, the ever-increasing reliance on wireless connectivity has raised users' expectations of Quality of Service and seamless availability of wireless connectivity everywhere.
Cloud Radio Access Network (C-RAN) is one example of a network architecture in which a centralized cloud-based access network provides wireless connectivity to wireless terminals. However, C-RAN deployments rely on expensive deployments of fiber optic infrastructure to connect base stations with each other and with a central network controller. Furthermore, such an architecture requires planning, and deployments can be relatively slow due to the labor and resources required to lay down fiber. As a result, C-RAN and similar solutions are expensive, and cannot be quickly deployed (or taken down) to meet short term increase in demand of wireless services. Furthermore, when such an deployment reaches its maximum capacity, incremental deployment is often not possible without having to significantly alter the existing infrastructure.
The techniques described in the present document can be used in wireless network embodiments to overcome such problems. In one example aspect, network nodes may be deployed using short range, high speed millimeter wavelength (mmwave) links. Such installations have minimal footprint and power requirements and can be deployed and taken down to quickly meet time and geography-specific demand for wireless services.
In another beneficial aspect, the present technology may be used to deploy networks that provide short links between base stations, or network nodes, thereby providing reduced latency, jitter and fronthaul traffic loading in wireless networks.
In another beneficial aspect, the disclosed techniques may be used to manage a soft handover whereby a user equipment (UE) and N neighboring Base stations (typically N=3) constitute a cooperative multi-point (COMP) service zone.
In another beneficial aspect, embodiments may benefit from increased network performance without any change or replacement of existing antennas on towers, e.g., does require setting new mmwave links or computing platforms. The inventor's rough calculations have shown that it may be possible for embodiments to increase network capacity by at least a factor of two and at least 5db Signal to Interference and Noise Ratio (SINR) improvement.
Some embodiments of the disclosed distributed COMP technology may be used to address both intra-cell and inter-cell interference, or alternatively inter-sector interference and cell edge poor coverage, using a computer platform that processes jointly all three sectors of all towers in a cluster. One advantage is that the physical front end, e.g., antennas on tower, may not have to be changed, and yet the techniques may be embodied for boosting performance.
As further described in the present document, in some embodiments, distributed COMP may include groups of cell towers in which all cell towers carry the functionality of a Remote Radio Head (RRH) while one of them carry the computation for the cluster and is connected to the network for TCP/IP traffic. In other words, there is no need for a fronthaul to the network. Cluster formation may be performed using one of the techniques described in the present document. A cluster takes advantage of shared resource management and load balancing.
Embodiments of the disclosed technology provide various improvements to the operation of wireless networks and equipment, including:
1) Accurate geometry extraction and multipath attributes acquisition, based on instantaneous measurements over a limited band and over a short period of time. For example, the sparse channel representation technique, as described in Section 3, provides a computationally efficient was of modeling and predicting channels. The computations can be performed by a network device on behalf of multiple base stations and thus the network device can control transmissions from/to the multiple base stations so that wireless devices can move freely between coverage areas of the base stations without any interference from transmissions from/to other base stations in the distributed cooperative zone of base stations.
2) Accurate channel prediction on same band or on a different adjacent band based on instantaneous measurements over a limited band and over a short period of time, as described in Sections 2-5. The sparse channel measurement may be performed using very few reference signal transmissions and thus channel conditions in multiple neighboring cells can be quickly acquired and used at a network-side server that controls operation of distributed base stations in a cooperative manner.
3) Use of predicted channel state information for centralized & distributed MU-MIMO precoding. For example, Sections 2, 4 and 5 describe certain techniques for predicting channels at different time instances, frequencies and spatial positions.
4) Use of predicted channel state information to determine Modulation and Coding Scheme (MCS) attributes (Resource block bit loading-modulation order and forward error correction codes).
5) Use of predicted channel State information to determine retransmission to meet delivery reliability criteria. For example, the network server may obtain accurate channel estimates at future times or other frequencies and, without waiting for ACK feedbacks, is able to decide retransmission strategy based on the channel conditions. For example, channel condition may be compared with a threshold and retransmission may be used when channel condition falls below the threshold.
6) Base Station clustering & front haul network organization for defining CoMP regions & Soft handoff between CoMP regions, as described in Section 2.
7) Pilot arrangement to minimize pilot contamination, as described in Section 6. The central awareness of channels for all base stations in a zone or a cluster allows the cluster controller on the network side to arrange pilots from different base stations to be non-overlapping in terms of their transmission resources.
8) Signal processing to separate pilot mixtures and contamination mitigation.
Embodiments of the disclosed technology include distributed COMP architectures that implement a separation of a base station's functionality of transmission and reception of radio frequency (RF) signals between UEs and the functionality of channel estimation, prediction, precoding and retransmission management. Furthermore, millimeter (mm) wave links may be established between the RF functionality sites and remote or network-side computing servers for exchanging information related to ongoing operation of the cellular network.
The following calculations may be used for resource planning in the network.
#Nodes=9n{circumflex over ( )}2−3n+1
#Clusters=3n{circumflex over ( )}2−3n+1
#Links=6(3n{circumflex over ( )}2−3n+1)
3n(n−1)+1=(R/D){circumflex over ( )}2
The table below shows example values which may be used in some embodiments.
The embodiments described in the present document may be used to achieve wide scale COMP, as is described herein. For example, the clustering approach may be used on regional basis, while the entire network may include clustered and COMP-based operating cell, while some other cells may be operating in the conventional manner.
One limitation with present day implementations of wireless technologies is that wireless networks are not able to achieve full rank operation and manage interference among various transmission points (e.g., base stations). For example, in embodiments in which dual polarization is used for signal transmissions, only about 40% gain in efficiency is achieved over single polarization transmission due to imperfections in estimation and/or use of transmission rank of the channel. As a result, often, in practical implementations, transmissions are performed using single user MIMO (SU-MIMO) even in cases where MU-MIMO operation ins possible. As further described throughout the present document, the techniques described herein may be used to achieve the following operational advantages:
(1) Use MU-MIMO in a true sense—i.e., every time it is theoretically possible, then it is practically used
(2) Predict channel at a future time or at a different frequency accurately, using, for example, reciprocity and sparse channel computation techniques described in the present document. Using the predicted channel, scheduling operation may be improved in the selection of correct resource blocks for a UE, and also in selecting a modulation scheme for the selected resource blocks.
(3) Coordinate operation of towers (base stations) to minimize interference and improve signal to noise ratio (SNR) per layer of transmission.
One advantage of the depicted embodiment is that the connection between base stations and the mobile core network does not need any changes and can continue to operate as before. The base stations are communicating with each other through a separate link for exchanging information regarding channel conditions and UE related information such as the h coefficients that can be used to generate the pre-coding coefficients.
In the example shown in
In some embodiments, and as shown in CoMP server computations in
In some embodiments, and as shown in
In some embodiments, the information exchanged by the base stations over the separate link is on the order of 400 Mbps/20 MHz/layer and can be easily accommodated by a wireless transmission link between base stations using, for example, orthogonal time frequency space (OTFS) modulated signals. In other embodiments, the separate link may be a mmwave link or an 10G optical fiber link.
One advantage of the depicted system is that each base station or tower only needs to connect to a first-hop neighbor, thereby simplifying operation of the b2b link between base station. Furthermore, the COMP server base station only has to be able to coordinate downstream and upstream traffic for its nearest neighbors, a task that is within reasonable limits of computation to make it practical.
In some embodiments, traffic on the downlink may be characterized by the following steps:
In some embodiments, the uplink traffic may be characterized by as follows:
In some embodiments, and as shown in
As shown in
The function splits shown in the example in
In some embodiments, network entry starts with the UE transmitting a random access preamble in the PRACH. The eNB is required to answer within a configurable random access response window (as shown in
To mitigate the impact on UL power-limited UEs, the forced ACK may be combined with subframe bundling as shown in
For example, as shown in
On the transmit side, the controller 3300 may determine a layer association 7 for signal to be transmitted. The layer association 7 may be processed by a multiplexing block 3306 such that all outgoing traffic of a same layer is mapped together on a layer for transmission. The multiplexing block 3306 may also perform the tasks of mapping outgoing user data to antennas, specifying precoder to be used for each stream of data and a modulation scheme (bitloading) to be used for each layer or stream. The layered streams of data may then be precoded in the SSP block 3304 and transmitted via antennas 3302 (in some examples, a single antenna may be used).
A wireless channel, between a transmitting antenna of a device to a receiving antenna of another device, may be described by a geometric model of rays originating from the transmitting device, as shown in
More formally, let αi, τi, θi and υi represent the complex gain, delay, AoA and Doppler of ray i, respectively. Then, for Nr rays (or reflectors), the wireless channel response at time t, space s and frequency f is
where, the space dimension is used for multiple receive antennas.
The channel response, H, may be considered to be a super-position of all received rays. This patent document describes a method for extracting the values αi, τi, θi and υi from H(t, s, f), under the assumptions that Nr is small (for example, typical situations may have between zero to 10 reflectors). Obtaining these values, gives a sparse and compact channel representation of equation (0)), regardless of how large the frequency, time and space dimensions are.
For example, a channel with 3 NLoS reflectors, received over 16 antennas, 512 frequency tones and 4 time samples, will be described by 3×4=12 values (gain, delay, angle and Doppler for 3 reflectors) instead of 16×512×4=32768 values.
Furthermore, with the knowledge of the values of αi, τi, θi and υi, a covariance matrix for the channel, RHH, can be constructed for any desired frequency, time and space instances. This can be applied for predicting the channel response in any one of these dimensions.
This section describes two methods for constructing the covariance matrix under the sparse channel assumptions. Both methods use convex optimization techniques. Variations of these methods, or alternative methods may be used as well.
It is assumed that the channel response, H, is given for Nt, Ns and Nf time, space and frequency grid points, respectively. This channel response may be obtained from known reference signals transmitted from one device to the other.
The following algorithm solves the optimization problem of finding the complex values of vectors in delay, angular and Doppler dimensions, which after transformation to frequency, time and space, will give a channel response, which is the closest to the empirical measurement of the channel response H, under the assumption that the number of elements with non-negligible energy in these vectors is small (sparse).
More specifically, let's define grids of Mτ, Mθ and Mυpoints over the delay, angular and Doppler dimensions, respectively. These grids represent the desired detection resolution of these dimensions. Let λτ, λθ and λυbe vectors of complex values over these grids. The constructed channel response is
The general optimization problem minimizes ∥λτ∥1, ∥λθ∥1 and ∥λυ∥1, subject to
where ∥·∥1 is the L1 norm and ε represents a small value (which may correspond to the SNR of the channel).
The above optimization problem may be too complex to solve directly. To reduce the complexity, one possible alternative is to solve the problem sequentially for the different dimensions. Any order of the dimensions may be applied.
For example, embodiments may start with the delay dimension and solve the optimization problem of minimizing ∥λτ∥1, subject to
For the solution, we detect the delay indexes with non-negligible energy, mτ∈T, such that |λτ(mτ)|2≥ελ, where ελ represents an energy detection threshold (which may correspond to the SNR of the channel). Then, embodiments may continue to solve for the next dimension, for example, the angular dimension. For the next optimization problem, we reduce the delay dimension from Mτ indexes, to the set of indexes, T. Thus, we solve the optimization problem of minimizing ∥λτ,θ∥1, subject to
Note, that the size of the optimization vector is now T·Mθ and mτ,θ is an index to this vector, corresponding to delay indexes in T and angular indexes in Mθ. For this solution, some embodiments may detect the delay-angular indexes with non-negligible energy, mτ,θ∈TΘ, such that |λτ,θ(mτ,θ)|2≥ελ and continue to the final dimension, Doppler. Here, embodiments may solve the optimization problem of minimizing ∥λτ,θ,υ∥1, subject to
The size of the optimization vector is now TΘ·Mυand mτ,θ,υis an index to this vector, corresponding to delay indexes in T, angular indexes in Θ and Doppler indexes in Mυ. Finally, embodiments may detect the Doppler indexes with non-negligible energy, mτ,θ,υ∈TΘΥ, such that |λτ,θ,υ(mτ,θ,v)|2≥ελ. The final information, representing the sparse channel, is now a small set of |TΘΥ| values.
Now, for any selection of time, space and frequency grids, denoted by the indexes t′, s′ and f′, we can use this representation to construct a covariance for the channel as
The following algorithm solves the optimization problem of finding the most likelihood covariance matrix, for an empirical channel measurement. Let's consider the function r(·), which translates a covariance from the delay, angular or Doppler dimensions, to frequency, space or time dimensions
r(α,κ,x)=sinc(κ·x)·ej2παx
The covariance of the channel is a Toeplitz matrix generated by the function:
In the above equation, Mτ, Mθ and Mυare the desired resolutions in delay, angular and Doppler dimensions, κτ, κθ and κυare constants and f, s and t are indexes in the frequency, space and time grids. The variables λτ, λθ and λυare the unknown non-negative weights that needs to be determined for each element in these three dimensions. To find them, some embodiments may solve the optimization problem of finding the covariance R that maximizes the probability of getting the empirical channel response H. More formally, find
One possible method for solving this, is to use convex optimization techniques for an equivalent minimization problem. The assumption of a sparse channel representation is not used explicitly to formalize the optimization problem. However, the geometric physical model of the channel in delay, angular and Doppler dimensions, implicitly implies of a sparse channel representation.
To reduce the complexity of solving such an optimization problem, it is possible to perform the optimization sequentially over the dimensions (in any order), in a similar way to the one described for method 1. First, we solve for one of the dimensions, for example delay. We find the delay covariance
that maximizes the probability
Then, some embodiments may detect the delay indexes with non-negligible energy, τ∈T, such that |λτ(τ)|2≥ελ, where ελ represents an energy detection threshold (which may correspond to the SNR of the channel). Then, some embodiments may continue to solve for the next dimension, for example, the angular dimension. Some embodiments may find the delay-angular covariance matrix as follows:
that maximizes the probability
Again, some embodiments may detect the delay-angular indexes with non-negligible energy, τ, θ∈TΘ, such that |λτθ(τ, θ)|2≥Eλ and continue to solve for the final dimension, Doppler. Some embodiments may find the delay-angular-Doppler covariance matrix
that maximizes the probability
Finally, some embodiments may detect the delay-angular-Doppler indexes with non-negligible energy, τ, θ, υ∈TΘΥ, such that |λτθυ(τ, θ, υ)|2≥ελ and use them to construct a covariance for the channel for any selection of frequency, space and time grids, denoted by the indexes f′, s′ and t′
The optimization problems, solved for a grid size of M points in one of the dimensions can be iteratively solved by constructing the M points in a tree structure. For example, M=8 can be constructed as a tree of 3 levels, as shown in
It will be appreciated by practitioners of the art that it is possible to use the detection tree for both method 1 and 2, described above.
Once the reflectors are detected (method 1), or the covariance weights are determined (method 2), a covariance matrix can be constructed for any frequency, space and time grids. If we denote these grids as a set of elements, Y, and denote X as a subset of Y, and representing the grid elements for an instantaneous measurement of the channel as HX, then the prediction filter may be computed as
C=RYX·(RXX)−1
and the predicted channel is computed as
ĤY=C·HX
The matrices RYX and RXX are a column decimated, and a row-column decimated, versions of the channel constructed covariance matrix. These matrices are decimated to the grid resources represented by X.
The described techniques may be used for predicting the wireless channels in a Time Division Duplex (TDD) or a Frequency Division Duplex (FDD) system. Such a system may include base stations (BS) and multiple user equipment (UE). This technique is suitable for both stationary and mobile UE. Generally, these techniques are used to compute a correct covariance matrix representing the wireless channels, based on a sparse multi-dimensional geometric model, from a relatively small number of observations (in frequency, time and space). From this covariance matrix, a prediction filter is computed and applied to some channel measurements, to predict the channels in some or all the frequency, space and time dimensions. The predicted channels for the UE, along with other predicted channels for other UE, may be used to generate a precoded downlink transmission from one BS to multiple UE (Multi-User MIMO, Mu-MIMO), or from several BS to multiple UE (also known as CoMP—Coordinated Multi-Point or distributed Mu-MIMO).
Note, that although most of the computational load, described in the following paragraphs, is attributed to the BS (or some other network-side processing unit), some of it may be performed, in alternative implementations, in the UE.
In this scenario, the BS predicts the wireless channels from its antennas to the UE in a future time instance. This may be useful for generating a precoded downlink transmission. The UE may transmit at certain time instances reference signals to the BS, from which the BS will estimate the wireless channels response. Note, that typically, a small number of time instances should be sufficient, which makes it a method, suitable for mobile systems. Then, the estimated channel responses (whole or partial), are used with one of the described methods, to determine the covariance matrix of the channels and compute a prediction filter. This processing may take place in the base station itself, or at a remote or a network-side processing unit (also known as “cloud” processing). The prediction filter may be applied to some of the channel responses already received, or to some other estimated channel responses, to generate a prediction of the wireless channels, at a future time instance and over the desired frequency and space grids.
In this scenario too, the BS predicts the wireless channels from its antennas to the UE in a future time instance. However, the UE to BS uplink transmissions and the BS to UE downlink transmissions are over different frequency bands. The generation of the of prediction filter is similar to TDD systems. The UE may transmit at certain time instances reference signals to the BS, from which the BS will estimate the wireless channels response. Then, the estimated channel responses (whole or partial), are used with one of the described methods, to determine the covariance matrix of the channels and compute a prediction filter. In parallel, at any time instance, the BS may transmit reference signals to the UE. The UE will feedback to the BS through its uplink, some the received reference signals (all or partial), as raw or processed information (implicit/explicit feedback). The BS will generate, if needed, an estimated channel response for the downlink channel, from the information received from the UE and apply the prediction filter to it. The result is a predicted channel at the downlink frequency band and at a future time instance.
It is useful for the BS to know the quality of the prediction of the channels in order to determine correctly which modulation and coding (MCS) to use for its precoded transmission. The more accurate the channels are represented by the computed covariance matrix; the higher prediction quality is achieved, and the UE will have a higher received SNR.
This section covers multiple access and precoding protocols that are used in typical OTFS systems.
Orthogonal Multiple Access
Currently the common technique used for multiple access is orthogonal multiple access. This means that the hub breaks it's time and frequency resources into disjoint pieces and assigns them to the UEs. An example is shown in
The advantage of orthogonal multiple access is that each UE experience its own private channel with no interference. The disadvantage is that each UE is only assigned a fraction of the available resource and so typically has a low data rate compared to non-orthogonal cases.
Precoding Multiple Access
Recently, a more advanced technique, precoding, has been proposed for multiple access. In precoding, the hub is equipped with multiple antennas. The hub uses the multiple antennas to create separate beams which it then uses to transmit data over the entire bandwidth to the UEs. An example is depicted in
The advantage of precoding it that each UE receives data over the entire bandwidth, thus giving high data rates. The disadvantage of precoding is the complexity of implementation. Also, due to power constraints and noisy channel estimates the hub cannot create perfectly disjoint beams, so the UEs will experience some level of residual interference.
Introduction to Precoding
Precoding may be implemented in four steps: channel acquisition, channel extrapolation, filter construction, filter application.
Channel acquisition: To perform precoding, the hub determines how wireless signals are distorted as they travel from the hub to the UEs. The distortion can be represented mathematically as a matrix: taking as input the signal transmitted from the hubs antennas and giving as output the signal received by the UEs, this matrix is called the wireless channel.
Channel prediction: In practice, the hub first acquires the channel at fixed times denoted by s1, s2, . . . , sn. Based on these values, the hub then predicts what the channel will be at some future times when the pre-coded data will be transmitted, we denote these times denoted by t1, t2, . . . , tm.
Filter construction: The hub uses the channel predicted at t1, t2, . . . , tm to construct precoding filters which minimize the energy of interference and noise the UEs receive.
Filter application: The hub applies the precoding filters to the data it wants the UEs to receive.
Channel Acquisition
This section gives a brief overview of the precise mathematical model and notation used to describe the channel.
Time and frequency bins: the hub transmits data to the UEs on a fixed allocation of time and frequency. This document denotes the number of frequency bins in the allocation by Nf and the number of time bins in the allocation by Nt.
Number of antennas: the number of antennas at the hub is denoted by Lh, the total number of UE antennas is denoted by Lu.
Transmit signal: for each time and frequency bin the hub transmits a signal which we denote by φ(f, t)∈L
Receive signal: for each time and frequency bin the UEs receive a signal which we denote by y(f, t)∈L
White noise: for each time and frequency bin white noise is modeled as a vector of iid Gaussian random variables with mean zero and variance N0. This document denotes the noise by w(f, t)∈L
Channel matrix: for each time and frequency bin the wireless channel is represented as a matrix and is denoted by H(f, t)∈L
The wireless channel can be represented as a matrix which relates the transmit and receive signals through a simple linear equation:
y(f,t)=H(f,t)φ(f,t)+w(f,t) (1)
for f=1, . . . , Nf and t=1, . . . , Nt.
Two common ways are typically used to acquire knowledge of the channel at the hub: explicit feedback and implicit feedback.
Explicit Feedback
In explicit feedback, the UEs measure the channel and then transmit the measured channel back to the hub in a packet of data. The explicit feedback may be done in three steps.
Pilot transmission: for each time and frequency bin the hub transmits a pilot signal denoted by p(f, t)∈L
Channel acquisition: for each time and frequency bin the UEs receive the pilot signal distorted by the channel and white noise:
H(f,t)p(f,t)+w(f,t), (2)
for f=1, . . . , Nf and t=1, . . . , Nt. Because the pilot signal is known by the UEs, they can use signal processing to compute an estimate of the channel, denoted by Ĥ(f, t).
Feedback: the UEs quantize the channel estimates Ĥ(f, t) into a packet of data. The packet is then transmitted to the hub.
The advantage of explicit feedback is that it is relatively easy to implement. The disadvantage is the large overhead of transmitting the channel estimates from the UEs to the hub.
Implicit Feedback
Implicit feedback is based on the principle of reciprocity which relates the uplink channel (UEs transmitting to the hub) to the downlink channel (hub transmitting to the UEs).
Specifically, denote the uplink and downlink channels by Hup and H respectively, then:
H(f,t)=AHupT(f,t)B, (3)
for f=1, . . . , Nf and t=1, . . . , Nt. Where HupT(f, t) denotes the matrix transpose of the uplink channel. The matrices A∈L
H(f,t)=HupT(f,t) (4)
The principle of reciprocity can be used to acquire channel knowledge at the hub. The procedure is called implicit feedback and consists of three steps.
Reciprocity calibration: the hub and UEs calibrate their hardware so that equation (4) holds.
Pilot transmission: for each time and frequency bin the UEs transmits a pilot signal denoted by p(f, t)∈L
Channel acquisition: for each time and frequency bin the hub receives the pilot signal distorted by the uplink channel and white noise:
Hup(f,t)p(f,t)+w(f,t) (5)
for f=1, . . . , Nf and t=1, . . . , Nt. Because the pilot signal is known by the hub, it can use signal processing to compute an estimate of the uplink channel, denoted by (f, t). Because reciprocity calibration has been performed the hub can take the transpose to get an estimate of the downlink channel, denoted by Ĥ(f, t).
The advantage of implicit feedback is that it allows the hub to acquire channel knowledge with very little overhead; the disadvantage is that reciprocity calibration is difficult to implement.
Channel Prediction
Using either explicit or implicit feedback, the hub acquires estimates of the downlink wireless channel at certain times denoted by s1, s2, . . . , sn using these estimates it must then predict what the channel will be at future times when the precoding will be performed, denoted by t1, t2, . . . , tm.
There are tradeoffs when choosing the feedback times s1, s2, . . . , sn.
Latency of extrapolation: Refers to the temporal distance between the last feedback time, sn, and the first prediction time, t1, determines how far into the future the hub needs to predict the channel. If the latency of extrapolation is large, then the hub has a good lead time to compute the pre-coding filters before it needs to apply them. On the other hand, larger latencies give a more difficult prediction problem.
Density: how frequent the hub receives channel measurements via feedback determines the feedback density. Greater density leads to more accurate prediction at the cost of greater overhead.
There are many channel prediction algorithms in the literature. They differ by what assumptions they make on the mathematical structure of the channel. The stronger the assumption, the greater the ability to extrapolate into the future if the assumption is true. However, if the assumption is false then the extrapolation will fail. For example:
Polynomial extrapolation: assumes the channel is smooth function. If true, can extrapolate the channel a very short time into the future≈0.5 ms.
Bandlimited extrapolation: assumes the channel is a bandlimited function. If true, can extrapolated a short time into the future≈1 ms.
MUSIC extrapolation: assumes the channel is a finite sum of waves. If true, can extrapolate a long time into the future≈10 ms.
Precoding Filter Computation and Application
Using extrapolation, the hub computes an estimate of the downlink channel matrix for the times the pre-coded data will be transmitted. The estimates are then used to construct precoding filters. Precoding is performed by applying the filters on the data the hub wants the UEs to receive. Before going over details we introduce notation.
Channel estimate: for each time and frequency bin the hub has an estimate of the downlink channel which we denote by Ĥ(f, t)∈L
Precoding filter: for each time and frequency bin the hub uses the channel estimate to construct a precoding filter which we denote by W(f, t)∈L
Data: for each time and frequency bin the UE wants to transmit a vector of data to the UEs which we denote by x(f, t)∈L
Hub Energy Constraint
When the precoder filter is applied to data, the hub power constraint is an important consideration. We assume that the total hub transmit energy cannot exceed NfNtLh. Consider the pre-coded data:
W(f,t)x(f,t), (6)
for f=1, . . . , Nf and t=1, . . . , Nt. To ensure that the pre-coded data meets the hub energy constraints the hub applies normalization, transmitting:
λW(f,t)x(f,t), (7)
for f=1, . . . , Nf and t=1, . . . , Nt. Where the normalization constant λ is given by:
Receiver SNR
The pre-coded data then passes through the downlink channel, the UEs receive the following signal:
λH(f,t)W(f,t)x(f,t)+w(f,t), (9)
for f=1, . . . , Nf and t=1, . . . , Nt. The UE then removes the normalization constant, giving a soft estimate of the data:
for f=1, . . . , Nf and t=1, . . . , Nt. The error of the estimate is given by:
The error of the estimate can be split into two terms. The term H(f, t)W(f, t)−x(f, t) is the interference experienced by the UEs while the term
gives the noise experienced by the UEs.
When choosing a pre-coding filter there is a tradeoff between interference and noise. We now review the two most popular pre-coder filters: zero-forcing and regularized zero-forcing.
Zero Forcing Precoder
The hub constructs the zero forcing pre-coder (ZFP) by inverting its channel estimate:
WZF(f,t)=(Ĥ*(f,t)Ĥ(f,t))−1Ĥ*(f,t), (12)
for f=1, . . . , Nf and t=1, . . . , Nt. The advantage of ZPP is that the UEs experience little interference (if the channel estimate is perfect then the UEs experience no interference). The disadvantage of ZFP is that the UEs can experience a large amount of noise. This is because at time and frequency bins where the channel estimate Ĥ(f, t) is very small the filter WZF(f, t) will be very large, thus causing the normalization constant λ to be very small giving large noise energy.
Regularized Zero-Forcing Pre-Coder (rZFP)
To mitigates the effect of channel nulls (locations where the channel has very small energy) the regularized zero forcing precoder (rZFP) is constructed be taking a regularized inverse of its channel estimate:
WrZF(f,t)=(Ĥ*(f,t)Ĥ(f,t)+αI)−1Ĥ*(f,t), (13)
for f=1, . . . , Nf and t=1, . . . , Nt. Where α>0 is the normalization constant. The advantage of rZFP is that the noise energy is smaller compared to ZPF. This is because rZFP deploys less energy in channel nulls, thus the normalization constant λ is larger giving smaller noise energy. The disadvantage of rZFP is larger interference compared to ZFP. This is because the channel is not perfectly inverted (due to the normalization constant), so the UEs will experience residual interference.
As described above, there are three components to a precoding system: a channel feedback component, a channel prediction component, and a pre-coding filter component. The relationship between the three components is displayed in
OTFS Precoding System
Various techniques for implementing OTFS precoding system are discussed. Some disclosed techniques can be used to provide the ability to shape the energy distribution of the transmission signal. For example, energy distribution may be such that the energy of the signal will be high in regions of time frequency and space where the channel information and the channel strength are strong. Conversely, the energy of the signal will be low in regions of time frequency and space where the channel information or the channel strength are weak.
Some embodiments may be described with reference to three main blocks, as depicted in
Channel prediction: During channel prediction, second order statistics are used to build a prediction filter along with the covariance of the prediction error.
Optimal precoding filter: using knowledge of the predicted channel and the covariance of the prediction error: the hub computes the optimal precoding filter. The filter shapes the spatial energy distribution of the transmission signal.
Vector perturbation: using knowledge of the predicted channel, precoding filter, and prediction error, the hub perturbs the transmission signal. By doing this the hub shapes the time, frequency, and spatial energy distribution of the transmission signal.
Review of OTFS Modulation
A modulation is a method to transmit a collection of finite symbols (which encode data) over a fixed allocation of time and frequency. A popular method used today is Orthogonal Frequency Division Multiplexing (OFDM) which transmits each finite symbol over a narrow region of time and frequency (e.g., using subcarriers and timeslots). In contrast, Orthogonal Time Frequency Space (OTFS) transmits each finite symbol over the entire allocation of time and frequency. Before going into details, we introduce terminology and notation.
We call the allocation of time and frequency a frame. We denote the number of subcarriers in the frame by Nf. We denote the subcarrier spacing by df. We denote the number of OFDM symbols in the frame by Nt. We denote the OFDM symbol duration by dt. We call a collection of possible finite symbols an alphabet, denoted by A.
A signal transmitted over the frame, denoted by φ, can be specified by the values it takes for each time and frequency bin:
φ(f,t)∈, (14)
for f=1, . . . , Nf and t=1, . . . , Nt.
OTFS Modulation
Suppose a transmitter has a collection of NfNt QAM symbols that the transmitter wants to transmit over a frame, denoted by:
x(f,t)∈A, (15)
for f=1, . . . , Nf and t=1, . . . , Nt. OFDM works by transmitting each QAM symbol over a single time frequency bin:
φ(f,t)=x(f,t), (16a)
for f=1, . . . , Nf and t=1, . . . , Nt. The advantage of OFDM is its inherent parallelism, this makes many computational aspects of communication very easy to implement. The disadvantage of OFDM is fading, that is, the wireless channel can be very poor for certain time frequency bins. Performing pre-coding for these bins is very difficult.
The OTFS modulation is defined using the delay Doppler domain, which is relating to the standard time frequency domain by the two-dimensional Fourier transform.
The delay dimension is dual to the frequency dimension. There are Nτ delay bins with Nτ=Nf. The Doppler dimension is dual to the time dimension. There are Nν Doppler bins with Nν=Nt.
A signal in the delay Doppler domain, denoted by ϕ, is defined by the values it takes for each delay and Doppler bin:
ϕ(τ,ν)∈, (16b)
for τ=1, . . . , Nτ and ν=1, . . . , Nν.
Given a signal ϕ in the delay Doppler domain, some transmitter embodiments may apply the two-dimensional Fourier transform to define a signal φ in the time frequency domain:
φ(f,t)=(Fϕ)(f,t), (17)
for f=1, . . . , Nf and t=1, . . . , Nt. Where F denotes the two-dimensional Fourier transform.
Conversely, given a signal φ in the time frequency domain, transmitter embodiments could apply the inverse two-dimensional Fourier transform to define a signal ϕ in the delay Doppler domain:
ϕ(τ,ν)=(F−1φ)(τ,ν), (18)
for τ=1, . . . , Nτ and ν=1, . . . , Nν.
The advantage of OTFS is that each QAM symbol is spread evenly over the entire time frequency domain (by the two-two-dimensional Fourier transform), therefore each QAM symbol experience all the good and bad regions of the channel thus eliminating fading. The disadvantage of OTFS is that the QAM spreading adds computational complexity.
MMSE Channel Prediction
Channel prediction is performed at the hub by applying an optimization criterion, e.g., the Minimal Mean Square Error (MMSE) prediction filter to the hub's channel estimates (acquired by either implicit or explicit feedback). The MMSE filter is computed in two steps. First, the hub computes empirical estimates of the channel's second order statistics. Second, using standard estimation theory, the hub uses the second order statistics to compute the MMSE prediction filter. Before going into details, we introduce notation:
We denote the number of antennas at the hub by Lh. We denote the number of UE antennas by Lu. We index the UE antennas by u=1, . . . , Lu. We denote the number frequency bins by Nf. We denote the number of feedback times by npast. We denote the number of prediction times by nfuture.
For each UE antenna, the channel estimates for all the frequencies, hub antennas, and feedback times can be combined to form a single NfLhnpast dimensional vector. We denote this by:
Ĥpast(u)∈N
Likewise, the channel values for all the frequencies, hub antennas, and prediction times can be combined to form a single NfLhnfuture dimensional vector. We denote this by:
Hfuture(u)∈N
In typical implementations, these are extremely high dimensional vectors and that in practice some form of compression should be used. For example, principal component compression may be one compression technique used.
Empirical Second Order Statistics
Empirical second order statistics are computed separately for each UE antenna in the following way:
At fixed times, the hub receives through feedback N samples of Ĥpast(u) and estimates of Hfuture(u). We denote them by: Ĥpast(u)i and Ĥfuture(u)i for i=1, . . . , N.
The hub computes an estimate of the covariance of Ĥpast(u), which we denote by {circumflex over (R)}past(u):
The hub computes an estimate of the covariance of Hfuture(u), which we denote by {circumflex over (R)}future(u):
The hub computes an estimate of the correlation between Hfuture(u) and Ĥpast(u), which we denote by {circumflex over (R)}past,future(u):
In typical wireless scenarios (pedestrian to highway speeds) the second order statistics of the channel change slowly (on the order of 1-10 seconds). Therefore, they should be recomputed relatively infrequently. Also, in some instances it may be more efficient for the UEs to compute estimates of the second order statistics and feed these back to the hub.
MMSE Prediction Filter
Using standard estimation theory, the second order statistics can be used to compute the MMSE prediction filter for each UE antenna:
C(u)={circumflex over (R)}future,past(u){circumflex over (R)}past−1(u), (24)
Where C(u) denotes the MMSE prediction filter. The hub can now predict the channel by applying feedback channel estimates into the MMSE filter:
Ĥfuture(u)=C(u)Ĥpast(u). (25)
Prediction Error Variance
We denote the MMSE prediction error by ΔHfuture(u), then:
Hfuture(u)=Ĥfuture(u)+ΔHfuture(u). (26)
We denote the covariance of the MMSE prediction error by Rerror(u), with:
Rerror(u)=[ΔHfuture(u)ΔHfuture(u)*]. (27)
Using standard estimation theory, the empirical second order statistics can be used to compute an estimate of Rerror(u):
{circumflex over (R)}error(u)=C(u){circumflex over (R)}past(u)C(u)*−C(u){circumflex over (R)}future,past(u)*−{circumflex over (R)}future,past(u)C(u)*+Rfuture(u) (28)
Simulation Results
We now present simulation results illustrating the use of the MMSE filter for channel prediction. Table 1 gives the simulation parameters and
Fifty samples of Ĥpast and Ĥfuture were used to compute empirical estimates of the second order statistics. The second order statistics were used to compute the MMSE prediction filter.
Block Diagrams
In some embodiments, the prediction is performed independently for each UE antenna. The prediction can be separated into two steps:
1) Computation of the MMSE prediction filter and prediction error covariance: the computation can be performed infrequently (on the order of seconds). The computation is summarized in
2) Channel prediction: is performed every time pre-coding is performed. The procedure is summarized in
Optimal Precoding Filter
Using MMSE prediction, the hub computes an estimate of the downlink channel matrix for the allocation of time and frequency the pre-coded data will be transmitted. The estimates are then used to construct precoding filters. Precoding is performed by applying the filters on the data the hub wants the UEs to receive. Embodiments may derive the “optimal” precoding filters as follows. Before going over details we introduce notation.
Frame (as defined previously): precoding is performed on a fixed allocation of time and frequency, with Nf frequency bins and Nt time bins. We index the frequency bins by: f=1, . . . , Nf. We index the time bins by t=1, . . . , Nt.
Channel estimate: for each time and frequency bin the hub has an estimate of the downlink channel which we denote by Ĥ(f, t)∈L
Error correlation: we denote the error of the channel estimates by ΔH(f, t), then:
H(f,t)=Ĥ(f,t)+ΔH(f,t), (29)
We denote the expected matrix correlation of the estimation error by RΔH (f, t)∈L
RΔH(f,t)=[ΔH(f,t)*ΔH(f,t)]. (30)
The hub can be easily compute these using the prediction error covariance matrices computed previously: {circumflex over (R)}error(u) for u=1, . . . , Lu.
Signal: for each time and frequency bin the UE wants to transmit a signal to the UEs which we denote by s(f, t)∈L
Precoding filter: for each time and frequency bin the hub uses the channel estimate to construct a precoding filter which we denote by W(f, t)∈L
White noise: for each time and frequency bin the UEs experience white noise which we denote by n(f, t)∈L
Hub Energy Constraint
When the precoder filter is applied to data, the hub power constraint may be considered. We assume that the total hub transmit energy cannot exceed NfNtLh. Consider the pre-coded data:
W(f,t)s(f,t), (31)
To ensure that the pre-coded data meets the hub energy constraints the hub applies normalization, transmitting:
λW(f,t)s(f,t), (32)
Where the normalization constant A is given by:
Receiver SINR
The pre-coded data then passes through the downlink channel, the UEs receive the following signal:
λH(f,t)W(f,t)s(f,t)+n(f,t), (34)
The UEs then removes the normalization constant, giving a soft estimate of the signal:
The error of the estimate is given by:
The error can be decomposed into two independent terms: interference and noise. Embodiments can compute the total expected error energy:
Optimal Precoding Filter
We note that the expected error energy is convex and quadratic with respect to the coefficients of the precoding filter. Therefore, calculus can be used to derive the optimal precoding filter:
Accordingly, some embodiments of an OTFS precoding system use this filter (or an estimate thereof) for precoding.
Simulation Results
We now present a simulation result illustrating the use of the optimal precoding filter. The simulation scenario was a hub transmitting data to a single UE. The channel was non line of sight, with two reflector clusters: one cluster consisted of static reflectors, the other cluster consisted of moving reflectors.
The simulation results depicted in
Example Block Diagrams
Precoding is performed independently for each time frequency bin. The precoding can be separated into three steps:
[1] Computation of error correlation: the computation be performed infrequently (on the order of seconds). The computation is summarized in
[2] Computation of optimal precoding filter: may be performed every time pre-coding is performed. The computation is summarized in
[3] Application of the optimal precoding filter: may be performed every time pre-coding is performed. The procedure is summarized in
OTFS Vector Perturbation
Before introducing the concept of vector perturbation, we outline the application of the optimal pre-coding filter to OTFS.
OTFS Optimal Precoding
In OTFS, the data to be transmitted to the UEs are encoded using QAMs in the delay-Doppler domain. We denote this QAM signal by x, then:
x(τ,ν)∈AL
for τ=1, . . . , Nτ and ν=1, . . . , Nν. A denotes the QAM constellation. Using the two-dimensional Fourier transform the signal can be represented in the time frequency domain. We denote this representation by X:
X(f,t)=(Fx)(f,t), (40)
for f=1, . . . , Nf and t=1, . . . , Nt. F denotes the two-dimensional Fourier transform. The hub applies the optimal pre-coding filter to X and transmit the filter output over the air:
λWopt(f,t)X(f,t) (41)
for f=1, . . . , Nf and t=1, . . . , Nt. λ denotes the normalization constant. The UEs remove the normalization constant giving a soft estimate of X:
for f=1, . . . , Nf and t=1, . . . , Nt. The term w(f, t) denotes white noise. We denote the error of the soft estimate by E:
E(f,t)=Xsoft(f,t)−X(f,t), (43)
for f=1, . . . , Nf and t=1, . . . , Nt. The expected error energy was derived earlier in this document:
Where:
We call the positive definite matrix Merror(f, t) the error metric.
Vector Perturbation
In vector perturbation, the hub transmits a perturbed version of the QAM signal:
x(τ,ν)+p(τ,ν), (46)
for τ=1, . . . , Nτ and ν=1, . . . , Nν. Here, p(τ,ν) denotes the perturbation signal. The perturbed QAMs can be represented in the time frequency domain:
X(f,t)+P(f,t)=(Fx)(f,t)+(Fp)(f,t), (47)
for f=1, . . . , Nf and t=1, . . . , Nt. The hub applies the optimal pre-coding filter to the perturbed signal and transmits the result over the air. The UEs remove the normalization constant giving a soft estimate of the perturbed signal:
X(f,t)+P(f,t)+E(f,t), (48)
for f=1, . . . , Nf and t=1, . . . , Nt. Where E denotes the error of the soft estimate. The expected energy of the error is given by:
expected error energy=Σf=1N
The UEs then apply an inverse two dimensional Fourier transform to convert the soft estimate to the delay Doppler domain:
x(τ,ν)+p(τ,ν)+e(τ,ν), (50)
for τ=1, . . . , Nτ and ν=1, . . . , Nν. The UEs then remove the perturbation p(τ, ν) for each delay Doppler bin to recover the QAM signal x.
Collection of Vector Perturbation Signals
One question is: what collection of perturbation signals should be allowed? When making this decision, there are two conflicting criteria:
1) The collection of perturbation signals should be large so that the expected error energy can be greatly reduced.
2) The collection of perturbation signals should be small so the UE can easily remove them (reduced computational complexity):
x(τ,ν)+p(τ,ν)→x(τ,ν) (51)
Coarse Lattice Perturbation
An effective family of perturbation signals in the delay-Doppler domain, which take values in a coarse lattice:
p(τ,ν)∈BL
for τ=1, . . . , Nτ and ν=1, . . . , Nν. Here, B denotes the coarse lattice. Specifically, if the QAM symbols lie in the box: [−r, r]×j[−r, r] we take as our perturbation lattice B=2r+2rj. We now illustrate coarse lattice perturbation with an example.
Consider QPSK (or 4-QAM) symbols in the box [−2,2]×j[−2,2]. The perturbation lattice is then B=4+4j.
The UE receives the perturbed QPSK symbol. The UE then removes the perturbation to recover the QPSK symbol. To do this, the UE first searches for the coarse lattice point closest to the received signal.
The UE subtracts the closest lattice point from the received signal, thus recovering the QPSK symbol 1+1j.
Finding Optimal Coarse Lattice Perturbation Signal
The optimal coarse lattice perturbation signal, popt, is the one which minimizes the expected error energy:
popt=argminpΣf=1N
The optimal coarse lattice perturbation signal can be computed using different methods. A computationally efficient method is a form of Thomlinson-Harashima precoding which involves applying a DFE filter at the hub.
Coarse Lattice Perturbation Example
We now present a simulation result illustrating the use of coarse lattice perturbation. The simulation scenario was a hub antenna transmitting to a single UE antenna. Table 2 displays the modulation parameters. Table 3 display the channel parameters for this example.
Because this is a SISO (single input single output) channel, the error metric Merror(f, t) is a positive scaler for each time frequency bin. The expected error energy is given by integrating the product of the error metric with the perturbed signal energy:
expected error energy=Σf=1N
The simulation illustrates the gain from using vector perturbation: shaping the energy of the signal to avoid time frequency regions where the error metric is high.
Block Diagrams
Vector perturbations may be performed in three steps. First, the hub perturbs the QAM signal. Next, the perturbed signal is transmitted over the air using the pre-coding filters. Finally, the UEs remove the perturbation to recover the data.
Computation of error metric: the computation can be performed independently for each time frequency bin. The computation is summarized in
Computation of perturbation: the perturbation is performed on the entire delay Doppler signal. The computation is summarized in
Application of the optimal precoding filter: the computation can be performed independently for each time frequency bin. The computation is summarized in
UEs removes perturbation: the computation can be
Spatial Tomlinson Harashima Precoding
This section provides additional details of achieving spatial precoding and the beneficial aspects of using Tomlinson Harashima precoding algorithm in implementing spatial precoding in the delay Doppler domain. The embodiments consider a flat channel (with no frequency or time selectivity).
Review of Linear Precoding
In precoding, the hub wants to transmit a vector of QAMs to the UEs. We denote this vector by x∈L
An estimate of the downlink channel, denoted by: Ĥ∈L
The matrix covariance of the channel estimation error, denoted by: RΔH∈L
From this information, the hub computes the “optimal” precoding filter, which minimizes the expected error energy experienced by the UEs:
By applying the precoding filter to the QAM vector the hub constructs a signal to transmit over the air: λWoptx∈L
λHWoptx+w,
Where w∈L
x+e,
where e∈L
expected error energy=x*Merrorx
where Merror is a positive definite matrix computed by:
Review of Vector Perturbation
The expected error energy can be greatly reduced by perturbing the QAM signal by a vector v∈L
x+v+e
Again, the expected error energy can be computed using the error metric:
expected error energy=(x+v)*Merror(x+v)
The optimal perturbation vector minimizes the expected error energy:
vopt=argminv(x+v)*Merror(x+v).
Computing the optimal perturbation vector is in general NP-hard, therefore, in practice an approximation of the optimal perturbation is computed instead. For the remainder of the document we assume the following signal and perturbation structure:
The QAMs lie in the box [−1,1]×j[−1,1].
The perturbation vectors lie on the coarse lattice: (2+2j)L
Spatial Tomlinson Harashima Precoding
In spatial THP a filter is used to compute a “good” perturbation vector. To this end, we make use of the Cholesky decomposition of the positive definite matrix Merror:
Merror=U*DU,
where D is a diagonal matrix with positive entries and U is unit upper triangular. Using this decomposition, the expected error energy can be expressed as:
expected error energy=(U(x+v))*D(U(x+v))=z*Dz=Σn=1L
where z=U(x+v). We note that minimizing the expected error energy is equivalent to minimizing the energy of the z entries, where:
z(Lu)=x(Lu)+v(Lu),
z(n)=x(n)+v(n)+Σm=n+1L
for n=1, 2, . . . , Lu−1. Spatial THP iteratively choses a perturbation vector in the following way.
v(Lu)=0
Suppose v(n+1), v(n+2), . . . , v(Lu) have been chosen, then:
v(n)=−(x(n)+Σm=n+1L
where denotes projection onto the coarse lattice. We note that by construction the coarse perturbation vector bounds the energy of the entries of z by two.
Simulation Results
We now present the results of a simple simulation to illustrate the use of spatial THP. Table 4 summarizes the simulation setup.
The error energy is low when the signal transmitted to UE1 and UE2 are similar. Conversely, the error energy is high when the signals transmitted to the UEs are dissimilar. We can expect this pattern to appear when two UEs are spatially close together; in these situations, it is advantageous to transmit the same message to both UEs.
The error energy has the shape of an ellipses. The axes of the ellipse are defined by the eigenvectors of Merror.
A large number data of PAM vectors was generated and spatial THP was applied.
This section overviews channel estimation for OTFS systems, and in particular, aspects of channel estimation and scheduling for a massive number of users. A wireless system, with a multi-antenna base-station and multiple user antennas, is shown in
In some embodiments, and when the channels are not static and when the number of users is very large, some of the challenges of such a precoded system include:
Typical solutions in systems, which assume a low number of users and static channels, are to let each user transmit known pilot symbols (reference signals) from each one of its antennas. These pilots are received by all the base-station antennas and used to estimate the channel. It is important that these pilot symbols do not experience significant interference, so that the channel estimation quality is high. For this reason, they are typically sent in an orthogonal way to other transmissions at the same time. There are different methods for packing multiple pilots in an orthogonal (or nearly-orthogonal) way, but these methods are usually limited by the number of pilots that can be packed together (depending on the channel conditions) without causing significant interference to each other. Therefore, it becomes very difficult to have an efficient system, when the number of user antennas is high and the channels are not static. The amount of transmission resources that is needed for uplink pilots may take a considerable amount of the system's capacity or even make it unimplementable. For prediction of the channel, it is typically assumed that the channel is completely static and will not change from the time it was estimated till the end of the downlink transmission. This assumption usually causes significant degradation in non-static channels.
It is assumed that the downlink and uplink channels are reciprocal and after calibration it is possible to compensate for the difference in the uplink-downlink and downlink-uplink channel responses. Some example embodiments of the calibration process using reciprocity are further discussed in Section 5.
Embodiments of the disclosed technology include a system and a method for packing and separating multiple non-orthogonal pilots, as well as a method for channel prediction. In such a system, it is possible to pack together a considerably higher number of pilots comparing to other commonly used methods, thus allowing an accurate prediction of the channel for precoding.
Second-Order Training Statistics
The system consists of a preliminary training step, in which all users send uplink orthogonal pilots to the base-station. Although these pilots are orthogonal, they may be sent at a very low rate (such as one every second) and therefore do not overload the system too much. The base-station receives a multiple of NSOS such transmissions of these pilots, and use them to compute the second-order statistics (covariance) of each channel.
The computation of the second-order statistics for a user antenna u is defined as:
To accommodate for possible future changes in the channel response, the second-order statistics may be updated later, after the training step is completed. It may be recomputed from scratch by sending again NSOS orthogonal pilots, or gradually updated. One possible method may be to remove the first column of H(u) and attach a new column at the end and then re-compute the covariance matrix again.
The interval at which these orthogonal pilots need to be repeated depends on the stationarity time of the channel, e.g., the time during which the second-order statistics stay approximately constant. This time can be chosen either to be a system-determined constant, or can be adapted to the environment. In particular, users can determine through observation of downlink broadcast pilot symbols changes in the second-order statistics, and request resources for transmission of the uplink pilots when a significant change has been observed. In another embodiment, the base-station may use the frequency of retransmission requests from the users to detect changes in the channel, and restart the process of computing the second-order statistics of the channel.
To reduce the computational load, it is possible to use principal component analysis (PCA) techniques on RHH(u). We compute {λ(u)}, the K(u) most dominant eigenvalues of RHH(u), arranged in a diagonal matrix D(u)=diag(λ1(u), λ2(u), . . . , λK
Non-Orthogonal Pilots
The non-orthogonal pilots (NOP), P(u), for user antenna u, may be defined as a pseudo-random sequence of known symbols and of size NNOP, over a set of frequency grid elements. The base-station can schedule many users to transmit their non-orthogonal pilots at the same subframe using overlapping time and frequency resources. The base-station will be able to separate these pilots and obtain a high-quality channel estimation for all the users, using the method describes below.
Define the vector Y of size (L·NNOP)×1, as the base-station received signal over all its antennas, at the frequency grid elements of the shared non-orthogonal pilots. Let {tilde over (V)}(u) be the eigenvectors matrix V(u) decimated along its first dimension (frequency-space) to the locations of the non-orthogonal pilots.
The base-station may apply a Minimum-Mean-Square-Error (MMSE) estimator to separate the pilots of every user antenna:
Herein, ⊙ is defined as the element-by-element multiplication. For a matrix A and vector B, the A⊙B operation includes replicating the vector B to match the size of the matrix A before applying the element-by-element multiplication.
If principal component analysis (PCA) is not used, the covariance matrices can be computed directly as:
RYY(u)=(P(u)[P(u)]H)⊙RHH(u)
RXY(u)=(1[P(u)]H)⊙RHH(u)
and invert it. Note that it is possible to apply PCA here as well by finding the dominant eigenvalues of RYY(DR
Note that HNOP(u) is the channel response over the frequency grid-elements of the non-orthogonal pilots for the L base-station received antennas. It may be also interpolated along frequency to obtain the channel response over the entire bandwidth.
Prediction Training
The method described in the previous section for separating non-orthogonal pilots is applied to train different users for prediction. In this step, a user sends uplink non-orthogonal pilots on consecutive subframes, which are divided to 3 different sections, as shown in the example in
1. Past—the first Npast subframes. These subframes will later be used to predict future subframes.
2. Latency—the following Nlatency subframes are used for the latency required for prediction and precoding computations.
3. Future—the last Nfuture subframes (typically one), where the channel at the downlink portion will be later predicted.
Each user, is scheduled NPR times to send uplink non-orthogonal pilots on consecutive Npast+Nlatency+Nfuture subframes. Note that in one uplink symbol in the subframe, both orthogonal and non-orthogonal pilots may be packed together (although the number of orthogonal pilots will be significantly lower than the number of non-orthogonal pilots). The base-station applies the pilot separation filter for the non-orthogonal pilots of each user and computes HNOP(u). To reduce storage and computation, the channel response may be compressed using the eigenvector matrix computed in the second-order statistics step
HK(u)=({tilde over (V)}(u))H·HNOP(u)
For subframes, which are part of the “Past” section, store HK(u) as columns in the matrix Hpast,(i)(u), where i=1, 2, . . . , NPR. Use all or part of the non-orthogonal pilots to interpolate the channel over the whole or part of the downlink portion of the “Future” subframes, compress it using {tilde over (V)}(u) and store it as Hfuture,(i)(u). Compute the following covariance matrices:
Rpast,(i)(u)=Hpast,(i)(u)·(Hpast,(i)(u))H
Rfuture,(i)(u)=Hfuture,(i)(u)·(Hfuture,(i)(u))H
Rfuture_past,(i)(u)=Hfuture,(i)(u)·(Hpast,(i)(u))H
After all NPR groups of prediction training subframes have been scheduled, compute the average covariance matrices for each user
Finally, for each user compute the MMSE prediction filter
CPR(u)=Rfuture_past(u)·(Rpast(u))−1
and its error variance for the precoder
RE(u)=Rfuture(u)−CPR(u)·(Rfuture_past(u))H.
Scheduling a Downlink Precoded Transmission
For each subframe with a precoded downlink transmission, the base-station should schedule all the users of that transmission to send uplink non-orthogonal pilots for Npast consecutive subframes, starting Npast+Nlatency subframes before it, as shown in FIG. 83. The base-station will separate the non-orthogonal pilots of each user, compress it and store the channel response as HK,past(u). Then, it will apply the prediction filter to get the compressed channel response for the future part
HK,future(u)=CPR(u)·HK,past(u)
Finally, the uncompressed channel response is computed as
Hfuture(u)={tilde over (V)}(u)·HK,future(u)
The base-station may correct for differences in the reciprocal channel by applying a phase and amplitude correction, α(f), for each frequency grid-element
Hfuture_reciprocity(u)(f)=α(f)·Hfuture(u)(f)
Then, use Hfuture_reciprocity(u) and RE(u) of the participating users to compute the precoder for the downlink transmission.
Scheduling of the Uplink Pilots
If during a frame there are multiple orthogonal resources available for pilot transmission (e.g., different timeslots or different frequency grid elements), then the set of uplink pilots that needs to be transmitted can be divided into sets such that each set is transmitted on a different resource. The criterion of for the division into sets can be, e.g., the achievable pilot SINR. The transmission of non-orthogonal pilots leads to a reduction in the achievable pilot SINR, which is the more pronounced the stronger the alignment of the vector spaces containing the correlation matrices from different users is. Thus, arranging users in sets such that two pilots with very similar correlation matrices are not transmitted at the same time improves performance. However, other criteria are possible as well. For example, for users that have only a low SINR during data transmission, achieving a high pilot SINR might be wasteful; thus, achieving an optimal “matching” of the pilot SINR to the data SINR might be another possible criterion.
The embodiments of the disclosed technology described in this section may be characterized, but not limited, by the following features:
This section covers scheduling pilots to reduce transmission overhead and improve the throughput of a wireless communication system. One possible FWA system design is based on separating users based on their angular power spectra. For example, users can operate in parallel if they do not create “significant” interference in each other's “beams.” A beam may for example be a Luneburg beam. A precoding vector can also be associated with a beam pattern. However, for ease of explanation, the word “precoder pattern” is used in the present description. Consider as an example a system with 8 beams in a 90-degree sector, such that any two adjacent beams have overlapping beam patterns, while beams whose difference of indices is at least 2 are orthogonal to each other. If there is a pure line of sight (LoS), or a small angular spread around the LoS direction, then a spatial reuse factor of 2 may be possible. For example, beams 1, 3, 5, and 7 can operate in parallel (and similarly beam 2, 4, 6, 8). However, most channels provide a larger angular spread than can be handled by such a configuration, so that only beams with a wider angular separation may use the same time/frequency resources; e.g., a reuse factor on the order of 4 may be achieved. This means that only 2 users can operate on the same time-frequency resources within one sector, so that the overall performance gain compared to traditional systems is somewhat limited.
Considerably better spatial reuse can be achieved when the user separation is based on instantaneous channel state information, using joint receive processing of the multiple beam signals, and joint precoding, for the uplink and downlink, respectively. To take the example of the uplink, with N antenna (beam) ports, N signals can be separated, so that N users can be active at the same time (and analogously for the downlink). The simplest way to achieve this is zero-forcing, though it may suffer from poor performance in particular if users are close together (in mathematical terms, this occurs if their channel vectors are nearly linearly dependent). More sophisticated techniques, such as turbo equalization in the uplink, and Tomlinson-Harashima Precoding (THP) in the downlink can improve the performance further. Such implementations can increase signal to interference plus noise ratio (SINR) for the users, though they may not increase the degrees of freedom.
However, while these methods have great advantages, they rely on the knowledge of the instantaneous channel state information (CSI) for the processing, while the beam-based transmission can be performed simply by the time-averaged (for FWA) or second order (for mobile) systems CSI. The problem is aggravated by two facts:
1) while N users can be served in parallel (since they are separated by their different instantaneous CSI), the pilots cannot be separated this way (because the CSI is not yet known when the pilots are transmitted—it is a “chicken and egg” problem). Thus, pilots can be separated based on their average or second-order statistics.
2) OTFS modulation may have a higher pilot overhead compared to, e.g., OFDMA, because of the spreading of the information over the whole time-frequency plane, such that each user attempts to determine the CSI for the whole bandwidth.
Example System Model and Basic Analysis
A. Assumptions for the Analysis
An example system is described and for ease of explanation, the following assumptions are made:
1) Luneburg lens with 8 beams. Adjacent beams have overlap, beams separated by at least 1 other beam have a pattern overlap separation of better than 30 dB. However, in general, any number of beams may be used.
2) For the uplink, no use of continuous pilots. Channels might be estimated either based on the pilots embedded in the data packets. Alternatively, placing a packet in a queue for, say 4 ms, to allow transmission of uplink pilots before the transmission of data can improve channel estimation performance.
3) For the downlink, every UE observes broadcast pilots, which, in this example, are sent periodically or continuously, and extrapolates the channel for the next downlink frame. It then might send this information, in quantized form, to the BS (for the case that explicit channel state feedback is used).
4) The discussion here only considers the basic degrees of freedom for the pilot tones, not the details of overhead associated with delay-Doppler versus time-frequency multiplexing. In some implementations, both may give approximately the same overhead.
5) A frame structure with 1 ms frame duration is used. Different users may transmit in different frames. It is assumed that in the uplink and for the precoded pilots of the downlink, two pilots are transmitted per user, one at the beginning of the frame, and one at the end of the frame, so that interpolation can be done. For the broadcast pilots in the downlink, this may not be done, since it will be transmitted once per frame anyway, so that interpolation and extrapolation is implicitly possible.
6) A system bandwidth of 10 MHz is assumed.
B. Efficiency of an Example System
The following presents a first example calculation of the pilot overhead when the pilots in all beams are kept completely orthogonal. For the example, first compute the degrees of freedom for the pilot for each user. With 10 MHz bandwidth and 1 ms frame duration, and two polarizations, there are in general 10,000 “resolvable bins” (degrees of freedom) that can be used for either data transmission or pilot tone transmission. The propagation channel has 200 degrees of freedom (resolvable delay bin 100 ns and 5 microseconds maximum excess delay means 50 delay coefficients characterize the channel, plus two resolvable Doppler bins within each channel, on each of two polarizations). Thus, the pilot tones for each user constitute an overhead of 2% of the total transmission resources. Due to the principle of OTFS of spreading over the whole system bandwidth and frame duration, the pilot tone overhead does not depend on the percentage of resources assigned to each user, but is a percentage of taken over all resources. This implies a high overhead when many users with small number of bytes per packet are active.
If completely orthogonalizing the users in the spatial and polarization domains, then the pilot overhead gets multiplied with the number of beams and polarizations. In other words, reserve a separate delay-Doppler (or time-frequency) resource for the pilot of each beam, which ensures that there is no pilot contamination. The broadcast pilots in the downlink need therefore 16% of the total resources (assuming communication in a sector) or 64% (for a full circular cell). The following examples will mostly concentrate on a single sector.
Similarly, for the uplink pilots, orthogonal pilots may be used for each of the users, in each of the beams. This results in a 16% overhead per user; with multiple users, this quickly becomes unsustainable.
The overhead for digitized feedback from the users can also be considerable. Since there are 200 channel degrees of freedom, quantization with 24 bit (12 bits each on I and Q branch) results in 4.8 Mbit/s for each user. Equivalently, if assuming on average 16 QAM (4 bit/s/Hz spectral efficiency), 1200 channel degrees of freedom are used up for the feedback of quantized information from a single user. This implies that the feedback of the digitized information is by a factor 6 less spectrally efficient than the transmission of an analog uplink pilot whose information can be used. Furthermore, the feedback is sent for the channel state information (CSI) from each BS antenna element to the customer premises equipment (CPE) or user device. Even though the feedback can be sent in a form that allows joint detection, in other words, the feedback info from users in different beams can be sent simultaneously, the overall effort for such feedback seems prohibitively large.
In addition, it is useful to consider the overhead of the embedded pilots for the downlink, where they are transmitted precoded in the same way as the data, and thus are used for the demodulation. By the nature of zero-forcing precoding, pilots can be transmitted on each beam separately. Thus, the overhead for the embedded downlink pilots is about 2% of the resources times the average number of users per beam.
For explicit feedback, there is yet another factor to consider, namely the overhead for the uplink pilots that accompany the transmission of the feedback data. This tends to be the dominant factor. Overhead reduction methods are discussed in the next section.
Overhead Reduction Methods
From the above description, it can be seen that overhead reduction is useful. The main bottlenecks indeed are the downlink broadcast pilots and the uplink pilots, since these pilots have to be sent on different time-frequency (or delay/Doppler) resources in different beams. However, under some circumstances, overhead reduction for the feedback packets is important as well. Before going into details, it is worth repeating why transmitters cannot transmit pilots on all beams all the time. Neither the UL pilots nor the broadcast DL pilots are precoded. To separate the pilots from/to different users, transmitters would have to beamform, but in order to beamform, a transmitter should know the channel, e.g., have decided pilots. Thus, a continuous transmission of pilots leads to “pilot contamination”, e.g., the signals from/to users employing the same pilots interfere with each other and lead to a reduced pilot SINR. Since the pilot quality determines the capability of coherently decoding the received data signal, reduction of the pilot SINR is—to a first approximation—as detrimental as reduction of the data SINR. While countermeasures such as joint equalization and decoding are possible, they greatly increase complexity and still result in a performance loss.
One effective method of reducing pilot contamination is minimum mean square error (MMSE) filtering, which achieves separation of users with the same pilot tones by projection of the desired users' pilot onto the null-space of the channel correlation matrix of the interfering user. This reduces interference, though at the price of reduced signal power of the desired user. This method can be combined with any and all of the methods described below, and, in some situations, such a combined method will achieve the best performance. In some embodiments, linearly dependent pilot tones for the different users (instead of sets of users that use the same pilots within such a set, while the pilots in different sets are orthogonal to each other) may be used. Again, such a whitening approach can be used in conjunction with the methods described here.
A. Pilot Scheduling
The previous derivations assumed that the downlink broadcast and uplink pilots in different beams are on orthogonal resources, in order to reduce the overhead. Such an arrangement may not be needed when the angular spectra of the users are sufficiently separated. The simplest assumption is that each user has only a very small angular spread; then users that are on beams without overlaps (beam 1,3,5, . . . etc.) can be transmitted simultaneously. For a larger angular spread, a larger spacing between the beams is used. Still, if, e.g., every 4th beam can be used, then the overall overhead for the downlink broadcast pilots reduces, e.g., from 32% to 16% in one sector. Equally importantly, the overhead remains at 16% when moving from a sector to a 360 degree cell.
However, this consideration still assumes that there is a compact support of the angular power spectrum, and there is no “crosstalk”, e.g., between a beam at 0 degree and one at 60 degree. Often, this is not the case. In the presence of scattering objects, the sets of directions of contributions from/to different user devices can be quite different, and not simply a translation (in angle domain) of each other. If simply basing the beam reuse on the “worst case”, one might end up with complete orthogonalization. Thus, for every deployment, it is useful to assess individually what the best pattern is for a spatial reuse of the pilots. This is henceforth called “pilot scheduling”.
Before describing some examples of pilot scheduling embodiments, note that it is based on the knowledge of the power transfer matrix (PTM). The PTM may be a K×M matrix, where M is the number of beams at the BS, and K is the number of UEs. The (I,j)th entry of the PMT is then the amount of power (averaged over small-scale fading or time) arriving at the j-th beam when the i-th UE transmits with unit power (one simplification we assume in this exemplary description is that the PTM is identical for the two polarizations, which is reasonable, e.g., when there is sufficient frequency selectivity such that OTFS averages out small-scale fading over its transmission bandwidth; in any case generalization to having different PMT entries for different polarization ports is straightforward). For example, in the uplink, the receiver (base station) should know when a particular user transmits a pilot tone, in which beams to anticipate appreciable energy. This might again seem like a “chicken and egg” problem, since the aim of the pilot transmission is to learn about the channel between the user and the BS. However, the PTM is based on the knowledge of the average or second order channel state information (CSI). Since this property of a channel changes very slowly (on the order of seconds for mobile systems, on the order of minutes or more for FWA), learning the PTM does not require a significant percentage of time-frequency resources. While provisions should be taken in the protocol for suitable mechanisms, those pose no fundamental difficulty, and the remainder of the report simply assumes that PTM is known.
1) Pilot scheduling for the uplink: as mentioned above, the PTM contains information about the amount of power that is transferred from the ith user to the jth beam. Now, given the PTM, the question is: when can two uplink pilots be transmitted on the same time-frequency resources?
The answer may depend on the subsequent data transmission, for example, if the criterion is: “is the loss of capacity resulting from the imperfect beamforming vectors is less than the spectral efficiency gain of the reduced pilot overhead”. Conventional techniques do not consider such a criterion. This aspect of inquiry can be used in many advantageous ways:
a) It is not necessary to have highly accurate (contamination-free) pilots if the subsequent data transmission uses a low-order QAM anyways.
b) The pilot scheduling depends on the receiver type. First, different receivers allow different modulation schemes (even for the same SINR). Second, a receiver with iterative channel estimation and data decoding might be able to deal with more pilot contamination, since it processes the decoded data (protected by forward error correction FEC) to improve the channel estimates and reduce contamination effects.
c) The pilot scheduling, and the pilot reuse, may change whenever the transmitting users change. A fixed scheduling, such as beams 1, 5, 9, . . . etc. may be highly suboptimum.
d) Given the high overhead for uplink pilots, allowing considerable pilot contamination, and use of associated low SINR and modulation coding scheme (MCS), is reasonable, in particular for small data packets.
e) For an FWA system, it may be reasonable to allow uplink transmission without embedded pilots, basing the demodulation solely on the average channel state. However, due to the clock drift, a few pilots for phase/timing synchronizations may still be used, but no pilots may be used for channel re-estimation. For those short packets, a reduced-order MCS may be used. Alternatively, the short packets could be transmitted on a subband of the time-frequency resources, where the subband could even be selected to provide opportunistic scheduling gain.
The optimum scheduler may be highly complicated, and may change whenever the combination of considered user devices changes. Due to the huge number of possible user combinations in the different beams, it may not even possible to compute the optimum scheduler for each combination in advance and then reuse. Thus, a simplified (and suboptimum) scheduler may have to be designed.
2) Pilot scheduling for the downlink: The scheduler for the downlink broadcast pilots has some similarities to the uplink pilots, in that it is based on the PTM. However, one difference is worth noting: the scheduler has to provide acceptable levels of pilot contamination for all users in the system, since all users are monitoring the broadcast pilots and extrapolate the channel in order to be able to feed back the extrapolated channel when the need arises. Thus, the reuse factor of the broadcast pilots may be large (meaning there is less reuse) than for the uplink pilots. For the computation of the pilot schedule, a few things may be taken into account:
a) the schedule may only be changed when the active user devices change, e.g., a modem goes to sleep or wakes up. This happens on a much rarer basis than the schedule change in the uplink, which happens whenever the actually transmitting user devices change.
b) In the downlink pilots, it may not be exactly known what pilot quality will be required at what time (e.g., the required SINR), since the transmitting user schedule is not yet known (e.g., when the pilots are transmitted continuously). Thus, it may be assumed that data transmission could occur without interference (e.g., all other beams are silent because there are no data to transmit), so that the data transmission for the user under consideration happens with the MCS that is supported by the SNR.
c) It is possible that one (or a few) user devices become a “bottleneck”, in the sense that they require a large reuse factor when all other users might allow dense reuse. It is thus useful to consider the tradeoff of reducing the pilot quality to these bottleneck user devices and reducing the MCS for the data transmission, as this might lead to an increase of sum spectral efficiency, and may be performed by taking minimum (committed) service quality constraints into account.
Since broadcast pilots are always transmitted from the BS, and can be only either transmitted or not transmitted (there is no power control for them), the number of possible combinations is manageable (2{circumflex over ( )}8), and it is thus possible to compute the SINR at all users in the cell for all pilot schedules, and check whether they result in acceptable SNR at all users, and pick the best one. As outlined above, there is no need to recompute the schedule, except when the set of active user devices changes. When considering a combination of this scheme with MMSE receivers, scheduling should be based on the SINR that occurs after the MMSE filtering.
B. Exploiting the Properties of FWA
One way for reducing the overhead is to exploit the special properties of FWA channels, namely that the instantaneous channel is the average channel plus a small perturbation. This can be exploited both for reducing the reuse factor, and for more efficient quantization.
1) Reducing the reuse factor: The goal of the pilot tones is to determine the CSI for each user device with a certain accuracy. Let us consider the uplink: for the i-th user in the j-th beam, the CSI can be written as Havij+ΔHij; the power ratio (ΔHij/Havij)2 is the temporal Rice factor for this particular link Kij. Now any pilot contamination based on Havij is known and can be eliminated by interference cancellation. Thus, denoting the kj-th entry of the PTM Ckj, then a naïve assessment of the pilot contamination would say that the achievable pilot SIR in the j-th beam is Cij/Ckj. However, by first subtracting the known contribution Havkj from the overall received signal, KkjCij/Ckj can be achieved. Having thus improved the SIR for each user, the system can employ a much smaller reuse factor (that is, reduce overhead). In practice this method can probably reduce the reuse factor by about a factor of 2. The same approach can also be applied in the downlink. The improvement that can be achieved will differ from user device to user device, and the overall reuse factor improvement will be determined by the “bottleneck links” (the ones requiring the largest reuse factor). Some embodiments can sacrifice throughput on a few links if that helps to reduce the pilot reuse factor and thus overhead, as described above. When combining this method with MMSE filtering, the procedure may occur in two steps: first, the time-invariant part of the channel is subtracted. The time-variant part is estimated with the help of the MMSE filtering (employing the channel correlation matrix of the time-variant part), and then the total channel is obtained as the sum of the time-invariant and the thus-estimated time-variant channel.
2) Improved quantization: Another question is the level of quantization that is to be used for the case that explicit feedback is used. Generally, the rule is that quantization noise is 6 dB for every bit of resolution. The 12 bit resolution assumed above for the feedback of the CSI thus amply covers the desired signal-to-quantization-noise ratio and dynamic range. However, in a fixed wireless system, implementations do not need a large dynamic range margin (the received power level stays constant except for small variations), and any variations around the mean are small. Thus, assume a temporal Rice factor of 10 dB, and an average signal level of −60 dBm. This means that the actual fluctuations of the signal have a signal power of −70 dBm. 4-bit quantization provides −24 dB quantization noise, so that the quantization noise level is at −94 dBm, providing more than enough SIR. Embodiments can thus actually reduce the amount of feedback bits by a factor of 3 (from 12-bit as assumed above to 4 bits) without noticeable performance impact.
3) Adaptivity of the methods: The improvements described above use the decomposition of the signal into fixed and time-varying parts, and the improvements are the larger the larger the temporal Rice factor is. Measurements have shown that the temporal Rice factor varies from cell to cell, and even UE to UE, and furthermore might change over time. It is thus difficult to determine in advance the reduction of the reuse factor, or the suitable quantization. For the reduction of the reuse factor, variations of the Rice factor from cell to cell and between user devices such as UEs can be taken care of as a part of the pilot scheduling design, as described above. Changes in the temporal Rice factor (e.g., due to reduced car traffic during nighttime, or reduction of vegetation scatter due to change in wind speed) might trigger a new scheduling of pilots even when the active user set has not changed. For the quantization, the protocol should not contain a fixed number of quantization bits, but rather allow an adaptive design, e.g., by having the feedback packet denote in the preamble how many bits are used for quantization.
C. Reduction Methods for Small Packet Size
The most problematic situation occurs when a large number of users, each with a small packet, are scheduled within one frame. Such a situation is problematic no matter whether it occurs in the uplink or the downlink, as the pilot overhead in either case is significant. This problem can be combatted in two ways (as alluded to above)
1) reduce the bandwidth assigned to each user. This is a deviation from the principle of full-spreading OTFS, but well aligned with other implementations of OTFS that can assign a subband to a particular user, and furthermore to various forms of OFDMA.
The two design trade-offs of the approach are that (i) it may use a more sophisticated scheduler, which now considers frequency selectivity as well, and (ii) it is a deviation from the simple transmission structure described above, where different users are designed different timeslots and/or delay/Doppler bins. Both of these issues might be solved by a multi-subband approach (e.g., 4 equally spaced subbands), though this may trade off some performance (compared to full OTFS) and retains some significant pilot overhead, since at least CSI in the 4 chosen subbands has to be transmitted.
2) transmit the small packets without any pilots, relying on the average CSI for suppression of inter-beam interference. It is noteworthy that for the downlink, an implementation can sacrifice SIR (due to pilot contamination) on some links without disturbing others. Imagine that precise CSI for UE j is available, while it is not available for UE k. An implementation can thus ensure that the transmission for k lies in the exact null-space of j, since the CSI vector hj=[h1j; h2j; . . . ] is known accurately, and thus its nullspace can be determined accurately as well. So, if the link to j wants to send a big data packet for which the use of a high-order MCS is essential, then the system can invest more resources (e.g., reduce pilot imprecision) for this link, and reap the benefits.
3) For the uplink, the approach 2 may not work: in order to have high SINR for the signal from the j-th user, it is advantageous to suppress the interference from all other users that are transmitting in parallel. Thus, instead one approach may be to provide orthogonalization in time/frequency (or delay/Doppler) between the group of users that needs low pilot contamination (usually large packets, so that the efficiency gain from transmitting pilots outweighs the overhead), and another group of users (the ones with small packets) that do not transmit pilots (or just synchronization pilots) and thus are efficient, yet have to operate with lower-order MCS due to the pilot contamination. It must be noted that methods 2 and 3 only work for FWA systems, where one can make use of the average CSI to get a reasonable channel estimate without instantaneous pilots. When migrating to a mobile system, it is recommended to move to approach 1.
Examples for the Achievable Gain
This section describes some examples of the gain that can be achieved by the proposed methods versus a baseline system. It should be noted that the gain will be different depending on the number of users, the type of traffic, and particularly the directional channel properties. There are examples where the simple orthogonalization scheme provides optimum efficiency, so that no gain can be achieved, and other examples where the gain can be more than an order of magnitude. The section will use what can be considered “typical” examples. Ray tracing data for typical environments and traffic data from deployed FWA or similar systems, for example, can be used to identify a typical or representative system.
A. Gain of Pilot Scheduling
One purpose of pilot scheduling is to place pilots on such time-frequency resources that they either do not interfere significantly with each other, or that the capacity loss by using more spectral resources for pilots is less than the loss one would get from pilot contamination. In a strictly orthogonal system, there is no pilot contamination, but 16% of all spectral resources must be dedicated to the downlink pilots, and a fraction 0.16*Nupb of the resources for the uplink pilots, where Nupb is the number of users per beam in the uplink. For a full 360 degree cell, the numbers are 64% and 0.64*Nupb.
A possibly simplest form of pilot scheduling is just a reuse of the pilot in every P-th beam, where P is chosen such that the interference between two beams separated by P is “acceptable” (either negligible, or with a sufficiently low penalty on the capacity of the data stream). This scheme achieves a gain of 36/P in a completely homogeneous environment. For a suburban LoS type of environment, P is typically 4, so that the pilot overhead can be reduced by a factor of 9 (almost an order of magnitude) for a 360 degree cell. Equivalently, for the uplink pilots, the number of users in the cell can be increased by a factor of 9 (this assumes that the overhead for the uplink pilots dominates the feedback overhead, as discussed above).
Simple scheduling may work only in an environment with homogeneous channel and user conditions. It can be seen that a single (uplink) user with angular spread covering P0 beams would entail a change in the angular reuse factor to P0 (assuming that a regular reuse pattern for all users is used), thus reducing the achievable gain. The more irregular the environment, the more difficult it is to find a reasonable regular reuse factor, and in the extreme case, complete orthogonalization might be necessary for regular reuse patterns, while an irregular scheduling that simply finds the best combination of users for transmitting on the same spectral resources, could provide angular reuse factors on the order of 10. However, in an environment with high angular dispersion (e.g., microcell in a street canyon), where radiation is incident on the BS from all directions, even adaptive scheduling cannot provide significant advantages over orthogonalization.
In conclusion, pilot scheduling provides an order-of-magnitude reduction in pilot overhead, or equivalently an order of magnitude larger number of users that can be accommodated for a fixed pilot overhead, compared to full orthogonalization. Compared to simple (regular) pilot reuse, environment-adaptive scheduling retains most of the possible gains, while regular scheduling starts to lose its advantages over complete orthogonalization as the environment becomes more irregular.
B. Exploiting FWA Properties for Pilot Scheduling
The exploitation of FWA properties can be more easily quantified if we retain the same reuse factor P as we would have with a “regular” scheme, but just make use of the better signal-to-interference ratio of the pilots (e.g., reduced pilot contamination). As outlined in Sec. 3.2, the reduction in the pilot contamination is equal to the temporal Rice factor. Assuming 15 dB as a typical value, and assuming a high-enough SNR that the capacity of the data transmission is dominated by pilot contamination, the SINR per user is thus improved by 15 dB. Since 3 dB SNR improvement provide 1 bit/s/Hz increase in spectral efficiency, this means that for each user, capacity is increased by 5 bit/s/Hz. Assuming 32 QAM as the usual modulation scheme, an implementation can double the capacity through this scheme.
A different way to look at the advantages is to see how much the number of users per beam can be increased, when keeping the pilot SIR constant. This can depend on the angular spectrum of the user devices. However, with a 15 dB suppression of the interference, one can conjecture that (with suitable scheduling), a reuse factor of P=2, and possibly even P=1, is feasible. This implies that compared to the case where an implementation does not use this property, a doubling or quadrupling of the number of users is feasible (and even more in highly dispersive environments)
In summary, exploiting the FWA properties for pilot scheduling doubles the capacity, or quadruples the number of users
C. Exploiting the FWA Properties for Reduction of Feedback Overhead
As outlined above, exploiting the FWA properties allow to reduce the feedback from 12 bit to 4 bit, thus reducing overhead by a factor of 3. Further advantages can be gained if the time-variant part occurs only in the parts of the impulse response with small delay, as has been determined experimentally. Then the feedback can be restricted to the delay range over which the time changes occur. If, for example, this range is 500 ns, then the feedback effort is reduced by a further factor of 10 (500 ns/5 microsec). In summary, the reduction of the feedback overhead can be by a factor of 3 to 30.
This section covers using second order statistics of a wireless channel to achieve reciprocity in frequency division duplexing (FDD) systems. FDD systems may have the following challenges in implementing such a precoded system:
In some embodiments, the base-station may send, before every precoded downlink transmission, reference signals (pilots) to the user equipment. The users will receive them and send them back to the base-station as uplink data. Then, the base-station will estimate the downlink channel and use it for precoding. However, this solution is very inefficient because it takes a large portion of the uplink capacity for sending back the received reference signals. When the number of users and/or base-station antennas grow, the system might not even be implementable. Also, the round-trip latency, in non-static channels, may degrade the quality of the channel prediction.
Second-Order Statistics Training
For simplicity, the case with a single user antenna and the L base-station antennas is considered, but can be easily extended to any number of users. The setup of the system is shown in
To achieve this, the system preforms a preliminary training phase, consisting of multiple sessions, where in each session i=1, 2, . . . , Ntraining, the following steps are taken:
Herein, (·)H is the Hermitian operator. For the case that the channel has non-zero-mean, both the mean and the covariance matrix should be determined. When the training sessions are completed, the base-station averages out the second order statistics:
Then, it computes the prediction filter and the covariance of the estimation error:
Cprediction=RDL,UL·(RUL)−1
RE=RDL−Cprediction·(RDL,UL)H
The inversion of RUL may be approximated using principal component analysis techniques. We compute {λ}, the K most dominant eigenvalues of RUL, arranged in a diagonal matrix D=diag(λ1, λ2, . . . , λK) and their corresponding eigenvectors matrix V. Typically, K will be in the order of the number of reflectors along the wireless path. The covariance matrix can then be approximated by RUL≈V·D·(V)H and the inverse as RUL−1≈V·D−1·(V)H.
Note, that there is a limited number of training sessions and that they may be done at a very low rate (such as one every second) and therefore will not overload the system too much.
To accommodate for possible future changes in the channel response, the second-order statistics may be updated later, after the training phase is completed. It may be recomputed from scratch by initializing again new Ntraining sessions, or by gradually updating the existing statistics.
The interval at which the training step is to be repeated depends on the stationarity time of the channel, e.g., the time during which the second-order statistics stay approximately constant. This time can be chosen either to be a system-determined constant, or can be adapted to the environment. Either the base-station or the users can detect changes in the second-order statistics of the channel and initiate a new training phase. In another embodiment, the base-station may use the frequency of retransmission requests from the users to detect changes in the channel, and restart the process of computing the second-order statistics of the channel.
Scheduling a Downlink Precoded Transmission
For each subframe with a precoded downlink transmission, the base-station should schedule all the users of that transmission to send uplink reference signals Nlatency subframes before. The base-station will estimate the uplink channel responses and use it to predict the desired downlink channel responses
HDL=Cprediction·HUL
Then, the downlink channel response HDL and the prediction error covariance RE will be used for the computation of the precoder.
This section covers using second order statistics of a wireless channel to achieve efficient channel estimation. Channel knowledge is a critical component in wireless communication, whether it is for a receiver to equalize and decode the received signal, or for a multi-antenna transmitter to generate a more efficient precoded transmission.
Channel knowledge is typically acquired by transmitting known reference signals (pilots) and interpolating them at the receiver over the entire bandwidth and time of interest. Typically, the density of the pilots depends on characteristics of the channel. Higher delay spreads require more dense pilots along frequency and higher Doppler spreads require more dense pilots along time. However, the pilots are typically required to cover the entire bandwidth of interest and, in some cases, also the entire time interval of interest.
Embodiments of the disclosed technology include a method based on the computation of the second-order statistics of the channel, where after a training phase, the channel can be estimated over a large bandwidth from reference signals in a much smaller bandwidth. Even more, the channel can also be predicted over a future time interval.
Second-Order Statistics Training for Channel Estimation
The system preforms a preliminary training phase, consisting of multiple sessions, where in each session i=1, 2, . . . , Ntraining, the following steps are taken:
The receiver receives the reference signals and estimates the channel over their associated bandwidth, resulting in channel responses H1(i) and H2(i).
The receiver computes the second-order statistics of these two parts:
R1(i)=H1(i)·(H1(i))H
R2,1(i)=H2(i)·(H1(i))H
R2(i)=H2(i)·(H2(i))H
Herein, (·)H is the Hermitian operator. For the case that the channel has non-zero-mean, both the mean and the covariance matrix should be determined. When the training sessions are completed, the base-station averages out the second-order statistics in a manner similar to that described in Section 6.
Efficient Channel Estimation
After the training phase is completed, the transmitter may only send reference signals corresponding to BW1. The receiver, estimated the channel response H1 and use it to compute (and predict) and channel response H2 over BW2 using the prediction filter:
H2=Cprediction·H1.
Embodiments of the disclosed technology include systems that comprise a central base station with multiple spatial antennas and multiple users, which can be configured to transmit or receive simultaneously to or from multiple users, over the same time and frequency resources, using spatial separation. In an example, this may be achieved by creating multiple beams, where each beam is focused towards a specific user. To avoid cross-interference between the beams, each beam should be nulled at the directions of the other users, thus maximizing the SINR for each user.
In some embodiments, the configuration of these beams may depend on the detection of the main radiation pattern coming from the users, or in other words, their angles-of-arrival, which is shown in the example in
In some embodiments, a method for detecting the angle-of-arrival of users in a wireless system comprises processing the received uplink transmissions from users to create multiple beams, wherein these beams have minimal cross-interference between them, and subsequently, transmitting or receiving to or from the multiple users.
In some embodiments, each beam has its energy focused towards the angle-of-arrival of a specific user and has minimal energy towards the angle-of-arrival of the other users.
In some embodiments, the angle-or-arrival is derived from uplink reference signals.
In some embodiments, the beams are created for a selected subset of the users under some angular separation criterion.
Detecting the Aliased Angle-of-Arrival
In some embodiments, and when the spacing between the spatial antennas is larger than half of the wavelength of the uplink transmission, there is an aliasing phenomenon, where the angle-of-arrival folds into a smaller range than 180 degrees. If the downlink has a different frequency (FDD), then the true angle is needed for correct beam configuration.
Embodiments of the disclosed technology include an algorithm for detecting whether the angle-of-arrival is aliased or not. If it is known to be aliased, the angle can be unfolded. In an example, the algorithm measures the angle-of-arrival at two different frequencies within the band. For OFDM, the different frequencies may be different subcarriers, preferably at the band edges. Then, a function of these two detected angles is compared to the ratio of these two frequencies, which enables the detection of aliasing because the function results in different outcomes for aliased and non-aliased angles. In an example, the function may be the ratio of these angles.
In some embodiments, a method for detecting aliased angle-of-arrival in a system with antenna spacing larger than half of the wavelength comprises measuring the angle-of-arrival at two different frequencies and comparing a function of these two measurements to the ratio of the frequencies.
In some embodiments, the true angle-of-arrival is derived from a detected aliased angle-of-arrival.
In some embodiments, the beam creation for the downlink is based on the true angle-of-arrival.
One practical problem faced by COMP operation is the availability of backhaul bandwidth to wireless devices, such as base station towers, for communicating both user data and control messages that need to be exchanged for COMP operation. For example, in embodiments in which millimeter wave wireless connections are used by network devices for backhaul communication, 2 Gbps is a typical bandwidth available under present-day technologies. A significant amount of this bandwidth may be utilized for COMP coordination messages.
As further described in the present document, a fractional COMP method may be used in some embodiments to alleviate the bandwidth pressure on backhaul. Wireless devices services by a network device may be logically divided into a first fraction of devices that are serviced using a joint COMP mode in which backhaul message exchange occurs for coordination among multiple transmissions towers or network nodes or network devices and a remaining fraction in which user devices or the corresponding transmission resources such as frequency bands, are controlled by the corresponding network device. In the latter mode, no backhaul COMP messaging bandwidth may be needed for scheduling or channel quality related messaging, thereby substantially alleviating bandwidth pressure on the backhaul link. For example, in some cases, network devices may be able to operate with zero backhaul bandwidth utilization for the locally managed mode for the remaining fraction of devices.
In some embodiments, and with reference to the scheduler interface shown in
In some embodiments, these functionality could be offloaded to the channel estimation (ChEst) and pairing estimation (PairingEst) embodiments described in the present document. This would typically be implemented in a manner that was cognizant of the scheduler timing constraints.
As shown in
In some embodiments, the scheduler interface for precoding can be configured to support a per-UE state for precoding with the following characteristics and rules:
In some embodiments, channel estimation can be performed using the sounding reference signal (SRS), with the following (non-limiting) characteristics:
In some embodiments, the coefficient computation block shown in
In some embodiments, the DL coefficient calculations could be moved into the PHY layer to advantageously reduce this latency.
In some embodiments, the API PHY interfaces (e.g., as shown in
Some embodiments of the disclosed technology can be configured to account for uplink antenna usage and the impact on downlink precoding. In an example, to meet the objective of MU-MIMO with 1 layer for each multiplexed UE, the uplink SRS can use a single antenna; e.g., 2 UEs each with one precoded layer in the DL for 2T2R (2×2 MIMO), or 4 UEs each with one precoded layer in the DL for 4T4R (4×4 MIMO).
Some embodiments of the disclosed technology can be configured to alleviate potential problems of using Transmission Mode 4 (TM4) for MU-MIMO precoding. Existing systems currently face the following potential problems:
The embodiments described herein can provide the following potential solutions:
The method 10000 includes, at operation 10004, providing, by the network device, wireless connectivity to the one or more wireless devices. For example, the wireless connectivity may allow the wireless devices to send and receive wireless signals that include user and control data. The wireless connectivity may provide access to the wireless devices to wireless services and application servers such as for voice, video and other user level communication with other devices via the wireless network.
In some embodiments, the network device jointly manages transmission resources for a first wireless device due to the COMP management status being a joint COMP status and the network device locally manages transmission resources for a second wireless device due to the COMP management status being a local COMP status. Some examples of calculations performed for COMP management of wireless devices are described with reference to
In some embodiments, wireless devices at cell edge are given the joint COMP status and wireless device towards cell center are given the local COMP status.
Alternatively, or in addition, in some embodiments, wireless devices that exhibit high mobility may be managed locally by giving them local COMP status while wireless device that are less mobile may be managed in a distributed manner by giving them joint COMP status. To help with this, wireless devices may be classified into two or more mobility classes (e.g., high mobility, low mobility) based on an order of mobility. A first number of classes may be assigned local COMP status and a second number of classes lower in the order (having lower mobility) will be given a joint COMP status.
For example, the amount of backhaul traffic generated for scheduling and other management of a high mobility device may be more than similar traffic for a low mobility device having a mobility less than the high mobility device. Therefore, backhaul bandwidth pressure may be effectively addressable by moving a high mobility device under full local control.
In some embodiments, in addition to the above discussed statuses of relative location in a cell and mobility, other operational features may be taken into account when classifying wireless devices into local or joint COMP status. For example, a number of antennas of a wireless device has an impact on the amount of channel quality information generated for or by the wireless device and this could be used as another status for deciding whether to manage locally or in COMP manner.
In some embodiments, a weighted average of various such criteria may be used in deciding which wireless devices to manage locally and which ones to manage using COMP framework. This decision may also depend on how much backhaul bandwidth is currently being used or available and availability of computational resources locally to a base station.
In some embodiments, the following technical solutions can be implemented:
1. A wireless communication method, comprising determining, by a network device, a cooperative multipoint (COMP) management status of wireless devices served by the network device; and providing, by the network device, wireless connectivity to the one or more wireless devices, wherein the network device jointly manages transmission resources for a first wireless device due to the COMP management status being a joint COMP status and the network device locally manages transmission resources for a second wireless device due to the COMP management status being a local COMP status.
2. The method of solution 1, wherein the COMP management status is determined using channel quality measurements for the one or more wireless devices.
3. The method of any of solution 1 or 2, wherein the transmission resources for the first wireless device include resource blocks used for transmissions in both directions between the network device and the first wireless device.
4. The method of any of solutions 1 to 3, further including grouping the wireless devices into groups of wireless devices; mapping the groups of wireless devices to layers of transmission; and blending the layers of transmission together into multi-user multi-input multi-output (MU-MIMO) signals for transmission or reception using multiple antenna.
5. The method of solution 4, wherein the mapping is performed by predicting channel for user devices at a future time and/or a different frequency band and by calculating modulation coding scheme for the user devices.
6. The method of solution 1, wherein the COMP status of wireless devices is determined based on a signal to noise ratio (SNR) estimate or a signal to interference plus noise (SINR) estimate.
7. The method of solution 1, wherein wireless devices at cell edge are given the joint COMP status and wireless device towards cell center are given the local COMP status.
8. The method of any of solutions 1 to 7, wherein the providing wireless connectivity includes exchanging COMP messages for the joint COMP status with neighboring network devices on a backhaul connection on which a fractional bandwidth is used for the COMP messages compared to that used for providing wireless connectivity for entirety of wireless device serviced by the network node.
9. The method of solution 8, wherein the backhaul connection is a microwave wireless connection.
10. A wireless communication system, comprising an arrangement of a plurality of network nodes in which each network node is configured to provide wireless connectivity to wireless devices using a mode that includes a joint cooperative multipoint (COMP) mode and a local COMP mode, wherein, in the joint COMP mode, transmission resources for wireless devices are managed cooperatively with other network nodes, and wherein, in the local COMP mode, transmission resources for wireless devices are managed locally, without explicit coordination with other network nodes. Some embodiments of such a system are described with reference to
11. The system of solution 10, wherein the mode further includes a deferred COMP mode in which transmission resources for wireless device are managed based on control commands received from a remote controller.
12. The system of solution 10, wherein a different mode is used for wireless devices at cellular edge and wireless devices not at cellular edge. Other possible classifications of the wireless devices into joint or local management status are described with reference to
13. The system of solution 12, wherein wireless devices are determined to be at cellular edge based on signal to noise ratio (SNR) calculations.
14. A wireless communication apparatus comprising one or more processors configured to implement a method recited in any one or more of solutions 1 to 9.
15. A non-transitory computer readable medium storing instructions, which when executed by at least one computing device, perform any of the methods in solutions 1 to 9.
The disclosed and other embodiments, modules and the functional operations described in this document can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structures disclosed in this document and their structural equivalents, or in combinations of one or more of them. The disclosed and other embodiments can be implemented as one or more computer program products, i.e., one or more modules of computer program instructions encoded on a computer readable medium for execution by, or to control the operation of, data processing apparatus. The computer readable medium can be a machine-readable storage device, a machine-readable storage substrate, a memory device, a composition of matter effecting a machine-readable propagated signal, or a combination of one or more them. The term “data processing apparatus” encompasses all apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them. A propagated signal is an artificially generated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus.
A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a standalone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program does not necessarily correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
The processes and logic flows described in this document can be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
Processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read-only memory or a random access memory or both. The essential elements of a computer are a processor for performing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Computer readable media suitable for storing computer program instructions and data include all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto optical disks; and CD ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
While this patent document contains many specifics, these should not be construed as limitations on the scope of an invention that is claimed or of what may be claimed, but rather as descriptions of features specific to particular embodiments. Certain features that are described in this document in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable sub-combination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a sub-combination or a variation of a sub-combination. Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results.
Only a few examples and implementations are disclosed. Variations, modifications, and enhancements to the described examples and implementations and other implementations can be made based on what is disclosed.
This patent document is a 371 of International Patent Application No. PCT/US2020/032073, filed 8 May 2020, which claims priority to and benefits of U.S. Provisional Patent Application No. 62/845,037 filed 8 May 2019 and U.S. Provisional Patent Application No. 62/852,240 filed 23 May 2019. The entire content of the before-mentioned patent applications are incorporated by reference as part of the disclosure of this patent document.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2020/032073 | 5/8/2020 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2020/227619 | 11/12/2020 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
9130709 | Hu et al. | Sep 2015 | B2 |
9820205 | Gaal et al. | Nov 2017 | B2 |
20100285810 | Ko | Nov 2010 | A1 |
20130310098 | Baligh | Nov 2013 | A1 |
20160037420 | Gaal | Feb 2016 | A1 |
20180020470 | Zhuang | Jan 2018 | A1 |
Number | Date | Country |
---|---|---|
2016124979 | Aug 2016 | WO |
Entry |
---|
International Search Report and Written Opinion for PCT/US2020/32073, dated Jul. 16, 2020, 9 pages. |
Extended European Search Report for co-pending EP Application No. 20802625.2, mailed Nov. 25, 2022 15 pages. |
Zhang et al: “Transmission Mode Selection for Downlink Coordinated Multipoint Systems”, IEEE Transactions on Vehicular Technology, IEEE, USA, vol. 62, No. 1, Jan. 1, 2013, 7 Pages. |
Su et al: “The Value of Channel Prediction in CoMP Systems with Large Backhaul Latency” , IEEE Transactions on Communications, IEEE Service Center, Piscataway, NJ. USA, vol. 61, No. 11, Nov. 1, 2013, 14 pages. |
Number | Date | Country | |
---|---|---|---|
20220224464 A1 | Jul 2022 | US |
Number | Date | Country | |
---|---|---|---|
62852240 | May 2019 | US | |
62845037 | May 2019 | US |