Embodiments of the present disclosure generally relate to wireless communication technology, and more particularly to beam management in a wireless communication system.
Wireless communication systems are widely deployed to provide various telecommunication services, such as telephony, video, data, messaging, broadcasts, and so on. Wireless communication systems may employ multiple access technologies capable of supporting communication with multiple users by sharing available system resources (e.g., time, frequency, and power). Examples of wireless communication systems may include fourth generation (4G) systems, such as long term evolution (LTE) systems, LTE-advanced (LTE-A) systems, or LTE-A Pro systems, and fifth generation (5G) systems which may also be referred to as new radio (NR) systems.
As the number of users (e.g., user equipment (UE)) and base stations (BSs) increases, interference among users becomes more and more serious and the scheduling of BS resources becomes quite complicated.
Among the existing beamforming algorithms, nonlinear and objective function optimization algorithms have a high complexity and are difficult to deploy in practice, while linear algorithms have low complexity but unsatisfactory performance. These problems are becoming increasingly evident as the number of users and BSs increases. The industry desires solutions that can provide high performance while maintaining low computational complexity.
Some embodiments of the present disclosure provide a user equipment (UE). The UE may include: a transceiver; and a processor coupled to the transceiver. The processor may be configured to: receive pilot signals from a plurality of base stations (BSs); measure channel state information (CSI) between the UE and each of the plurality of BSs; generate a CSI matrix based on the measured CSI between the UE and the plurality of BSs; encode the CSI matrix; and transmit the encoded CSI matrix to one of the plurality of BSs.
The processor may be further configured to select the one of the plurality of BSs based on signal strengths or distances between the UE and the plurality of BSs. The CSI matrix may indicate: channel amplitude information associated with the plurality of BSs; channel phase information associated with the plurality of BSs; and a normalization factor associated with the channel amplitude information.
To encode the CSI matrix, the processor may be configured to: quantize the CSI matrix according to an accuracy associated with a codebook; and compare the quantized CSI matrix with elements in the codebook to determine an index for the quantized CSI matrix. To compare the quantized CSI matrix with the elements in the codebook, the processor may be configured to determine a similarity of the quantized CSI matrix and the elements in the codebook by one of the following: calculating a Minkowski distance between the quantized CSI matrix and a corresponding element in the codebook; calculating a cosine similarity between the quantized CSI matrix and the corresponding element in the codebook; calculating a Pearson correlation coefficient between the quantized CSI matrix and the corresponding element in the codebook; calculating a Mahalanobis distance between the quantized CSI matrix and the corresponding element in the codebook; calculating a Jaccard coefficient between the quantized CSI matrix and the corresponding element in the codebook; and calculating a Kullback-Leibler divergence between the quantized CSI matrix and the corresponding element in the codebook. To encode the CSI matrix, the processor may be further configured to compress the index for the quantized CSI matrix, and wherein transmitting the encoded CSI matrix comprises transmitting the compressed index to the one of the plurality of BSs.
Some embodiments of the present disclosure provide a BS. The BS may include: a transceiver; and a processor coupled to the transceiver. The processor may be configured to: receive, from a UE served by the BS, information associated with channel state information (CSI) between the UE and a plurality of BSs including the BS; transmit the information associated with the CSI to a cloud apparatus; receive a beamforming matrix from the cloud apparatus in response to the transmission of the information associated with the CSI; and perform a beamforming operation according to the beamforming matrix.
The CSI between the UE and each of the plurality of BSs may indicate: amplitude information related to a channel between the UE and a corresponding BS; phase information related to the channel between the UE and the corresponding BS; and a normalization factor associated with the amplitude information. The processor may be further configured to add an ID of the BS to the information associated with the CSI before the transmission.
Some embodiments of the present disclosure provide a cloud apparatus. The cloud apparatus may include: a transceiver; and a processor coupled to the transceiver. The processor may be configured to: receive first information associated with channel state information (CSI) between a plurality of user equipment (UE) and a plurality of base stations (BSs), wherein the cloud apparatus manages the plurality of BSs and each of the plurality of UEs accesses a corresponding BS of the plurality of BSs; generate a beamforming matrix based on the first information by a beamforming model deployed on the cloud apparatus; and transmit the beamforming matrix to the plurality of BSs.
Some embodiments of the present disclosure provide a method for wireless communication performed by a user equipment (UE). The method may include: receiving pilot signals from a plurality of base stations (BSs); measuring channel state information (CSI) between the UE and each of the plurality of BSs; generating a CSI matrix based on the measured CSI between the UE and the plurality of BSs; encoding the CSI matrix; and transmitting the encoded CSI matrix to one of the plurality of BSs.
Some embodiments of the present disclosure provide a method for wireless communication performed by a BS. The method may include: receiving, from a UE served by the BS, information associated with channel state information (CSI) between the UE and a plurality of BSs including the BS; transmitting the information associated with the CSI to a cloud apparatus; receiving a beamforming matrix from the cloud apparatus in response to the transmission of the information associated with the CSI; and performing a beamforming operation according to the beamforming matrix.
Some embodiments of the present disclosure provide a method for wireless communication performed by a cloud apparatus. The method may include: receiving first information associated with channel state information (CSI) between a plurality of user equipment (UE) and a plurality of base stations (BSs), wherein the cloud apparatus manages the plurality of BSs and each of the plurality of UEs accesses a corresponding BS of the plurality of BSs; generating a beamforming matrix based on the first information by a beamforming model deployed on the cloud apparatus; and transmitting the beamforming matrix to the plurality of BSs.
Some embodiments of the present disclosure provide an apparatus. The apparatus may be a UE, a BS, or a cloud apparatus. According to some embodiments of the present disclosure, the apparatus may include: at least one non-transitory computer-readable medium having stored thereon computer-executable instructions; at least one receiving circuitry; at least one transmitting circuitry; and at least one processor coupled to the at least one non-transitory computer-readable medium, the at least one receiving circuitry and the at least one transmitting circuitry, wherein the at least one non-transitory computer-readable medium and the computer executable instructions may be configured to, with the at least one processor, cause the apparatus to perform a method according to some embodiments of the present disclosure.
In order to describe the manner in which the advantages and features of the disclosure can be obtained, a description of the disclosure is rendered by reference to specific embodiments thereof, which are illustrated in the appended drawings. These drawings depict only exemplary embodiments of the disclosure and are not therefore to be considered limiting of its scope.
The detailed descriptions section of the appended drawings are intended as a description of the preferred embodiments of the present disclosure and are not intended to represent the only form in which the present disclosure may be practiced. It should be understood that the same or equivalent functions may be accomplished by different embodiments that are intended to be encompassed within the spirit and scope of the present disclosure.
Reference will now be made in detail to some embodiments of the present disclosure, examples of which are illustrated in the accompanying drawings. To facilitate understanding, embodiments are provided under specific network architecture and new service scenarios, such as the 3rd generation partnership project (3GPP) 5G (NR), 3GPP long-term evolution (LTE) Release 8, and so on. It is contemplated that along with the developments of network architectures and new service scenarios, all embodiments in the present disclosure are also applicable to similar technical problems; and moreover, the terminologies recited in the present disclosure may change, which should not affect the principles of the present disclosure.
For example, in the context of the present disclosure, user equipment (UE) may include computing devices, such as desktop computers, laptop computers, personal digital assistants (PDAs), tablet computers, smart televisions (e.g., televisions connected to the Internet), set-top boxes, game consoles, security systems (including security cameras), vehicle on-board computers, network devices (e.g., routers, switches, and modems), or the like. According to some embodiments of the present disclosure, the UE may include a portable wireless communication device, a smart phone, a cellular telephone, a flip phone, a device having a subscriber identity module, a personal computer, a selective call receiver, or any other device that is capable of sending and receiving communication signals on a wireless network. In some embodiments of the present disclosure, the UE includes wearable devices, such as smart watches, fitness bands, optical head-mounted displays, or the like. Moreover, the UE may be referred to as a subscriber unit, a mobile, a mobile station, a user, a terminal, a mobile terminal, a wireless terminal, a fixed terminal, a subscriber station, a user terminal, or a device, or described using other terminology used in the art. The present disclosure is not intended to be limited to the implementation of any particular UE.
In the context of the present disclosure, a base station (BS) may also be referred to as an access point, an access terminal, a base, a base unit, a macro cell, a Node-B, an evolved Node B (eNB), a gNB, a Home Node-B, a relay node, or a device, or described using other terminology used in the art. The BS is generally a part of a radio access network that may include one or more controllers communicably coupled to one or more corresponding BSs. The present disclosure is not intended to be limited to the implementation of any particular BS.
In the context of the present disclosure, the UE may communicate with a BS via uplink (UL) communication signals. The BS may communicate with UE(s) via downlink (DL) communication signals.
The wireless communication system 100 may be compatible with any type of network that is capable of sending and receiving wireless communication signals. For example, the wireless communication system 100 is compatible with a wireless communication network, a cellular telephone network, a time division multiple access (TDMA)-based network, a code division multiple access (CDMA)-based network, an orthogonal frequency division multiple access (OFDMA)-based network, an LTE network, a 3GPP-based network, a 3GPP 5G network, a satellite communications network, a high altitude platform network, and/or other communications networks. The present disclosure is not intended to be limited to the implementation of any particular wireless communication system architecture or protocol.
As shown in
BS 103 may be a macro BS (MBS) or a logical center (e.g., anchor point) managing BSs 102A and 102B. In some examples, the BS(s) 102 may also be referred to a micro BS, a Pico BS, a Femto BS, a low power node (LPN), a remote radio-frequency head (RRH) or described using other terminology used in the art.
The coverage of BS(s) 102 (e.g., coverage 111 of BS 102A and coverage 112 of BS 102B) may be in the coverage 113 of BS 103. BS 103 and BS(s) 102 can exchange data, signaling (e.g., control signaling), or both with each other via a backhaul link. BS 103 may be used as a distributed anchor. BS(s) 102 may have connections with users, e.g., UE(s) 101. Each UE 101 may be served by a BS 102. For example, referring to
To meet the demand for fast-growing wireless data services, wireless communication system 100 may support massive multiple-input multiple-output (MIMO) technology which has significant advantages in terms of enhanced spectrum and energy efficiency, supporting large data and providing high-speed and reliable data communication.
However, this may bring about some problems. For example, there may be multiuser interference. For instance, the performance of a user may significantly degrade due to such interference from other users. To tackle this problem, interference reduction or cancellation techniques, such as maximum likelihood multiuser detection for the uplink, dirty paper coding (DPC) techniques for the downlink, or interference alignment, may be employed. However, these techniques are complicated and have high computational complexity.
The acquisition of channel state information (CSI) may also be problematic. For instance, in order to achieve a high spatial multiplexing gain, the BS may need to process the received signals coherently. This requires accurate and timely acquisition of the CSI, which can be challenging, especially in high mobility scenarios.
The more antennas a BS is equipped with, the more degrees of freedom are offered, and hence, more users can simultaneously communicate in the same time-frequency resource. As a result, a huge sum of throughput can be obtained. With large antenna arrays, conventional signal processing techniques become prohibitively complex due to high signal dimensions. It would be hard to obtain a huge multiplexing gain with low-complexity signal processing and low-cost hardware implementation.
In some examples, linear or nonlinear techniques such as weighted minimum mean square error (WMMSE) and DPC may be employed and channel capacity can be enhanced by effectively reusing space resources. However, the computational complexity of these algorithms grows significantly with the number of network variables due to the large number of complex operations involved in the iterations. Although these traditional iterative algorithms can achieve satisfying performance, they cannot meet the requirements of real-time applications.
In some examples, deep learning (also known as “machine learning (ML)”) methods may be employed to satisfy both the low complexity and high performance requirements by simple linear or nonlinear transformation operators. In some works of deep learning-based resource allocation, the models are designed to be simple, which causes model representation capability to decrease as the complexity of the system increases. Simply increasing the number of model layers (e.g., depth) cannot effectively improve model performance, and may also cause model performance to degrade due to gradient disappearance and/or explosion. Embodiments of the present disclosure provide enhanced deep learning models to solve the above issues.
In some examples, the deep learning models for beamforming can be classified into supervised learning, unsupervised learning and reinforcement learning.
Supervised learning is to fit labeled data, which in this scenario is to fit the beamforming results calculated by specific mathematical methods. However, supervised learning has two drawbacks. One is that model training requires labeled data, which is costly, and it is difficult for the model to outperform the performance achieved by the mathematical methods. The other is that as the scale of the scene increases, the value of each element in the beamforming matrix will gradually decrease, and the error value on which the model training depends will therefore become smaller, making the model difficult to train and degrading the performance. Known unsupervised learning models may not require labeled data, but may suffer from the problem of poor applicability of the models described above in large-scale scenarios. Known reinforcement learning models, in order to simplify the model design, mostly use the codebook as the output. This makes the model performance largely dependent on the design of the codebook, which is artificially set, increasing the cost of model deployment.
Embodiments of the present disclosure provide solutions to solve the above issues. For example, a deep learning based beamforming method that can well balance real-time and performance is provided. The method may use channel state information (CSI) as the model input, and the model may directly output the final beamforming results, which can be used by the system directly and is better than selecting from determined beams.
Moreover, compared with known artificial intelligence (AI) based beamforming methods, embodiments of the present disclosure may use AI for beamforming design directly, thereby reducing performance loss caused by multi-level settings. Embodiments of the present disclosure may take into account the performance of the beam on the basis of fast and accurate beam management. Embodiments of the present disclosure may be applied to, but are not limited to, a massive MIMO network. More details on the embodiments of the present disclosure will be illustrated in the following text in combination with the appended drawings.
In some embodiments of the present disclosure, a deep learning model for beamforming is applied to balance real-time and performance. Unsupervised learning is employed to train the model, to reduce the training cost and improve the performance of the model in the face of large-scale scenarios. In addition, the structural design of existing deep learning models has the problem of gradient disappearance in large-scale scenarios. To solve this problem, an Inception structure is employed to design the beamforming model based on unsupervised learning. The Inception structure extends the width of the model, and can use a shortcut to connect two layers that are far apart to alleviate the gradient disappearance problem in the case that the model deepens.
The beamforming model may be deployed on a cloud apparatus (e.g., a computing unit of the apparatus). The cloud apparatus can be an MBS or a logical center (anchor point) for cell resource allocation, for example, BS 103 shown in
The application scenario may include a cloud (e.g., as a logical center for cell resource allocation) and several BSs connected to users (e.g., UEs), each of which may be served by a BS. The cloud (e.g., BS 103 shown in
The details will be described in the following text.
In some embodiments of the present disclosure, a UE (e.g., UE 101 in
The UE may select a BS from the reachable BSs as its serving BS. For example, referring to
In some examples, the UE may select a BS with the strongest signal strength (e.g., reference signal received power (RSRP)) as its serving BS. If there are two or more BSs having the same strongest signal strength, the UE may select the one nearest to the UE. In some examples, the UE may select a BS with the closest distance to the UE as the serving BS. If there are two or more BSs having the same closest distance, the UE may select the one with the strongest signal strength. If there are two or more BSs with the same strongest signal strength and the same closest distance to the UE, the UE may randomly select a BS from the two or more BSs. When the user is on the move, BS switching may be performed according to mobile user switching scheme A3 as specified according to the 3GPP specification.
The UE may generate a CSI matrix H based on the measured CSI. The UE may normalize the amplitude with a normalization factor C (also referred to as “amplitude scaling factor”) to obtain a normalized amplitude A. The CSI matrix may indicate the normalized amplitude A, the phase B and the normalization factor C. The UE may transmit the generated CSI matrix to its serving BS.
For example, assuming that the UE receives pilot signals from N BSs, the left part of
In some embodiments of the present disclosure, the UE may encode the generated CSI matrix(es). For example, the UE may quantize a CSI matrix according to an accuracy associated with a codebook, and compare the quantized CSI matrix with elements in the codebook to determine an index for the CSI matrix. The codebook may also be stored at the cloud. In this way, the computational efficiency can be improved and the size of the codebook can also be reduced.
In some embodiments of the present disclosure, quantizing a CSI matrix may include quantizing the elements, e.g., the normalized amplitude and the phase, in the CSI matrix. In some embodiments of the present disclosure, comparing the quantized CSI matrix with elements in the codebook may include comparing the quantized CSI matrix elements with elements in the codebook to determine respective indexes for the quantized CSI matrix elements.
The comparison may be performed according to a similarity computation algorithm, including but not limited to: Minkowski distance; cosine similarity; Pearson correlation coefficient; Mahalanobis distance; Jaccard coefficient; or Kullback-Leibler divergence. For example, when it is determined that a CSI matrix element is most similar to a codebook element, the CSI matrix element can be indicated by the index of the codebook element. A CSI matrix can be indicated by indexes of its elements. For example, the indexes of its elements may be concatenated as the index for the CSI matrix. In some embodiments of the present disclosure, transmitting the generated CSI matrix to the serving BS may include transmitting the index(es) for the CSI matrix(es) to the serving BS.
In some embodiments of the present disclosure, to reduce reporting overhead, the UE may compress the index for the CSI matrix. For example, a lossless data compression may be performed to compress the index(es) for the CSI matrix(es). The lossless data compression algorithm may include, but not limited to, run-length encoding, LZF algorithm, Huffman coding, LZ77 algorithm, and LZ78 algorithm. In some examples, the UE may expand all CSI matrix indexes into one dimension according to a predefined order (e.g., an order associated with the BSs), and then perform the lossless data compression. In some embodiments of the present disclosure, transmitting the generated CSI matrix to the serving BS may include transmitting the compressed index(es) to the serving BS.
In some embodiments of the present disclosure, the codebook and the compression algorithm may be determined based on the actual application situation and a priori knowledge. In some examples, the codebook, the compression algorithm, or both may be exchanged between the serving BS and the UE via radio resource control (RRC) signaling. In some examples, the codebook, the compression algorithm, or both may be predefined, for example, in a standard(s). In principle, the network and the UE should have the same compression algorithm and codebook, and therefore the CSI matrix generated at the UE side can be understood by the network side. For example, the network (e.g., the cloud or the BS) may decode the encoded CSI received based on the codebook.
In some embodiments of the present disclosure, a BS (e.g., BS 102A) may collect, from the UE(s) (e.g., UE 101A) served by the BS, information associated with the CSI between the UE(s) and the reachable BS(s) of the UE(s) (e.g., the index(es) of the CSI matrix(es)). The BS may transmit the collected information to the cloud, which manages the BS. In some embodiments of the present disclosure, the BS may add an ID of the BS to the collected information, for example, at the beginning of the collected information, before the transmission. The BS may receive a beamforming matrix from the cloud in response to the collected information. The BS may then perform a beamforming operation according to the beamforming matrix.
In some embodiments of the present disclosure, the cloud may manage a plurality of BSs, each of which may serve a plurality of UEs. The cloud may receive, information associated with the CSI between the plurality of UEs and the plurality of BSs (e.g., the indexes of CSI matrixes). The cloud may combine the received CSI into a global CSI matrix. The received CSI from the plurality of BSs may be arranged according to a predefined order (e.g., an order associated with the plurality of BSs) to form the global CSI matrix. For instance, the CSI may be arranged according to the IDs of the BSs.
Assuming that the cloud manages N BSs (e.g., BS1-BSN), which serves M UEs (e.g., UE1-UEM), the right part of
A beamforming model may be deployed in the cloud (e.g., a computing unit of the cloud). The design and training of the beamforming model will be described in detail in the following text. The cloud may generate a beamforming matrix based on the CSI (e.g., the global CSI matrix) from the plurality of BSs and may transmit the beamforming matrix to the plurality of BSs. For example, the global CSI matrix may be input into the beamforming model, which may output the beamforming matrix. In this way, the cloud can calculate the beamforming matrix in real time based on the global CSI matrix.
In some embodiments of the present disclosure, the cloud may split the beamforming matrix into a plurality of beamforming sub-matrixes. Each of the plurality of beamforming sub-matrixes may be associated with a corresponding BS of the plurality of BSs. Transmitting the beamforming matrix to the plurality of BSs may include transmit a beamforming sub-matrix of the plurality of beamforming sub-matrixes to a corresponding BS of the plurality of BSs. A BS can perform a beamforming operation according to the corresponding beamforming sub-matrix.
In some embodiments of the present disclosure, the cloud may periodically receive the CSI transmitted by the plurality of BSs. The cloud may perform the above operations (e.g., generating a global CSI matrix, generating the beamforming matrix, and transmitting the same to the BSs) in response to the reception of the CSI.
In some embodiments of the present disclosure, the deployed beamforming model may be updated according to a certain criterion. For example, the deployed beamforming model may be updated periodically (e.g., once a week or month). For example, the deployed beamforming model may be updated dynamically, for example, based on a performance decline of the beamforming model. For instance, when the performance decline of the beamforming model reaches a certain threshold, e.g., a certain percentage (e.g., 80%) of the performance achieved by the WMMSE algorithm, the beamforming model may be updated. Updating the beamforming model may include fine-tuning the parameter(s) of the beamforming model (e.g., a weight of a layer of the beamforming model).
The details of the design and training process of the beamforming model will be described in the following text.
Before beamforming, the cloud may construct an elaborated convolutional neural network to compose the beamforming model. Each layer of the beamforming model may be assigned with a corresponding weight, which can be updated during the training process by back propagation.
In some embodiments of the present disclosure, the beamforming model may include at least one Inception structure. The number of the Inception structures can be determined by the actual application scenario. The Inception structure may convert the original single structure of each layer into a spliced combination of multidimensional structures to enhance the model's ability to extract features.
An Inception structure may include multilayers, such as convolutional layers, batch normalization layers, and activation layers. The activation layers may be included in the convolutional layers and the batch normalization layers of the Inception structure. The Inception structure may include at least two branches, each of which may include at least one convolutional layer. The number of branches is also referred to as the width of the structure. The width is provided by the convolutional layers can reduce the calculation cost of the model. The number of the convolutional layers in a branch may also be referred to as the depth of the branch or Inception structure. The numbers of the convolutional layers in different branches can be the same or different. The convolutional layers in the Inception structure (either within a different or the same branch) may have the same or different convolutional kernel sizes. The number of branches and various layers in the Inception structure and the parameters of the various layers (e.g., convolution kernel size including, for example, 1×1, 2×2, 3×3, or 4×4) can be determined by the actual application scenario.
An Inception structure may include at most one pooling layer for filtering input data which is included in one of the at least two branches. The number of the pooling layers (e.g., 0 or 1) in the Inception structure and the parameters of a pooling layer (e.g., pooling layer size and holding window size) can be determined by the actual application scenario.
An Inception structure may include a shortcut. The presence of shortcuts can alleviate the gradient disappearance problem to a certain extent and make the model perform better. In some examples, the shortcut may connect an input of the inception structure block and an output of the inception structure block. In some examples, the shortcut may connect two internal functional layers of the inception structure block. The number of the shortcuts in the Inception structure and the connection relationship of the shortcut can be determined by the actual application scenario. In some examples, when the internal structure of the inception structure is simple, the performance gain brought by a shortcut may be not obvious, and thus can be omitted.
In some embodiments of the present disclosure, the beamforming model may include an output activation layer for outputting the beamforming matrix. The output activation layer can ensure that the beamforming matrix satisfies a power constraint of the plurality of the BSs while ensuring that the nonlinearity is not lost. In some embodiments of the present disclosure, the beamforming model may output a system rate (e.g., the sum-rate of the plurality of UEs) according to the loss function, which can be used as the basis for determining the completion of the training.
The exemplary beamforming model 300 may be input with input 311 and may output outputs 313 and 315. The input 311 may be CSI such as a global CSI matrix(es). In some examples, the input may be processed as a two-dimensional matrix of two channels passing through a convolutional layer with, for example, a 2×2 convolutional kernel, and then through an activation layer (for example, included in the convolutional layer), the output of which may pass through the Inception structures. Outputs 313 and 315 may be a negative system rate and the beamforming matrix, respectively.
After constructing the beamforming model, the beamforming model may be trained offline using collected CSI. For example, the collected CSI (e.g., global CSI matrixes collected before the deployment of the model) may be divided into a training set and a test set. For instance, 70%, 80% or 90% of the collected CSI may be used as the training set while the remaining collected CSI may be used as the test set.
The training set may be iteratively fed into the beamforming model. The parameters of the beamforming model (e.g., weights of layers of the model) may be updated by back propagation during the training of the model, which may continuously increase the system rate (or reduce the negative system rate) obtained from the model prediction beamforming matrix until an end condition is satisfied. For example, the training set may be input into the beamforming model in batches. For instance, the training set may include about 100,000 global CSI matrixes, every 64 global CSI matrixes may be arranged as a batch to be input into the beamforming model. For each batch, the weights of layers of the model may be updated by back propagation. The cloud may iteratively input the training set into the beamforming model until the end condition is met. For example, after all of the training set is input into the beamforming model (which may also be referred to as a single iteration) and the end condition is not satisfied, the cloud may start another iteration until the end condition is met.
There are several options for setting the end condition. In some examples, the end condition may be determined in response to at least one of the following: the number of iterations reaching a training threshold; and an improvement on the system rate being less than or equal to an improvement threshold. For example, the system rate no longer increases or the loss function no longer decreases may mean that the algorithm of the model converges.
In some embodiments of the present disclosure, in response to the end condition being met, the cloud may determine whether the performance of the beamforming model satisfies a performance demand. The perform demand can be a performance value of the beamforming model relative to a mathematical methods (e.g., the WMMSE algorithm or the zero force (ZF) algorithm). For example, the cloud may input the test set into the beamforming model to determine a model performance, when the model performance reaches (i.e., greater than or equal to) a certain percentage (e.g., 80%) of the WMMSE algorithm, it is determined that the model performance satisfies the performance demand.
The cloud may determine the completion of the training in response to determining that the performance of the beamforming model satisfies the performance demand. Then, the cloud may deploy the trained beamforming model for determining a beamforming management scheme for the plurality of BSs.
In response to determining that the performance of the beamforming model fails to satisfy the performance demand, the cloud may, in some examples, update the parameters of the beamforming model to satisfy the performance demand. For instance, the weights of the layers of the model may be fine-tuned. In some examples, the cloud may reconstruct the beamforming model and train the reconstructed beamforming model to satisfy the performance demand.
In some other embodiments of the present disclosure, all the collected CSI may be used for training, and the model training is completed in response to an end condition being met.
The following text describes an example of unsupervised learning-based beam management. For convenience, a single base station massive MIMO scenario is taken as an example to illustrate this.
A transmitter at a BS equipped with P antennas may serve K UEs (e.g., UE1 to UEK), each with Q receive antennas. The channel between a UEk (∈UE1 to UEK) and the BS can be denoted as a matrix Hk∈C[Q×P] which may include channel gains between different transceiver antenna-pairs. The received signal at UEk can be denoted as:
where sk∈C[P×M] represents the transmitted vector, M represents the number of data streams transmitted by the BS, and nk∈C[Q×1] represents the white Gaussian noise vector at UEk with covariance Rn
The transmit vector sk can be e denoted as the data vector x1, . . . , xM∈C[Q×M] passing through M linear filters:
where matrix Vk=[v1, . . . , vM], vm∈C[F×M] is a beamforming matrix for user k and xm is the input vector. It is assumed that the data streams which are received by each UE are independent such that E[xmxmH]=IM and E[xlxmH]=0, for t≠m.
In this example, the Saleh-Valenzuela mmWave channel model for Hk with one line-of-sight (LoS) path and (L-1) non-LoS (NLoS) paths is adopted. Thus, Hk can be represented as:
where d represents antenna spacing, Nt represents the number of transmitting antennas, Nr represents the number of receiving antennas, αl represents the path loss and phase shift of the Ith path, αr represents the array response or steering vector of the receiver, αt represents the array response or steering vector of the transmitter, λ represents the wavelength of the carrier frequency, ∂l and θl represent the arrival and departure angles of the Ith path, respectively, modeled as uniformly distributed within
Persons skilled in the art would understand that other channel models can also be employed.
The objective of the beamforming model is to maximize the weighted sum-rate of all UEs in the system by designing the beamforming matrixes V1, . . . Vk. Therefore, the utility maximization problem can be formulated as:
Rk represents the spectral efficiency of UEk, uk≥0 represents the corresponding weight and can be setting according to the actual scenario, and Pmax represents the maximum power supported by the BS.
The input of the model is the matrix H indicating CSI between the UEs and all BSs under the management of the cloud, for example, the global CSI matrix as described above. An output of the model may be the beamforming matrix V. The loss function can be represented as:
where Θ represents the parameters of the model.
The model may include a lambda layer (e.g., “Lambda layer rate” in
where b is a gain factor that ensures the signal in each sample to satisfy the transmit power constraint.
After several iterations of training, the model loss function may, for example, no longer decrease. After the model passes on the test set, the model training is completed.
The unsupervised learning-based beamforming model proposed in the subject disclosure is advantageous in various aspects. For example, compared with supervised learning models, the training cost of the proposed model is low and the training process is easier and simple. In addition, compared with known unsupervised learning models, the proposed model has a novel and better model structure design, and can maintain better system performance in large-scale scenarios.
From
The performance of the ResNet neural network model is also inferior to the model of the present disclosure because of the single structure of each layer of the ResNet model and the reduced ability to capture the structure of the scene in large-scale scenarios. Therefore, the structure of the present disclosure is better.
Table 1 below shows the spectral efficiency and computational performance for different schemes.
From the above table, it can be concluded that the computational time consumed by the ML model increases slowly as the scene size increases, while the time consumed by the WMMSE algorithm increases rapidly, so the ML-based model can better guarantee real-time computational results.
It can also be concluded that the performance of all algorithms decreases with the increase of the number of users. However, in a maximum scenario (for example, UE=10, Nt=64, and Nr=16), the model proposed in the present disclosure still achieves a performance of more than 96% of the WMMSE algorithm, while the performance of other models is already less than 70%. This shows that the model structure proposed in the present disclosure can well guarantee the performance of the model in large-scale scenarios.
To sum up, the present disclosure comprehensively considers the beamforming management method and the service process design, and improves the performance in massive MIMO, which has universality and can be applied more practically.
Referring to
In operation 713, the UE may measure CSI between the UE and each of the plurality of BSs. The CSI may include, for example, channel amplitude and channel phase information associated with respective BSs.
In operation 715, the UE may generate a CSI matrix based on the measured CSI between the UE and the plurality of BSs. In some embodiments, the CSI matrix may indicate channel amplitude information associated with the plurality of BSs, and channel phase information associated with the plurality of BSs. The UE may select a suitable normalization factor for channel amplitude information associated with each BS. The CSI matrix may further indicate the normalization factor associated with the channel amplitude information.
In operation 717, the UE may encode the CSI matrix. For example, the UE may quantize the CSI matrix according to an accuracy associated with a codebook. The codebook may be shared by the UE and the network. The UE may quantize the elements, e.g., the amplitude and phase information, in the CSI matrix.
Then, the UE may compare the quantized CSI matrix with the codebook to determine an index for the quantized CSI matrix. For example, the UE may compare the quantized elements in the CSI matrix with the elements in the codebook to determine indexes for the quantized elements in the CSI matrix, which may be used as the index for the quantized CSI matrix.
In some embodiments, to compare the quantized CSI matrix with the elements in the codebook, the UE may determine a similarity of the quantized CSI matrix (e.g., quantized elements in the CSI matrix) and the elements in the codebook. Various methods can be employed for determining the similarity. For example, a Minkowski distance, a cosine similarity, a Pearson correlation coefficient, a Mahalanobis distance, a Jaccard coefficient, or a Kullback-Leibler divergence between the quantized CSI matrix and a corresponding element in the codebook may be calculated.
In some embodiments, to reduce overhead, the UE may compress the index for the quantized CSI matrix. Various data compression algorithm can be employed. For example, a lossless data compression algorithm, such as run-length encoding, LZF algorithm, Huffman coding, LZ77 algorithm, or LZ78 algorithm can be employed. The network and the UE should have the same understanding of the data compression algorithm. For example, the data compression algorithm can be predefined or communicated between the UE and the network via, for example, RRC signaling.
In operation 719, the UE may transmit the encoded CSI matrix to one of the plurality of BSs, for example, the serving BS of the UE. In some embodiments, transmitting the encoded CSI matrix may include transmitting the index for the quantized CSI matrix or the compressed index to the serving BS.
It should be appreciated by persons skilled in the art that the sequence of the operations in exemplary procedure 700 may be changed and some of the operations in exemplary procedure 700 may be eliminated or modified, without departing from the spirit and scope of the disclosure.
Referring to
In operation 813, the BS may transmit the information associated with the CSI to a cloud apparatus (e.g., MBS 103 in
In operation 815, the BS may receive a beamforming matrix from the cloud apparatus in response to the transmission of the information associated with the CSI. In operation 817, the BS may perform a beamforming operation according to the beamforming matrix.
It should be appreciated by persons skilled in the art that the sequence of the operations in exemplary procedure 800 may be changed and some of the operations in exemplary procedure 800 may be eliminated or modified, without departing from the spirit and scope of the disclosure.
Referring to
In some embodiments, the CSI between the plurality of UEs and the plurality of BSs may indicate: amplitude information related to a channel between a corresponding UE and a corresponding BS; phase information related to the channel between the corresponding UE and the corresponding BS; and a normalization factor associated with the amplitude information. In some examples, the cloud apparatus may combine the received information into a global CSI matrix. For instance, the cloud apparatus may decode the indexes of the encoded CSI matrixes and may form a global CSI matrix based on the decoded information.
In operation 913, the cloud apparatus may generate a beamforming matrix based on the first information by a beamforming model deployed on the cloud apparatus. The cloud apparatus may update the deployed beamforming model according to various policies.
In some examples, the cloud apparatus may update the deployed beamforming model periodically. In some examples, the cloud apparatus may update the deployed beamforming model according to the performance of the model, for example, when a performance decline of the beamforming model reaches a threshold. For example, the cloud apparatus may store the first information and calculation result (e.g., system rate) periodically, and use a mathematical method to evaluate the performance of the current model in an offline state. When the model performance drops to a threshold, the cloud apparatus may fine-tune the model parameters, such as weights.
In some embodiments, before deployment of a beamforming model, the cloud apparatus may design the beamforming model according to the actual application scenario. The cloud apparatus may train the model with pre-collected first information.
For example, the cloud apparatus may construct the beamforming model for determining a beamforming management scheme for the plurality of BSs. The cloud apparatus may train the beamforming model based on a plurality of the first information (e.g., a plurality of global CSI matrix). In some examples, the plurality of the first information may be the training set as described above. The cloud apparatus may deploy the trained beamforming model on the cloud apparatus in response to a completion of the training.
In some embodiments, the beamforming model may include an inception structure block, which may include at least two branches and at most one pooling layer for filtering input data. The at most one pooling layer may be included in one of the at least two branches. Each branch may include at least one convolutional layer. In some embodiments, the inception structure block may further include a shortcut. In some examples, the shortcut may connect an input of the inception structure block and an output of the inception structure block. In some examples, the shortcut may connect two internal functional layers of the inception structure block. In some embodiments, the beamforming model may further include an output activation layer (e.g., Lambda layer V in
In some embodiments, training the beamforming model may include iteratively inputting the plurality of the first information into the beamforming model until an end condition is met. In some embodiments, for each iteration, the cloud apparatus may input the plurality of the first information in batches into the beamforming model, and may update a parameter(s) of the beamforming model (e.g., weights of the functional layers) according to a back propagation algorithm to improve a sum-rate of the plurality of UEs.
In some embodiments, the end condition may include one of the following: an improvement on a sum-rate of the plurality of UEs being less than or equal to an improvement threshold; and the number of iterations reaching a training threshold. In some embodiments, in response to the end condition being met, the cloud apparatus may determine whether a performance of the beamforming model satisfies a performance demand. The cloud apparatus may determine the completion of the training in response to determining that the performance of the beamforming model satisfies the performance demand. In some examples, the performance demand determination may be performed based on the test set as described above.
In some embodiments, in response to determining that the performance of the beamforming model fails to satisfy the performance demand, the cloud apparatus may perform at least one of the following: updating a parameter(s) of the beamforming model (e.g., weights of the functional layers) to satisfy the performance demand; and reconstructing the beamforming model and train the reconstructed beamforming model to satisfy the performance demand.
In operation 915, the cloud apparatus may transmit the beamforming matrix to the plurality of BSs. In some embodiments, the cloud apparatus may split the beamforming matrix into a plurality of beamforming sub-matrixes, each of which may be associated with a corresponding BS of the plurality of BSs. Transmitting the beamforming matrix to the plurality of BSs may include transmitting a beamforming sub-matrix of the plurality of beamforming sub-matrixes to a corresponding BS of the plurality of BSs. A BS can then perform a beamforming operation according to the received beamforming sub-matrix.
It should be appreciated by persons skilled in the art that the sequence of the operations in exemplary procedure 900 may be changed and some of the operations in exemplary procedure 900 may be eliminated or modified, without departing from the spirit and scope of the disclosure.
As shown in
Although in this figure, elements such as the at least one transceiver 1002 and processor 1006 are described in the singular, the plural is contemplated unless a limitation to the singular is explicitly stated. In some embodiments of the present application, the transceiver 1002 may be divided into two devices, such as a receiving circuitry and a transmitting circuitry. In some embodiments of the present application, the apparatus 1000 may further include an input device, a memory, and/or other components.
In some embodiments of the present application, the apparatus 1000 may be a UE. The transceiver 1002 and the processor 1006 may interact with each other to perform the operations with respect to the UE described in
In some embodiments of the present application, the apparatus 1000 may further include at least one non-transitory computer-readable medium.
For example, in some embodiments of the present disclosure, the non-transitory computer-readable medium may have stored thereon computer-executable instructions to cause the processor 1006 to implement the method with respect to the UE as described above. For example, the computer-executable instructions, when executed, cause the processor 1006 interacting with transceiver 1002 to perform the operations with respect to the UE described in
In some embodiments of the present disclosure, the non-transitory computer-readable medium may have stored thereon computer-executable instructions to cause the processor 1006 to implement the method with respect to the BS (e.g., BS 102 in
In some embodiments of the present disclosure, the non-transitory computer-readable medium may have stored thereon computer-executable instructions to cause the processor 1006 to implement the method with respect to the cloud apparatus (e.g., MBS 103 in
Those having ordinary skill in the art would understand that the operations or steps of a method described in connection with the aspects disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, a hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art. Additionally, in some aspects, the operations or steps of a method may reside as one or any combination or set of codes and/or instructions on a non-transitory computer-readable medium, which may be incorporated into a computer program product.
While this disclosure has been described with specific embodiments thereof, it is evident that many alternatives, modifications, and variations may be apparent to those skilled in the art. For example, various components of the embodiments may be interchanged, added, or substituted in other embodiments. Also, all of the elements of each figure are not necessary for the operation(s) of the disclosed embodiments. For example, one of ordinary skill in the art of the disclosed embodiments would be enabled to make and use the teachings of the disclosure by simply employing the elements of the independent claims. Accordingly, embodiments of the disclosure as set forth herein are intended to be illustrative, not limiting. Various changes may be made without departing from the spirit and scope of the disclosure.
In this document, the terms “includes,” “including,” or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that includes a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. An element proceeded by “a,” “an,” or the like does not, without more constraints, preclude the existence of additional identical elements in the process, method, article, or apparatus that includes the element. Also, the term “another” is defined as at least a second or more. The term “having” and the like, as used herein, are defined as “including.” Expressions such as “A and/or B” or “at least one of A and B” may include any and all combinations of words enumerated along with the expression. For instance, the expression “A and/or B” or “at least one of A and B” may include A, B, or both A and B. The wording “the first,” “the second” or the like is only used to clearly illustrate the embodiments of the present application, but is not used to limit the substance of the present application.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/CN2021/119102 | 9/17/2021 | WO |