Various example embodiments relate to an apparatus comprising at least one processor.
Further example embodiments relate to a method associated with an apparatus comprising at least one processor.
In wireless communication systems, channel estimation can be used to determine properties of a radio channel, e.g., to process data received over the radio channel. In some approaches, an objective of channel estimation is to determine, e.g., extract a channel vector ‘H’ from a received signal vector ‘Y’, e.g., in order to accurately decode a transmitted data signal ‘X’. In some approaches, channel prediction can be used, wherein properties of the radio channel are predicted.
Various example embodiments of the disclosure are set out by the independent claims.
Some examples relate to an apparatus, comprising at least one processor, and at least one memory storing instructions that, when executed by the at least one processor, cause the apparatus to: provide a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, provide a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction, determine the second channel prediction using at least the second machine learning model. In some examples, this may contribute to a comparatively precise channel prediction.
In some examples, the signal may, e.g., be a signal as transmitted by at least one transmitter, e.g., over a radio channel and received by a receiver, e.g., of a wireless communication system, or a signal derived from such received signal.
In some examples, the receiver may be a receiver for a wireless, e.g., cellular, communication system or network.
In some examples, the receiver may be a receiver for a terminal device, e.g., user equipment, for a wireless communication network, or Unmanned Aerial Vehicle (UAV), e.g., as a complementary assistance part of the conventional wireless communication network.
In some examples, the receiver may be a receiver for a network device, e.g., base station, e.g., gNB, for a wireless communication network.
In some examples, the signal may, e.g., be a signal as can be obtained by simulation, e.g., of a receiver, e.g., of a wireless communication system, or a signal derived from such received signal.
In some examples, the first channel prediction is a first, for example comparatively coarse, channel prediction for the radio channel associated with the received signal. In some examples, the second channel prediction is a second, for example comparatively fine, channel prediction for the radio channel associated with the received signal.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to at least temporarily provide the first information and the second channel prediction as the second information to the second machine learning model. In some examples, this enables to provide a feedback loop, e.g., with respect to the second machine learning model.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide first training data for training of at least the first machine learning model, wherein the first training data comprises an estimable part, a predictable part, and associated labels, train at least the first machine learning model based on the first training data and a first loss function.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide historic data of the at least one signal, to partition the historic data into the estimable part and the predictable part. Thus, in some examples, the historic data, which has, for example, been determined or generated in the past, may be used, e.g., for training of at least one of the machine learning models.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide second training data for training of the second machine learning model, wherein the second training data comprises the first training data and at least one of: a) the first channel prediction, or b) the second channel prediction, train the second machine learning model based on the second training data and a second loss function, which is different from the first loss function.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: use the first training data and the first channel prediction as obtained by the first machine learning model in a first training phase for training the second machine learning model, use the first training data and the second channel prediction as obtained by the second machine learning model in a subsequent second training phase for training the second machine learning model. In some examples, the second training phase may follow, for example directly, the first training phase. In some examples, the second training phase does not follow directly the first training phase.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: collect historic data of the at least one signal associated with at least two different domains of the following domains: a) a time domain, or b) a frequency domain, or c) a spatial domain, determine training data for training at least one of the first machine learning model or the second machine learning model based on the collected historic data.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide at least one machine learning system comprising an instance of the first machine learning model and an instance of the second machine learning model, provide a plurality of prediction paths to obtain a plurality of second channel predictions using the at least one machine learning system, each of the plurality of second channel predictions associated with a respective one of the plurality of prediction paths, combine at least two of the plurality of second channel predictions.
In some examples, the at least one machine learning system may, e.g., implement a respective instance of the first machine learning model and the second machine learning model.
In some examples, more than one machine learning system may be provided, wherein, for example, at least two machine learning systems may comprise different instances of the first machine learning model and the second machine learning model. In some examples, the at least two machine learning systems may be implemented by providing a single hardware and/or software basis for implementing the first and second machine learning models, and by providing different sets of parameters and/or hyper parameters for the respective machine learning models, the different sets of parameters and/or hyper parameters, e.g., associated with a respective one of the different machine learning systems.
In some examples, e.g., when providing a single hardware and/or software basis for implementing the first and second machine learning models, processing of aspects related to the different machine learning systems (e.g., performing at least one of training or inference) may, e.g., be performed in a time division duplexed manner.
In some examples, e.g., when providing a several hardware and/or software bases for implementing the first and second machine learning models, processing of aspects related to the different machine learning systems (e.g., performing at least one of training or inference) may, e.g., be performed in a at least partly temporally overlapping fashion, e.g., in parallel, e.g., simultaneously.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide a plurality of machine learning systems, train at least two machine learning systems of the plurality of machine learning systems differently from each other. In some examples, training the at least two machine learning systems of the plurality of machine learning systems differently from each other may comprise at least one of: a) using different training data for the training, or b) using different parameters, e.g., hyper parameters, for the at least two machine learning systems.
In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide the at least two of the plurality of second channel predictions to at least one further machine learning system.
In some examples, at least one of a) the first machine learning model or b) the second machine learning model is a dense neural network.
In some examples, at least one of a) the first machine learning model or b) the second machine learning model may comprise two or more layers.
In some examples, at least one of a) the first machine learning model or b) the second machine learning model may comprise two fully connected layers.
Some examples relate to an apparatus, comprising means for: providing a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, providing a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction, determining the second channel prediction using at least the second machine learning model.
In some examples, the means for providing the first machine learning model, providing the second machine learning model, determining the second channel prediction may comprise at least one processor, and at least one memory storing instructions that, when executed by the at least one processor, cause an apparatus to perform at least one of the aforementioned aspects of providing the first machine learning model, providing the second machine learning model, determining the second channel prediction.
In some examples, the means for providing the first machine learning model, providing the second machine learning model, determining the second channel prediction may, e.g., comprise circuitry configured to perform the aforementioned aspects.
Some examples relate to a method, comprising: providing a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, providing a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction, determining the second channel prediction using at least the second machine learning model.
Some examples relate to a device for a communication system, e.g., a wireless, e.g., cellular, communication system, the device comprising at least one apparatus according to the example embodiments.
Some examples relate to a computer program comprising instructions which, when executed by an apparatus, cause the apparatus to perform at least some aspects of the method according to the example embodiments.
In some examples, the computer program may be provided on a computer readable storage medium SM, e.g., a non-transitory computer readable medium.
Some examples relate to a data carrier signal carrying and/or characterizing the computer program according to the example embodiments.
Some examples, see
In some examples,
In some examples, the signal SIG may comprise reference signals, e.g., sounding reference signals, SRS, or demodulation reference signals, DMRS, e.g., according to some accepted specification.
In some examples, the receiver RX may be a receiver for a wireless, e.g., cellular, communication system or network.
In some examples, the receiver RX may be a receiver for a terminal device, e.g., user equipment, 10 for a wireless communication network.
In some examples, the receiver RX may be a receiver for a network device, e.g., base station, e.g., gNB, (not shown) for a wireless communication network.
In some examples, the signal SIG may, e.g., be a signal as can be obtained by simulation, e.g., of a receiver, e.g., of a wireless communication system, or a signal derived from such received signal.
In some examples,
In some examples,
In some examples,
In some examples,
In some examples,
In some examples,
In some examples,
In some examples,
In some examples,
In some examples, more than one machine learning system may be provided, wherein, for example, at least two machine learning systems may comprise different instances of the first machine learning model MLM-1 (
In some examples, e.g., when providing a single hardware and/or software basis for implementing the first and second machine learning models, e.g., using components 102, 104, 106 of the apparatus 100 according to some examples, processing of aspects related to the different machine learning systems (e.g., performing at least one of training or inference) may, e. g., be performed in a time division duplexed manner. In some examples, e.g., when providing several hardware and/or software bases for implementing the first and second machine learning models, e.g., using components 102, 104, 106 of the apparatus 100 according to some examples, processing of aspects related to the different machine learning systems (e.g., performing at least one of training or inference) may, e.g., be performed in a at least partly temporally overlapping fashion, e.g., in parallel, e.g., simultaneously.
In some examples,
In some examples,
In some examples,
In some examples, at least one of a) the first machine learning model MLM-1 or b) the second machine learning model MLM-2 may comprise two or more layers.
In some examples, at least one of a) the first machine learning model or b) the second machine learning model may comprise two fully connected layers.
In some other examples, at least one of the a) the first machine learning model or b) the second machine learning model may comprise another type and/or topology, e.g., convolutional neural network, CNN, or other type.
Some examples,
In some examples,
In some examples,
In some examples, a combination of the configurations 100, 100′ is also possible, e.g., providing the configuration 100 of
Some examples,
Some examples,
In some examples, the device 10 may, e.g., be a user equipment or other mobile device.
In the following, further example aspects and example embodiments are disclosed, which, in some examples, may be combined with each other and/or with at least one of the aforementioned example aspects or example embodiments.
As already explained above with reference to
In some examples,
In some examples, the first information I-1 may be provided in the form of an observation vector, e.g., according to yest=hest+z, wherein vector hest characterizes the ground truth channel response as an estimable part (e.g., after removing known information associated with the signal SIG, such as, e.g., components of the reference signals, e.g., SRS), and wherein vector z characterizes a noise term, e.g. characterizing Additive White Gaussian Noise (AWGN).
Note that in some examples, e.g., throughout the following description, ĥest denotes an estimation of the ground truth estimable channel response hest, and ĥpdt denotes a prediction of the ground truth predictable channel response hpdt.
In some examples, the first channel prediction CH-PRED-1 as provided by the first machine learning model MLM-1 is a scalar value, e.g., according to ĥpdt,0=WhpTyest, wherein WhpT characterizes a vector a multiplication of which with the observation vector yest yields the first channel prediction CH-PRED-1, e.g., the scalar value ĥpdt,0, which in some examples can be considered as a rough first channel prediction. In other words, in some examples, the first machine learning model MLM-1 can be configured and/or trained to perform or approximate the multiplication of the observation vector yest with the vector WhpT.
In some examples,
Thus, in some examples, the first training data DAT-TRAIN-1 (
In some examples, e.g., based on a given channel data set, which may, e.g., be obtained by collecting historic data associated with the signal SIG (
In some examples, the first machine learning model MLM-1 can be trained using the observation vector yest and the label vector ylab, e.g., as the first training data DAT-TRAIN-1, see also block 222 of
e. g., so that the training of the first machine learning model MLM-1 effects “determining” the vector Whp (e.g., by learning parameters of the first machine learning model MLM-1) for minimizing the loss as defined by the first loss function LF-1, e.g., based on the observations yest and labels ylab, e.g., according to a machine learning process. In some examples, e.g., when the training of the first machine learning model MLM-1, e.g., a searching procedure of the vector Whp, reaches convergence, e.g., according to a predetermined degree (e.g., characterized by the first loss function LF-1 yielding values below a first threshold), the first machine learning model MLM-1 can be considered to be trained, and in some examples, the parameters of the so trained first machine learning model MLM-1 may be stored, e.g., for future inference, see, for example,
In some examples
In some examples, the first, e.g., rough, channel prediction ĥpdt,o may be used, e.g., treated, as a virtual observation, e.g., for determining a further, e.g., new, observation vector for the second machine learning model MLM-2, e. g., characterized by {tilde over (y)}est=[yestT ĥpdt,i−1]T, where ĥpdt,i−1 denotes a virtual observation of an ith iteration (i≥1), and ĥpdt,0, e. g., as explained above, is an initial value, e.g., for the iteration i=1.
In some examples, physically true observations as, e.g., characterized by the observation vector yest, e.g., as opposed to the first channel prediction ĥpdt,0, are dominant in the further observation vector {tilde over (y)}est, e.g., in terms of element numbers of the further observation vector {tilde over (y)}est.
In some examples, the further observation vector {tilde over (y)}est may, e.g., together with the labels ylab, be used to train the second machine learning model MLM-2, e.g., as the second training data DAT-TRAIN-2, see, for example, block 230 of
In some examples, the second loss function LF-2 may, e.g., be characterized by
wherein index i denotes an i-th iteration, (i≥1). Thus, using the second loss function LF-2, in some examples, the second machine learning model MLM-2 may be trained to determine, e.g., implement, e.g., learn and compute, a matrix Whh,i, e.g., for minimizing the loss as characterized by the second loss function LF-2, e.g., with the available observations and labels of corresponding iterations. In some examples, the parameters of the second machine learning model MLM-2 may be stored for future inference, e.g., when the training of the second machine learning model MLM-2 reaches its convergence.
In some examples, using the second machine learning model MLM-2, which e.g. implements the matrix Whh,i, e.g., based on learned parameters of the second machine learning model MLM-2 for an ith iteration, a channel estimation ĥestT and a channel prediction ĥpdt,i can be obtained, e.g. as characterized by [ĥestT ĥpdt,i]T=Whh,i[yestT ĥpdt,i−1]T. In other words, in some examples, the vector [yestT ĥpdt,i−1]T may form the second information I-2, and the second machine learning model MLM-2 is configure to determine the vector [ĥestT ĥpdt,i]T based on the second information, wherein, for example, element ĥpdt,i may represent the second channel prediction CH-PRED-2.
In some examples, the second channel prediction CH-PRED-2, e. g., element ĥpdt,i, may be fed back, e.g., as a refined virtual observation, see the switch SW1 of
In some examples, a further, e.g., new, iteration of training the second machine learning model MLM-2 may be performed based on the feedback, e.g., to further improve a channel prediction, e.g., with the updated observations, e.g., the updated channel prediction ĥpdt,i and the for example original, observation yest.
In some examples, e.g., in an inference or testing stage or phase, e.g., subsequent to the training of the machine learning models MLM-1, MLM-2, the trained machine learning models MLM-1, MLM-2 may adapt at least one of the vector whp or the matrix whh,i, e.g., to any observation yest provided as input information.
In some examples, both a channel estimation ĥest and a channel prediction ĥpdt,i may be obtained using the machine learning models MLM-1, MLM-2, e.g., if i iterations (i≥1) are used.
In some examples, the prediction at the output of the second machine learning model MLM-2 may be improved by implementing the matrix Whh,i, e.g., because an element-wise cross-correlation as characterized by the matrix Whh may be exploited to finetune the prediction, e.g., with respect to the estimation ĥest.
In some examples, the second machine learning model MLM-2 may, e.g., comprise a topology as disclosed in the following reference: Yejian Chen; Jafar Mohammadi; Stefan Wesemann; Thorsten Wild; “Turbo-AI, Part I: Iterative Machine Learning Based Channel Estimation for 2D Massive Arrays,” in Proc. 2021 IEEE 93rd Veh. Technol. Conf. (VTC′ 21 Spring), Helsinki, Finland, April 2021.
As mentioned above, in some examples,
In some examples, similarly, the second machine learning model MLM-2 may, e.g., be a second dense neural network, e.g., comprising two fully connected layers. In some examples, the second machine learning model MLM-2 may be configured to receive as input L2+L many complex numbers, e. g., characterizing the observation vector associated with the second information I-2 and, optionally (e.g., for a training), the labels LAB.
In some examples, at least one of the first machine learning model MLM-1 or the second machine learning model MLM-2 may use an activation function of the SoftMax type.
As mentioned above, in some examples, historic data DAT-HIST of the at least one signal SIG associated with at least two different domains of the following domains: a) a time domain D-TIM, or b) a frequency domain D-FREQ, or c) a spatial domain D-SPAT may be collected, see
In some examples,
In some examples,
In some examples, using samples in the prediction path from frequency and spatial domains a2, a3 (e.g., additionally to the time domain a1) may improve a prediction, e.g., due to an increasing correlation.
In some examples,
In some examples, training data, e.g., a training data set, e.g., comprising an observation data set and a label data set, may be determined by, for example, repeatedly, picking out data in accordance with the prediction path “1”, “2”, . . . , “13” as illustrated by the example of
In some examples, the machine learning models MLM-1, MLM-2 (
In some examples,
In some examples,
In some examples, the first information I-1′ provided as input to the instances MLM-1′, MLM-2′ of the first machine learning system ML-SYS-1 may, e.g., be associated with first frequency resources, e.g., one or more first subcarriers, whereas the first information I-1″ provided as input to the instances MLM-1″, MLM-2″ of the second machine learning system ML-SYS-2 may, e.g., be associated with second frequency resources, e.g., one or more second subcarriers, which are different from the first frequency resources. In some examples, the first subcarrier(s) and the second subcarrier(s) may be neighboring subcarriers.
In some examples, e.g., as neighboring subcarriers in the frequency domain may have a comparatively high correlation, a combination, e.g., an averaging, of both channel predictions, e.g., the respective second channel prediction ĥpdt,1, ĥpdt,2, may introduce a diversity gain, e.g., with at least one of a) independent noise realizations, or b) independent channel estimation errors, or c) independent channel prediction errors. In some examples, the combination (see, for example also block 262 of
In some examples, the first information I-1′, I-1″ provided as input to machine learning systems ML-SYS-1, ML-SYS-2 may, e.g., be associated with at least two different domains (e.g., temporal, spatial, frequency), e.g., at least similar to the example prediction path of
In some examples,
In some examples, a diversity order of 2 or higher, which in some examples may, e.g., be attained by combining the outputs of several machine learning systems ML-SYS-1, ML-SYS-2, may improve a prediction performance, e.g., for both low speed cases (e.g., associated with a comparatively low mobility, e.g., of the receiver RX) and high speed cases (e.g., associated with a comparatively high mobility).
In some examples, subcarrier combining may be performed, e.g., using a configuration as depicted by
In some examples, as mentioned above, reference signals, e.g., sounding reference signals, SRS, may be used, e.g., to provide observations, e.g., as input data to the machine learning model(s) MLM-1, MLM-2. In some examples, other reference signals may be used alternatively or additionally, e.g., demodulation reference signals (DMRS), e.g., in cases with a channel reciprocity of uplink/downlink channels (e.g., with time division duplexing, TDD, scenarios). In some examples, using different types of reference signals, such as, e.g., SRS and DMRS, a prediction can be further enhanced or improved.
Returning to
Note that, in some examples, the machine learning models MLM-1, MLM-2 of at least some, for example all, machine learning systems as depicted by
In some examples,
In some examples, the respective output of the machine learning systems ML-SYS-n, . . . , ML-SYS-n+5 is combined by a combiner COMB (see, for example, also block 262 of
In some examples, the configuration of
In some examples,
In some examples,
In some examples, the principle according to example embodiments enables to use reference signals, e.g., of the SRS and/or DMRS type, wherein, for example, comparatively precise second channel predictions may be obtained based on an SRS interval of, e.g., 10 ms.
In some examples, the principle according to example embodiments enables to perform at least one of precoding or beamforming, e.g., for a downlink direction, e.g., on a symbol-level, which, in some examples, may, for example significantly, impact, e.g., improve, both multiuser and single user MIMO, so that, in some examples, a downlink throughput may be comparatively close to an achievable downlink capacity.
Some examples,
In some examples,
Some examples,
In some examples, the principle according to example embodiments enables to address channel aging. In some examples, channel aging means a nonvanishing delay between a first time instant, when, for example, a reception of a reference signal, e.g., SRS, in an uplink direction, and a respective channel estimation happen, and a second time instant, when a downlink transmission happens, e.g., based on the channel estimation. In some examples, channel aging can lead to certain inconsistencies with respect to a radio channel, e.g., a channel response in the real world can deviate significantly from a conventionally estimated or predicted channel (as, e.g., obtained through at least one of uplink channel state information, UL CSI, or uplink sounding reference signals, UL SRS).
In some examples, channel aging can be attributed to a processing latency, e.g., in UL channel estimation and downlink, DL, precoding computation and to a signaling latency, e.g., limited by an SRS reporting interval. In some conventional approaches, channel aging can thus lead to performance degradation, e.g., since an outdated channel may introduce mismatched DL beamforming vector or precoding matrix and start to constrain a DL throughput. In some examples, channel aging may happen, when a user of a mobile device 10 (
| Number | Date | Country | Kind |
|---|---|---|---|
| 23214930.2 | Dec 2023 | EP | regional |