APPARATUS COMPRISING AT LEAST ONE PROCESSOR

Information

  • Patent Application
  • 20250193054
  • Publication Number
    20250193054
  • Date Filed
    December 06, 2024
    10 months ago
  • Date Published
    June 12, 2025
    4 months ago
Abstract
An apparatus, including at least one processor, and at least one memory storing instructions that, when executed by the at least one processor, cause the apparatus to: provide a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, provide a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily including the first channel prediction, determine the second channel prediction using at least the second machine learning model.
Description
FIELD OF THE DISCLOSURE

Various example embodiments relate to an apparatus comprising at least one processor.


Further example embodiments relate to a method associated with an apparatus comprising at least one processor.


BACKGROUND

In wireless communication systems, channel estimation can be used to determine properties of a radio channel, e.g., to process data received over the radio channel. In some approaches, an objective of channel estimation is to determine, e.g., extract a channel vector ‘H’ from a received signal vector ‘Y’, e.g., in order to accurately decode a transmitted data signal ‘X’. In some approaches, channel prediction can be used, wherein properties of the radio channel are predicted.


SUMMARY

Various example embodiments of the disclosure are set out by the independent claims.


Some examples relate to an apparatus, comprising at least one processor, and at least one memory storing instructions that, when executed by the at least one processor, cause the apparatus to: provide a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, provide a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction, determine the second channel prediction using at least the second machine learning model. In some examples, this may contribute to a comparatively precise channel prediction.


In some examples, the signal may, e.g., be a signal as transmitted by at least one transmitter, e.g., over a radio channel and received by a receiver, e.g., of a wireless communication system, or a signal derived from such received signal.


In some examples, the receiver may be a receiver for a wireless, e.g., cellular, communication system or network.


In some examples, the receiver may be a receiver for a terminal device, e.g., user equipment, for a wireless communication network, or Unmanned Aerial Vehicle (UAV), e.g., as a complementary assistance part of the conventional wireless communication network.


In some examples, the receiver may be a receiver for a network device, e.g., base station, e.g., gNB, for a wireless communication network.


In some examples, the signal may, e.g., be a signal as can be obtained by simulation, e.g., of a receiver, e.g., of a wireless communication system, or a signal derived from such received signal.


In some examples, the first channel prediction is a first, for example comparatively coarse, channel prediction for the radio channel associated with the received signal. In some examples, the second channel prediction is a second, for example comparatively fine, channel prediction for the radio channel associated with the received signal.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to at least temporarily provide the first information and the second channel prediction as the second information to the second machine learning model. In some examples, this enables to provide a feedback loop, e.g., with respect to the second machine learning model.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide first training data for training of at least the first machine learning model, wherein the first training data comprises an estimable part, a predictable part, and associated labels, train at least the first machine learning model based on the first training data and a first loss function.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide historic data of the at least one signal, to partition the historic data into the estimable part and the predictable part. Thus, in some examples, the historic data, which has, for example, been determined or generated in the past, may be used, e.g., for training of at least one of the machine learning models.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide second training data for training of the second machine learning model, wherein the second training data comprises the first training data and at least one of: a) the first channel prediction, or b) the second channel prediction, train the second machine learning model based on the second training data and a second loss function, which is different from the first loss function.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: use the first training data and the first channel prediction as obtained by the first machine learning model in a first training phase for training the second machine learning model, use the first training data and the second channel prediction as obtained by the second machine learning model in a subsequent second training phase for training the second machine learning model. In some examples, the second training phase may follow, for example directly, the first training phase. In some examples, the second training phase does not follow directly the first training phase.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: collect historic data of the at least one signal associated with at least two different domains of the following domains: a) a time domain, or b) a frequency domain, or c) a spatial domain, determine training data for training at least one of the first machine learning model or the second machine learning model based on the collected historic data.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide at least one machine learning system comprising an instance of the first machine learning model and an instance of the second machine learning model, provide a plurality of prediction paths to obtain a plurality of second channel predictions using the at least one machine learning system, each of the plurality of second channel predictions associated with a respective one of the plurality of prediction paths, combine at least two of the plurality of second channel predictions.


In some examples, the at least one machine learning system may, e.g., implement a respective instance of the first machine learning model and the second machine learning model.


In some examples, more than one machine learning system may be provided, wherein, for example, at least two machine learning systems may comprise different instances of the first machine learning model and the second machine learning model. In some examples, the at least two machine learning systems may be implemented by providing a single hardware and/or software basis for implementing the first and second machine learning models, and by providing different sets of parameters and/or hyper parameters for the respective machine learning models, the different sets of parameters and/or hyper parameters, e.g., associated with a respective one of the different machine learning systems.


In some examples, e.g., when providing a single hardware and/or software basis for implementing the first and second machine learning models, processing of aspects related to the different machine learning systems (e.g., performing at least one of training or inference) may, e.g., be performed in a time division duplexed manner.


In some examples, e.g., when providing a several hardware and/or software bases for implementing the first and second machine learning models, processing of aspects related to the different machine learning systems (e.g., performing at least one of training or inference) may, e.g., be performed in a at least partly temporally overlapping fashion, e.g., in parallel, e.g., simultaneously.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide a plurality of machine learning systems, train at least two machine learning systems of the plurality of machine learning systems differently from each other. In some examples, training the at least two machine learning systems of the plurality of machine learning systems differently from each other may comprise at least one of: a) using different training data for the training, or b) using different parameters, e.g., hyper parameters, for the at least two machine learning systems.


In some examples, the instructions, when executed by the at least one processor, cause the apparatus to: provide the at least two of the plurality of second channel predictions to at least one further machine learning system.


In some examples, at least one of a) the first machine learning model or b) the second machine learning model is a dense neural network.


In some examples, at least one of a) the first machine learning model or b) the second machine learning model may comprise two or more layers.


In some examples, at least one of a) the first machine learning model or b) the second machine learning model may comprise two fully connected layers.


Some examples relate to an apparatus, comprising means for: providing a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, providing a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction, determining the second channel prediction using at least the second machine learning model.


In some examples, the means for providing the first machine learning model, providing the second machine learning model, determining the second channel prediction may comprise at least one processor, and at least one memory storing instructions that, when executed by the at least one processor, cause an apparatus to perform at least one of the aforementioned aspects of providing the first machine learning model, providing the second machine learning model, determining the second channel prediction.


In some examples, the means for providing the first machine learning model, providing the second machine learning model, determining the second channel prediction may, e.g., comprise circuitry configured to perform the aforementioned aspects.


Some examples relate to a method, comprising: providing a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, providing a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction, determining the second channel prediction using at least the second machine learning model.


Some examples relate to a device for a communication system, e.g., a wireless, e.g., cellular, communication system, the device comprising at least one apparatus according to the example embodiments.


Some examples relate to a computer program comprising instructions which, when executed by an apparatus, cause the apparatus to perform at least some aspects of the method according to the example embodiments.


In some examples, the computer program may be provided on a computer readable storage medium SM, e.g., a non-transitory computer readable medium.


Some examples relate to a data carrier signal carrying and/or characterizing the computer program according to the example embodiments.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A schematically depicts a simplified block diagram according to some examples,



FIG. 1B schematically depicts a simplified block diagram according to some examples,



FIG. 2 schematically depicts a simplified block diagram according to some examples,



FIG. 3 schematically depicts a simplified flow chart according to some examples,



FIG. 4 schematically depicts a simplified flow chart according to some examples,



FIG. 5 schematically depicts a simplified flow chart according to some examples,



FIG. 6 schematically depicts a simplified block diagram according to some examples,



FIG. 7 schematically depicts a simplified flow chart according to some examples,



FIG. 8 schematically depicts a simplified flow chart according to some examples,



FIG. 9 schematically depicts a simplified flow chart according to some examples,



FIG. 10 schematically depicts a simplified flow chart according to some examples,



FIG. 11 schematically depicts a simplified flow chart according to some examples,



FIG. 12 schematically depicts a simplified flow chart according to some examples,



FIG. 13A schematically depicts aspects of training of a machine learning system according to some examples,



FIG. 13B schematically depicts aspects of using a machine learning system according to some examples,



FIG. 14 schematically depicts aspects of a prediction path according to some examples,



FIG. 15 schematically depicts aspects of training data according to some examples,



FIG. 16 schematically depicts aspects of combining according to some examples,



FIG. 17 schematically depicts aspects of combining according to some examples,



FIG. 18 schematically depicts aspects of resources according to some examples,



FIG. 19 schematically depicts a block diagram according to some examples.





DESCRIPTION OF SOME EXAMPLE EMBODIMENTS

Some examples, see FIG. 1A, 2, 3, relate to an apparatus 100, comprising at least one processor 102, and at least one memory 104 storing instructions 106 that, when executed by the at least one processor 102, cause the apparatus 100 to: provide 200 (FIG. 3) a first machine learning model MLM-1 configured to receive first information I-1 associated with at least one signal SIG and to provide a first channel prediction CH-PRED-1 based on the first information I-1, provide 202 a second machine learning model MLM-2 configured to receive second information I-2 and to provide a second channel prediction CH-PRED-2 based on the second information I-2, the second information I-2 comprising the first information I-1 and at least temporarily comprising the first channel prediction CH-PRED-1, determine 204 the second channel prediction CH-PRED-2 using at least the second machine learning model MLM-2. In some examples, this may contribute to a comparatively precise channel prediction.


In some examples, FIG. 2, the signal SIG may, e.g., be a signal as transmitted by at least one transmitter (not shown), e.g., over a radio channel and received by a receiver RX, e.g., of a wireless communication system, or a signal derived from such received signal SIG.


In some examples, the signal SIG may comprise reference signals, e.g., sounding reference signals, SRS, or demodulation reference signals, DMRS, e.g., according to some accepted specification.


In some examples, the receiver RX may be a receiver for a wireless, e.g., cellular, communication system or network.


In some examples, the receiver RX may be a receiver for a terminal device, e.g., user equipment, 10 for a wireless communication network.


In some examples, the receiver RX may be a receiver for a network device, e.g., base station, e.g., gNB, (not shown) for a wireless communication network.


In some examples, the signal SIG may, e.g., be a signal as can be obtained by simulation, e.g., of a receiver, e.g., of a wireless communication system, or a signal derived from such received signal.


In some examples, FIG. 2, the first channel prediction CH-PRED-1 is a first, for example comparatively coarse, channel prediction for the radio channel associated with the received signal SIG (e.g., the radio channel through which the signal SIG has been transmitted). In some examples, the second channel prediction CH-PRED-2 is a second, for example comparatively fine, channel prediction for the radio channel associated with the received signal SIG. In some examples, the second channel prediction CH-PRED-2 may be used for a future processing, e.g., of signals received (and/or to be transmitted) over the radio channel, e.g., similar to the signal SIG.


In some examples, FIG. 4, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to at least temporarily provide 210 the first information I-1 and the second channel prediction CH-PRED-2 as the second information I-2 to the second machine learning model MLM-2 (FIG. 2). In some examples, this enables to provide a feedback loop, e.g., with respect to the second machine learning model MLM-2. The optional block 212 of FIG. 4 symbolizes determining, according to some examples, the second channel prediction CH-PRED-2 using at least the second machine learning model MLM-2.


In some examples, FIG. 5, 6, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to: provide 220 first training data DAT-TRAIN-1 for training of at least the first machine learning model MLM-1, wherein the first training data DAT-TRAIN-1 comprises an estimable part TRAIN-EST, a predictable part TRAIN-PRED, and associated labels TRAIN-LAB, train 222 at least the first machine learning model MLM-1 based on the first training data DAT-TRAIN-1 and a first loss function LF-1.


In some examples, FIG. 7, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to: provide 225 historic data DAT-HIST of the at least one signal SIG (or a signal derived from the at least one signal SIG), to partition 227 the historic data DAT-HIST into the estimable part TRAIN-EST and the predictable part TRAIN-PRED. Thus, in some examples, the historic data DAT-HIST, which has, for example, been determined or generated in the past, may be used, e.g., for training of at least one of the machine learning models MLM-1, MLM-2.


In some examples, FIG. 8, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to: provide 230 second training data DAT-TRAIN-2 for training of the second machine learning model MLM-2 (FIG. 2), wherein the second training data DAT-TRAIN-2 comprises the first training data DAT-TRAIN-1 and at least one of: a) the first channel prediction CH-PRED-1, or b) the second channel prediction CH-PRED-2, train 232 the second machine learning model MLM-2 based on the second training data DAT-TRAIN-2 and a second loss function LF-2, which is different from the first loss function LF-1.


In some examples, FIG. 9, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to: use 240 the first training data DAT-TRAIN-1 and the first channel prediction CH-PRED-1 as obtained by the first machine learning model MLM-1 in a first training phase PH-TRAIN-1 for training the second machine learning model MLM-2, use 242 the first training data DAT-TRAIN-1 and the second channel prediction CH-PRED-2 as obtained by the second machine learning model MLM-2 in a subsequent second training phase PH-TRAIN-2 for training the second machine learning model MLM-2. In some examples, the second training phase PH-TRAIN-2 may follow, for example directly, the first training phase PH-TRAIN-2. In some examples, the second training phase does not follow directly the first training phase.


In some examples, FIG. 10, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to: collect 250 historic data DAT-HIST (e.g., in some examples, the historic data mentioned above) of the at least one signal SIG associated with at least two different domains of the following domains: a) a time domain D-TIM, or b) a frequency domain D-FREQ, or c) a spatial domain D-SPAT, determine 252 training data DAT-TRAIN for training at least one of the first machine learning model MLM-1 or the second machine learning model MLM-2 based on the collected historic data. In some examples, at least parts of the first training data DAT-TRAIN-1 and/or of the second training data DAT-TRAIN-2 may be determined based on the training data DAT-TRAIN as obtained according to block 252 of FIG. 10. Further details and examples regarding historic data associated with different domains D-TIM, D-FREQ, D-SPAT are provided further below with reference to FIG. 14.


In some examples, FIG. 11, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to: provide 260 at least one machine learning system ML-SYS (see also FIG. 2) comprising an instance of the first machine learning model MLM-1 and an instance of the second machine learning model MLM-2, provide 261 a plurality PP of prediction paths to obtain a plurality CH-PRED-2′ of second channel predictions using the at least one machine learning system ML-SYS, each of the plurality CH-PRED-2′ of second channel predictions associated with a respective one of the plurality PP of prediction paths, combine 262 at least two of the plurality of second channel predictions. In some examples, this enables to further refine second channel predictions. Further details regarding aspects of combining according to further examples are explained further below with reference to FIG. 16, 17.


In some examples, FIG. 2, the at least one machine learning system ML-SYS may, e.g., implement a respective instance of the first machine learning model MLM-1 and the second machine learning model MLM-2.


In some examples, more than one machine learning system may be provided, wherein, for example, at least two machine learning systems may comprise different instances of the first machine learning model MLM-1 (FIG. 2) and the second machine learning model MLM-2. In some examples, the at least two machine learning systems may be implemented by providing a single hardware and/or software basis for implementing the first and second machine learning models, e.g., using components 102, 104, 106 of the apparatus 100 according to some examples, and by providing different sets of parameters (e.g., weights, and the like) and/or hyper parameters for the respective machine learning models, the different sets of parameters and/or hyper parameters, e.g., associated with a respective one of the different machine learning systems.


In some examples, e.g., when providing a single hardware and/or software basis for implementing the first and second machine learning models, e.g., using components 102, 104, 106 of the apparatus 100 according to some examples, processing of aspects related to the different machine learning systems (e.g., performing at least one of training or inference) may, e. g., be performed in a time division duplexed manner. In some examples, e.g., when providing several hardware and/or software bases for implementing the first and second machine learning models, e.g., using components 102, 104, 106 of the apparatus 100 according to some examples, processing of aspects related to the different machine learning systems (e.g., performing at least one of training or inference) may, e.g., be performed in a at least partly temporally overlapping fashion, e.g., in parallel, e.g., simultaneously.


In some examples, FIG. 12, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to: provide 265 a plurality ML-SYS' of machine learning systems, train 267 at least two machine learning systems of the plurality ML-SYS' of machine learning systems differently from each other. In some examples, training 267 the at least two machine learning systems of the plurality of machine learning systems differently from each other may comprise at least one of: a) using different training data for the training, or b) using different parameters, e.g., hyper parameters, for the at least two machine learning systems.


In some examples, FIG. 11, the instructions 106, when executed by the at least one processor 102, cause the apparatus 100 to: provide 263 the at least two of the plurality CH-PRED-2′ of second channel predictions to at least one further machine learning system. Further details and examples regarding providing second channel predictions to at least one further machine learning system are provided further below with reference to FIG. 17.


In some examples, FIG. 2, at least one of a) the first machine learning model MLM-1 or b) the second machine learning model MLM-2 is a neural network.


In some examples, at least one of a) the first machine learning model MLM-1 or b) the second machine learning model MLM-2 may comprise two or more layers.


In some examples, at least one of a) the first machine learning model or b) the second machine learning model may comprise two fully connected layers.


In some other examples, at least one of the a) the first machine learning model or b) the second machine learning model may comprise another type and/or topology, e.g., convolutional neural network, CNN, or other type.


Some examples, FIG. 1B, relate to an apparatus 100′, comprising means 102′ for: providing 200 (FIG. 3) a first machine learning model MLM-1 configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, providing 202 a second machine learning model MLM-2 configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction, determining 204 the second channel prediction using at least the second machine learning model.


In some examples, FIG. 1B, the means 102′ for providing the first machine learning model, providing the second machine learning model, determining the second channel prediction may comprise at least one processor 102 (FIG. 1A), and at least one memory 104 storing instructions 106 that, when executed by the at least one processor 102, cause the apparatus 100′ to perform at least one of the aforementioned aspects of providing 200 the first machine learning model, providing 202 the second machine learning model, determining 204 the second channel prediction.


In some examples, FIG. 1B, the means 102′ for providing the first machine learning model, providing the second machine learning model, determining the second channel prediction may, e.g., comprise circuitry 104′ configured to perform at least one of the aforementioned aspects 200, 202, 204.


In some examples, a combination of the configurations 100, 100′ is also possible, e.g., providing the configuration 100 of FIG. 1A with some circuitry 104′ (FIG. 1B), e.g., for hardware acceleration of at least one of the machine learning models MLM-1, MLM-2.


Some examples, FIG. 3, relate to a method, comprising: providing 200 a first machine learning model MLM-1 configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information, providing 202 a second machine learning model MLM-2 configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction, determining 204 the second channel prediction using at least the second machine learning model MLM-2.


Some examples, FIG. 2, relate to a device 10 for a communication system, e.g., a wireless, e.g., cellular, communication system, the device 10 comprising at least one apparatus 100, 100′ according to the example embodiments.


In some examples, the device 10 may, e.g., be a user equipment or other mobile device.


In the following, further example aspects and example embodiments are disclosed, which, in some examples, may be combined with each other and/or with at least one of the aforementioned example aspects or example embodiments.



FIG. 13A schematically depicts aspects of a machine learning system ML-SYS according to some examples, e.g., related to a training phase. FIG. 13B schematically depicts aspects of the machine learning system ML-SYS according to some examples, e.g., related to an inference phase.


As already explained above with reference to FIG. 2, the machine learning system ML-SYS comprises the first machine learning model MLM-1, e.g., a first dense neural network, and the second machine learning model MLM-2, e.g., a second dense neural network.


In some examples, FIG. 13A, 13B, the first machine learning model MLM-1 can be configured, e.g., trained, to receive first information I-1 associated with the at least one signal SIG (FIG. 2) and to provide the first channel prediction CH-PRED-1 based on the first information I-1.


In some examples, the first information I-1 may be provided in the form of an observation vector, e.g., according to yest=hest+z, wherein vector hest characterizes the ground truth channel response as an estimable part (e.g., after removing known information associated with the signal SIG, such as, e.g., components of the reference signals, e.g., SRS), and wherein vector z characterizes a noise term, e.g. characterizing Additive White Gaussian Noise (AWGN).


Note that in some examples, e.g., throughout the following description, ĥest denotes an estimation of the ground truth estimable channel response hest, and ĥpdt denotes a prediction of the ground truth predictable channel response hpdt.


In some examples, the first channel prediction CH-PRED-1 as provided by the first machine learning model MLM-1 is a scalar value, e.g., according to ĥpdt,0=WhpTyest, wherein WhpT characterizes a vector a multiplication of which with the observation vector yest yields the first channel prediction CH-PRED-1, e.g., the scalar value ĥpdt,0, which in some examples can be considered as a rough first channel prediction. In other words, in some examples, the first machine learning model MLM-1 can be configured and/or trained to perform or approximate the multiplication of the observation vector yest with the vector WhpT.


In some examples, FIG. 13A, e.g., for a training of the first machine learning model MLM-1, labels LAB may be provided to the first machine learning model MLM-1, e.g., additionally to the first information I-1, e.g., in the form of a label vector ylab=[hestT hpdt]T, wherein hest characterizes the estimable part, as mentioned above, e.g., similar to element TRAIN-EST of FIG. 6, and wherein hpdt characterizes a predictable part, e.g., similar to element TRAIN-PRED of FIG. 6.


Thus, in some examples, the first training data DAT-TRAIN-1 (FIG. 6) for training of the first machine learning model MLM-1 (FIG. 13A) may, e.g., comprise the observation vector yest and the label vector ylab.


In some examples, e.g., based on a given channel data set, which may, e.g., be obtained by collecting historic data associated with the signal SIG (FIG. 2), a training data structure may be provided for training of the first machine learning model MLM-1, the training data structure having the estimable part hest (also see element TRAIN-EST of FIG. 6, for example), which is assumed to have been received in the past, e.g., at time instants t0, t−1, t−2, etc., and the predictable part hpdt (also see element TRAIN-PRED of FIG. 6), which in some examples is assumed to happen in the future, e.g. at a time instant t1. Note that, as, in some examples, all the data of the given channel data set has been already physically generated in the past, in some examples, the respective portions of data hest, hpdt may be, e.g., virtually, sorted out, e.g., to fulfill the timing criteria in the sense of the estimable part and the predictable part, also see the partitioning block 227 of FIG. 7.


In some examples, the first machine learning model MLM-1 can be trained using the observation vector yest and the label vector ylab, e.g., as the first training data DAT-TRAIN-1, see also block 222 of FIG. 5, and the first loss function LF-1. In some examples, a conventional training procedure for training a neural network based on the first training data and the first loss function may be used for the training 222. In some examples, the first loss function LF-1 may, e.g., be characterized by








min

w

h

p









w

h

p

T



y
est


-

h
pdt




2


,




e. g., so that the training of the first machine learning model MLM-1 effects “determining” the vector Whp (e.g., by learning parameters of the first machine learning model MLM-1) for minimizing the loss as defined by the first loss function LF-1, e.g., based on the observations yest and labels ylab, e.g., according to a machine learning process. In some examples, e.g., when the training of the first machine learning model MLM-1, e.g., a searching procedure of the vector Whp, reaches convergence, e.g., according to a predetermined degree (e.g., characterized by the first loss function LF-1 yielding values below a first threshold), the first machine learning model MLM-1 can be considered to be trained, and in some examples, the parameters of the so trained first machine learning model MLM-1 may be stored, e.g., for future inference, see, for example, FIG. 13B.


In some examples FIG. 13A, a rough channel prediction, e.g., in the sense of the first channel prediction CH-PRED-1, can be determined, e.g., computed, e.g., based on the first machine learning model MLM-1, e.g., as ĥpdt,0=WhpTyest.


In some examples, the first, e.g., rough, channel prediction ĥpdt,o may be used, e.g., treated, as a virtual observation, e.g., for determining a further, e.g., new, observation vector for the second machine learning model MLM-2, e. g., characterized by {tilde over (y)}est=[yestT ĥpdt,i−1]T, where ĥpdt,i−1 denotes a virtual observation of an ith iteration (i≥1), and ĥpdt,0, e. g., as explained above, is an initial value, e.g., for the iteration i=1.


In some examples, physically true observations as, e.g., characterized by the observation vector yest, e.g., as opposed to the first channel prediction ĥpdt,0, are dominant in the further observation vector {tilde over (y)}est, e.g., in terms of element numbers of the further observation vector {tilde over (y)}est.


In some examples, the further observation vector {tilde over (y)}est may, e.g., together with the labels ylab, be used to train the second machine learning model MLM-2, e.g., as the second training data DAT-TRAIN-2, see, for example, block 230 of FIG. 8. In some examples, FIG. 13A, 13B, the further observation vector {tilde over (y)}est may form the second information I-2, as explained above.


In some examples, the second loss function LF-2 may, e.g., be characterized by








min

W

hh
,
i









W


h

h

,
i


[




y
est







h
ˆ


pdt
,

i
-
1






]

-

[




h
est






h
pdt




]




2


,




wherein index i denotes an i-th iteration, (i≥1). Thus, using the second loss function LF-2, in some examples, the second machine learning model MLM-2 may be trained to determine, e.g., implement, e.g., learn and compute, a matrix Whh,i, e.g., for minimizing the loss as characterized by the second loss function LF-2, e.g., with the available observations and labels of corresponding iterations. In some examples, the parameters of the second machine learning model MLM-2 may be stored for future inference, e.g., when the training of the second machine learning model MLM-2 reaches its convergence.


In some examples, using the second machine learning model MLM-2, which e.g. implements the matrix Whh,i, e.g., based on learned parameters of the second machine learning model MLM-2 for an ith iteration, a channel estimation ĥestT and a channel prediction ĥpdt,i can be obtained, e.g. as characterized by [ĥestT ĥpdt,i]T=Whh,i[yestT ĥpdt,i−1]T. In other words, in some examples, the vector [yestT ĥpdt,i−1]T may form the second information I-2, and the second machine learning model MLM-2 is configure to determine the vector [ĥestT ĥpdt,i]T based on the second information, wherein, for example, element ĥpdt,i may represent the second channel prediction CH-PRED-2.


In some examples, the second channel prediction CH-PRED-2, e. g., element ĥpdt,i, may be fed back, e.g., as a refined virtual observation, see the switch SW1 of FIG. 13A, 13B. Thus, in some examples, the second information I-2 may at least temporarily comprise the first information I-1, e.g., in the form of the observation vector yest, and the first channel prediction CH-PRED-1, e.g., in the form of the element ĥpdt,0, and the second information I-2 may at least temporarily comprise the first information I-1, e.g., in the form of the observation vector Jest, and the second channel prediction CH-PRED-2, e.g., in the form of the element ĥpdt,i, e.g., with i>0, e.g., depending on a state of the switch SW1.


In some examples, a further, e.g., new, iteration of training the second machine learning model MLM-2 may be performed based on the feedback, e.g., to further improve a channel prediction, e.g., with the updated observations, e.g., the updated channel prediction ĥpdt,i and the for example original, observation yest.


In some examples, e.g., in an inference or testing stage or phase, e.g., subsequent to the training of the machine learning models MLM-1, MLM-2, the trained machine learning models MLM-1, MLM-2 may adapt at least one of the vector whp or the matrix whh,i, e.g., to any observation yest provided as input information.


In some examples, both a channel estimation ĥest and a channel prediction ĥpdt,i may be obtained using the machine learning models MLM-1, MLM-2, e.g., if i iterations (i≥1) are used.


In some examples, the prediction at the output of the second machine learning model MLM-2 may be improved by implementing the matrix Whh,i, e.g., because an element-wise cross-correlation as characterized by the matrix Whh may be exploited to finetune the prediction, e.g., with respect to the estimation ĥest.


In some examples, the second machine learning model MLM-2 may, e.g., comprise a topology as disclosed in the following reference: Yejian Chen; Jafar Mohammadi; Stefan Wesemann; Thorsten Wild; “Turbo-AI, Part I: Iterative Machine Learning Based Channel Estimation for 2D Massive Arrays,” in Proc. 2021 IEEE 93rd Veh. Technol. Conf. (VTC′ 21 Spring), Helsinki, Finland, April 2021.


As mentioned above, in some examples, FIG. 13A, 13B, the first machine learning model MLM-1 may, e.g., be a first dense neural network, e.g., comprising two fully connected layers. In some examples, the first machine learning model MLM-1 is configured to receive as input the first information I-1 and, optionally, the labels LAB (e.g., for training). In some examples, the first information I-1 may be characterized by L many complex numbers, e.g., characterizing the observations yest, and the labels LAB may be characterized by a, e.g., one, complex number. In some examples, the first machine learning model MLM-1 is configured to provide at its output the second channel prediction CH-PRED-2, e.g., in the form of L many complex numbers, e.g., characterizing the elements ĥestT and ĥpdt,i.


In some examples, similarly, the second machine learning model MLM-2 may, e.g., be a second dense neural network, e.g., comprising two fully connected layers. In some examples, the second machine learning model MLM-2 may be configured to receive as input L2+L many complex numbers, e. g., characterizing the observation vector associated with the second information I-2 and, optionally (e.g., for a training), the labels LAB.


In some examples, at least one of the first machine learning model MLM-1 or the second machine learning model MLM-2 may use an activation function of the SoftMax type.


As mentioned above, in some examples, historic data DAT-HIST of the at least one signal SIG associated with at least two different domains of the following domains: a) a time domain D-TIM, or b) a frequency domain D-FREQ, or c) a spatial domain D-SPAT may be collected, see FIG. 10, and training data DAT-TRAIN for training at least one of the first machine learning model MLM-1 or the second machine learning model MLM-2 may be determined based on the collected historic data.


In some examples, FIG. 2, 13A, 13B, channel prediction may be performed using the machine learning system ML-SYS, which does not rely on, e.g., only one, i.e., a single one of the aforementioned domains, e.g., the time domain. Rather, in some examples, channel prediction may be performed associated with at least two different domains.


In some examples, FIG. 14, a prediction path through various, presently for example three, domains, e.g., time, see axis a1 of FIG. 14, frequency, see axis a2, and space, see axis a3, may be defined, wherein different elements of the prediction path associated with the respective domains are denoted with circles “1”, “2”, . . . , “13” connected via arrows which are not provided with references signs for the sake of intelligibility.


In some examples, using samples in the prediction path from frequency and spatial domains a2, a3 (e.g., additionally to the time domain a1) may improve a prediction, e.g., due to an increasing correlation.


In some examples, FIG. 14, e.g., for a given data set, e.g., obtained based on historic data of received reference signals, e.g., SRS, e.g., with an example SRS interval of 10 milliseconds, a, for example fixed, pattern may be determined, e.g., sorted out, e.g., for observations (e.g., for a noisy version of symbol 1 to symbol 12, see the circles “1” to “12”) and labels (e.g., for a clean version of symbol 1 to symbol 13), e.g., with a reference current time instant t ref.


In some examples, training data, e.g., a training data set, e.g., comprising an observation data set and a label data set, may be determined by, for example, repeatedly, picking out data in accordance with the prediction path “1”, “2”, . . . , “13” as illustrated by the example of FIG. 14, e.g., with a fixed pattern as characterized by the prediction path, e.g., at different current time instants.



FIG. 15 schematically depicts such training data as can be obtained in some examples, wherein bracket B1 comprises the observation data set, and wherein bracket B2 comprises the label data set, wherein the elements z1, . . . , zL symbolise respective noise terms, e.g., of AWGN type.


In some examples, the machine learning models MLM-1, MLM-2 (FIG. 2) may be used to process data associated with at least two different prediction paths, see, for example blocks 260, 261 of FIG. 11.


In some examples, FIG. 16, two machine learning systems ML-SYS-1, ML-SYS-2 are provided to process data associated with two different prediction paths. In some examples, both machine learning systems ML-SYS-1, ML-SYS-2 comprise respective instances MLM-1′, MLM-2′, MLM-1″, MLM-2″ of the first machine learning models MLM-1, MLM-2, e.g., as depicted by FIG. 13A, 13B. Note that an optional feedback path from an output to an input of the respective second machine learning model instance MLM-2′, MLM-2″ and the respective switch SW1 (FIG. 13A, 13B) is not depicted by FIG. 16 in detail, but rather symbolically by the dashed arrows fb1, fb2, for the sake of intelligibility.


In some examples, FIG. 16, a first prediction path may be implemented using the first machine learning system ML-SYS-1, and a second prediction path may be implemented using the second machine learning system ML-SYS-2.


In some examples, the first information I-1′ provided as input to the instances MLM-1′, MLM-2′ of the first machine learning system ML-SYS-1 may, e.g., be associated with first frequency resources, e.g., one or more first subcarriers, whereas the first information I-1″ provided as input to the instances MLM-1″, MLM-2″ of the second machine learning system ML-SYS-2 may, e.g., be associated with second frequency resources, e.g., one or more second subcarriers, which are different from the first frequency resources. In some examples, the first subcarrier(s) and the second subcarrier(s) may be neighboring subcarriers.


In some examples, e.g., as neighboring subcarriers in the frequency domain may have a comparatively high correlation, a combination, e.g., an averaging, of both channel predictions, e.g., the respective second channel prediction ĥpdt,1, ĥpdt,2, may introduce a diversity gain, e.g., with at least one of a) independent noise realizations, or b) independent channel estimation errors, or c) independent channel prediction errors. In some examples, the combination (see, for example also block 262 of FIG. 11) may be effected by adding the respective second channel predictions ĥpdt,1, ĥpdt,2, see the symbolic adder ad1 of FIG. 16, optionally using weighting factors wf1, wf2, wherein, in some examples, wf1=wf2=0.5.


In some examples, the first information I-1′, I-1″ provided as input to machine learning systems ML-SYS-1, ML-SYS-2 may, e.g., be associated with at least two different domains (e.g., temporal, spatial, frequency), e.g., at least similar to the example prediction path of FIG. 14.


In some examples, FIG. 16, more than two machine learning systems ML-SYS-1, ML-SYS-2 may be provided, wherein their respective outputs, e.g., respective second channel predictions, may be combined, e.g., in the sense of a, for example weighted, sum ĥpdt.


In some examples, a diversity order of 2 or higher, which in some examples may, e.g., be attained by combining the outputs of several machine learning systems ML-SYS-1, ML-SYS-2, may improve a prediction performance, e.g., for both low speed cases (e.g., associated with a comparatively low mobility, e.g., of the receiver RX) and high speed cases (e.g., associated with a comparatively high mobility).


In some examples, subcarrier combining may be performed, e.g., using a configuration as depicted by FIG. 16 or similar.


In some examples, as mentioned above, reference signals, e.g., sounding reference signals, SRS, may be used, e.g., to provide observations, e.g., as input data to the machine learning model(s) MLM-1, MLM-2. In some examples, other reference signals may be used alternatively or additionally, e.g., demodulation reference signals (DMRS), e.g., in cases with a channel reciprocity of uplink/downlink channels (e.g., with time division duplexing, TDD, scenarios). In some examples, using different types of reference signals, such as, e.g., SRS and DMRS, a prediction can be further enhanced or improved.



FIG. 17 schematically depicts aspects of combining respective outputs of a plurality of machine learning systems according to some examples. Depicted is a first plurality of, presently for example six, machine learning systems ML-SYS-n, . . . , ML-SYS−n+5, only two of which are depicted for the sake of clarity. In some examples, the first plurality of machine learning systems ML-SYS-n, . . . , ML-SYS-n+5 is configured to provide a respective channel prediction for a time instant t+1, e.g., based on input information, e.g., in the sense of the first information I-1 (FIG. 13A, 13B), e.g., observations, associated with preceding time instants t-1, t, also see the dashed rounded rectangle RR1 of the schematic time frequency resource pattern depicted by FIG. 18, e.g., denoting samples e1, e2, e3, e4, wherein the horizontal axis t characterizes time, and wherein the vertical axis f characterizes frequency.


Returning to FIG. 17, also depicted is a second plurality of, presently for example six, machine learning systems ML-SYS-m, . . . , ML-SYS-m+5, only one of which (e.g., system ML-SS-m) is depicted for the sake of clarity. In some examples, the second plurality of machine learning systems ML-SYS-m,, ML-SYS-m+5 is configured to provide a respective channel prediction for a time instant t+2, e.g., based on input information, e.g., in the sense of the first information I-1 (FIG. 13A, 13B), e.g., observations, associated with time instants t, t+1, also see the dashed rounded rectangle RR2, e.g., denoting samples e2, e4, e5, e6, of FIG. 18, wherein second channel predictions as may be obtained by the second plurality of machine learning systems ML-SYS-m, . . . , ML-SYS-m+5, e.g., for the time instant t+2, are symbolized with reference signs e7, e8 in FIG. 18.


Note that, in some examples, the machine learning models MLM-1, MLM-2 of at least some, for example all, machine learning systems as depicted by FIG. 17 may have a similar or identical topology and/or configuration.


In some examples, FIG. 17, the input I-1-n, . . . , I−1-n+5 to the various, presently for example six, machine learning systems ML-SYS-n,., ML-SYS-n+5, may comprise a number of observations, e.g., characterizing the same, for example twelve, symbols, however, with a respective different sequence of symbols for the different machine learning systems ML-SYS-n, . . . , ML-SYS-n+5.


In some examples, the respective output of the machine learning systems ML-SYS-n, . . . , ML-SYS-n+5 is combined by a combiner COMB (see, for example, also block 262 of FIG. 11), which may, e.g., be configured to perform a vectorization such that an observation vector OV may be provided as input to the further machine learning systems ML-SYS-m,., ML-SYS-m+5, wherein a first number n1 of elements of the observation vector OV characterize an estimation of the time instant t, and wherein a second number n2 of elements of the observation vector OV characterize the second channel predictions CH-PRED-2-n, . . . , CH-PRED-2-n+5 as output by the machine learning systems ML-SYS-n, . . . , ML-SYS-n+5, wherein, in some examples, the second channel predictions CH-PRED-2-n, . . . , CH-PRED-2-n+5 may e.g. be considered to represent virtual observations.


In some examples, the configuration of FIG. 17 can be considered as a concatenated prediction structure, wherein a plurality of second channel predictions CH-PRED-2-n, . . . , CH-PRED-2-n+5 as, e.g., obtained by the first plurality of machine learning systems ML-SYS-n, . . . , ML-SYS-n+5 is provided to at least one further machine learning system ML-SYS-m, ML-SYS-m+5, also see block 263 of FIG. 11.


In some examples, FIG. 17, 18, each of the machine learning systems ML-SYS-n, . . . , ML-SYS-n+5 may be configured to use, e.g., exploit, the samples e1, e2, e3, e4, e.g., in time instant t-1 to t, see the dashed rounded rectangle RR1, e.g., to determine a second channel prediction, e.g., for the time instant t+1, see the elements e5, e6. In some examples, the further machine learning systems ML-SYS-m, ML-SYS-m+5 may be configured to use, e.g., exploit, the samples e2, e4, e5, e6 e.g., in time instant t to t+1, see the dashed rounded rectangle RR2, e.g., to determine a second channel prediction, e.g., for the time instant t+2, see the elements e7, e8.


In some examples, FIG. 2, the principle according to example embodiments may be used for devices 10 for a wireless, e.g., cellular, communications system, e.g., based on and/or adhering to some accepted specification, e.g., standard, such as, e.g., 5G or 6G or beyond 6G (“B6G”) or some other accepted specification or standard.


In some examples, the principle according to example embodiments enables to use reference signals, e.g., of the SRS and/or DMRS type, wherein, for example, comparatively precise second channel predictions may be obtained based on an SRS interval of, e.g., 10 ms.


In some examples, the principle according to example embodiments enables to perform at least one of precoding or beamforming, e.g., for a downlink direction, e.g., on a symbol-level, which, in some examples, may, for example significantly, impact, e.g., improve, both multiuser and single user MIMO, so that, in some examples, a downlink throughput may be comparatively close to an achievable downlink capacity.


Some examples, FIG. 19, relate to a computer program PRG comprising instructions INSTR which, when executed by an apparatus, e.g., apparatus 100 or 100′ (FIG. 1A, 1B, 2), cause the apparatus to perform at least some aspects of the method according to the example embodiments.


In some examples, FIG. 19, the computer program PRG may be provided on a computer readable storage medium SM, e.g., a non-transitory computer readable medium.


Some examples, FIG. 19, relate to a data carrier signal DCS carrying and/or characterizing the computer program PRG according to the example embodiments.


In some examples, the principle according to example embodiments enables to address channel aging. In some examples, channel aging means a nonvanishing delay between a first time instant, when, for example, a reception of a reference signal, e.g., SRS, in an uplink direction, and a respective channel estimation happen, and a second time instant, when a downlink transmission happens, e.g., based on the channel estimation. In some examples, channel aging can lead to certain inconsistencies with respect to a radio channel, e.g., a channel response in the real world can deviate significantly from a conventionally estimated or predicted channel (as, e.g., obtained through at least one of uplink channel state information, UL CSI, or uplink sounding reference signals, UL SRS).


In some examples, channel aging can be attributed to a processing latency, e.g., in UL channel estimation and downlink, DL, precoding computation and to a signaling latency, e.g., limited by an SRS reporting interval. In some conventional approaches, channel aging can thus lead to performance degradation, e.g., since an outdated channel may introduce mismatched DL beamforming vector or precoding matrix and start to constrain a DL throughput. In some examples, channel aging may happen, when a user of a mobile device 10 (FIG. 2) moves, e.g., with increasing mobility. In some conventional approaches, a communication through the downlink may, e.g., not become better, until a next SRS arrives, e.g., to refresh DL beamforming or precoding weights. In some examples, the principle according to example embodiments enables to mitigate at least some of the potential disadvantages that may result from channel aging, e.g., by providing a, for example second, channel prediction.

Claims
  • 1. An apparatus, comprising: at least one processor; andat least one memory storing instructions that, when executed with the at least one processor, cause the apparatus to: provide a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information;provide a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction; anddetermine the second channel prediction using at least the second machine learning model.
  • 2. The apparatus according to claim 1, wherein the instructions, when executed with the at least one processor, cause the apparatus to at least temporarily provide the first information and the second channel prediction as the second information to the second machine learning model.
  • 3. The apparatus according to claim 1, wherein the instructions, when executed with the at least one processor, cause the apparatus to: provide first training data for training of at least the first machine learning model, wherein the first training data comprises an estimable part, a predictable part, and associated labels; andtrain at least the first machine learning model based on the first training data and a first loss function.
  • 4. The apparatus according to claim 3, wherein the instructions, when executed with the at least one processor, cause the apparatus to: provide historic data of the at least one signal; andpartition the historic data into the estimable part and the predictable part.
  • 5. The apparatus according to claim 3, wherein the instructions, when executed with the at least one processor, cause the apparatus to: provide second training data for training of the second machine learning model, wherein the second training data comprises the first training data and at least one of: the first channel prediction, or the second channel prediction; andtrain the second machine learning model based on the second training data and a second loss function, which is different from the first loss function.
  • 6. The apparatus according to claim 3, wherein the instructions, when executed with the at least one processor, cause the apparatus to: use the first training data and the first channel prediction as obtained with the first machine learning model in a first training phase for training the second machine learning model; anduse the first training data and the second channel prediction as obtained with the second machine learning model in a subsequent second training phase for training the second machine learning model.
  • 7. The apparatus according to claim 1, wherein the instructions, when executed with the at least one processor, cause the apparatus to: collect historic data of the at least one signal associated with at least two different domains of the following domains: a time domain, or a frequency domain, or a spatial domain; anddetermine training data for training at least one of the first machine learning model or the second machine learning model based on the collected historic data.
  • 8. The apparatus according to claim 1, wherein the instructions, when executed with the at least one processor, cause the apparatus to: provide at least one machine learning system comprising an instance of the first machine learning model and an instance of the second machine learning model;provide a plurality of prediction paths to obtain a plurality of second channel predictions using the at least one machine learning system, each of the plurality of second channel predictions associated with a respective one of the plurality of prediction paths; andcombine at least two of the plurality of second channel predictions.
  • 9. The apparatus according to claim 8, wherein the instructions, when executed with the at least one processor, cause the apparatus to: provide a plurality of machine learning systems; andtrain at least two machine learning systems of the plurality of machine learning systems differently from each other.
  • 10. The apparatus according to claim 8, wherein the instructions, when executed with the at least one processor, cause the apparatus to: provide the at least two of the plurality of second channel predictions to at least one further machine learning system.
  • 11. The apparatus according to claim 1, wherein at least one of the first machine learning model or the second machine learning model is a dense neural network.
  • 12. (canceled)
  • 13. A method, comprising: providing a first machine learning model configured to receive first information associated with at least one signal and to provide a first channel prediction based on the first information;providing a second machine learning model configured to receive second information and to provide a second channel prediction based on the second information, the second information comprising the first information and at least temporarily comprising the first channel prediction; anddetermining the second channel prediction using at least the second machine learning model.
  • 14. A device for a communication system, the device comprising at least one apparatus according to claim 1.
  • 15. A non-transitory program storage device readable with an apparatus, tangibly embodying a program of instructions executable with the apparatus to perform at least some aspects of the method according to claim 13.
Priority Claims (1)
Number Date Country Kind
23214930.2 Dec 2023 EP regional