This application relates to the field of communications technologies, and particularly, to a training data set obtaining method and apparatus, a wireless transmission method and apparatus, and a communications device.
Generalization means that a neural network can still obtain a proper output for data not encountered in a training (learning) process.
According to a first aspect, a training data set obtaining method is provided. The method includes:
According to a second aspect, a training data set obtaining apparatus is provided, including:
According to a third aspect, a wireless transmission method is provided. The method includes:
According to a fourth aspect, a wireless transmission apparatus is provided, including:
According to a fifth aspect, a communications device is provided. The communications device includes a processor, a memory, and a program or an instruction stored in the memory and executable on the processor, where when the program or the instruction is executed by the processor, the steps of the method according to the first aspect or the steps of the method according to the third aspect are implemented.
According to a sixth aspect, a communications device is provided, including a processor and a communications interface, where the processor is configured to determine, based on degrees of contribution of transmission conditions to an optimization objective of a neural network, data volumes of training data under the transmission conditions; and obtain, based on the data volumes of the training data under the transmission conditions, the training data under the transmission conditions, to form a training data set for training the neural network, where the degrees of contribution of the transmission conditions to the optimization objective of the neural network represent degrees of impact of the transmission conditions on a value of the optimization objective of the neural network.
According to a seventh aspect, a communications device is provided, including a processor and a communications interface, where the processor is configured to perform an operation of wireless transmission based on a neural network model, to implement the wireless transmission, where the neural network model is obtained by performing training by using a training data set in advance, and the training data set is obtained based on the training data set obtaining method according to the first aspect.
According to an eighth aspect, a non-transitory computer-readable storage medium is provided. The non-transitory computer-readable storage medium stores a program or an instruction, and when the program or the instruction is executed by a processor, the steps of the method according to the first aspect or the steps of the method according to the third aspect are implemented.
According to a ninth aspect, a chip is provided. The chip includes a processor and a communications interface, the communications interface is coupled to the processor, and the processor is configured to execute a program or an instruction to implement the method according to the first aspect or the method according to the third aspect.
According to a tenth aspect, a computer program/program product is provided. The computer program/program product is stored in a non-transient storage medium, and the program/program product is executed by at least one processor to implement the steps of the training data set obtaining method according to the first aspect or the steps of the wireless transmission method according to the third aspect.
The following clearly describes the technical solutions in the embodiments of this application with reference to the accompanying drawings in the embodiments of this application. Apparently, the described embodiments are merely some but not all of the embodiments of this application. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of this application fall within the protection scope of this application.
The specification and claims of this application, and terms “first” and “second” are used to distinguish similar objects, but are not used to describe a specific sequence or order. It should be understood that, the terms used in this way is exchangeable in a proper case, so that the embodiments this application can be implemented in another order except those shown or described herein. In addition, objects distinguished by “first” and “second” usually belong to one type, and a quantity of objects is not limited. For example, there may be one or more first objects. In addition, “and/or” in this specification and the claims represents at least one of the connected objects. The character “I” usually indicates an “or” relationship between associated objects.
It should be noted that, the technologies described in the embodiments of this application are not limited to the Long Term Evolution (LTE)/LTE-Advanced (LTE-A) system, and may be further applied to other wireless communication systems such as Code Division Multiple Access (CDMA), Time Division Multiple Access (TDMA), Frequency Division Multiple Access (FDMA), Orthogonal Frequency Division Multiple Access (OFDMA), Single Carrier Frequency Division Multiple Access (SC-FDMA), and other systems. The terms “system” and “network” may be used interchangeably in the embodiments of this application. The described technologies can be applied to the systems and radio technologies mentioned above, and can also be applied to other systems and radio technologies. The following descriptions describe a new radio (NR) system for an example purpose, and NR terms are used in most of the following descriptions, but these technologies may also be applied to an application other than the NR system application, such as a 6-th Generation (6G) communication system.
Currently, to implement a generalization capability for a changeable wireless transmission environment, training may be performed based on mixed data to obtain one common neural network, and parameters of the neural network do not need to switch as the environment changes. However, the neural network cannot achieve optimal performance under each transmission condition.
A training data set obtaining method and apparatus, a wireless transmission method and apparatus, and a communications device provided in the embodiments of this application are described in detail below with reference to the accompanying drawings by using some embodiments and application scenarios thereof.
Step 201: Determine, based on degrees of contribution of transmission conditions to an optimization objective of a neural network, data volumes of training data under the transmission conditions.
The degrees of contribution of the transmission conditions to the optimization objective of the neural network represent degrees of impact of the transmission conditions on a value of the optimization objective of the neural network.
It may be understood that, for the transmission conditions oriented to a wireless transmission environment, in this embodiment of this application, the degrees of contribution of the transmission conditions to the optimization objective of the neural network may be determined in advance by testing the degrees of impact of the transmission conditions on the value of the optimization objective of the neural network, that is, by testing degrees of impact of values of the transmission conditions on the value of the optimization objective. Usually, a higher degree of impact indicates a larger value of a degree of contribution of a corresponding transmission condition; otherwise, a lower degree of impact indicates a smaller value of a degree of contribution.
For example, when the optimization objective is expressed as a function of the transmission conditions, if the function is an increasing function about the transmission conditions (for example, a throughput is an SNR increasing function), a transmission condition causing the optimization objective to have a large value has a high degree of contribution; or if the function is a decreasing function about the transmission conditions (for example, an NMSE is an SNR decreasing function), a transmission condition causing the optimization objective to have a small value has a high degree of contribution.
Optionally, for any transmission condition, a degree of contribution of the transmission condition to the optimization objective may be determined based on an achievable optimal value of the optimization objective under the transmission condition. That is to say, a degree of contribution of one given transmission condition to the optimization objective may be measured using a magnitude of an achievable optimal value of the optimization objective under the given transmission condition, and if the achievable optimal value is larger, it indicates that the given transmission condition has a higher degree of impact on the optimization objective, and also has a larger corresponding degree of contribution.
Based on the determined degrees of contribution of the transmission conditions, in this embodiment of this application, the data volumes of the training data corresponding to the transmission conditions may be respectively adjusted based on magnitudes of values of the degrees of contribution corresponding to the transmission conditions, to cause a data volume of training data under each transmission condition to be associated with a degree of contribution of the transmission condition to the optimization objective. The data volumes are reference data volumes of training data that needs to be prepared for the transmission conditions, or are referred to as specified data volumes, and when a training data set is actually obtained, and the training data is prepared for the transmission conditions, reference needs to be made to the data volumes.
It may be understood that, to enable the neural network to have a better generalization capability for various transmission conditions, volumes of data under transmission conditions having high degrees of contribution may be reduced, to reduce degrees of impact of these transmission conditions; and volumes of data under transmission conditions having low degrees of contribution may be increased, to increase degrees of impact of these transmission conditions. That is to say, impact of the transmission conditions on the optimization objective is balanced through data volumes. The data volume may be an absolute value of the data volume, and may also be a relative proportion of the data volume.
The transmission conditions are parameters involved in an actual wireless transmission environment, such as a transmission medium, a transmission signal, and a transmission environment. Optionally, types of the transmission conditions include at least one of:
It may be understood that, types of the transmission conditions involved in the embodiments of this application may include but are not limited to one of or a combination of more than one of the types of the transmission conditions listed above.
The interference strength may represent, for example, a strength of an inter-cell co-channel interference, or a magnitude of another interference; the channel parameter is, for example, the number of channel taps (or LOS or NLOS scenario), a delay (or maximum delay), Doppler (or maximum Doppler), an angle of arrival (including horizontal and vertical) range, an angle of departure (including horizontal and vertical) range, or a channel-related coefficient; the cell type is, for example, an indoor cell, an outdoor cell, a macro cell, a micro cell, or a pico cell; the inter-site distance may be, for example, classified into an inter-site distance less than 200 meters, from 200 to 500 meters, or more than 500 meters; the weather and environment factors may be, for example, information such as temperature and/or humidity of a network environment in which training data is located; the antenna configuration information of the transmit end or receive end may be, for example, an antenna quantity and/or an antenna polarization manner; and the UE capability/type may be, for example, Redcap UE and/or normal UE.
Step 202: Obtain, based on the data volumes of the training data under the transmission conditions, the training data under the transmission conditions, to form a training data set for training the neural network.
It may be understood that, in this embodiment of this application, based on the obtained data volumes of the training data under the transmission conditions, that is, based on the obtained reference data volumes, the training data under the transmission conditions is obtained by using the reference data volumes corresponding to the transmission conditions as a reference or setting, to cause the data volumes of the finally obtained training data under the transmission conditions to match the reference or setting. Finally, the obtained training data under the transmission conditions is non-uniformly mixed, to obtain a data set, that is, a training data set, and the training data set may be used to train the foregoing neural network or neural network model in the wireless transmission environment.
Optionally, the obtaining the training data under the transmission conditions, to form a training data set for training the neural network includes: collecting, based on the data volumes of the training data under the transmission conditions, data under the transmission conditions and calibrating the data, to form the training data set under the transmission conditions; or collecting specified volumes of data under the transmission conditions, and selecting, based on the data volumes of the training data under the transmission conditions, partial data from the specified volumes of data and calibrating the partial data, or replenishing the specified volumes of data and calibrating replenished data, to form the training data set under the transmission conditions.
It may be understood that, in this embodiment of this application, when the training data set is constructed according to the determined data volumes of the training data under the transmission conditions, the data volumes required for the transmission conditions may be first calculated, and then data under the transmission conditions is obtained according to the data volumes. Alternatively, a large volume of data under the transmission conditions may be obtained first, that is, a specified volume of data is obtained, then data volumes required for the transmission conditions are calculated, and then selection or replenishment is performed from the foregoing large volume of data obtained in advance.
For the latter case, it is assumed that a total data volume obtained in advance under a k-th transmission condition is Mk, and a required data volume determined through calculation is Nk. If Mk≥Nk, Nk pieces of data need to be randomly selected from the Mk pieces of data and placed into the training data set. If Mk<Nk, Nk−Mk pieces of data need to be obtained, and Nk pieces of data are replenished and then placed into the training data set. The required data volumes are data volumes of the training data under the transmission conditions determined according to the previous step.
After the data under the transmission conditions is obtained, the data needs to be calibrated, that is, a label is added to data under each transmission condition, and the added label is a transmission environment truth value corresponding to the data. For example, in a DMRS channel estimation scenario, a label added to a DMRS signal of data under a transmission condition is a channel truth value corresponding to the DMRS signal.
It should be noted that this embodiment of this application may be applied to any scenario in which functions of one or more modules in the existing wireless transmission network can be replaced with machine learning, that is, when machine learning is used to train the neural network, the training data set may be constructed by using the training data set obtaining method of this embodiment of this application. The application scenario is, for example, pilot design, channel estimation, signal detection, user pairing, HARQ, and positioning at a physical layer, resource allocation, switching, and mobility management at a higher layer, and scheduling or slicing at a network layer. A wireless transmission application scenario is not limited in this embodiment of this application.
In this embodiment of this application, when a training data set is constructed in a communication system based on artificial intelligence, data under a plurality of transmission conditions is selected in different proportions according to degrees of contribution of data under different transmission conditions to an optimization objective (or referred to as an objective function or a loss function) of a neural network, to construct a mixed training data set, so that a generalization capability of the neural network can be effectively improved.
Optionally, the determining, based on degrees of contribution of transmission conditions to an optimization objective of a neural network, data volumes of training data under the transmission conditions includes: sorting the degrees of contribution of the transmission conditions; and executing, based on equal-proportion mixing, at least one of reducing a data volume of training data under a transmission condition corresponding to a higher degree of contribution in the sorting or increasing a data volume of training data under a transmission condition corresponding to a lower degree of contribution in the sorting.
It may be understood that, in this embodiment of this application, when the data volumes of the training data under the transmission conditions are determined, degrees of contribution of data under different transmission conditions to the optimization objective (or objective function or loss function) of the neural network during equal-proportion mixing may be first sorted. Then, when the mixed training data set is constructed, a low degree of contribution and a high degree of contribution may be determined based on the foregoing sorting, to determine a transmission condition with the low degree of contribution and a transmission condition with the high degree of contribution, and based on the premise of ensuring that data under all the transmission conditions is plentiful, a data volume under the transmission condition with the low degree of contribution may be increased, and/or a data volume under the transmission condition with the high degree of contribution may be reduced.
A threshold may be set (that is, a threshold is preset), and a proportion of a data volume under any transmission condition to the total data volume is caused to be not less than the threshold, to satisfy the foregoing “premise of ensuring that data under all the transmission conditions is plentiful”.
In this embodiment of this application, by sorting the degrees of contribution of the transmission conditions and adjusting a data volume under a corresponding transmission condition according to the sorting, an adjustment policy on the data volume under the corresponding transmission condition (including whether the data volume needs to be increased or reduced, an increase or reduction amplitude, and the like) can be determined more explicitly and accurately, thereby making the efficiency higher and the result more accurate.
Optionally, the executing at least one of reducing a data volume of training data under a transmission condition corresponding to a higher degree of contribution in the sorting or increasing a data volume of training data under a transmission condition corresponding to a lower degree of contribution in the sorting includes: executing, according to a rule, at least one of reducing a data volume of training data under a transmission condition corresponding to a higher degree of contribution in the sorting or increasing a data volume of training data under a transmission condition corresponding to a lower degree of contribution in the sorting, where the rule includes: that a larger value of the higher degree of contribution indicates a larger amplitude of the reducing; and that a smaller value of the lower degree of contribution indicates a larger amplitude of the increasing.
It may be understood that, in this embodiment of this application, when the data volume under the transmission condition with the low degree of contribution is increased, and the data volume under the transmission condition with the high degree of contribution is reduced, an aim is to gradually reduce the data volumes of the training data under the transmission conditions as the degrees of contribution under the transmission conditions are gradually increased. Therefore, a lower degree of contribution of a transmission condition indicates a larger data volume increased correspondingly under the transmission condition; and a higher degree of contribution of a transmission condition indicates a larger data volume reduced correspondingly under the transmission condition.
In this embodiment of this application, data volumes of training data under corresponding transmission conditions are increased or reduced proportionally according to magnitudes of values of degrees of contribution, so that as the degrees of contribution of the transmission conditions are gradually increased, the data volumes of the training data under the transmission conditions are gradually reduced, thereby better balancing impact of the transmission conditions on the final neural network and better facilitating an improvement in the generalization capability of the neural network.
Optionally, in a case that a result of the sorting is an ascending order, the data volumes of the training data under the transmission conditions decrease in a direction of the sorting; and in a case that a result of the sorting is a descending order, the data volumes of the training data under the transmission conditions increase in the direction of the sorting.
It may be understood that, when the data volumes of the training data under the transmission conditions are determined based on the foregoing embodiment, if a sorting manner of the degrees of contribution is from small to large, proportions of data under corresponding transmission conditions to total data may be in any decrease manner, for example, linear decrease, arithmetic decrease, geometric decrease, exponential function decrease, or power function decrease. Otherwise, if a sorting manner of the degrees of contribution is from large to small, proportions of data under corresponding transmission conditions to total data may be in any increase manner, for example, linear increase, arithmetic increase, geometric increase, exponential function increase, or power function increase.
Optionally, the executing at least one of reducing a data volume of training data under a transmission condition corresponding to a higher degree of contribution in the sorting or increasing a data volume of training data under a transmission condition corresponding to a lower degree of contribution in the sorting includes: determining a reference degree of contribution according to the sorting, and comparing a degree of contribution of a transmission condition with the reference degree of contribution; and
It may be understood that, in this embodiment of this application, when the data volumes of the training data under the transmission conditions are adjusted according to the sorting, one intermediate contrast reference quantity of degrees of contribution may be first determined according to the sorting, and is referred to as a reference degree of contribution. Optionally, the reference degree of contribution is a median of the sorting, or a degree of contribution at a specified position in the sorting, or an average of the degrees of contribution in the sorting, or a degree of contribution in the sorting closest to the average. The average may be an arithmetic average, a geometric average, a harmonic average, a weighted average, a square average, an exponential average, or the like.
Then, a degree of contribution of each transmission condition is compared with the reference degree of contribution sequentially. If a degree of contribution of an i-th transmission condition is greater than the reference degree of contribution, the i-th transmission condition is determined as the high degree of contribution in the foregoing embodiment, and a data volume under the i-th transmission condition is reduced; otherwise, if a degree of contribution of an i-th transmission condition is less than a median degree of contribution, the i-th transmission condition is determined as the low degree of contribution in the foregoing embodiment, and a data volume under the i-th transmission condition is increased.
In this embodiment of this application, by determining an intermediate contrast reference quantity of degrees of contribution, it is only necessary to compare another degree of contribution with the contrast reference quantity, to determine according to a comparison result to increase or reduce a data volume under a corresponding transmission condition, so that the algorithm is simple and the operation quantity is small.
Optionally, the determining, based on degrees of contribution of transmission conditions to an optimization objective of a neural network, data volumes of training data under the transmission conditions includes: determining, based on probability densities of the transmission conditions in an actual application, weighting coefficients corresponding to the transmission conditions; and determining, based on the degrees of contribution of the transmission conditions to the optimization objective and with reference to the weighting coefficients, the data volumes of the training data under the transmission conditions.
It may be understood that, when proportions of data under different transmission conditions to the total data volume are determined, weighted terms may be designed based on actual probability densities under the different transmission conditions, a data volume under a condition with a high probability density is increased, and a data volume under a condition with a low probability density is reduced. For example, assuming that a probability density of a k-th SNR is pk, a corresponding weighted term is f(pk), and f(pk) is an increase function about pk. A data volume of the k-th SNR updated by the weighted term is considered as f(pk)·Nk. A probability density is to reflect occurrence probabilities of transmission conditions in different environments. Transmission conditions do not exist with an equal probability in different environments, occurrence probabilities of some transmission conditions are slightly high, and occurrence probabilities of some transmission conditions are slightly low.
Optionally, the weighting coefficients and the probability densities are in a function increasing relationship. That is to say, a relationship between a weighted term and a probability density may be any increase function relationship. That is to say, it is necessary to ensure that a weighted term of a high probability density is large and a weighted term of a low probability density is small.
In this embodiment of this application, designing a weighted term according to an actual probability density of a transmission condition can better adapt to an actual environment.
Optionally, the method further includes: sending the training data set to a target device, where the target device is configured to train the neural network based on the training data set.
It may be understood that, the training data set obtained in this embodiment of this application may be used to train the neural network, and the training data set obtaining method in this embodiment of this application may be applied to a data transmission scenario in which a data collection end and a neural network training end may not be at a same execution end. In this embodiment of this application, a training data set is built after data is collected and calibrated according to the determined proportion of the mixed data set, and the training data set is fed back to another device that needs to execute a neural network training process, that is, a target device. The target device is a second device different from the current device, and may be a terminal or a network-side device, configured to complete training of the neural network model by using the obtained training data set.
In this embodiment of this application, by sending the obtained training data set to the second device different from the current device, data sharing and joint training can be implemented between a plurality of devices, thereby effectively reducing an operation quantity of a single device and effectively improving operation efficiency.
Optionally, the sending the training data set to a target device includes: directly sending the training data set to the target device, or sending the training data set on which specified conversion is performed to the target device, where the specified conversion includes at least one of specific quantization, specific compression, or processing that is performed according to a neural network agreed or configured in advance.
It may be understood that, in this application, when the training data set is sent to another device, a sending manner may be direct sending or indirect sending. The indirect sending means that training data in the training data set is converted and then fed back. For example, to-be-sent training data may be processed and then sent in a specific quantization manner or a specific compression manner or according to a neural network that is agreed or configured in advance.
To describe technical solutions of the embodiments of this application, description is made below using an example, but the scope that this application claims to protect is not limited.
An application scenario in which the optimization objective (or objective function or loss function) of the neural network is an indicator that needs to be minimized such as a mean square error (MSE) or normalized mean square error (NMSE) and a transmission condition is a signal to noise ratio (SNR) is used as an example. Given that data under K signal to noise ratios is to be mixed, a total data volume is denoted as Nall, a k-th signal to noise ratio is denoted as SNRk, and a data volume under the k-th signal to noise ratio is denoted as Nk.
First, the data under the K signal to noise ratios is sorted:
Assuming that the data is mixed at an equal proportion (that is, Nk/Nall=1/K), a degree of contribution of data under the k-th SNR to the foregoing indicator that needs to be minimized is denoted as Ck. SNRs are sorted in ascending order (or descending order) of Ck.
Then, mixing proportions are determined:
That optimized indicators are an MSE and an NMSE is used as an example. Generally, a degree of contribution of data under a low SNR to the MSE or NMSE is large, and a degree of contribution of data under a high SNR to the MSE or NMSE is small. Therefore, in the finally determined mixing proportions, a proportion of data under a low SNR to the total data volume is lowest. As an SNR increases, a corresponding data volume also increases.
Optionally, when proportions of data under different transmission conditions to the total data volume are determined, weighted terms may be designed based on actual probability densities under the different transmission conditions. Assuming that a probability density of a k-th SNR is pk, a corresponding weighted term is f(pk), and f(pk) is an increase function about pk. A data volume of the k-th SNR updated by the weighted term is considered as f(pk)−Nk.
In addition, an application scenario in which the optimization objective (or objective function or loss function) of the neural network is an indicator that needs to be minimized such as a signal to interference plus noise ratio (SINR), a spectral efficiency, or a throughput and a transmission condition is a signal to noise ratio (SNR) is used as an example. Given that data under K SNRs is to be mixed, a total data volume is denoted as Nall, a k-th signal to noise ratio is denoted as SNRk, and a data volume under the k-th signal to noise ratio is denoted as Nk.
First, the data under the K signal to noise ratios is sorted:
Assuming that the data is mixed at an equal proportion (that is, Nk/Nall=1/K), a degree of contribution of data under the k-th SNR to the foregoing indicator that needs to be maximized is denoted as Ck. SNRs are sorted in ascending order (or descending order) of Ck.
Then, mixing proportions are determined:
(1) For the k-th SNR, Ck is larger, a value of Nk is smaller through adjustment. In ascending order of Ck, corresponding values of Nk are from large to small and a decrease rule may be any decrease rule; or in descending order of Ck, corresponding values of Nk are from small to large and an increase rule may be any increase rule.
(2) After the data volume is adjusted, it is determined that a proportion of data under any SNR to the total data is not less than a threshold γ.
That optimized indicators are an SINR, a spectral efficiency, or a throughput is used as an example. Generally, a degree of contribution of data under a low SNR to the SINR, the spectral efficiency, and the throughput is small, and a degree of contribution of data under a high SNR to the SINR, the spectral efficiency, and the throughput is large. Therefore, a proportion of data under a low SNR to the total data volume is highest. As an SNR decreases, a corresponding data volume also decreases.
Optionally, when proportions of data under different transmission conditions to the total data volume are determined, weighted terms may be designed based on actual probability densities under the different transmission conditions. Assuming that a probability density of a k-th SNR is pk, a corresponding weighted term is f(pk), and f(pk) is an increase function about pk. A data volume of the k-th SNR updated by the weighted term is considered as f(pk)·Nk.
It should be noted that an execution body of the training data set obtaining method provided in this embodiment of this application may be a training data set obtaining apparatus, or a control module configured to execute the training data set obtaining method in the training data set obtaining apparatus. In the embodiments of this application, a training data set obtaining apparatus provided in the embodiments of this application is described using an example in which a training data set obtaining apparatus executes the training data set obtaining method.
A structure of a training data set obtaining apparatus of an embodiment of this application is shown in
The first processing module 301 is configured to determine, based on degrees of contribution of transmission conditions to an optimization objective of a neural network, data volumes of training data under the transmission conditions; and
Optionally, the first processing module is configured to:
Optionally, types of the transmission conditions include at least one of:
Optionally, the second processing module is configured to:
Optionally, when being configured to execute at least one of reducing a data volume of training data under a transmission condition corresponding to a higher degree of contribution in the sorting or increasing a data volume of training data under a transmission condition corresponding to a lower degree of contribution in the sorting, the first processing module is configured to:
Optionally, in a case that a result of the sorting is an ascending order, the data volumes of the training data under the transmission conditions decrease in a direction of the sorting; and in a case that a result of the sorting is a descending order, the data volumes of the training data under the transmission conditions increase in the direction of the sorting.
Optionally, when being configured to execute at least one of reducing a data volume of training data under a transmission condition corresponding to a higher degree of contribution in the sorting or increasing a data volume of training data under a transmission condition corresponding to a lower degree of contribution in the sorting, the first processing module is configured to:
Optionally, the reference degree of contribution is a median of the sorting, or a degree of contribution at a specified position in the sorting, or an average of the degrees of contribution in the sorting, or a degree of contribution in the sorting closest to the average.
Optionally, the first processing module is further configured to:
Optionally, the weighting coefficients and the probability densities are in a function increasing relationship.
Optionally, the apparatus further includes:
Optionally, the sending module is configured to:
The training data set obtaining apparatus in this embodiment of this application may be an apparatus, an apparatus with an operating system, or an electronic device, or may be a component, an integrated circuit, or a chip in a terminal or network-side device. The apparatus or electronic device may be a mobile terminal, or may be a non-mobile terminal, or may include but is not limited to the foregoing listed types of the network-side device 102. For example, the mobile terminal may include but is not limited to the foregoing listed types of the terminal 101, and the non-mobile terminal may be a server, a network attached storage (NAS), a personal computer (PC), a television (TV), a teller machine, or a self-service machine, which is not limited in the embodiments of this application.
The training data set obtaining apparatus provided in this embodiment of this application can implement each process implemented in the method embodiment of
An embodiment of this application further provides a wireless transmission method. The method may be performed by a terminal and/or a network-side device. The terminal may be the terminal 101 shown in
Step 401: Perform an operation of wireless transmission based on a neural network model, to implement the wireless transmission.
The neural network model is obtained by performing training by using a training data set in advance, and the training data set is obtained based on the training data set obtaining method according to the foregoing embodiments.
It may be understood that, in this embodiment of this application, a training data set may be obtained (or proportions of data under transmission conditions may be obtained) according to the foregoing training data set obtaining method embodiments in advance, and the neural network model is obtained by using the training data set to train an initially built neural network. Then, the neural network model is applied to a wireless transmission operation process of this embodiment of this application, and wireless transmission of this embodiment of this application is finally implemented through operations.
A wireless transmission application environment of this embodiment of this application may be applied to any wireless transmission environment in which functions of one or more modules in the existing wireless transmission network can be replaced with machine learning, that is, when machine learning is used in wireless transmission to train the neural network, a training data set may be constructed by using the foregoing training data set obtaining method embodiment of this application, and the neural network model is obtained by performing training by using the training data set and then is used for wireless transmission. The wireless transmission application environment is, for example, pilot design, channel estimation, signal detection, user pairing, HARQ, and positioning at a physical layer, resource allocation, switching, and mobility management at a higher layer, and scheduling or slicing at a network layer. A wireless transmission application scenario is not limited in this embodiment of this application.
In this embodiment of this application, data under a plurality of transmission conditions is selected in different proportions according to degrees of contribution of data under different transmission conditions to an optimization objective (or referred to as an objective function or a loss function) of a neural network, to construct a non-uniformly mixed training data set, and training is performed based on the non-uniformly mixed training data set, to obtain one common neural network that is used for actual wireless transmission under different transmission conditions, so that the neural network obtained through training can achieve higher performance under each transmission condition.
Optionally, before the performing an operation of wireless transmission based on a neural network model, the wireless transmission method further includes:
It may be understood that, before a wireless transmission operation is performed by using a neural network model, training needs to be first performed by using a training data set to obtain the neural network model. For example, a training stage of the neural network may be offline performed, and an execution body may be a network-side device, or a terminal-side device, or a network-side device and terminal-side device combination. Optionally, the network-side device is an access network device, a core network device, or a data network device. That is to say, the network-side device in this embodiment of this application may include one or more of a network-side device in an access network, a core network device, and a data network device (DN). The network-side device in the access network may be a base station, or a node responsible for AI training on a RAN side, or is not limited to the type of the access network device 1021 listed in
When the execution body is a network-side device, the training may be centralized training based on a single network-side device, or may be distributed training (for example, federated learning) based on a plurality of network-side devices. When the execution body is a terminal-side device, the training may be centralized training based on a single terminal, or may be distributed training (for example, federated learning) based on a plurality of terminals. When the execution body is a network-side device and terminal-side device combination, a single network-side device may be combined with a plurality of terminal devices, or a single terminal device may be combined with a plurality of network-side devices, or a plurality of network-side devices may be combined with a plurality of terminal-side devices. The execution body of the training process is not limited in this application.
Optionally, the wireless transmission method further includes: sharing, in a process of the distributed training, proportions of the training data under the transmission conditions between bodies executing the distributed training.
It may be understood that, in this embodiment of this application, when the neural network model is trained by using a plurality of network-side devices, or a plurality of terminal devices, or a network-side device and terminal device combination, proportions of training data under a set of transmission conditions are shared between execution devices.
In this embodiment of this application, by share proportions of training data between execution bodies, the execution bodies can train the neural network model without sharing their own data, so that a problem that a single device has an insufficient calculation capability or training capability, or data cannot be shared between devices (related to a privacy problem), or cost of transmitting a large volume of data is very large can be resolved.
Optionally, in a case that the distributed training is joint distributed training of a plurality of network-side devices, any network-side device of the plurality of network-side devices performs calculation to determine the proportions of the training data under the transmission conditions, and sends the proportions to other network-side devices different from the any network-side device of the plurality of network-side devices through a first specified type of interface signaling.
It may be understood that, when a combination of a plurality of network-side devices performs distributed training on the neural network model, all the network-side devices share a same data proportion. In addition, one network-side device of the plurality of network-side devices may perform calculation to obtain proportions of training data under the transmission conditions, and share the proportions to other network-side devices of the plurality of network-side devices through network-side interface signaling. The network-side interface signaling is of a first specified type that is preset. Optionally, the first specified type of interface signaling includes Xn interface signaling, N1 interface signaling, N2 interface signaling, N3 interface signaling, N4 interface signaling, N5 interface signaling, N6 interface signaling, N7 interface signaling, N8 interface signaling, N9 interface signaling, N10 interface signaling, N11 interface signaling, N12 interface signaling, N13 interface signaling, N14 interface signaling, N15 interface signaling, or N22 interface signaling. Data proportion sharing may be implemented between base stations through Xn interface signaling by using an Xn interface, and data proportion sharing may be implemented between core network devices through a corresponding type of interface signaling by using an N1, N2, N3, N4, N5, N6, N7, N8, N9, N10, N11, N12, N13, N14, N15, or N22 interface between the core network devices.
For example, one network-side device may perform calculation to determine data proportion information, and then share the data proportion information to other network-side devices through (including but not limited to) the Xn interface signaling.
Optionally, in a case that the distributed training is joint distributed training of a plurality of terminals, any terminal of the plurality of terminals performs calculation to determine the proportions of the training data under the transmission conditions, and sends the proportions to other terminals different from the any terminal of the plurality of terminals through a second specified type of interface signaling.
It may be understood that, when a combination of a plurality of terminals performs distributed training on the neural network model, all the terminals share a same data proportion. In addition, one terminal of the plurality of terminals may perform calculation to obtain proportions of training data under the transmission conditions, and share the proportions to other terminals of the plurality of terminals through terminal interface signaling. The terminal interface signaling is of a second specified type that is preset. Optionally, the second specified type of interface signaling includes PC5 interface signaling or sidelink interface signaling.
For example, one terminal device performs calculation to determine data proportion information, and then shares the data proportion information to other terminal devices through (including but not limited to) the PC5 interface signaling.
Optionally, in a case that the distributed training is joint distributed training of network-side device(s) and terminal(s), any network-side device or any terminal of the network-side device(s) and the terminal(s) performs calculation to determine the proportions of the training data under the transmission conditions, and sends the proportions to other network-side devices or terminals different from the any network-side device or the any terminal of the network-side device(s) and the terminal(s) through a third specified type of signaling.
It may be understood that, when the execution body is a network-side device and terminal-side device combination, all devices share a same data proportion. In addition, one terminal (or network-side device) in the network-side device and terminal-side device combination may perform calculation to obtain proportions of training data under the transmission conditions, and share the proportions to other devices of the network-side device and terminal-side device combination through a specified type of interface signaling. The specified type of interface signaling is a third specified type of interface signaling that is preset.
Optionally, the third specified type of signaling includes radio resource control (RRC), physical downlink control channel (PDCCH) layer-1 signaling, physical downlink shared channel (PDSCH), medium access control control element (MAC CE), system information block (SIB), Xn interface signaling, N1 interface signaling, N2 interface signaling, N3 interface signaling, N4 interface signaling, N5 interface signaling, N6 interface signaling, N7 interface signaling, N8 interface signaling, N9 interface signaling, N10 interface signaling, N11 interface signaling, N12 interface signaling, N13 interface signaling, N14 interface signaling, N15 interface signaling, N22 interface signaling, physical uplink control channel (PUCCH) layer-1 signaling, physical uplink shared channel (PUSCH), physical random access channel's message1 (PRACH's MSG1), PRACH's MSG3, PRACH's MSG A, PC5 interface signaling, or sidelink interface signaling.
That is to say, between a plurality of network-side devices and terminal-side devices participating in training, data proportion information may be shared through signaling (including but not limited to) such as RRC, PDCCH layer-1 signaling, PDSCH, MAC CE, SIB, Xn interface signaling, N1 interface signaling, N2 interface signaling, N3 interface signaling, N4 interface signaling, N5 interface signaling, N6 interface signaling, N7 interface signaling, N8 interface signaling, N9 interface signaling, N10 interface signaling, N11 interface signaling, N12 interface signaling, N13 interface signaling, N14 interface signaling, N15 interface signaling, N22 interface signaling, PUCCH layer-1 signaling, PUSCH, PRACH's MSG1, PRACH's MSG3, PRACH's MSG A, PC5 interface signaling, or sidelink interface signaling.
Optionally, the method further includes: obtaining real-time data under the transmission conditions and adjusting a trained neural network model online based on the real-time data.
It may be understood that, in this embodiment of this application, a network is pre-trained based on a large volume of data collected offline according to the foregoing embodiments, to make the network convergent, and then fine-tuning is performed on parameters of the pre-trained neural network by using online data of the transmission environment collected in real time, so that the neural network adapts to the actual environment. The fine-tuning may be considered as a training process performed by using the parameters of the pre-trained neural network as initialization. Parameters of some layers may be frozen at a fine-tuning stage. Generally, layers close to an input end are frozen, and layers close to an output end are activated. In this way, it can be ensured that the network still can be convergent. If the data volume at the fine-tuning stage is smaller, it is suggested that more layers are frozen, and fine-tuning is performed on only a few layers close to the output end.
Optionally, when the trained neural network is taken to an actual wireless environment and is subject to fine-tuning or fine-tuning, mixing proportions of data under different transmission conditions at a model training stage may continue to be used. Alternatively, when the trained neural network is taken to an actual wireless environment and is subject to fine-tuning, fine-tuning may be performed directly by using data generated in an actual wireless environment without controlling a data proportion.
In this embodiment of this application, based on completing offline training of the neural network model, online fine-tuning is performed on the trained neural network model, so that the neural network can better adapt to an actual environment.
Optionally, the obtaining real-time data under the transmission conditions and adjusting a trained neural network model online based on the real-time data includes: obtaining the real-time data under the transmission conditions based on the proportions of the training data under the transmission conditions; and skipping, if a proportion of real-time data under any transmission condition of the transmission conditions is greater than a proportion of training data under the any transmission condition, inputting data that is in the real-time data under the any transmission condition and whose proportion is the proportion of the real-time data minus the proportion of the training data under the any transmission condition into the trained neural network model in a process of adjusting the trained neural network model online.
It may be understood that, according to the foregoing embodiments, when the trained neural network model is taken to an actual wireless environment and is subject to fine-tuning or fine-tuning, mixing proportions of data under different transmission conditions at a model training stage may continue to be used. That is to say, proportions or data volumes of real-time data under the transmission conditions at an online fine-tuning stage are determined according to the proportions of the training data under the transmission conditions at the training stage, and corresponding volumes of real-time data under the transmission conditions are obtained according to the proportions or data volumes of the real-time data. When the mixing proportions at the training stage continue to be used, and when a proportion of data under a transmission condition from an actual wireless environment exceeds a proportion under the transmission condition at the model training stage, surplus data is not input into the network to perform fine-tuning.
In this embodiment of this application, when the data proportion at the training stage continues to be used, data under the transmission condition exceeding the proportion is not input into the neural network model to perform fine-tuning, so that unbalanced impact of a data volume exceeding the proportion can be avoided.
Optionally, the obtaining the real-time data under the transmission conditions includes: collecting data of at least one of a network-side device or a terminal under the transmission conditions online as the real-time data under the transmission conditions; and the adjusting the trained neural network model online includes: adjusting, based on the data of the at least one of the network-side device or the terminal under the transmission conditions, the trained neural network model online through the network-side device or the terminal.
It may be understood that, similar to the training stage of the neural network, at the fine-tuning or fine-tuning stage of the neural network model in this embodiment of this application, if the data proportion at the training stage continues to be used, the execution body may alternatively be a network-side device and/or terminal-side device at the input end of the neural network. That is to say, when an execution body is a network-side device, real-time data of the network-side device under the transmission conditions may be online obtained; when an execution body is a terminal device, real-time data of the terminal side under the transmission conditions may be online obtained; and when execution bodies not only include a network-side device but also include a terminal, real-time data under the transmission conditions needs to be obtained for the two execution bodies.
Then, during actual online fine-tuning, the network-side device or terminal performs online fine-tuning on the neural network model according to its own corresponding real-time data, to update network parameters.
Optionally, in a case that the network-side device or the terminal does not obtain the proportions of the training data under the transmission conditions at a training stage, before the obtaining the real-time data under the transmission conditions based on the proportions of the training data under the transmission conditions, the wireless transmission method further includes:
It may be understood that, at the online fine-tuning stage of the neural network model, if an execution body does not obtain data proportion information at the training stage, the execution body needs to first obtain, according to different execution body types and a target execution body type oriented to the training stage, the data proportion information through (including but not limited to) Xn interface signaling, or interface signaling such as Xn interface signaling, N1 interface signaling, N2 interface signaling, N3 interface signaling, N4 interface signaling, N5 interface signaling, N6 interface signaling, N7 interface signaling, N8 interface signaling, N9 interface signaling, N10 interface signaling, N11 interface signaling, N12 interface signaling, N13 interface signaling, N14 interface signaling, N15 interface signaling, and N22 interface signaling, or PC5 interface signaling, or sidelink interface signaling, or interface signaling such as RRC, PDCCH layer-1 signaling, MAC CE, or SIB, and then perform fine-tuning or fine-tuning.
In this embodiment of this application, based on a case that an execution body does not obtain data proportion information at the training stage, the data proportion information is obtained from another execution body through a specified type of interface signaling, so that the data proportion information can be shared at the online fine-tuning stage, to ensure the generalization capability of the neural network.
To describe technical solutions of the embodiments of this application, description is made below using an example, but the scope that this application claims to protect is not limited.
When the training data set is constructed and obtained, data under all transmission conditions may be first mixed at an equal proportion, to determine data volumes or data proportions under the transmission conditions at the equal proportion. Then, all the transmission conditions in the mixing are sorted according to degrees of contribution to an optimization objective of a neural network. Then, based on the premise of ensuring that data under all the transmission conditions is plentiful, a data volume under a transmission condition with a low degree of contribution is increased, and a data volume under a transmission condition with a high degree of contribution is reduced, to determine proportions of data under the transmission conditions, and construct a mixed training data set according to the proportions.
When data volumes under different transmission conditions are determined, “ensuring that data under all the transmission conditions is plentiful” may mean that one threshold is set, and a proportion of a data volume under any transmission condition to a total data volume is not less than the threshold.
When “the data volume under the transmission condition with the low degree of contribution is increased, and the data volume under the transmission condition with the high degree of contribution is reduced”, a lower degree of contribution indicates a larger data volume increased, and a higher degree of contribution indicates a larger data volume reduced. If a sorting manner of the degrees of contribution is from small to large when the degrees of contribution are sorted, proportions of data under corresponding transmission conditions to total data may be in any decrease manner, for example, linear decrease, arithmetic decrease, geometric decrease, exponential function decrease, or power function decrease. Otherwise, if a sorting manner of the degrees of contribution is from large to small, proportions of data under corresponding transmission conditions to total data may be in any increase manner, for example, linear increase, arithmetic increase, geometric increase, exponential function increase, or power function increase.
A feasible implementation of determining proportions of data under transmission conditions is shown in
A median (middle value) of degrees of contribution is found after sorting, and a degree of contribution of each transmission condition is compared with the median sequentially.
If a degree of contribution of an i-th transmission condition is greater than the median of the degrees of contribution, based on the premise of ensuring that data volumes under all transmission conditions are plentiful, the data volume under the i-th transmission condition is reduced, where a reduction amplitude is directly proportional to a difference between the degree of contribution of the i-th transmission condition and the median (middle value) of the degrees of contribution; and
After a training data set is obtained, using the training data set, a neural network model is trained cyclically and iteratively offline and is made convergent, to obtain a trained neural network model.
Then, data in an actual wireless network is collected in real time, and fine-tuning is performed on parameters of the pre-trained neural network model, so that the neural network model can adapt to an actual environment. Online fine-tuning may be considered as a retraining process performed by using the parameters of the pre-trained neural network as initialization.
For example, an application scenario is channel estimation of a demodulation reference signal (DMRS) in wireless transmission, one system including N_RB RBs is considered, and each RB includes N_SC subcarriers and N_Sym symbols. That is to say, the system includes a total of N_RE=N_RB*N_SC*N_Sym time frequency resources. N_SC_DMRS DMRSs and N_Sym_DMRS DMRSs are respectively placed in a frequency domain and a time domain of each RB, and are used for channel estimation. That is to say, the DMRSs occupy a total of N_RE_DMRS=N_RB*N_SC_DMRS*N_Sym_DMRS time frequency resources. A receiving end restores, based on received DMRS signals in positions of the N_RE_DMRS time frequency resources, channel estimation on all the N_RE time frequency resources.
A structure of a neural network for implementing the foregoing process is shown in
When the neural network is trained, training data is a labeled DMRS information pair, that is, one DMRS signal sample (including N_RE_DMRS symbols) corresponds to one label (the label is a channel truth value corresponding to the current DMRS signal sample, and has a total of N_RE_DMRS symbols). During training, a large quantity of labeled DMRS information pairs are used to adjust parameters of the neural network, to minimize a normalized mean square error NMSE between an output of the neural network based on a DMRS signal sample and a label thereof.
Given that the training data includes DMRS information pairs obtained under K SNRs, a k-th signal to noise ratio is denoted as SNRk, and a data volume under the k-th signal to noise ratio is Nk, including a total of Nall=Σk=0KNk pieces of training data. Assuming that the neural network is trained based on federated learning, one network-side device and a plurality of terminal-side devices participate in federated learning.
First, a proportion of data under each SNR during construction of a mixed training data set is determined on the network-side device.
(1) According to degrees of contribution of transmission conditions to the optimization objective, data under the K signal to noise ratios is sorted:
Assuming that the data is mixed at an equal proportion (that is, Nk/Nall=1/K), a degree of contribution of data under the k-th SNR to the foregoing NMSE is denoted as Ck. SNRs are sorted in ascending order (or descending order) of Ck. In this embodiment, a degree of contribution of data under a low SNR to the NMSE is large, and a degree of contribution of data under a high SNR to the NMSE is small. A median of all the degrees of contribution in the sorting is denoted as
(2) Mixing proportions are determined:
Then, the network-side device sends, through interface signaling such as RRC, PDCCH layer-1 signaling, MAC CE, or SIB, data proportion information under each SNR to all terminals participating in joint training, to perform federated learning of the neural network, that is, offline training.
After the offline training of the neural network is completed, online fine-tuning is performed on the trained neural network in an actual wireless network. Because data proportion information is shared to all the terminals at the offline training stage, the data proportion information may continue to be used at the online fine-tuning stage. When a proportion of data under an SNR from an actual wireless environment exceeds a proportion under the SNR at the model training stage, surplus data is not input into the neural network to perform fine-tuning.
This embodiment of this application can improve the generalization capability of the neural network model obtained through training in a changing wireless environment.
It should be noted that an execution body of the wireless transmission method provided in this embodiment of this application may be a wireless transmission apparatus, or a control module configured to execute the wireless transmission method in the wireless transmission apparatus. In the embodiments of this application, a wireless transmission apparatus provided in the embodiments of this application is described using an example in which a wireless transmission apparatus executes the wireless transmission method.
A structure of a wireless transmission apparatus of an embodiment of this application is shown in
The neural network model is obtained by performing training by using a training data set in advance, and the training data set is obtained based on the training data set obtaining method according to the foregoing embodiments.
Optionally, the wireless transmission apparatus further includes:
Optionally, the network-side device is an access network device, a core network device, or a data network device.
Optionally, the wireless transmission apparatus further includes:
Optionally, in a case that the distributed training is joint distributed training of a plurality of network-side devices, the fourth processing module is configured by any network-side device of the plurality of network-side devices to perform calculation to determine the proportions of the training data under the transmission conditions, and send the proportions to other network-side devices different from the any network-side device of the plurality of network-side devices through a first specified type of interface signaling.
Optionally, the first specified type of interface signaling includes Xn interface signaling, N1 interface signaling, N2 interface signaling, N3 interface signaling, N4 interface signaling, N5 interface signaling, N6 interface signaling, N7 interface signaling, N8 interface signaling, N9 interface signaling, N10 interface signaling, N11 interface signaling, N12 interface signaling, N13 interface signaling, N14 interface signaling, N15 interface signaling, or N22 interface signaling.
Optionally, in a case that the distributed training is joint distributed training of a plurality of terminals, the fourth processing module is configured by any terminal of the plurality of terminals to perform calculation to determine the proportions of the training data under the transmission conditions, and send the proportions to other terminals different from the any terminal of the plurality of terminals through a second specified type of interface signaling.
Optionally, the second specified type of interface signaling includes PC5 interface signaling or sidelink interface signaling.
Optionally, in a case that the distributed training is joint distributed training of network-side device(s) and terminal(s), the fourth processing module is configured by any network-side device or any terminal of the network-side device(s) and the terminal(s) to perform calculation to determine the proportions of the training data under the transmission conditions, and send the proportions to other network-side devices or terminals different from the any network-side device or the any terminal of the network-side device(s) and the terminal(s) through a third specified type of signaling.
Optionally, the third specified type of signaling includes RRC, PDCCH layer-1 signaling, PDSCH, MAC CE, SIB, Xn interface signaling, N1 interface signaling, N2 interface signaling, N3 interface signaling, N4 interface signaling, N5 interface signaling, N6 interface signaling, N7 interface signaling, N8 interface signaling, N9 interface signaling, N10 interface signaling, N11 interface signaling, N12 interface signaling, N13 interface signaling, N14 interface signaling, N15 interface signaling, N22 interface signaling, PUCCH layer-1 signaling, PUSCH, PRACH's MSG1, PRACH's MSG3, PRACH's MSG A, PC5 interface signaling, or sidelink interface signaling.
Optionally, the wireless transmission apparatus further includes:
Optionally, the fine-tuning module is configured to:
Optionally, when being configured to obtain the real-time data under the transmission conditions, the fine-tuning module is configured to:
Optionally, the wireless transmission apparatus further includes:
The wireless transmission apparatus in this embodiment of this application may be an apparatus, an apparatus with an operating system, or an electronic device, or may be a component, an integrated circuit, or a chip in a terminal or network-side device. The apparatus or electronic device may be a mobile terminal, or may be a non-mobile terminal, or may include but is not limited to the foregoing listed types of the network-side device 102. For example, the mobile terminal may include but is not limited to the foregoing listed types of the terminal 101, and the non-mobile terminal may be a server, a network attached storage (NAS), a personal computer (PC), a television (TV), a teller machine, or a self-service machine, which is not limited in the embodiments of this application.
The wireless transmission apparatus in the embodiments of this application may be an apparatus with an operating system. The operating system may be an Android operating system, may be an iOS operating system, or may be another possible operating system, and is not limited in the embodiments of this application.
The wireless transmission apparatus provided in this embodiment of this application can implement each process implemented in the wireless transmission method embodiments of
As shown in
An embodiment of this application further provides a communications device. The communications device may be a terminal or a network-side device. The communications device includes a processor and a communications interface, where the processor is configured to determine, based on degrees of contribution of transmission conditions to an optimization objective of a neural network, data volumes of training data under the transmission conditions; and obtain, based on the data volumes of the training data under the transmission conditions, the training data under the transmission conditions, to form a training data set for training the neural network, where the degrees of contribution of the transmission conditions to the optimization objective of the neural network represent degrees of impact of the transmission conditions on a value of the optimization objective of the neural network. It should be noted that the communications device embodiment corresponds to the foregoing training data set obtaining method embodiment, and each implementation process and implementation manner of the foregoing network side device embodiment is applicable to the communications device embodiment, and can achieve the same technical effect.
An embodiment of this application further provides a communications device. The communications device may be a terminal or a network-side device. The communications device includes a processor and a communications interface, where the processor is configured to perform an operation of wireless transmission based on a neural network model, to implement the wireless transmission, where the neural network model is obtained by performing training by using a training data set in advance, and the training data set is obtained based on the training data set obtaining method according to the foregoing embodiments. It should be noted that the communications device embodiment corresponds to the foregoing wireless transmission method embodiment, and each implementation process and implementation manner of the foregoing network side device embodiment is applicable to the communications device embodiment, and can achieve the same technical effect.
For example,
A person skilled in the art may understand that, the terminal 1000 may further include a power supply (such as a battery) for supplying power to the components. The power supply may be logically connected to the processor 1010 by using a power management system, thereby implementing functions such as charging, discharging, and power consumption management by using the power management system. A terminal structure shown in
It should be understood that, in this embodiment of this application, the input unit 1004 may include a graphics processing unit (GPU) 10041 and a microphone 10042. The graphics processing unit 10041 performs processing on image data of a static picture or a video that is obtained by an image acquisition device (for example, a camera) in a video acquisition mode or an image acquisition mode. The display unit 1006 may include a display panel 10061, for example, a display panel 10061 configured in a form such as a liquid crystal display or an organic light-emitting diode. The user input unit 1007 includes a touch panel 10071 and another input device 10072. The touch panel 10071 is also referred to as a touchscreen The touch panel 10071 may include two parts: a touch detection apparatus and a touch controller. The another input device 10072 may include, but not limited to, a physical keyboard, a functional key (such as a volume control key or a switch key), a track ball, a mouse, and a joystick, which are not described herein in detail.
In this embodiment of this application, the radio frequency unit 1001 receives downlink data from a network side device, and then delivers the downlink data to the processor 1010 for processing; and in addition, sends uplink data to the network side device. Generally, the radio frequency unit 1001 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
The memory 1009 may be configured to store a software program or instruction and various data. The memory 1009 may mainly include a program or instruction storage area and a data storage area. The program or instruction storage region may store an operating system, an application or instruction required by at least one function (for example, a voice playing function or an image playing function), and the like. In addition, the memory 1009 may include a high-speed random access memory, or may further include a non-volatile memory, or the like. The nonvolatile memory may be a read-only memory (ROM), a programmable read-only memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), or a flash memory. For example, the nonvolatile memory is at least one magnetic disk storage device, a flash memory device, or other non-volatile solid-state storage devices. Volatile memories may be a random access memory (RAM), a static random access memory (SRAM), a dynamic random access memory (DRAM), a synchronous dynamic random access memory (SDRAM), a double data rate synchronous dynamic random access memory (DDRSDRAM), an enhanced synchronous dynamic random access memory (ESDRAM), a synchronous link dynamic random access memory (SLDRAM), and a direct rambus dynamic random access memory (DRRAM). The memory 1009 in this embodiment of this application includes but is not limited to these memories and any other suitable types of memories.
The processor 1010 may include one or more processing units. Optionally, the processor 1010 may integrate an application processor and a modem processor. The application processor mainly processes an operating system, a user interface, an application or instruction, and the like. The modem processor mainly processes wireless communication, and is, for example, a baseband processor. It may be understood that the foregoing modem processor may not be integrated into the processor 1010.
The processor 1010 is configured to determine, based on degrees of contribution of transmission conditions to an optimization objective of a neural network, data volumes of training data under the transmission conditions; and obtain, based on the data volumes of the training data under the transmission conditions, the training data under the transmission conditions, to form a training data set for training the neural network, where the degrees of contribution of the transmission conditions to the optimization objective of the neural network represent degrees of impact of the transmission conditions on a value of the optimization objective of the neural network.
In this embodiment of this application, when a training data set is constructed in a communication system based on artificial intelligence, data under a plurality of transmission conditions is selected in different proportions according to degrees of contribution of data under different transmission conditions to an optimization objective (or referred to as an objective function or a loss function) of a neural network, to construct a mixed training data set, so that a generalization capability of the neural network can be effectively improved.
Optionally, the processor 1010 is further configured to sort the degrees of contribution of the transmission conditions; and execute, based on equal-proportion mixing, at least one of reducing a data volume of training data under a transmission condition corresponding to a higher degree of contribution in the sorting or increasing a data volume of training data under a transmission condition corresponding to a lower degree of contribution in the sorting.
In this embodiment of this application, by sorting the degrees of contribution of the transmission conditions and adjusting a data volume under a corresponding transmission condition according to the sorting, an adjustment policy on the data volume under the corresponding transmission condition (including whether the data volume needs to be increased or reduced, an increase or reduction amplitude, and the like) can be determined more explicitly and accurately, thereby making the efficiency higher and the result more accurate.
Optionally, the processor 1010 is further configured to: collect, based on the data volumes of the training data under the transmission conditions, data under the transmission conditions and calibrate the data, to form the training data set under the transmission conditions; or collect specified volumes of data under the transmission conditions, and select, based on the data volumes of the training data under the transmission conditions, partial data from the specified volumes of data and calibrating the partial data, or replenish the specified volumes of data and calibrate replenished data, to form the training data set under the transmission conditions.
Optionally, the processor 1010 is further configured to execute, according to a rule, at least one of reducing a data volume of training data under a transmission condition corresponding to a higher degree of contribution in the sorting or increasing a data volume of training data under a transmission condition corresponding to a lower degree of contribution in the sorting, where the rule includes: that a larger value of the higher degree of contribution indicates a larger amplitude of the reducing; and that a smaller value of the lower degree of contribution indicates a larger amplitude of the increasing.
In this embodiment of this application, data volumes of training data under corresponding transmission conditions are increased or reduced proportionally according to magnitudes of values of degrees of contribution, so that as the degrees of contribution of the transmission conditions are gradually increased, the data volumes of the training data under the transmission conditions are gradually reduced, thereby better balancing impact of the transmission conditions on the final neural network and better facilitating an improvement in the generalization capability of the neural network.
Optionally, the processor 1010 is further configured to determine a reference degree of contribution according to the sorting, and compare a degree of contribution of a transmission condition with the reference degree of contribution; and determine, if the degree of contribution of the transmission condition is greater than the reference degree of contribution, that the degree of contribution of the transmission condition is the higher degree of contribution, and reduce a data volume of training data under the transmission condition, or otherwise, determine that the degree of contribution of the transmission condition is the lower degree of contribution, and increase the data volume of the training data under the transmission condition.
In this embodiment of this application, by determining an intermediate contrast reference quantity of degrees of contribution, it is only necessary to compare another degree of contribution with the contrast reference quantity, to determine according to a comparison result to increase or reduce a data volume under a corresponding transmission condition, so that the algorithm is simple and the operation quantity is small.
Optionally, the processor 1010 is further configured to determine, based on probability densities of the transmission conditions in an actual application, weighting coefficients corresponding to the transmission conditions; and determine, based on the degrees of contribution of the transmission conditions to the optimization objective and with reference to the weighting coefficients, the data volumes of the training data under the transmission conditions.
In this embodiment of this application, designing a weighted term according to an actual probability density of a transmission condition can better adapt to an actual environment.
Optionally, the radio frequency unit 1001 is configured to send the training data set to a target device, where the target device is configured to train the neural network based on the training data set.
In this embodiment of this application, by sending the obtained training data set to the second device different from the current device, data sharing and joint training can be implemented between a plurality of devices, thereby effectively reducing an operation quantity of a single device and effectively improving operation efficiency.
Optionally, the radio frequency unit 1001 is configured to directly send the training data set to the target device, or send the training data set on which specified conversion is performed to the target device; and the processor 1010 is further configured to perform the specified conversion on the training data set, where the specified conversion includes at least one of specific quantization, specific compression, or processing that is performed according to a neural network agreed or configured in advance.
Optionally, the processor 1010 is further configured to perform an operation of wireless transmission based on a neural network model, to implement the wireless transmission, where the neural network model is obtained by performing training by using a training data set in advance, and the training data set is obtained based on the training data set obtaining method according to the foregoing training data set obtaining method embodiments.
In this embodiment of this application, data under a plurality of transmission conditions is selected in different proportions according to degrees of contribution of data under different transmission conditions to an optimization objective (or referred to as an objective function or a loss function) of a neural network, to construct a non-uniformly mixed training data set, and training is performed based on the non-uniformly mixed training data set, to obtain one common neural network that is used for actual wireless transmission under different transmission conditions, so that the neural network obtained through training can achieve higher performance under each transmission condition.
Optionally, the processor 1010 is further configured to perform, based on the training data set in any one of training manners, training to obtain the neural network model, where the training manners include:
Optionally, the radio frequency unit 1001 is further configured to share, in a process of the distributed training, proportions of the training data under the transmission conditions between bodies executing the distributed training.
In this embodiment of this application, by share proportions of training data between execution bodies, the execution bodies can train the neural network model without sharing their own data, so that a problem that a single device has an insufficient calculation capability or training capability, or data cannot be shared between devices (related to a privacy problem), or cost of transmitting a large volume of data is very large can be resolved.
Optionally, the processor 1010 is further configured, in a case that the distributed training is joint distributed training of a plurality of network-side devices, by any network-side device of the plurality of network-side devices to perform calculation to determine the proportions of the training data under the transmission conditions; and
Optionally, the processor 1010 is further configured, in a case that the distributed training is joint distributed training of a plurality of terminals, by any terminal of the plurality of terminals to perform calculation to determine the proportions of the training data under the transmission conditions; and
Optionally, the processor 1010 is further configured, in a case that the distributed training is joint distributed training of network-side device(s) and terminal(s), by any network-side device or any terminal of the network-side device(s) and the terminal(s) to perform calculation to determine the proportions of the training data under the transmission conditions; and
Optionally, the processor 1010 is further configured to obtain real-time data under the transmission conditions and adjust a trained neural network model online based on the real-time data.
In this embodiment of this application, based on completing offline training of the neural network model, online fine-tuning is performed on the trained neural network model, so that the neural network can better adapt to an actual environment.
Optionally, the processor 1010 is further configured to obtain the real-time data under the transmission conditions based on the proportions of the training data under the transmission conditions; and skip, if a proportion of real-time data under any transmission condition of the transmission conditions is greater than a proportion of training data under the any transmission condition, inputting data that is in the real-time data under the any transmission condition and whose proportion is the proportion of the real-time data minus the proportion of the training data under the any transmission condition into the trained neural network model in a process of adjusting the trained neural network model online.
In this embodiment of this application, when the data proportion at the training stage continues to be used, data under the transmission condition exceeding the proportion is not input into the neural network model to perform fine-tuning, so that unbalanced impact of a data volume exceeding the proportion can be avoided.
Optionally, the input unit 1004 is configured to collect data of at least one of a network-side device or a terminal under the transmission conditions online as the real-time data under the transmission conditions; and
Optionally, in a case that the communications device is a network-side device, the radio frequency unit 1001 is further configured to obtain the proportions of the training data under the transmission conditions from a network-side device at the training stage through any network-side interface signaling of Xn interface signaling, N1 interface signaling, N2 interface signaling, N3 interface signaling, N4 interface signaling, N5 interface signaling, N6 interface signaling, N7 interface signaling, N8 interface signaling, N9 interface signaling, N10 interface signaling, N11 interface signaling, N12 interface signaling, N13 interface signaling, N14 interface signaling, N15 interface signaling, and N22 interface signaling, or obtain the proportions of the training data under the transmission conditions from a terminal at the training stage through any signaling of PUCCH layer-1 signaling, PUSCH, PRACH's MSG1, PRACH's MSG3, and PRACH's MSG A; and
In this embodiment of this application, based on a case that an execution body does not obtain data proportion information at the training stage, the data proportion information is obtained from another execution body through a specified type of interface signaling, so that the data proportion information can be shared at the online fine-tuning stage, to ensure the generalization capability of the neural network.
For example,
The frequency band processing apparatus may be located in the baseband apparatus 1103. The method performed by the network side device in the foregoing embodiments may be implemented in the baseband apparatus 1103. The baseband apparatus 1103 includes a processor 1104 and a memory 1105.
The baseband apparatus 1103 may include, for example, at least one baseband plate. A plurality of chips are disposed on the baseband plate. As shown in
The baseband apparatus 1103 may further include a network interface 1106, configured to exchange information with the radio frequency apparatus 1102. The interface is, for example, a common public radio interface (CPRI).
For example, the access network device in this embodiment of the present application further includes: an instruction or a program stored in the memory 1105 and executable on the processor 1104. The processor 1104 invokes the instruction or program in the memory 1105 to perform the method performed by the modules shown in
For example,
In this embodiment of this application, the core network device 1200 further includes: a computer program stored in the memory 1203 and executable on the processor 1201. When the computer program is executed by the processor 1201, the method performed by the modules shown in
In
The processor 1201 is responsible for the management of the bus architecture and normal processing, and the memory 1203 may store data used when the processor 1201 performs an operation.
An embodiment of this application further provides a non-transitory computer-readable storage medium. The non-transitory computer-readable storage medium stores a program or an instruction. When the program or instruction is executed by a processor, each process of the foregoing training data set obtaining method embodiment is implemented or each process of the foregoing wireless transmission method embodiment is implemented, and the same technical effect can be achieved. To avoid repetition, details are not described herein again.
The processor is a processor in the terminal or network-side device described in the foregoing embodiments. The non-transitory computer-readable storage medium includes a computer read-only memory (ROM), a random access memory (RAM), a magnetic disk, an optical disc, or the like.
An embodiment of this application further provides a chip. The chip includes a processor and a communications interface. The communications interface is coupled to the processor. The processor is configured to execute a program or an instruction, to implement each process of the foregoing training data set obtaining method embodiment or implement each process of the foregoing wireless transmission method embodiment, and can achieve the same technical effect. To avoid repetition, details are not described herein again.
It should be noted that, the chip mentioned in the embodiments of this application may also be referred to as a system-level chip, a system chip, a chip system, a system on chip, or the like.
It should be noted that the terms “include”, “include”, or any other variation thereof in this specification is intended to cover a non-exclusive inclusion, which specifies the presence of stated processes, methods, objects, or apparatuses, but do not preclude the presence or addition of one or more other processes, methods, objects, or apparatuses. Without more limitations, elements defined by the sentence “including one” does not exclude that there are still other same elements in the processes, methods, objects, or apparatuses. In addition, it should be noted that the range of the method and apparatus in the implementation manners of this application is not limited to performing functions in the shown or discussed order, and may further include performing functions in a basically simultaneous manner or in reverse order according to the involved functions. For example, the described method may be performed in an order different from the described order, and steps may be further added, omitted, or combined. In addition, features described with reference to some examples may be combined in other examples.
Through the descriptions of the foregoing implementations, a person skilled in the art may clearly understand that the method according to the foregoing embodiments may be implemented by means of software and a necessary general hardware platform, and certainly, may alternatively be implemented by hardware. Based on such an understanding, the technical solutions of this application essentially or the part contributing to the prior art may be implemented in a form of a computer software product. The computer software product is stored in a storage medium (such as a ROM/RAM, a magnetic disk, or an optical disc) and includes several instructions for instructing a terminal (which may be a mobile phone, a computer, a server, a network device, or the like) to perform the methods described in the embodiments of this application.
The embodiments of this application have been described above with reference to the accompanying drawings. This application is not limited to the embodiments described above, and the embodiments described above are merely exemplary and not limitative. Those of ordinary skill in the art may make various variations under the teaching of this application without departing from the spirit of this application and the protection scope of the claims, and such variations shall all fall within the protection scope of this application.
Number | Date | Country | Kind |
---|---|---|---|
202110513732.0 | May 2021 | CN | national |
This application is a Bypass Continuation Application of International Patent Application No. PCT/CN2022/092144, filed May 11, 2022, and claims priority to Chinese Patent Application No. 202110513732.0, filed May 11, 2021, the disclosures of which are hereby incorporated by reference in their entireties.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2022/092144 | May 2022 | US |
Child | 18388635 | US |