This application relates to the field of communication technologies, and specifically, to a model training processing method and apparatus, a terminal, and a network side device.
With the development of communication technologies, a communication scenario based on artificial intelligence (Artificial Intelligence, AI) has been introduced in wireless communication. Currently, in many scenarios of wireless communication based on AI, it is difficult to obtain a large volume of labeled data. Without a large volume of labeled data, it is impossible to train a suitable model through supervised learning, resulting in low reliability in communication.
According to a first aspect, a model training processing method is provided, including:
According to a second aspect, a model training processing method is provided, including:
According to a third aspect, a model training processing apparatus is provided, including:
According to a fourth aspect, a model training processing apparatus is provided, including:
According to a fifth aspect, a model training processing apparatus is provided, including:
According to a sixth aspect, a terminal is provided, including a processor and a memory, where the memory stores a program or an instruction runnable on the processor, and the program or the instruction, when executed by the processor, implements the steps of the method according to the first aspect.
According to a seventh aspect, a terminal is provided, including a processor and a communication interface.
When the terminal is a first device, the processor is configured to obtain first information, the first information including first data; and process the first data by using a first model to obtain second data, where both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the terminal is a first device, the communication interface is configured to send second information to the first device, the second information including a first model, and the first model being used by the first device to obtain second data based on first data, where both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the terminal is a first device, the communication interface is configured to send first information to the first device, the first information including first data, and the first data being used by the first device to obtain second data based on a first model, where both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
According to an eighth aspect, a network side device is provided, including a processor and a memory, where the memory stores a program or an instruction runnable on the processor, and the program or the instruction, when executed by the processor, implements the steps of the method according to the second aspect.
According to a ninth aspect, a network side device is provided, including a processor and a communication interface.
When the network side device is a first device, the processor is configured to obtain first information, the first information including first data; and process the first data by using a first model to obtain second data, where both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the network side device is a first device, the communication interface is configured to send second information to the first device, the second information including a first model, and the first model being used by the first device to obtain second data based on first data, where both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the network side device is a first device, the communication interface is configured to send first information to the first device, the first information including first data, and the first data being used by the first device to obtain second data based on a first model, where both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
According to a tenth aspect, a readable storage medium is provided, storing a program or an instruction, where the program or the instruction, when executed by a processor, implements the steps of the method according to the first aspect, or implements the steps of the method according to the second aspect.
According to an eleventh aspect, a chip is provided, including a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement the steps of the method according to the first aspect, or implement the steps of the method according to the second aspect.
According to a twelfth aspect, a computer program/program product is provided, stored in a storage medium and executed by at least one processor to implement the steps of the method according to the first aspect, or implement the steps of the method according to the second aspect.
The technical solutions in embodiments of this application are clearly described in the following with reference to the accompanying drawings in the embodiments of this application. Apparently, the described embodiments are merely some rather than all of the embodiments of this application. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of this application fall within the protection scope of this application.
The terms “first”, “second”, and so on in this specification and claims of this application are intended to distinguish between similar objects but are not intended to describe a specific order or sequence. It should be understood that the terms used in such a way are interchangeable in proper circumstances, so that the embodiments of this application can be implemented in other sequences than those illustrated or described herein. In addition, the objects distinguished by “first” and “second” are usually of one type, and there is no limitation on quantities of the objects. For example, there may be one or more first objects. In addition, “and/or” in this specification and the claims indicate at least one of the connected objects, and the character “/” usually indicates an “or” relationship between the associated objects.
It should be noted that, the technologies described in the embodiments of this application are not limited to a Long Term Evolution (Long Term Evolution, LTE)/LTE-Advanced (LTE-Advanced, LTE-A) system, and can be further used in other wireless communication systems, such as Code Division Multiple Address (Code Division Multiple Access, CDMA), Time Division Multiple Access (Time Division Multiple Access, TDMA), Frequency Division Multiple Access (Frequency Division Multiple Access, FDMA), orthogonal frequency division multiple access (Orthogonal Frequency Division Multiple Access, OFDMA), single-carrier frequency division multiple access (Single-carrier Frequency-Division Multiple Access, SC-FDMA), and other systems. The terms “system” and “network” in the embodiments of this application are often used interchangeably, and the described technologies can be used not only for the above-mentioned systems and radio technologies, but also for other systems and radio technologies. The following description describes a New Radio (New Radio, NR) system for exemplary purposes, and uses NR terms in most of the following descriptions, but these technologies are also applicable to applications other than the NR system application, such as a 6th generation (6th Generation, 6G) communication system.
For case of understanding, some contents involved in the embodiments of this application are described below.
At present, artificial intelligence is widely applied in various fields. There are a plurality of implementations of AI modules, for example, a neural network, a decision tree, a support vector machine, and a Bayes classifier. In this application, a neural network is used as an example for description, but a specific type of the AI module is not limited.
The neural network includes neurons, and a schematic diagram of the neurons is shown in
Parameters of the neural network are optimized by using an optimization algorithm. Optimization algorithms are a type of algorithms that can help minimize or maximize an objective function (sometimes referred to as a loss function). The objective function is usually a mathematical combination of a model parameter and data. For example, data x and a label Y corresponding to the data x are given, and a neural network model f (.) is constructed. With this model, a predicted output f (x) may be obtained based on an input x, and a gap (f(x)−Y) between a predicted value and a real value may be calculated. This is a loss function. The objective of this application is to find proper W, b to minimize the value of the loss function. A smaller loss value indicates that the model is closer to the real situation.
Currently, common optimization algorithms are basically based on an error back propagation (error Back Propagation, BP) algorithm. A basic idea of the BP algorithm is that a learning process includes two processes: signal forward propagation and error back propagation. During forward propagation, an input sample is transmitted from an input layer, processed layer by layer by each hidden layer, and then transmitted to an output layer. If an actual output of the output layer does not match an expected output, an error back propagation stage is performed. The error back propagation is to transmit an output error in a form layer by layer back to the input layer through hidden layers, and distribute the error to all units at each layer, to obtain an error signal of the units at each layer. This error signal is used as a basis for correcting a weight of each unit. Such a weight adjustment process at each layer of signal forward propagation and error back propagation is performed cyclically. The process of continuously adjusting the weight is a learning and training process of the network. This process continues until an error outputted by the network is reduced to an acceptable degree or a preset quantity of times of learning is performed.
According to different types of solutions, selected AI algorithms and used models also vary. Currently, a main method for improving performance of a 5G network through AI is to enhance or replace an existing algorithm or processing module by using an algorithm and a model based on a neural network. In a specific scenario, an algorithm and a model based on a neural network can achieve better performance than a deterministic algorithm. Commonly used neural networks include a deep neural network, a convolutional neural network, a recurrent neural network, and the like. With the help of existing AI tools, neural networks can be built, trained, and verified.
It should be understood that training of an AI model requires support of a large volume of data. If the data volume is insufficient, the training process of the model may not converge, or a trained model may be overfitted. However, in many scenarios in wireless communication, labeled data cannot be obtained, or a volume of labeled data is small (due to collection overheads, transmission overheads, and the like). Therefore, a problem of model training when there is insufficient labeled data needs to be resolved in wireless communication. Therefore, a model training processing method of this application is provided.
A model training processing method provided in the embodiments of this application is described in detail below through some embodiments and application scenarios thereof with reference to the accompanying drawings.
As shown in
Step 301: A first device obtains first information, the first information including first data.
Step 302: The first device processes the first data by using a first model to obtain second data.
Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
In this embodiment of this application, the first device may be a network side device or a terminal, and the first data may be at least part of data for training the second model. The first data may be labeled data or unlabeled data. The first model may be understood as a model that enhances training data of the second model. For example, when the first data is labeled data, the first model is used for expanding the first data, to obtain the second data with a larger data volume. When the first data is unlabeled data, the first model is used for labeling the first data, so that more labeled training data can be obtained. In this way, after the first data is processed through the first model, more labeled training data can be obtained, thereby ensuring that there is enough labeled training data to train the second model, so that the second model can effectively converge during the training process, and performance of the second model is improved.
For example, in some embodiments, the first data is N pieces of labeled data. In this case, after the N pieces of labeled data are inputted into the first model, M pieces of labeled data (that is, the second data is M pieces of labeled data) may be outputted. In this case, M is greater than N, and usually, M is much greater than N.
For another example, in some embodiments, the first data is M pieces of unlabeled data. In this case, after the M pieces of unlabeled data are inputted into the first model, M pieces of labeled data (that is, the second data is M pieces of labeled data) may be outputted.
Optionally, the first information may be stored in the first device or the second device. In addition, the first model may be stored in the first device, or may be stored in the second device, and is sent by the second device to the first device. When the first device is a core network device, the second device may be a base station. When the first device is a base station (for example, a base station A), the second device may be a base station (for example, a base station B) or a terminal. When the first device is a terminal (for example, a terminal A), the second device may be a base station or a terminal (for example, a terminal B). It should be understood that, in this embodiment of this application, the first information and the first model are stored in different devices.
In this embodiment of this application, a first device obtains first information, the first information including first data; and the first device processes the first data by using a first model to obtain second data. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data. In this way, more labeled training data can be obtained by using the first model, so that the second model can effectively converge in a training process, thereby improving performance of the second model. Therefore, reliability of AI-based wireless communication can be improved in this embodiment of this application.
Optionally, in some implementations, the first model is stored in the second device. In this case, before the processing, by the first device, the first data by using a first model, the method further includes:
In this embodiment of this application, that the second information includes the first model may be understood as follows: The second information includes a parameter of the first model or address information of the first model, so that the first device can obtain the first model.
Further, in some embodiments, the second information further includes at least one of configuration information and first assistance information. The configuration information is used for indicating a usage manner of the first model. The first assistance information includes statistical information and environment information required for running the first model. The statistical information is used for representing a distribution feature of an input of the first model.
Optionally, the configuration information is used for indicating a usage method for the first model, and may include, for example, a data dimension or an input data format, an output dimension or an output data format, an input data volume, and an output data volume of the first model. The environment information may be understood as environment information related to a data augmentation algorithm of the first model. The environment information may include a software environment, a hardware environment, and the like required for running the model, for example, may include a software architecture, a hardware architecture, a power demand, a storage demand, and a computing power demand that need to be used. The statistical information may include distribution feature information such as a mean value and a variance of inputs of the model.
Optionally, in some embodiments, before the receiving, by the first device, second information from a second device, the method further includes:
In this embodiment of this application, when the first device needs to expand the first data, the second information may be obtained in a requested manner, thereby improving pertinence of obtaining the second information. Certainly, in other embodiments, the second device may alternatively proactively send the second information to the first device. For example, when the second device establishes a connection to the first device, the second device sends the second information to the first device. Alternatively, the second device may broadcast the second information, and the first device directly obtains the second information from broadcast information when needed.
Optionally, in some embodiments, after the processing, by the first device, the first data by using a first model to obtain second data, the method further includes:
In this embodiment of this application, the first device may train the second model to obtain the third model. The third model may be used on the first device or the second device.
It should be understood that the second model may be sent by the second device to the first device, or may be preconfigured in the first device in a protocol. This is not further limited herein.
Optionally, in some embodiments, if use of the third model is on the second device, after the training, by the first device, the second model based on the second data to obtain a third model, the method further includes:
In this embodiment of this application, the sending the third model may be understood as sending a parameter of the third model or sending address information of the third model. This is not further limited herein. In this way, when inference of a corresponding service is performed by using the trained third model, accuracy of the inference can be improved, thereby ensuring communication reliability.
It should be noted that, initial training data of the second model may include the first data, and may further include labeled third data. If the first data is labeled data, the training data used may include the first data, the second data, and the third data during the training of the second model. If the first data is unlabeled data, the training data used may include the second data and the third data during the training of the second model.
Optionally, based on different storage locations of the first information, corresponding manners of obtaining the first information are different. For example, in some embodiments, the obtaining, by a first device, first information includes either of the following:
In this embodiment of this application, when the first information is stored in the second device, the first device may receive the first information from the second device. When the first information is stored in the first device, the first device may obtain the first information locally.
Optionally, before the receiving, by the first device, the first information from a second device, the method further includes:
In this embodiment of this application, the first information is stored in the second device, and the first device needs to instruct the second device to send the first information. For example, the first device may schedule the second device to send the first information.
Optionally, in some embodiments, before the receiving, by the first device, the first information from a second device, the method further includes:
In this embodiment of this application, before the second device sends the first information, the second device may request the first device to send the first information. Subsequently, the second device may send the first information on a preconfigured resource, or the first device may dynamically schedule the second device to send the first information. For example, the second device is instructed, through the instruction information, to send the first information.
Optionally, in a case that the first device receives the first information from the second device, after the processing, by the first device, the first data by using a first model to obtain second data, the method further includes:
In this embodiment of this application, the process of training the second model is performed by the second device, and the first device needs to send the second data to the second device, for the second device to train the second model to obtain the third model. The training of the second model by the second device is similar to the training of the second model by the first device. For the definition of the training data, refer to the foregoing example. Details are not described herein again.
Further, in some embodiments, the third information further includes identification information, and the identification information is used for indicating that the second data is obtained based on the first model.
Optionally, in some embodiments, the first information further includes second assistance information, and the second assistance information is used for representing a distribution feature of the first data.
In this embodiment of this application, the second assistance information may include information representing distribution features such as a mean value and a variance of the first data.
For better understanding of this application, description is provided below through some examples.
In some embodiments, a device A sends a first model to a device B, the device B performs data enhancement by using the received first model and first data of the device B to obtain second data, and the device B then trains a second model based on the second data. As shown in
Step 401: A device B sends a first message to a device A, the first message being used for requesting a first model, a configuration parameter, and first assistance information.
Step 402: A device A sends the first model, the configuration parameter, and the first assistance information to the device B.
Step 403: The device B enhances first data based on the first model, the configuration parameter, and the first assistance information to obtain second data.
Step 404: The device B trains a second model based on the second data to obtain a third model.
In some embodiments, the device B sends the first data to the device A, the device A performs data enhancement by using the received first data and a first model trained by the device A, to obtain second data, and then the device A trains a second model based on the second data to obtain a third model. Finally, the third model is sent to the device B. As shown in
Step 501: A device B sends a second message to a device A, the second message being used for requesting to send first data.
Step 502: The device A sends a third message to the device B, the third message being used for instructing to send the first data.
Step 503: The device B sends the first data to the device A.
Step 504: The device A enhances the first data based on a first model of the device A, a configuration parameter, and first assistance information to obtain second data.
Step 505: The device A trains a second model based on the second data to obtain a third model.
Step 506: The device A sends the third model to the device B.
In some embodiments, the device B sends the first data to the device A, the device A performs data enhancement by using the received first data and a first model trained by the device A, to obtain second data, the device A then sends the second data to the device B, and the device B trains a second model by using the received second data. As shown in
Step 601: A device B sends a second message to a device A, the second message being used for requesting to send first data.
Step 602: The device A sends a third message to the device B, the third message being used for instructing to send the first data.
Step 603: The device B sends the first data to the device A.
Step 604: The device A enhances the first data based on a first model of the device A, a configuration parameter, and first assistance information to obtain second data.
Step 605: The device A sends the second data to the device B.
Step 606: The device B trains a second model based on the second data to obtain a third model.
Referring to
Step 701: A second device sends second information to a first device, the second information including a first model, and the first model being used by the first device to obtain second data based on first data.
Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Optionally, the second information further includes at least one of configuration information and first assistance information. The configuration information is used for indicating a usage manner of the first model. The first assistance information includes statistical information and environment information required for running the first model. The statistical information is used for representing a distribution feature of an input of the first model.
Optionally, before the sending, by the second device, second information to a first device, the method further includes:
Referring to
Step 801: A second device sends first information to a first device, the first information including first data, and the first data being used by the first device to obtain second data based on a first model.
Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Optionally, after the sending, by a second device, first information to a first device, the method further includes:
Optionally, after the sending, by a second device, first information to a first device, the method further includes:
Optionally, the third information further includes identification information, and the identification information is used for indicating that the second data is obtained based on the first model.
Optionally, after the receiving, by the second device, third information from the first device, the method further includes:
Optionally, before the sending, by a second device, first information to a first device, the method further includes:
Optionally, before the sending, by a second device, first information to a first device, the method further includes:
Optionally, the first information further includes second assistance information, and the second assistance information is used for representing a distribution feature of the first data.
The model training processing method provided in the embodiments of this application may be performed by a model training processing apparatus. In the embodiments of this application, an example in which the model training processing apparatus performs the model training processing method is used to describe the model training processing apparatus provided in the embodiments of this application.
Referring to
Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Optionally, the model training processing apparatus 900 further includes:
Optionally, the second information further includes at least one of configuration information and first assistance information. The configuration information is used for indicating a usage manner of the first model. The first assistance information includes statistical information and environment information required for running the first model. The statistical information is used for representing a distribution feature of an input of the first model.
Optionally, the model training processing apparatus 900 further includes:
Optionally, the model training processing apparatus 900 further includes:
Optionally, the model training processing apparatus 900 further includes:
Optionally, the obtaining module 901 includes either of the following:
Optionally, the model training processing apparatus 900 further includes:
Optionally, the receiving unit is further configured to: receive a second request message from the second device, the second request message being used by the second device to request to send the first information.
Optionally, the model training processing apparatus 900 further includes:
Optionally, the third information further includes identification information, and the identification information is used for indicating that the second data is obtained based on the first model.
Optionally, the first information further includes second assistance information, and the second assistance information is used for representing a distribution feature of the first data.
Referring to
Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Optionally, the second information further includes at least one of configuration information and first assistance information. The configuration information is used for indicating a usage manner of the first model. The first assistance information includes statistical information and environment information required for running the first model. The statistical information is used for representing a distribution feature of an input of the first model.
Optionally, the model training processing apparatus 1000 further includes:
Referring to
Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Optionally, the model training processing apparatus 1100 further includes:
Optionally, the model training processing apparatus 1100 further includes:
Optionally, the third information further includes identification information, and the identification information is used for indicating that the second data is obtained based on the first model.
Optionally, the model training processing apparatus 1100 further includes:
Optionally, the model training processing apparatus 1100 further includes:
Optionally, the third sending module 1101 is further configured to send a second request message to the first device, the second request message being used by the second device to request to send the first information.
Optionally, the first information further includes second assistance information, and the second assistance information is used for representing a distribution feature of the first data.
The model training processing apparatus in this embodiment of this application may be an electronic device, for example, an electronic device with an operating system, or may be a component in an electronic device, for example, an integrated circuit or a chip. The electronic device may be a terminal or a device other than the terminal. For example, the terminal may include, but is not limited to, the types of the terminal 11 listed above. The another device may be a server, a network attached storage (Network Attached Storage, NAS), or the like. This is not specifically limited in this embodiment of this application.
The model training processing apparatus provided in this embodiment of this application can implement the processes implemented in the method embodiments of
Optionally, as shown in
An embodiment of this application further provides a terminal, including a processor and a communication interface. When the terminal is a first device, the processor is configured to obtain first information, the first information including first data; and process the first data by using a first model to obtain second data. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the terminal is a first device, the communication interface is configured to send second information to the first device, the second information including a first model, and the first model being used by the first device to obtain second data based on first data. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the terminal is a first device, the communication interface is configured to send first information to the first device, the first information including first data, and the first data being used by the first device to obtain second data based on a first model. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
This embodiment of the terminal corresponds to the foregoing method embodiment of the terminal side. Implementation processes and implementations of the foregoing method embodiment are all applicable to this embodiment of the terminal, and the same technical effects can be achieved. Specifically,
The terminal 1300 includes, but is not limited to: at least some components in a radio frequency unit 1301, a network module 1302, an audio output unit 1303, an input unit 1304, a sensor 1305, a display unit 1306, a user input unit 1307, an interface unit 1308, a memory 1309, and a processor 1310.
A person skilled in the art may understand that, the terminal 1300 may further include a power supply (such as a battery) for supplying power to each component. The power supply may be logically connected to the processor 1310 by using a power management system, thereby implementing functions, such as charging, discharging, and power consumption management, by using the power management system. The terminal structure shown in
It should be understood that in this embodiment of this application, the input unit 1304 may include a graphics processing unit (Graphics Processing Unit, GPU) 13041 and a microphone 13042. The graphics processing unit 13041 processes image data of still pictures or videos captured by an image capture apparatus (such as a camera) in a video capture mode or an image capture mode. The display unit 1306 may include a display panel 13061. The display panel 13061 may be configured in the form of a liquid crystal display, an organic light-emitting diode, or the like. The user input unit 1307 includes at least one of a touch panel 13071 and another input device 13072. The touch panel 13071 is also referred to as a touchscreen. The touch panel 13071 may include two parts: a touch detection apparatus and a touch controller. The another input device 13072 may include, but is not limited to, a physical keyboard, a functional button (such as a sound volume control button or a power button), a trackball, a mouse, or a joystick. Details are not described herein.
In this embodiment of this application, after receiving downlink data from a network side device, the radio frequency unit 1301 may transmit the downlink data to the processor 1310 for processing. In addition, the radio frequency unit 1301 may send uplink data to the network side device. Usually, the radio frequency unit 1301 includes, but is not limited to, an antenna, an amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
The memory 1309 may be configured to store a software program or instruction and various data. The memory 1309 may mainly include a first storage area storing a program or an instruction and a second storage area storing data. The first storage area may store an operating system, an application program or an instruction required by at least one function (for example, a sound playing function or an image playing function), and the like. In addition, the memory 1309 may include a volatile memory or a non-volatile memory, or the memory 1309 may include both a volatile memory and a non-volatile memory. The non-volatile memory may be a read-only memory (Read-Only Memory, ROM), a programmable read-only memory (Programmable ROM, PROM), an erasable programmable read-only memory (Erasable PROM, EPROM), an electrically erasable programmable read-only memory (Electrically EPROM, EEPROM), or a flash memory. The volatile memory may be a random access memory (Random Access Memory, RAM), a static random access memory (Static RAM, SRAM), a dynamic random access memory (Dynamic RAM, DRAM), a synchronous dynamic random access memory (Synchronous DRAM, SDRAM), a double data rate synchronous dynamic random access memory (Double Data Rate SDRAM, DDRSDRAM), an enhanced synchronous dynamic random access memory (Enhanced SDRAM, ESDRAM), a synchronous link dynamic random access memory (Synch link DRAM, SLDRAM), and a direct rambus random access memory (Direct Rambus RAM, DRRAM). The memory 1309 in this embodiment of this application includes, but is not limited to, these memories and any other suitable types of memories.
The processor 1310 may include one or more processing units. Optionally, the processor 1310 integrates an application processor and a modem processor. The application processor mainly processes operations related to the operating system, a user interface, the application program, and the like. The modem processor mainly processes a wireless communication signal, and is, for example, a baseband processor. It may be understood that, the modem processor may alternatively not be integrated in the processor 1310.
When the terminal is a first device, the processor 1310 is configured to obtain first information, the first information including first data; and process the first data by using a first model to obtain second data. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the terminal is a first device, the radio frequency unit 1301 is configured to send second information to the first device, the second information including a first model, and the first model being used by the first device to obtain second data based on first data. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the terminal is a first device, the radio frequency unit 1301 is configured to send first information to the first device, the first information including first data, and the first data being used by the first device to obtain second data based on a first model. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
An embodiment of this application further provides a network side device, including a processor and a communication interface. When the network side device is a first device, the processor is configured to obtain first information, the first information including first data; and process the first data by using a first model to obtain second data. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the network side device is a first device, the communication interface is configured to send second information to the first device, the second information including a first model, and the first model being used by the first device to obtain second data based on first data. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
Alternatively, when the network side device is a first device, the communication interface is configured to send first information to the first device, the first information including first data, and the first data being used by the first device to obtain second data based on a first model. Both the first data and the second data are usable for training a second model, the second model is a service model, and the second data meets at least one of the following: in a case that the first data is unlabeled data, the second data is labeled data; and a data volume of the second data is greater than a data volume of the first data in a case that the first data is labeled data.
This network side device embodiment corresponds to the foregoing network side device method embodiment. Implementation processes and implementations of the foregoing method embodiment may all be applied to this network side device embodiment, and the same technical effects can be achieved.
Specifically, an embodiment of this application further provides a network side device. As shown in
The method performed by the network side device in the foregoing embodiment may be implemented in the baseband apparatus 1403. The baseband apparatus 1403 includes a baseband processor.
The baseband apparatus 1403 may include, for example, at least one baseband board. A plurality of chips are disposed on the baseband board. As shown in
The network side device may further include a network interface 1406. The interface is, for example, a common public radio interface (common public radio interface, CPRI).
Specifically, the network side device 1400 of this embodiment of the present invention further includes: an instruction or a program stored in the memory 1405 and runnable on the processor 1404. The processor 1404 calls the instruction or the program in the memory 1405 to perform the method performed by each module shown in
An embodiment of this application further provides a readable storage medium, storing a program or an instruction. The program or the instruction, when executed by a processor, implements the processes of the foregoing model training processing method embodiment, and the same technical effects can be achieved. To avoid repetition, details are not described herein again.
The processor is the processor in the terminal described in the foregoing embodiment. The readable storage medium includes a computer-readable storage medium, such as a computer read-only memory ROM, a random access memory RAM, a magnetic disk, or an optical disc.
An embodiment of this application further provides a chip, including a processor and a communication interface. The communication interface is coupled to the processor. The processor is configured to run a program or an instruction to implement the processes of the foregoing model training processing method embodiment, and the same technical effects can be achieved. To avoid repetition, details are not described herein again.
It should be understood that, the chip mentioned in this embodiment of this application may also be referred to as a system on a chip, a system chip, a chip system, a system-on-chip, or the like.
An embodiment of this application further provides a computer program/program product, stored in a storage medium and executed by at least one processor to implement the processes of the model training processing method embodiment, and the same technical effects can be achieved. To avoid repetition, details are not described herein again.
It should be noted that, in this specification, the term “include”, “comprise”, or any other variant thereof is intended to cover a non-exclusive inclusion, so that a process, method, article, or apparatus including a series of elements includes not only those elements, but also other elements not expressly listed, or elements inherent to such a process, method, article, or apparatus. Without more restrictions, the elements defined by the sentence “including a . . . ” do not exclude the existence of other identical elements in the process, method, article, or apparatus including the elements. In addition, it should be noted that, the scope of the methods and apparatuses in the implementations of this application is not limited to performing the functions in the order shown or discussed, but may further include performing the functions in a substantially simultaneous manner or in a reverse order depending on the functions involved. For example, the described methods may be performed in an order different from that described, and various steps may be added, omitted, or combined. In addition, features described with reference to some examples may be combined in other examples.
According to the descriptions in the foregoing implementations, a person skilled in the art may clearly learn that the method according to the foregoing embodiment may be implemented by means of software plus a necessary universal hardware platform, or certainly, by using hardware. However, in many cases, the former is a preferred implementation. Based on such an understanding, the technical solutions of this application essentially, or a part contributing to the related art, may be implemented in a form of a computer software product. The computer software product is stored in a storage medium (such as a ROM/RAM, a magnetic disk, or an optical disc), and includes several instructions for instructing a terminal (which may be a mobile phone, a computer, a server, an air conditioner, a network device, or the like) to perform the methods according to the embodiments of this application.
The embodiments of this application have been described above with reference to the accompanying drawings, but this application is not limited to the foregoing specific implementations. The foregoing specific implementations are only illustrative instead of restrictive. Under the inspiration of this application, without departing from the purpose of this application and the scope of protection of the claims, a person of ordinary skill in the art can still make many forms, which all fall within the protection of this application.
| Number | Date | Country | Kind |
|---|---|---|---|
| 202210489247.9 | May 2022 | CN | national |
This application is a continuation of International Patent Application No. PCT/CN2023/092028, filed on May 4, 2023, which claims priority to Chinese Patent Application No. 202210489247.9 filed in China on May 6, 2022, both of which are hereby incorporated by reference in their entireties.
| Number | Date | Country | |
|---|---|---|---|
| Parent | PCT/CN2023/092028 | May 2023 | WO |
| Child | 18935694 | US |