This application relates to the field of communication technologies, and in particular, to a model configuration method and an apparatus.
In a wireless communication network, for example, in a mobile communication network, services supported by the network are increasingly diversified, and therefore, requirements that need to be met are increasingly diversified. For example, the network needs to be capable of supporting an ultra-high rate, an ultra-low latency, and/or an ultra-large connection.
This feature makes network planning, network configuration, and/or resource scheduling increasingly complex. These new requirements, scenarios, and features bring unprecedented challenges to network planning, operations and maintenance, and efficient operation. To meet this challenge, an artificial intelligence technology may be introduced into the wireless communication network, to implement network intelligence. On this basis, how to configure an artificial intelligence model for an access network device in the network is a problem worth studying.
This application provides a model configuration method and an apparatus, to configure, for a first access network device, a first model that matches an application scenario or a performance level, and improve performance and a speed of model prediction.
According to a first aspect, a model configuration method is provided. The method is performed by operations, administration and maintenance (OAM), or may be a component (a processor, a chip, or the like) configured in the OAM, or may be a software module, or the like. The method includes:
According to the foregoing method, requirements on performance indicators of models that implement a same function are different in different application scenarios. In this application, the OAM stores the first mapping relationship, where the first mapping relationship is the mapping relationship between the model, the model application scenario, and the model function. Based on the mapping relationship, for a same model function, in different application scenarios, a model corresponding to a scenario may be allocated, so that the model meets a requirement on a performance indicator in the scenario corresponding to the model. In comparison with a case in which a same model is for model inference in any scenario, the solutions of this application can improve prediction performance, a prediction speed, and the like. Alternatively, in this application, the OAM establishes the first mapping relationship, where the first mapping relationship is the mapping relationship between the model, the model performance level (or the model performance indicator), and the model function. Because requirements on model performance indicators of models are different in different application scenarios, and there is a correspondence between a model performance indicator and a performance level, the OAM may configure different models for different model performance indicators in different application scenarios based on the first mapping relationship. In comparison with a case in which a same model is for model inference under a model performance indicator in any scenario, the solutions of this application can improve prediction performance, a prediction speed, and the like.
In a possible design, the first model request message indicates a model application scenario and a model function of the first model.
In a possible design, the first model request message indicates a model function and a model performance level of the first model, or indicates a model function and a model performance indicator of the first model.
In a possible design, the information about the first model indicates at least one of the following of the first model:
According to the foregoing method, the OAM sends the information about the first model to the first access network device. The first access network device may determine or restore the first model based on the information about the first model. Optionally, the information about the first model may include the training parameter, and the first access network device may subsequently continue to train the first model based on the training parameter, to meet various requirements on the model.
In a possible design, the method further includes: receiving first feedback information from the first access network device; and updating the first model based on the first feedback information.
In a possible design, the first feedback information indicates actual accuracy of prediction information output by the first model, and the updating the first model based on the feedback information includes: when the actual accuracy of the prediction information output by the first model is less than a threshold, retraining the first model, and updating the first model.
According to the foregoing method, when determining the first model, the first access network device may perform the model inference based on the first model, and an inference result may be referred to as the prediction information. The first feedback information is determined based on the prediction information, where the first feedback information includes the actual accuracy of the prediction information output by the first model, and the first feedback information is fed back to the OAM. The OAM updates the first model based on the first feedback information, to improve accuracy of the inference result of the first model.
In a possible design, the first feedback information indicates a scenario change indication, a model performance level change indication, or a model performance indicator change indication. That the first model is updated based on the first feedback information includes: selecting a second model for a terminal device based on the scenario change indication, the model performance level change indication, or the model performance indicator change indication; and sending information about the second model to the first access network device, where the second model is for updating the first model.
Similar to the foregoing method, a difference is that in this design, when the application scenario, the model performance level, or the model performance indicator changes, the first access network device sends, to the OAM, indication information indicating an application scenario change, a model performance level change, or a model performance indicator change, and the OAM reallocates a model to the first access network device based on a changed application scenario, model performance level, or model performance indicator for the model inference, it can be ensured as much as possible that the model allocated to the first access network device meets a requirement on the application scenario, the model performance level, the model performance indicator.
In a possible design, the method further includes: receiving a second model request message from a second access network device, where the second model request message indicates an index of the first model; determining the first model based on the index of the first model; and sending the information about the first model to the second access network device.
According to a second aspect, a model configuration method is provided. For beneficial effects of the method on a first access network device side corresponding to the first aspect, refer to the first aspect. The method is performed by a first access network device, or may be a component (a processor, a chip, or the like) configured in the first access network device, or may be a software module, or the like. The method includes:
In a possible design, before the sending a first model request message to OAM, the method further includes: determining whether an artificial intelligence AI-based enhanced service needs to be provided.
In a possible design, the information about the first model indicates at least one of the following of the first model:
In a possible design, the method further includes: performing model inference based on the first model to obtain prediction information.
In a possible design, the method further includes: sending the prediction information to a terminal device, where the prediction information includes a prediction result.
In a possible design, the prediction information further includes accuracy of predicting the prediction result.
In a possible design, the method further includes: receiving second feedback information from the terminal device, where the second feedback information indicates actual accuracy of the prediction information.
In a possible design, the method further includes: sending first feedback information to the OAM, where the first feedback information indicates the actual accuracy of the prediction information, an application scenario change indication, a model performance level change indication, or a model performance indicator change indication.
It should be noted that in this application, the first model may provide an artificial intelligence AI or machine learning ML inference service for the terminal device, or provide an AI or ML inference service for the first access network device. When the first model is configured to provide the AI or ML inference service for the terminal device, the method further includes: A first RAN node sends inference information (or referred to as prediction information) of the first model to the terminal device. The terminal device sends second feedback information to the first access network device based on the inference information, where the second feedback information may indicate actual accuracy of the inference information, and the first access network device may directly forward the second feedback information to the OAM. Alternatively, the first access network device obtains the actual accuracy of the inference information based on the second feedback information, and sends, to the OAM, the first feedback information indicating the actual accuracy of the inference information, or the like. The OAM may update the first model based on the actual accuracy of the inference information, to improve inference performance, an inference speed, and the like of the first model.
In a possible design, the method further includes: determining a target cell of the terminal device; and sending an index of the first model to a second access network device corresponding to the target cell.
In a possible design, the method further includes: sending first indication information to the second access network device corresponding to the target cell, where the first indication information indicates type information of input data of the first model.
In a possible design, the method further includes: sending, to the second access network device, historical data information needed for the model inference.
In a possible design, the method further includes: receiving second indication information from the second access network device, where the second indication information indicates at least one of the following: indicates that the terminal is handed over to the target cell, indicates a first access network device to delete the first model, or indicates a first access network device to release a corresponding computing resource; and deleting the first model to release the corresponding computing resource.
According to a third aspect, a model configuration method is provided. For beneficial effects of the method on a first access network device side corresponding to the first aspect, refer to the first aspect. The method is performed by a second access network device, or may be a component (a processor, a chip, or the like) configured in the second access network device, or may be a software module, or the like. The method includes:
receiving an index of a first model from a first access network device; sending a second model request message to operations, maintenance, and administration OAM, where the second model request message indicates the index of the first model; and receiving information about the first model from the OAM.
In a possible design, the information about the first model indicates at least one of the following of the first model: a model index, model structure information, a model parameter, a model input format, a model output format, a model performance indicator, a model application scenario, a model function, or a training parameter.
In a possible design, the method further includes: receiving first indication information from the first access network device, where the first indication information indicates type information of input data of the first model.
In a possible design, the method further includes: receiving historical data information needed for model inference from the first access network device, where in the model inference, the historical data information is for determining the input data of the first model.
In a possible design, the method further includes: sending second indication information to the first access network device, where the second indication information indicates at least one of the following: indicates that a terminal is handed over to a target cell, indicates the first access network device to delete the first model, or indicates the first access network device to release a corresponding computing resource.
According to a fourth aspect, an apparatus is provided. For beneficial effects, refer to the descriptions of the first aspect. The apparatus may be operations, administration and maintenance OAM, or an apparatus configured in the OAM, or an apparatus that can be used in matching with the OAM.
In a design, the apparatus may include units that are in one-to-one correspondence with the methods/operations/steps/actions described in the first aspect. The units may be implemented by using a hardware circuit, software, or a combination of a hardware circuit and software.
For example, the apparatus may include a processing unit and a communication unit, and the processing unit and the communication unit may perform corresponding functions in any design example of the first aspect.
The communication unit is configured to receive a first model request message from a first access network device. The processing unit is configured to determine a first model based on the first model request message and a first mapping relationship, where the first mapping relationship includes a mapping relationship between a model, a model application scenario, and a model function, or the first mapping relationship includes a mapping relationship between a model, a model performance level, and a model function. The communication unit is further configured to send information about the first model to the first access network device. For specific execution processes of the processing unit and the communication unit, refer to the first aspect. Details are not described herein again.
For example, the apparatus includes a memory, configured to implement the method described in the first aspect. The apparatus may further include a memory, configured to store instructions and/or data. The memory is coupled to the processor. When executing the program instructions stored in the memory, the processor can implement the method described in the first aspect. The apparatus may further include a communication interface, and the communication interface is used by the apparatus to communicate with another device. For example, the communication interface may be a transceiver, a circuit, a bus, a module, a pin, or another type of communication interface, and the another device may be a network device and the like. In a possible design, the apparatus includes:
For specific execution processes of the communication interface and the processor, refer to the descriptions of the first aspect. Details are not described again.
According to a fifth aspect, an apparatus is provided. For beneficial effects, refer to descriptions of the second aspect. The apparatus may be a first access network device, an apparatus configured in the first access network device, or an apparatus that can be used in matching with the first access network device. In a design, the apparatus may include units that are in one-to-one correspondence with the methods/operations/steps/actions described in the second aspect. The units may be implemented by using a hardware circuit, software, or a combination of a hardware circuit and software.
For example, the apparatus may include a processing unit and a communication unit, and the processing unit and the communication unit may perform corresponding functions in any design example of the second aspect.
The communication unit is configured to send a first model request message to operations, administration and maintenance OAM, where the first model request message indicates a model application scenario and a model function of a first model, or the first model request message indicates a model function and a model performance level of a first model; and receive information about the first model from the OAM. The processing unit is configured to determine the first model based on the information about the first model. For specific execution processes of the processing unit and the communication unit, refer to the second aspect. Details are not described herein again.
For example, the apparatus includes a memory, configured to implement the method described in the second aspect. The apparatus may further include a memory, configured to store instructions and/or data. The memory is coupled to the processor. When executing the program instructions stored in the memory, the processor can implement the method described in the second aspect. The apparatus may further include a communication interface, and the communication interface is used by the apparatus to communicate with another device. For example, the communication interface may be a transceiver, a circuit, a bus, a module, a pin, or another type of communication interface, and the another device may be a terminal and the like. In a possible design, the apparatus includes:
For specific execution processes of the communication interface and the processor, refer to the descriptions of the second aspect. Details are not described again.
According to a sixth aspect, an apparatus is provided. For beneficial effects, refer to descriptions of the third aspect. The apparatus may be a second access network device, an apparatus configured in the second access network device, or an apparatus that can be used in matching with the second access network device. In a design, the apparatus may include units that are in one-to-one correspondence with the methods/operations/steps/actions described in the third aspect. The units may be implemented by using a hardware circuit, software, or a combination of a hardware circuit and software.
For example, the apparatus may include a processing unit and a communication unit, and the processing unit and the communication unit may perform corresponding functions in any design example of the third aspect.
The communication unit is configured to receive an index of a first model from a first access network device; send a second model request message to operations, maintenance, and administration OAM, where the second model request message indicates the index of the first model; and receive information about the first model from the OAM. The processing unit is configured to determine the first model based on the information about the first model. For specific execution processes of the processing unit and the communication unit, refer to the third aspect. Details are not described herein again.
For example, the apparatus includes a memory, configured to implement the method described in the third aspect. The apparatus may further include a memory, configured to store instructions and/or data. The memory is coupled to the processor. When executing the program instructions stored in the memory, the processor can implement the method described in the third aspect. The apparatus may further include a communication interface, and the communication interface is used by the apparatus to communicate with another device. For example, the communication interface may be a transceiver, a circuit, a bus, a module, a pin, or another type of communication interface, and the another device may be a terminal and the like. In a possible design, the apparatus includes:
For specific execution processes of the communication interface and the processor, refer to the descriptions of the third aspect. Details are not described again.
According to a seventh aspect, an embodiment of this application further provides a computer-readable storage medium, including instructions. When the instructions are run on a computer, the computer is enabled to perform the method in any one of the first aspect, the second aspect, or the third aspect.
According to an eighth aspect, this application further provides a chip system. The chip system includes a processor, and may further include a memory, configured to implement the method in any one of the first aspect, the second aspect, or the third aspect. The chip system may include a chip, or may include a chip and another discrete component.
According to a ninth aspect, this application further provides a computer program product, including instructions. When the instructions are run on a computer, the computer is enabled to perform the method in any one of the first aspect, the second aspect, or the third aspect.
According to a tenth aspect, this application further provides a system. The system includes the apparatus in the fourth aspect and the apparatus in the fifth aspect; or the apparatus in the fourth aspect, the apparatus in the fifth aspect, and the apparatus in the sixth aspect.
The access network device may be a base station (base station), an evolved NodeB (evolved NodeB, eNodeB), a transmission reception point (TRP), a next generation NodeB (gNB) in a 5th generation (5G) mobile communication system, an access network device in an open radio access network (O-RAN), a next generation base station in a 6th generation (6G) mobile communication system, a base station in a future mobile communication system, an access node in a wireless fidelity (Wi-Fi) system, or the like; or may be a module or a unit, for example, may be a central unit (CU), a distributed unit (DU), a central unit control plane (CU-CP) module, or a central unit user plane (CU-UP) module, that completes a part of functions of a base station. The access network device may be a macro base station (for example, 110a in
In this application, an apparatus configured to implement a function of the access network device may be an access network device, or may be an apparatus that can support the access network device in implementing the function, for example, a chip system, a hardware circuit, a software module, or a hardware circuit and a software module. The apparatus may be installed in the access network device or may be used in matching with the access network device. In this application, the chip system may include a chip, or may include a chip and another discrete component. For ease of description, the following describes the technical solutions provided in this application by using an example in which the apparatus configured to implement the functions of the access network device is an access network device and the access network device is a RAN node.
Communication between an access network device and a terminal device complies with a specified protocol layer structure. The protocol layer structure may include a control plane protocol layer structure and a user plane protocol layer structure. For example, the control plane protocol layer structure may include functions of protocol layers such as a radio resource control (RRC) layer, a packet data convergence protocol (packet data convergence protocol, PDCP) layer, a radio link control (RLC) layer, a media access control (MAC) layer, and a physical layer. For example, the user plane protocol layer structure may include functions of protocol layers such as a PDCP layer, an RLC layer, a MAC layer, and a physical layer. In a possible implementation, a service data adaptation protocol (SDAP) layer may be further included above the PDCP layer.
Optionally, the protocol layer structure between the access network device and the terminal device may further include an artificial intelligence (AI) layer for transmitting data related to an AI function.
An access device may include a CU and a DU. A plurality of DUs may be controlled by one CU in a centralized manner. For example, an interface between the CU and the DU may be referred to as an F1 interface. A control plane (CP) interface may be F1-C, and a user plane (UP) interface may be F1-U. A specific name of each interface is not limited in this application. The CU and the DU may be divided based on protocol layers of a radio network. For example, functions of a PDCP layer and a protocol layer above the PDCP layer are set on the CU, and functions of protocol layers (for example, an RLC layer and a MAC layer) below the PDCP layer are set on the DU. For another example, functions of a protocol layer above a PDCP layer are set on the CU, and functions of the PDCP layer and protocol layers below the PDCP layer are set on the DU. This is not limited.
The foregoing division into processing functions of the CU and the DU based on the protocol layers is merely an example, and may alternatively be other division. For example, the CU or the DU may be defined to have functions of more protocol layers, and for another example, the CU or the DU may alternatively be defined to have a part of processing functions of the protocol layers. In a design, a part of functions of the RLC layer and functions of protocol layers above the RLC layer are set on the CU, and remaining functions of the RLC layer and functions of protocol layers below the RLC layer are set on the DU. In another design, division of functions of the CU or the DU may alternatively be performed based on service types or other system requirements. For example, division may be performed based on latencies. Functions whose processing time needs to satisfy a latency requirement are set on the DU, and functions whose processing time does not need to satisfy the latency requirement are set on the CU. In another design, the CU may alternatively have one or more functions of the core network. For example, the CU may be disposed on a network side to facilitate centralized management. In another design, a radio unit (RU) of the DU is disposed remotely. Optionally, the RU may have a radio frequency function.
Optionally, the DU and the RU may be distinguished at a physical layer (PHY). For example, the DU may implement higher-layer functions of the PHY layer, and the RU may implement lower-layer functions of the PHY layer. When the PHY layer is for sending, functions of the PHY layer may include at least one of the following functions: addition of cyclic redundancy check (CRC) code, channel encoding, rate matching, scrambling, modulation, layer mapping, precoding, resource mapping, physical antenna mapping, or radio frequency sending. When the PHY layer is for receiving, functions of the PHY layer may include at least one of the following functions: CRC check, channel decoding, de-rate matching, descrambling, demodulation, layer demapping, channel detection, resource demapping, physical antenna demapping, or radio frequency receiving. The higher-layer function of the PHY layer may include some functions of the PHY layer. For example, the some functions are closer to the MAC layer. The lower-layer function of the PHY layer may include some other functions of the PHY layer. For example, the part of functions are closer to the radio frequency function. For example, the higher-layer function of the PHY layer may include CRC code addition, channel coding, rate matching, scrambling, modulation, and layer mapping, and the lower-layer function of the PHY layer may include precoding, resource mapping, physical antenna mapping, and radio frequency sending functions. Alternatively, the higher-layer function of the PHY layer may include CRC code addition, channel coding, rate matching, scrambling, modulation, layer mapping, and precoding. The lower-layer function of the PHY layer may include resource mapping, physical antenna mapping, and radio frequency sending functions. For example, the higher-layer functions of the PHY layer may include the CRC check, the channel decoding, the de-rate matching, the decoding, the demodulation, and the layer demapping, and the lower-layer functions of the PHY layer may include functions of the channel detection, the resource demapping, the physical antenna demapping, and the radio frequency receiving. Alternatively, the higher-layer functions of the PHY layer may include the CRC check, the channel decoding, the de-rate matching, the decoding, the demodulation, the layer demapping, and the channel detection, and the lower-layer functions of the PHY layer may include functions of the resource demapping, the physical antenna demapping, and the radio frequency receiving.
For example, the functions of the CU may be implemented by one entity, or may be implemented by different entities. For example, the functions of the CU may be further divided. To be specific, a control plane and a user plane that are of the CU are separated and implemented by different entities, which are a control plane CU entity (that is, a CU-CP entity) and a user plane CU entity (that is, a CU-UP entity). The CU-CP entity and the CU-UP entity may be coupled to the DU, to jointly complete functions of an access network device.
Optionally, any one of the DU, the CU, the CU-CP, the CU-UP, and the RU may be a software module, a hardware structure, or a combination of a software module and a hardware structure. This is not limited. Different entities may exist in different forms, which is not limited. For example, the DU, the CU, the CU-CP, and the CU-UP are software modules, and the RU is a hardware structure. These modules and methods performed by these modules also fall within the protection scope of this application.
In a possible implementation, the access network device includes the CU-CP, the CU-UP, the DU, and the RU. For example, this application is executed by the DU, or the DU and the RU, or the CU-CP, the DU, and the RU, or the CU-UP, the DU, and the RU. This is not limited. Methods performed by the modules also fall within the protection scope of this application.
The terminal device may also be referred to as a terminal, user equipment (UE), a mobile station, a mobile terminal device, or the like. The terminal device is widely used in communication in various scenarios, for example, including but not limited to at least one of the following scenarios: device-to-device (D2D), vehicle-to-everything (V2X), machine-type communication (MTC), an internet of things (IoT), virtual reality, augmented reality, industrial control, self-driving, telemedicine, a smart grid, smart furniture, smart office, a smart wearable, smart transportation, a smart city, or the like. The terminal device may be a mobile phone, a tablet computer, a computer with a wireless transceiver function, a wearable device, a vehicle, an uncrewed aerial vehicle, a helicopter, an airplane, a ship, a robot, a robot arm, a smart home device, or the like. A specific technology and a specific device form that are used by the terminal device are not limited in this application.
In this application, an apparatus configured to implement a function of a terminal device may be a terminal device, or may be an apparatus that can support the terminal device in implementing the function, for example, a chip system, a hardware circuit, a software module, or a hardware circuit and a software module. The apparatus may be installed in the terminal device or may be used in matching with the terminal device. For ease of description, the following describes the technical solutions provided in this application by using an example in which the apparatus configured to implement the functions of the terminal device is a terminal device and the terminal device is UE.
The base station and the terminal device may be fixed or movable. The base station and/or the terminal device may be deployed on the land, including an indoor or outdoor scenario, and a handheld or vehicle-mounted scenario; or may be deployed on the water; or may be deployed on an airplane, a balloon, and an artificial satellite in the air. Application scenarios of the base station and the terminal device are not limited in this application. The base station and the terminal device may be deployed in a same scenario or different scenarios. For example, the base station and the terminal device are both deployed on the land. Alternatively, the base station is deployed on the land, and the terminal device is deployed on the water. Examples are not described one by one.
Roles of the base station and the terminal device may be relative. For example, a helicopter or an uncrewed aerial vehicle 120i in
In this application, an independent network element (for example, referred to as an AI network element or an AI node) may be introduced into the communication system shown in
In this application, as shown in
Optionally,
Alternatively, in a second design, as shown in
Alternatively, in a third design, as shown in
The AI model is a specific implementation of the AI function. The AI model represents a mapping relationship between input and output of the model. The AI model may be a neural network, a linear regression model, a decision tree model, a support vector machine (SVM), a Bayesian network, a Q-learning model, another machine learning model, or the like. In this application, the AI function may include at least one of the following: data collection (collecting training data and/or inference data), data preprocessing, model training (or referred to as model learning), model information release (model information configuration), model validation, model inference, or inference result release. The inference may also be referred to as prediction. In this application, the AI model may be referred to as a model for short.
Obtaining the AI model through learning by the model training node is equivalent to obtaining the mapping relationship between the input and the output of the model through learning by the model training node by using training data. The model inference node uses the AI model to perform inference based on the inference data provided by the data source and obtain an inference result. The method may also be described as follows: The model inference node inputs the inference data to the AI model, and obtains an output by using the AI model. The output is the inference result. The inference result may indicate a configuration parameter used (acted) by an object of action, and/or an operation acted by an object of action. The inference result may be centrally planned by an actor (actor) entity, and sent to one or more objects of action (for example, network entities) for action. Optionally, the actor entity or the actor object may feed back a parameter or a measurement amount collected by the actor entity or the actor object to the data source. This process may be referred to as performance feedback, and the fed-back parameter may be used as the training data or the inference data. Optionally, the actor entity or the actor object may further determine feedback information related to model performance based on the inference result output by the model inference node, and feed back the feedback information to the model inference node. The model inference node may feed back performance information of the model to the model training node based on the feedback information, so that the model training node optimizes or updates the deployed AI model. This process may be referred to as model feedback.
The AI model may be a neural network or another machine learning model. The neural network is used as an example. The neural network is a specific implementation form of a machine learning technology. According to a universal approximation theorem, the neural network may theoretically approximate to any continuous function, so that the neural network has a capability of learning any mapping. Therefore, the neural network can accurately perform abstract modeling on a complex high-dimensional problem.
An idea of the neural network is from a neuron structure of brain tissue. Each neuron performs a weighted summation operation on input values of the neuron, and outputs a result of the weighted summation through an activation function.
The neural network usually includes a multi-layer structure, and each layer may include one or more neurons. Increasing a depth and/or a width of the neural network can improve an expression capability of the neural network, and provide more powerful information extraction and abstract modeling capabilities for complex systems. The depth of the neural network may refer to a quantity of layers included in the neural network, and a quantity of neurons included in each layer may be referred to as a width of the layer.
In one solution, a RAN node supports a plurality of AI-assisted cases. For example, mobility enhancement, load balancing, and network energy saving. OAM is configured to train or store an AI model. The RAN node requests the AI model from the OAM and performs inference by using the AI model. How the OAM configures a corresponding model for the RAN node is a problem to be resolved in this application. In the descriptions of this application, the OAM may be replaced with another node that can train or store the AI model, for example, a cloud server, a core network device, or another possible device.
As shown in
Step 701: A first RAN node sends a first model request message to OAM, and correspondingly, the OAM receives the first model request message from the first RAN node.
For ease of distinguishing, a model request message sent by the first RAN node to the OAM is referred to as the first model request message. A model request message sent by a second RAN node to the OAM is referred to as a second model request message. The model request message is used by the RAN node to request a model from the OAM, and a name of the model request message is not limited. For example, the first model request message may also be referred to as a first message, and the second model request message may also be referred to as a second message. To distinguish between different RAN nodes, the RAN nodes are referred to as the first RAN node, the second RAN node, and the like. The first RAN node and the second RAN node are different RAN nodes. A relationship between the first RAN node and the second RAN node is not limited.
In a design, the first RAN node may determine a service scenario and a function requirement, and determine the first model request message based on the service scenario and the function requirement. The service scenario and the function requirement are a service scenario and a function requirement of a RAN node. For example, a first model requested by using the first model request message is for model inference related to the RAN node, for example, for cell energy saving. The first RAN node may determine the first model request message and the like based on the service scenario and the function requirement of the RAN node. Alternatively, the service scenario and the function requirement are a service scenario and a function requirement of UE. For example, a first model requested by using the first model request message is for model inference related to the UE, for example, for prediction of a moving trajectory of the UE. The first RAN node may determine the first model request message and the like based on the service scenario and the function requirement of the UE. Alternatively, the first RAN node may determine the first model request message and the like based on a service scenario and a function requirement of a node other than a RAN node and UE. This is not limited. The first RAN node may obtain the service scenario and the function requirement of the UE based on information reported by the UE. For example, the UE is in a self-driving scenario, and sends a quality of service (QOS) prediction service request to the RAN node. The RAN node may obtain the service scenario and the function requirement of the UE as self-driving and QoS prediction. Alternatively, the first RAN node may obtain the service scenario and the function requirement of the UE through analysis. For example, the RAN node finds that the UE moves to a cell edge and needs a mobility enhancement service, and the mobility enhancement service needs prediction of trajectory information of the UE, so that the service scenario and the function requirement of the UE are mobility enhancement and prediction of the trajectory information of the UE. Alternatively, the first RAN node may obtain the service scenario and the function requirement of the UE from another device, where the another device includes but is not limited to a minimization of drive tests (MDT) device and the like. Alternatively, the first RAN node stores the service scenario and the function requirement of the UE. For example, when accessing the first RAN node, the UE actively reports the service scenario and the function requirement of the UE to the first RAN node.
In this application, the service scenario includes but is not limited to at least one of the following: mobility enhancement, load balancing, virtual reality (VR), augmented reality (AR), vehicle-to-everything (self-driving), wireless robot cloud control, remote diagnosis with force feedback, a high-definition video (for example, an 8K video), or the like. For example, the mobility enhancement refers to providing an enhanced service for serving cell handover for the moving UE. For example, the RAN node predicts a target cell, handover time, and the like of the UE based on at least one of the moving trajectory, a measurement report, a cell load status, and the like of the UE. The target cell of the UE may refer to a serving cell to which the UE is to be handed over. A moving manner of the UE includes moving on a fixed route, moving on a non-fixed route, and the like. The mobility enhancement may include mobility enhancement for vehicle movement on a fixed route, mobility enhancement for vehicle movement on a non-fixed route, mobility enhancement for a pedestrian scenario, or the like. In the remote diagnosis, which is also referred to as remote consultation, a doctor performs, through a network, diagnosis or surgery on a patient who is at another place. The foregoing force feedback is haptic feedback of a force, and is a service that has a very high requirement on a latency, and a latency prediction service needs to be provided. For example, when a predicted latency exceeds a threshold, an information transmission solution needs to be adjusted or an alarm needs to be provided, to respond to an emergency in advance.
In this application, the function requirement includes but is not limited to at least one of the following: UE trajectory prediction, load prediction, signal to interference plus noise ratio (SINR) prediction, reference signal received power (RSRP) prediction, reference signal received quality (RSRQ) prediction, latency prediction, QoS prediction, throughput prediction, cell energy saving, or the like.
In a design, after the first RAN node determines the first model request message based on the service scenario and the function requirement, the first model request message indicates the service scenario and the function requirement, in other words, the first model request message indicates a model application scenario, a model function, and the like of the requested first model. The model application scenario corresponds to the foregoing service scenario, and the model function corresponds to the foregoing function requirement. An indication in this application may include an implicit indication, an explicit indication, or the like. This is not limited. For example, the first model request message may include the service scenario, the function requirement, and the like. Alternatively, the first RAN node may directly send the service scenario and the function requirement to the OAM or the like. For example, a combination of the service scenario and the function requirement that are sent by the first RAN node to the OAM may include: a mobility enhancement scenario and trajectory prediction, load balancing and trajectory prediction, self-driving and QoS prediction, VR/AR and QoS prediction, wireless robot cloud control and latency prediction, remote diagnosis with force feedback and latency prediction, and the like.
In another design, a process of determining the first model request message by the first RAN node based on the service scenario and the function requirement includes: The first RAN node determines a model function and a model performance indicator of the requested first model based on the service scenario and the function requirement. The model function corresponds to the foregoing function requirement. The model performance indicator includes at least one of the following: accuracy, a calculation amount, memory occupation, inference duration, or the like. The accuracy is a percentage of a quantity of samples that are correctly predicted to a total quantity of samples. The calculation amount is a quantity of calculation times needed by the model. An overall calculation amount of the model is equal to a sum of calculation amounts of operators in the model, and the calculation amount feeds back a requirement of the model on a hardware calculation unit. Alternatively, the calculation amount may be a size of a computing resource occupied when the model is running. The memory occupation refers to sizes of memory and a video RAM that are occupied when the model is running, or the memory occupation may be referred to as a size of the model or the like. The inference duration refers to duration needed by the model to infer a prediction result. The first model request message sent by the first RAN node to the OAM indicates the model function and the model performance indicator of the first model. For example, the first model request message includes the model function and the model performance indicator. Alternatively, the first RAN node may directly send the model function and the model performance indicator to the OAM. The model function may also be referred to as a model category. For example, the first RAN node learns, through analysis, that a QoS prediction service in the self-driving scenario needs to be provided for the UE, the first RAN node may continue analyzing a performance indicator needed for the QoS prediction service in the self-driving scenario, and the first RAN node sends the QoS prediction service and the corresponding performance indicator to the OAM. Alternatively, the first RAN node determines a model performance level based on the model performance indicator. There is a correspondence between the model performance level and the model performance indicator. For details, refer to the following descriptions. The first model request message sent by the first RAN node to the OAM indicates the model function and model performance level of the requested first model.
Optionally, before step 701, the method may further include: The first RAN node determines whether an AI-based enhanced service needs to be provided. If an AI-based enhanced service needs to be provided, step 701 is performed to send the first model request message to the OAM. If no AI-based enhanced service needs to be provided, the procedure shown in
Step 702: The OAM determines the first model based on the first model request message and a first mapping relationship.
In this application, to distinguish between different models, different models are respectively referred to as the first model, a second model, and the like. The model represents a mapping relationship between the input and an output of the model, and is for implementing a specific function. A name of the model is not limited. For example, the model may also be referred to as an AI model. The AI model may be a neural network, a linear regression model, a decision tree model, an SVM, a Bayesian network, a Q-learning model, or another machine learning model. The first mapping relationship represents a mapping, a correspondence, or the like. The first mapping relationship may also be referred to as a correspondence or the like. Similar cases are not described one by one subsequently.
In this application, the OAM may obtain the first mapping relationship. The first mapping relationship may be established by the OAM or another node. The OAM obtains the first mapping relationship via the another node.
In an example, the first mapping relationship includes a mapping relationship between a model, a model performance level, and a model function, and the model performance level may be referred to as a performance level for short.
For example, there is one-to-many correspondence between the model performance level and the model performance indicator. For example, the model performance indicator for model level classification includes the accuracy, and levels may be classified based on the accuracy. Higher accuracy indicates a higher model performance level. Alternatively, the model performance indicator for model level classification includes the inference duration, and levels may be classified based on the inference duration. Shorter inference duration indicates a higher model performance level. Alternatively, the model performance indicator for model level division includes at least two indicators of the accuracy, the calculation amount, the inference duration, and the memory occupation. For example, a plurality of indicators are weighted in a specific weighting manner to determine a weighted value, and levels are classified based on the weighted value. For models with different functions, manners for model performance level classification are the same or different. For example, model performance levels are classified based on model accuracy for a model for trajectory prediction and a model for SINR prediction. It is assumed that both the model for the trajectory prediction and the model for the SINR prediction include three model performance levels. As shown in Table 1, for the model for the trajectory prediction, accuracy corresponding to a model performance level 1 is [n1, n2], accuracy corresponding to a model performance level 2 is (n2, n3], and accuracy corresponding to a model performance level 3 is (n3, n4], where n4 is greater than or equal to n3, n3 is greater than or equal to n2, and n2 is greater than or equal to n1. As shown in Table 2, for the model for the SINR prediction, accuracy corresponding to a model performance level 1 is [m1, m2], and accuracy corresponding to a model performance level 2 is (m2, m3], where m3 is greater than or equal to m2, and m2 is greater than or equal to m1. For another example, as shown in Table 3, for the model for the SINR prediction, inference duration corresponding to a model performance level 1 is [t1, t2], inference duration corresponding to a model performance level 2 is (t2, t3], and inference duration corresponding to a model performance level 3 is (t3, t4], where t4 is greater than or equal to t3, t3 is greater than or equal to t2, and t2 is greater than or equal to t1.
In this design, the first RAN node may obtain the model application scenario, determine a corresponding model performance indicator based on the model application scenario, and determine a model performance level based on the model performance indicator. The first model request message sent to the OAM indicates the model performance level and a model function. The OAM determines a model, referred to as the first model, from the first mapping relationship based on the model performance level and the model function that are indicated by the first model request message. The first mapping relationship includes the mapping relationship between the model, the model performance level, and the model function. Reference may be made to the following descriptions in Table 6.
In an example, the first mapping relationship includes a mapping relationship between a model, a model application scenario, and a model function. For example, the first mapping relationship includes an AI model, a self-driving scenario, and QoS prediction.
For example, the first RAN node may obtain the model application scenario and the model function of the first model. Reference may be made to the descriptions in step 701. The first model request message sent by the first RAN node to the OAM indicates the model application scenario and the model function of the first model. The OAM may determine an AI model from the first mapping relationship based on the application scenario and the model function that are indicated by the first model request message. The AI model is referred to as the first model. For example, as shown in Table 4, the first mapping relationship includes four application scenarios (with indexes 1 to 4) and three model functions (with indexes are 1 to 3). The four application scenarios and the three model functions may form 12 mapping relationships, and indexes of the 12 mapping relationships are 1 to 12. For example, a mapping relationship 1 includes an application scenario 1, a model function 1, and a model 1, and a mapping relationship 2 includes an application scenario 1, a model function 2, and a model 2.
In the first mapping relationship shown in Table 4, for a same model function, corresponding models are different in different scenarios. For example, the model function 1 corresponds to the model 1, the model 4, the model 7, and the model 10 respectively in application scenarios 1 to 4. That the OAM establishes the first mapping relationship shown in Table 4 is used as an example. For a same model function, the OAM may separately collect corresponding training data in different scenarios and train corresponding AI models. For example, the OAM may collect, in the application scenario 1, training data corresponding to the model function 1, and train the model 1 by using the training data, so that the model 1 can meet a performance indicator of the application scenario 1. A training process may be machine learning, non-machine learning, reinforcement learning, or the like. This is not limited.
Alternatively, in this example, there is one-to-one correspondence between the model performance level and the model performance indicator, each model performance level corresponds to a unique model performance indicator, and each model performance indicator corresponds to a unique model performance level. For example, as shown in Table 5, model performance levels 1 to 3 correspond to model performance indicators 1 to 3 respectively. That the model performance indicator includes the accuracy, the calculation amount, the inference duration, and the memory occupation is used as an example. A model performance level 1 corresponds to a model performance indicator 1, where correspondingly, accuracy is at least 99%, a calculation amount is up to 3%, inference duration is up to 1s, and memory occupation is up to 500 KB.
In this example, the first RAN node may obtain the model performance indicator of the requested model, and the first model request message sent to the OAM indicates a model function and the model performance indicator of the requested model. When receiving the first model request message, the OAM may determine a model performance level based on the model performance indicator indicated by the first model request message. Alternatively, the first RAN node may determine the model performance level based on the model performance indicator, and the first model request message sent to the OAM indicates a model function and the model performance level. The OAM determines a model from the first mapping relationship based on the model function and the model performance level, where the model is referred to as the first model or the like. For example, as shown in Table 6, the first mapping relationship includes 12 mapping relationships with indexes 1 to 12. For example, a mapping relationship 1 includes a model 1, a model performance level 1, and a model function 1, and a mapping relationship 2 includes a model 2, a model performance level 1, and a model function 2.
It should be noted that in the foregoing example in which there is the one-to-one correspondence between the model performance level and the model performance indicator, the model performance level in the foregoing first mapping relationship may be replaced with the model performance indicator. Alternatively, in this application, the first mapping relationship may be established based on the model performance indicator. For example, the first mapping relationship includes a mapping relationship between a model, a model performance indicator, and a model function. In this design, the first model request message indicates the model function and the model performance indicator of the first model. The OAM determines a model from the first mapping relationship based on the model function and the model performance indicator that are indicated by the first model request message. The model may be referred to as the first model. For example, as shown in Table 7, the first mapping relationship includes 12 mapping relationships with indexes 1 to 12. For example, a mapping relationship 1 includes a model 1, a model performance indicator 1, and a model function 1, and a mapping relationship 2 includes a model 2, a model performance indicator 1, and a model function 2.
In the first mapping relationship shown in Table 6 or Table 7, models corresponding to a same model function are different under different model performance levels or model performance indicators. For example, reference is made to Table 6 or Table 7. For the model function 1, models corresponding to model performance levels 1 to 4 or model performance indicators 1 to 4 are the model 1, the model 4, the model 7, and the model 10 respectively. That the OAM establishes the first mapping relationship shown in Table 6 or Table 7 is used as an example. The OAM may separately collect corresponding training data under different model performance levels or model performance indicators, and train corresponding models. For example, the OAM may collect training data corresponding to the model function 1 under the model performance level 1 or the model performance indicator 1, and train the model 1.
As described above, the model performance indicator includes at least one of the following: the accuracy, the calculation amount, the memory occupation, the inference duration, or the like. In different scenarios, a same model function may have different requirements on model performance indicators. Optionally, the calculation amount in the model performance indicator is also referred to as computing resource occupation, and the memory occupation is also referred to as a model size. In a possible case, as shown in Table 8, in a VR/AR scenario and a V2X scenario, requirements on performance indicators of a model in QoS prediction are different. In another possible case, as shown in Table 9, in scenarios such as a vehicle on a fixed route, a vehicle on a non-fixed route, and a pedestrian scenario, requirements on performance indicators of a model in trajectory prediction are different.
In different scenarios, requirements on performance indicators of a model of a same function may be different. If a same model is for inference for a function in any scenario, it is difficult to meet a requirement on a performance indicator of a corresponding scenario. For example, in Table 8, for the AR/VR scenario and the V2X scenario, if a same model in QoS prediction is for model inference, it is difficult to meet a requirement on a performance indicator of each scenario. In the solution of this application, reference may be made to the first mapping relationship shown in Table 4 above. For a same model function, different models may be configured in different scenarios. For example, in the VR/AR scenario, a model is configured for the QoS prediction. In the V2X scenario, another model is configured for the QoS prediction. Alternatively, reference may be made to the mapping relationship shown in Table 6 above. Models with corresponding performance levels may be configured for requirements on performance indicators in different scenarios. Alternatively, reference may be made to the mapping relationship shown in Table 7 above, different models may be configured for different performance indicators, so that each model meets a requirement on a performance indicator of a corresponding application scenario.
Step 703: The OAM sends information about the first model to the first RAN node.
In this application, the information about the first model indicates at least one of the following of the first model: a model index (index), model structure information, a model parameter, a model input format, a model output format, an input data processing manner, an output data processing manner, a model performance indicator, a model application scenario, a model function, a training parameter, or the like.
The model index is for distinguishing between different models. The model structure information, the model parameter, the model input format, the model output format, and the like are used by the first RAN node to determine the first model. In other words, the first RAN node determines the first model based on the model structure information, the model parameter, the model input format, the model output format, and the like. The model structure information includes at least one of the following: a quantity of layers included in the model (also referred to as a depth of the model), a network type of each layer (for example, a fully connected layer network, a convolutional layer network, or a long short-term memory (LSTM) layer), a quantity of neurons included in each layer (also referred to as a width of the model), a connection relationship between layers, or the like. The model parameter may include at least one of the following: a weight of a neuron, an activation function of a neuron, an offset in an activation function, or the like. The model input format and output format may be formats that need to be met by input data and output data when model training or inference is performed on the model. The foregoing input data processing manner includes a manner of preprocessing raw data before the collected raw data is input into the model. For example, an imaginary part is separated from a real part of the raw data, normalization processing on the raw data is performed, or a phase and an amplitude of the raw data are separated. The foregoing output data processing manner includes a manner of processing the output data when the model outputs the data, for example, intercepting the output data. For the model performance indicator, the model application scenario, and the model function, refer to the foregoing descriptions. In this application, when determining the first model based on information such as a model structure, the model parameter, the model input format, and the model output format, the first RAN node may further continue to train the first model based on the foregoing training parameter. The training parameter may include indication information of a loss function and the like. For example, the first RAN node may obtain training data. The training data may be from the first RAN node, or from the UE, or from both the first RAN node and the UE, or from another node. The first RAN node trains the first model based on the training data. For example, the training data is input into the first model. A value of the loss function is determined based on output of the first model. If the value of the loss function is less than a threshold or meets a target requirement, training on the first model ends. If the value of the loss function is no less than the threshold or does not meet the target requirement, a parameter of the first model is updated, and training on the first model continues. That a parameter of the first model is updated may include updating at least one of the following information of the first model: a quantity of layers of a neural network, a width of a neural network, the connection relationship between the layers, the weight of the neuron, the activation function of the neuron, the offset in the activation function, or the like. A training process may be label learning, non-label learning, reinforcement learning, or the like. This is not limited. When the first model is trained, an initial model of the first model may be indicated by the information about the first model, or may be randomly generated, or may be agreed in a protocol. This is not limited.
Alternatively, the information about the first model indicates at least one of the following of the first model: a model, a model usage description, or a model index. The model includes a network architecture, a parameter weight, a training hyperparameter, an optimizer (optimizer), a batch size (batch size), a learning rate (learning rate), momentum (momentum), and the like. Optionally, the first RAN node may perform online training or the like based on the training hyperparameter. The model usage description includes at least one of a model input/output format, a model performance indicator, a model application scenario, and a model function. The model index is a number of the model, used to distinguish between different models, and the like.
During actual application, requirements on performance indicators of models that implement a same function are different in different application scenarios. For example, as shown in Table 8 or Table 9 above, for the QoS prediction and the trajectory prediction, requirements on performance indicators of models are different in different scenarios. In this application, the OAM establishes the first mapping relationship, where the first mapping relationship is the mapping relationship between the model, the model application scenario, and the model function. Based on the mapping relationship, for a same model function, in different application scenarios, a model corresponding to a scenario may be configured, so that the model meets a requirement on a performance indicator in the scenario corresponding to the model. In comparison with a case in which a same model is for model inference in any scenario, the solutions of this application can improve prediction performance, a prediction speed, and the like. Alternatively, in this application, the OAM establishes the first mapping relationship, where the first mapping relationship is the mapping relationship between the model, the model performance level, and the model function, or the first mapping relationship is the mapping relationship between the model, the model performance indicator, and the model function. Because requirements on performance indicators of models are different in different application scenarios, and there is a correspondence between a model performance indicator and a model performance level, the OAM may configure different models for different requirements on model performance indicators in different application scenarios based on the first mapping relationship. In comparison with a case in which a same model is for model inference under a model performance indicator in any scenario, the solutions of this application can improve prediction performance, a prediction speed, and the like.
Optionally, in addition to step 701 to step 703, the procedure shown in
Step 704: The first RAN node performs the model inference based on the first model to obtain prediction information.
For example, the first RAN node may collect the input data, allocate a computing resource, a memory resource, and the like for the model inference. The input data may be from the first RAN node and/or the UE, or from another node, or the like. This is not limited. Optionally, the first RAN node may collect the input data based on the model usage description. The input data is input into the first model, and is output as the prediction information. Alternatively, the input data is processed, and processed input data is input into the first model; and/or output data of the first model is processed, and processed output data is the prediction information. It may be understood that the input data and the output data meet requirements on an input format and an output format of the first model. The prediction information includes a prediction result. For example, when a model training node such as the OAM trains a model, the model training node classifies a collected data set into a training set, a validation set, and a test set. The model training node performs model training by using the training set, to obtain the model, and this process may be referred to as the model training. The model is verified by using the validation set, and this process may be referred to as model verification. The model verification is usually performed during the model training. For example, each time the model is trained for one or more epochs (epochs), the validation set may be used to verify a current model, to monitor a model training status, for example, to verify whether the current model is underfitted, overfitted, or has been converged, and determine whether to end the training. Optionally, in a model verification process, a hyperparameter of the model may be further adjusted. The hyperparameter may refer to at least one of the following parameters of the model: a quantity of layers of a neural network, a quantity of neurons, an activation function, a loss function, or the like. The model is tested by using the test set, and this process may be referred to as a model test. For example, after the model training is completed, the test set is for testing a trained model. For example, a generalization capability of the model is evaluated, whether the model meets a requirement is determined, whether the model is available is determined, or prediction accuracy of the model is obtained. In this application, the information that is about the first model and that is sent by the OAM to the first RAN node may further include the prediction accuracy and the like. Alternatively, when determining the first model, the first RAN node may test the first model by using locally collected test data, to obtain the prediction accuracy and the like. In addition to the prediction result inferred by the first model, optionally, the prediction information sent by the first RAN node to the UE may further include the prediction accuracy. For example, in the trajectory prediction, the prediction information sent by the first RAN node to the UE includes predicted trajectory information of the UE and accuracy (for example, 98%) of the predicted trajectory information.
In this application, if the prediction information is prediction information related to the first RAN node, the first RAN node determines first feedback information based on the prediction information. For example, the first RAN performs a corresponding action based on the prediction information. For example, in a network energy saving scenario, the prediction information is an energy saving policy, for example, carrier shutdown, symbol shutdown, cell shutdown, or deep sleep. After the first RAN node infers the energy saving policy by using the first model, the first RAN node may execute the energy saving policy. The first RAN node determines the first feedback information based on a change between performance (for example, a change between cell throughputs or QoS of the UE) before and after the prediction information is executed. Alternatively, the prediction information is prediction information related to the UE, and the first RAN node needs to send the prediction information to the UE. The UE determines second feedback information based on the prediction information. For example, the UE may determine the second feedback information based on a performance change that is after the prediction information is executed. For example, in a mobility enhancement scenario, the prediction information is a target cell to which the UE is to be handed over, handover time, and the like. The UE may perform a corresponding action based on the prediction information, and determine the second feedback information based on the performance change that is after the prediction information is executed. Alternatively, the UE compares the prediction information with actual information, to determine second feedback information. For example, in a trajectory prediction scenario, the UE may compare a predicted moving trajectory of the UE with an actual moving trajectory of the UE, to determine the second feedback information. The UE sends the second feedback information to the first RAN node. The first RAN node sends the first feedback information to the OAM based on the received second feedback information, the OAM updates the first model based on the first feedback information, and the like. For details, refer to descriptions of a first design in the following step 705 and step 706. Alternatively, when determining that an application scenario changes or a performance level (or a performance indicator) of a required model changes, the first RAN node may alternatively send, to the OAM, the first feedback information indicating that the scenario changes or the performance level changes. The OAM may reconfigure a model for the first RAN node based on a scenario change indication or a performance level change indication. The model may be referred to as a second model or the like. For example, when the UE changes from a walking scenario to a scenario in which a vehicle drives on a fixed route, the first RAN node predicts, by using a model in trajectory prediction requested in the walking scenario, a moving trajectory in the scenario in which the vehicle drives on the fixed route, and consequently, a decrease in prediction accuracy may be caused. In this application, when finding that a scenario of the UE changes, the first RAN node may indicate a scenario change to the OAM or the like, and the OAM may configure, for the first RAN node, a model that matches the current scenario, or the like. The foregoing scenario change indication or the performance level change indication may be discovered by the first RAN node by analyzing the UE, or may be reported by the UE to the first RAN node. For details, refer to descriptions of a second design in step 706. Alternatively, the first model is a model related to the first RAN node, and the first RAN may determine, by analyzing the first RAN, that the application scenario changes, the performance level changes, or the like.
Step 705: The first RAN node sends the prediction information to the UE. Correspondingly, the UE receives the prediction information from the first RAN node.
Step 706: The UE sends the second feedback information to the first RAN node. Correspondingly, the first RAN node receives the second feedback information from the UE.
In the first design, when receiving the prediction information, the UE may compare the prediction information sent by the first RAN node with the actual information, to determine accuracy of the prediction information. The accuracy is referred to as actual accuracy of the prediction information. For example, the prediction information is trajectory information. The UE may compare the predicted trajectory information with actual trajectory information of the UE, to determine the actual accuracy of the prediction information, and send the second feedback information to the first RAN node, where the second feedback information indicates the actual accuracy of the prediction information. The indication may be an implicit indication, an explicit indication, or the like. In other words, the UE sends the actual accuracy of the prediction information to the first RAN node. For example, the UE may feed back actual accuracy of prediction information in a period of time to the first RAN node, where the feedback may periodic feedback, feedback based on event triggering, or the like. For example, the UE may use 10 seconds as a periodicity, calculate the actual accuracy of the prediction information every 10 seconds, and feed back the actual accuracy to the first RAN node. Alternatively, the UE also calculates the accuracy of the prediction information every 10 seconds. When the accuracy is less than a threshold (for example, 80%), the UE sends the actual accuracy of the prediction information to the first RAN node.
In the second design, when finding that the application scenario of the UE changes or the performance level of the UE changes, the UE may alternatively send the second feedback information to the first RAN node. The second feedback information indicates that the application scenario changes or the performance level changes. In other words, the UE sends an application scenario change indication, the performance level change indication, or the like to the first RAN node.
Step 707: The first RAN node sends the first feedback information to the OAM. Correspondingly, the OAM receives the first feedback information from the first RAN node.
In a design, the first RAN node may send the first feedback information to the OAM periodically, or based on event triggering, or the like. The first feedback information may implicitly or explicitly indicate the actual accuracy of the prediction information. In other words, the first RAN node sends the actual accuracy of the prediction information to the OAM. Similar to the foregoing step 706, the first RAN node may feed back the actual accuracy of the prediction information in a period of time to the OAM. For example, the first RAN node may feed back the actual accuracy of the prediction information in a period of time to the OAM periodically, based on event triggering, or the like.
In another design, when an application scenario or a performance level related to the first model changes, the first RAN node sends the first feedback information to the OAM. The first feedback information indicates that the application scenario changes or the performance level changes. In other words, the first RAN node sends the application scenario change indication, the performance level change indication, or the like to the OAM.
Step 708: The OAM updates the first model based on the first feedback information.
In a design, when the first feedback information indicates the actual accuracy of the prediction information output by the first model, that the OAM updates the first model based on the first feedback information includes: When the actual accuracy of the prediction information output by the first model is less than a threshold, the first model is retrained, and the first model is updated. For example, the OAM may collect training data, where the training data is from the first RAN node and/or the UE, another node, or the like. The OAM may retrain the first model by using the collected training data. For a model training process, refer to the foregoing descriptions.
In another design, the first feedback information indicates a scenario change indication, a model performance level change indication, or a model performance indicator change indication. That the OAM updates the first model based on the first feedback information includes: A second model is determined based on the scenario change indication, the model performance level change indication, or the model performance indicator change indication. For example, the first mapping relationship includes a mapping relationship between a model, a model application scenario, and a model function, and the OAM may determine a changed scenario based on the indicated scenario change indication. A model corresponding to the changed scenario is determined based on the foregoing first mapping relationship, and is referred to as the second model. Alternatively, the first mapping relationship includes a mapping relationship between a model, a model performance level, and a model function, and the OAM may determine a changed model performance level based on the indicated model performance level change indication. Alternatively, the OAM may determine a changed model performance level based on the indicated model performance indicator change indication. A model corresponding to the changed model performance level is determined based on the foregoing mapping relationship, and is referred to as the second model. Alternatively, the first mapping relationship includes a mapping relationship between a model, a model performance indicator, and a model function. The OAM determines a model corresponding to an indicated changed model performance indicator based on the first mapping relationship, and the model is referred to as the second model.
Step 709: The OAM sends information about an updated model to the first RAN node. Correspondingly, the first RAN node receives the information about the updated model from the OAM.
For content included in the information about the updated model, refer to the content included in the information about the first model in the foregoing step 703. Details are not described herein again.
In this application, the first RAN node may send the first feedback information to the OAM periodically, based on event triggering, or the like. The OAM evaluates, based on the first feedback information, whether the first model needs to be updated, to update the first model in a timely manner, and improve prediction performance and a prediction speed of the first model.
After the first model is configured for the first RAN node by using the method in the procedure in
Step 801: The source RAN node determines a target cell.
In a design, the source RAN node may receive a measurement report reported by UE, where the measurement report includes identification information of the cell, quality information of the cell, and the like. The cell may be a current serving cell, a neighboring cell, and/or the like. This is not limited. In a design, the UE may periodically report the measurement report to the source RAN node. For example, the UE may periodically measure signal quality of a serving cell and/or signal quality of a neighboring cell, and then periodically report measured information to the source RAN node. Alternatively, the UE may periodically measure signal quality of a serving cell and/or signal quality of a neighboring cell, and report measured information or the like to the source RAN node when a specific condition is met.
In this application, the measurement report includes identification information of the cell. The identification information of the cell may include at least one of a cell global identifier (CGI), a physical cell identifier (PCI) and a frequency, a cell identity (cell ID), a non-public network identifier (NPN ID), a non-terrestrial network identifier (NTN ID), another cell identity, or the like of the cell. The CGI may include a public land mobile network (PLMN ID), a cell ID, and the like. Optionally, the identification information of the cell may further include a tracking area code (TAC) and/or identification information of a network device to which the cell belongs, for example, a global network device identifier.
In this application, the measurement report includes signal quality information of the cell. For example, the UE may obtain the quality information of the cell by measuring at least one of a downlink synchronization channel, a channel state information reference signal, a demodulation reference signal (DMRS), a cell-specific reference signal (CRS), a synchronization signal block (SSB), a synchronization signal/physical broadcast channel block, or another downlink signal. For example, the quality information of the cell may include at least one of received signal code power (RSCP), reference signal received power (RSRP), reference signal received quality (RSRQ), a signal noise ratio (SNR), a signal to interference plus noise ratio (SINR), a reference signal strength indication (RSSI), or other signal quality.
Optionally, the quality information of the cell may be at least one of a cell level, a beam level, a synchronization signal/physical broadcast channel block (SS/PBCH Block) level, a channel state information reference signal (CSI-RS) level, a numerology (numerology) level, a slicing (slicing) level, or a bandwidth part (BWP) level. A level of the quality information of the cell is a granularity at which the quality information of the cell is measured. For example, the quality information of the cell being at the cell level means that the UE measures each cell in a plurality of to-be-measured cells, to obtain quality of each cell. Alternatively, the quality of the cell being at the beam level means that the cell includes at least one beam, and the UE obtains the quality information of the cell by measuring a beam in the cell. For example, a cell includes three beams, and the UE may separately measure a beam 1, a beam 2, and a beam 3 in the three beams, to obtain quality information of the beam 1 and quality information of the beam 2 that meet a beam quality condition. Optionally, the UE may then obtain, for example, a larger value, an average, or a weighted summation based on the quality information of the beam 1 and the quality information of the beam 2, to obtain the quality information of the cell. Alternatively, the UE may report both the quality information of the beam 1 and the quality information of the beam 2. The quality information of the cell measured by the UE at another granularity is similar to the foregoing, and is not described again.
It should be noted that, if the cell quality information is the beam level, the SS/PBCH block level, the CSI-RS level, the air interface technology level, the slicing level, the BWP level, or the like, the identification information of the cell further includes at least one of identification information of a corresponding beam, identification information of an SS/PBCH block, identification information of a CSI-RS, identification information of an air interface technology, identification information of a slice, identification information of a BWP, or the like.
In a design, when receiving the measurement report, the source RAN node may determine, based on the measurement report, whether a serving cell needs to be handed over for the UE. For example, if quality of service of a source cell is lower than a specific threshold, it may be determined that the serving cell needs to be handed over for the UE. The source RAN node may select the target cell for the UE based on the received measurement report. For example, a cell whose signal quality is greater than the threshold and whose signal quality is the best may be selected from neighboring cells of the UE as the target cell.
In another design, the source RAN node may determine the target cell in an AI manner. The source RAN node may predict the target cell by using an AI model or a machine learning (ML) model. Input data of the AI model or the ML model includes at least one of the following: the measurement report of the UE, historical and/or real-time load of the neighboring cell, historical trajectory information of the UE, geographical coordinates of the UE at a current moment, a moving direction of the UE at a current moment, a moving speed of the UE, or the like. Output data of the AI model or the ML model includes at least one of the following: the target cell of the UE, a handover timestamp, and the like.
Step 802: The source RAN node sends model information to the target RAN node corresponding to the target cell.
In a design, the source RAN node may send the model information to the target RAN node, where the model information includes an index of a first model. Optionally, the model information may further include at least one of the following: an application scenario of the first model, a function of the first model, data that needs to be collected by the first model for model inference, or the like. For example, in an AI-based mobility enhancement solution, a RAN node that performs the model inference needs to request a historical handover success or failure case of the UE from a RAN node that the UE previously accesses. In this application, a process in which the source RAN node needs to notify the target RAN node of data that needs to be collected for the model inference may be further described as follows: The RAN node sends first indication information to the target RAN node, where the first indication information indicates type information of the input data of the first model. In this process, the source RAN node may notify the target RAN node in advance of the data that needs to be collected for the model inference in a model inference process, so that the target source RAN prepares for the model inference, to improve model inference efficiency.
Step 803: The source RAN node sends historical information of the UE to the target RAN node, where the historical information includes at least historical data, historical information, and/or the like collected by the source RAN node from the UE.
In a design, the historical information of the UE may be information needed for performing the model inference by using the first model or the like. In other words, when the target RAN node performs the model inference by using the first model, the historical information may be used as a part of the input data of the first model. The foregoing process may be further described as follows: The source RAN node sends historical data information needed for the model inference to the target RAN node, where the historical data information is for determining the input data of the first model. For example, the target RAN node may directly use the historical data as input of the first model to perform the model inference; or the target RAN node may process the historical data, and use processed historical data as input of the first model. This is not limited.
It should be noted that step 803 may be performed after step 802, or may be performed after step 804. This is not limited. In the procedure in
Step 804: The UE is handed over from the source cell corresponding to the source RAN node to the target cell corresponding to the target RAN node. After handover, the target cell serves as the serving cell of the UE.
Step 805: The target RAN node sends second indication information to the source RAN node, where the second indication information indicates at least one of the following: indicates that the UE is handed over to the target cell, indicates the source RAN node to delete the first model, or indicates the source RAN node to release a corresponding computing resource.
For example, the source RAN node allocates 1% of the resource to the UE for the model inference. After the UE is handed over to the target cell, the source RAN node may release the 1% of the computing resource allocated to the UE, delete the first model, and the like.
Step 806: The source RAN node deletes the first model to release the corresponding computing resource.
Step 807: The target RAN node sends a second model request message to OAM, where the second model request message indicates the index of the first model.
Step 808: The OAM may determine the first model based on the index of the first model.
In this application, there is a mapping relationship between the index of the first model and the first model. The OAM may determine the first model based on the index of the first model.
Step 809: The OAM sends information about the first model to the target RAN node.
Step 8010: The target RAN node performs the model inference based on the first model to obtain prediction information.
Step 8011: The target RAN node sends the prediction information to the UE.
Step 8012: The UE sends second feedback information to the target RAN node.
Step 8013: The target RAN node sends first feedback information to the OAM.
Step 8014: The OAM updates the first model.
Step 8015: The OAM sends information about an updated first model to the target RAN node.
For a specific performing process of steps 809 to 8015, refer to the descriptions in
In this application, when the UE performs cell handover, the source RAN node sends the index of the first model to the target RAN node, to ensure continuity of an AI service. In addition, after the UE is successfully handed over, the target RAN node notifies the source RAN node to delete model-related information to release the computing resource, so that memory resource occupation and computing resource occupation of the RAN node can be reduced.
It may be understood that, to implement functions in the foregoing methods, the OAM and the RAN node include corresponding hardware structures and/or software modules for performing the functions. A person skilled in the art should be easily aware that, with reference to units and method steps in the examples described in this application, this application can be implemented by hardware or a combination of hardware and computer software. Whether a function is performed in a form of hardware or hardware driven by computer software depends on particular application scenarios and design constraints of the technical solutions.
As shown in
When the communication apparatus 900 is configured to implement a function of the OAM in the method shown in
When the communication apparatus 900 is configured to implement a function of the first RAN node or the source RAN node in the method shown in
When the communication apparatus 900 is configured to implement a function of the target RAN node in the method shown in
For more detailed descriptions of the processing unit 910 and the transceiver unit 920, directly refer to the related descriptions of the method shown in
As shown in
When the communication apparatus 1000 is configured to implement the foregoing method, the processor 1010 is configured to implement the functions of the processing unit 910, and the interface circuit 1020 is configured to implement the functions of the transceiver unit 920.
When the communication apparatus is a module used in the OAM, the module in the OAM implements the functions of the OAM in the foregoing method. The module in the OAM receives information from another module (for example, a radio frequency module or an antenna) in the OAM, where the information is sent by the RAN node to the OAM; or the module in the OAM sends information to another module (for example, a radio frequency module or an antenna) in the OAM, where the information is sent by the OAM to the RAN node. The module in the OAM herein may be an OAM baseband chip, or may be another module and the like.
When the communication apparatus is a module used in the RAN node, the module in the RAN node implements functions of the RAN node in the foregoing method. The module in the RAN node receives information from another module (for example, a radio frequency module or an antenna) in the RAN node, where the information is sent by a terminal to the RAN node; or the module in the RAN node sends information to another module (for example, a radio frequency module or an antenna) in the RAN node, where the information is sent by the RAN node to a terminal. The module in the RAN node herein may be a RAN node baseband chip, or may be a DU or another module. The DU herein may be a DU in an open radio access network (O-RAN) architecture.
It may be understood that the processor in this application may be a central processing unit (CPU), or may be another general-purpose processor, a digital signal processor (DSP), or an application-specific integrated circuit (ASIC), a field programmable gate array (FPGA), or another programmable logic device, a transistor logic device, a hardware component, or any combination thereof. The general-purpose processor may be a microprocessor or any regular processor.
The memory in this application may be a random access memory, a flash memory, a read-only memory, a programmable read-only memory, an erasable programmable read-only memory, an electrically erasable programmable read-only memory, a register, a hard disk drive, a removable hard disk drive, a CD-ROM, or a storage medium of any other form well-known in the art.
For example, a storage medium is coupled to a processor, so that the processor can read information from the storage medium and write information into the storage medium. The storage medium may alternatively be a component of the processor. The processor and the storage medium may be disposed in an ASIC. In addition, the ASIC may be located in a base station or a terminal. Certainly, the processor and the storage medium may exist in a base station or terminal as discrete components.
All or a part of the methods in this application may be implemented by software, hardware, firmware, or any combination thereof. When software is for implementing embodiments, all or a part of embodiments may be implemented in a form of a computer program product. The computer program product includes one or more computer programs or instructions. When the computer programs or instructions are loaded and executed on the computer, the procedure or functions according to this application are completely or partially executed. The computer may be a general-purpose computer, a dedicated computer, a computer network, a network device, user equipment, a core network device, OAM, or another programmable apparatus. The computer program or the instructions may be stored in a computer-readable storage medium, or may be transmitted from a computer-readable storage medium to another computer-readable storage medium. For example, the computer program or the instructions may be transmitted from a website, computer, server, or data center to another website, computer, server, or data center in a wired or wireless manner. The computer-readable storage medium may be any usable medium that can be accessed by the computer, or a data storage device, for example, a server or a data center, integrating one or more usable media. The usable medium may be a magnetic medium, for example, a floppy disk, a hard disk drive, or a magnetic tape; or may be an optical medium, for example, a digital video disc; or may be a semiconductor medium, for example, a solid-state drive. The computer-readable storage medium may be a volatile or non-volatile storage medium, or may include two types of storage media: a volatile storage medium and a non-volatile storage medium.
In this application, unless otherwise stated or there is a logic conflict, terms and/or descriptions in different embodiments are consistent and may be mutually referenced, and technical features in different embodiments may be combined based on an internal logical relationship thereof, to form a new embodiment.
In this application, “at least one” means one or more, and “a plurality of” means two or more. “And/or” describes an association relationship between associated objects, and represents that three relationships may exist. For example, A and/or B may represent the following cases: A exists alone, both A and B exist, and B exists alone, where A and B may be singular or plural. In text descriptions of this application, the character “/” indicates an “or” relationship between associated objects. In a formula in this application, the character “/” indicates a “division” relationship between associated objects. “Including at least one of A, B, or C” may indicate: including A; including B; including C; including A and B; including A and C; including B and C; and including A, B, and C.
It may be understood that various numerals used in this application are merely differentiated for ease of description, but are not used to limit the scope of this application. Serial numbers of the foregoing processes do not mean an execution sequence, and the execution sequence of the processes needs to be determined based on functions and internal logic of the processes.
Number | Date | Country | Kind |
---|---|---|---|
202210096175.1 | Jan 2022 | CN | national |
This application is a continuation of International Application No. PCT/CN 2023/072691, filed on Jan. 17, 2023, which claims priority to Chinese Patent Application No. 202210096175.1, filed on Jan. 26, 2022. The disclosures of the aforementioned applications are herein incorporated by reference in their entireties.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2023/072691 | Jan 2023 | WO |
Child | 18784877 | US |