The disclosed embodiments relate generally to wireless communication, and, more particularly, to AI-ML model identification.
Artificial Intelligence (AI) and Machine Learning (ML) have permeated a wide spectrum of industries, ushering in substantial productivity enhancements. In the realm of mobile communications systems, these technologies are orchestrating transformative shifts. Mobile devices are progressively supplanting conventional algorithms with AI-ML models to improve performance, user experience, and reduce complexity/overhead.
In the conventional network of the 3rd generation partnership project (3GPP) 5G new radio (NR), by leveraging AI-ML technology to address challenges due to the increased complexity of foreseen deployments over the air interface, both for the network and UEs. The network and UEs interact with each other and exchange data and control signals by which in AI/ML Life cycle management (LCM) is realized. Due to the data volume and complexity of the AI-ML model, how to identify and create a general/common understanding among different network entities is challenging.
Improvements and enhancements are required to identify the AI-ML model across the wireless network.
Apparatus and methods are provided for AI-ML model/functionality identification. In one novel aspect, model ID is used to identify the AI-ML model. In one embodiment, the UE receives AI-ML model from an AI server in the wireless network, obtains related model information of the AI-ML model and provides the related model information of the received AI-ML model to the wireless network. In one embodiment, the AI server is an over-the-top (OTT) server. In one embodiment, the OTT server is UE vendor specific or chipset specific. In one embodiment, the OTT server can be trusted server or untrusted server. In one embodiment, the OTT server is owned by an operator. In one embodiment, the AI-ML model and the related model information of the AI-ML model are delivered together from the AI server to the UE. In one embodiment, the related model information of the AI-ML model is delivered from the AI server to a network function (NF) or a network server. In another embodiment, the related model information of the AI-ML model is delivered to the RAN node from the NF or the network server. In one embodiment, the UE receives the AI-ML model from the AI server through a user-plane (UP) traffic with the related model information of the AI-ML model. The UE starts model identification procedure to the network based on the AI-related information. In one embodiment, the UE obtains the related model information of the AI-ML model by performing at least one model identification procedure comprising a UE-vendor specification procedure and a chipset specific procedure. In one embodiment, the UE provides the related model information of the received AI-ML model to the RAN node through a radio resource control (RRC) procedure. In another embodiment, the UE provides the related model information of the received AI-ML model to a core network (CN) node through a Non-Access-Stratum (NAS) procedure. In one embodiment, the AI-ML model is identified by a model ID. In another embodiment, the UE receives a model index for each configured AI-ML model from the RAN during RRC connection, and wherein each model index is mapped to at least one model ID identifying an AI-ML model.
In one novel aspect, the radio access network (RAN) node of the wireless identifies the AI-ML model by model ID. In one embodiment, the RAN node receives a UE capability information with a first model ID for an AI-ML model in the wireless network, wherein the first model ID is maintained by an AI server, maps the first model ID to a second model ID, wherein the second model ID identifies the AI-ML model and is maintained by a network entity of the wireless network, wherein the first model ID and the second model ID are based on related model information of the AI-ML model, and performs an AI-ML model configuration for the UE based on the first model ID and the mapping of the first model ID and the second model ID. In one embodiment, the second model ID reuses the first model ID. In another embodiment, the second model ID is reconstructed based on the first model ID as a global model ID for AI-ML models from different AI server. In yet another embodiment, the RAN node assigns a model index for the AI-ML model, wherein the model index maps to the first model ID and the second model ID, and wherein the model index is used for the RRC CONNECTED UE to identify the AI-ML model.
This summary does not purport to define the invention. The invention is defined by the claims.
The accompanying drawings, where like numerals indicate like components, illustrate embodiments of the invention.
Reference will now be made in detail to some embodiments of the invention, examples of which are illustrated in the accompanying drawings.
Several aspects of telecommunication systems will now be presented with reference to various apparatus and methods. These apparatus and methods will be described in the following detailed description and illustrated in the accompanying drawings by various blocks, components, circuits, processes, algorithms, etc. (Collectively referred to as “elements”). These elements may be implemented using electronic hardware, computer software, or any combination thereof. Whether such elements are implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Please also note that terms such as transfer means uplink transfer and/or downlink transfer.
Wireless communication network 100 includes one or more fixed base infrastructure units forming a network distributed over a geographical region. One or more UEs, such as UE 101 are served by the base stations. The base unit may also be referred to as an access point, an access terminal, a base station, a Node-B, an eNode-B (eNB), a gNB, or by other terminology used in the art. As an example, base stations serve a number of mobile stations within a serving area, for example, a cell, or within a cell sector. In some systems, one or more base stations are coupled to a controller forming an access network that is coupled to one or more core networks. gNB 102, gNB 107 and gNB 108 are base stations in the wireless network, the serving area of which may or may not overlap with each other. gNB 102 is connected with gNB 107 via Xn interface 121. gNB 102 is connected with gNB 108 via Xn interface 122. gNB 107 is connected with gNB 108 via Xn interface 123. Core network (CN) entity 103 connects with gNB 102 and 107, through NG interface 125 and 126, respectively. Network entity CN 109 connects with gNB 108 via NG connection 127. Exemplary CN 103 and CN 109 connect to AI server 105 through internet 106. CN 103 and 109 includes core components such as user plane function (UPF) and core access and mobility management function (AMF). In one embodiment, AI server 105 is an over-the-top (OTT) server. In one scenario, the AI-ML model is stored in network-vendor OTT servers or UE-vendor OTT servers or chipset-vendor OTT servers.
The Model repository in UE, gNB, and CN is used for model storage for different functions and different versions. Data collection is capable of collecting data by the network nodes, management entity, or UE for the purpose of AI/ML model training, data analytics and inference. Model Identification is capable of identifying an AI/ML model for the common understanding between the NW and the UE. Model monitoring monitors the inference performance of the AI/ML model. Model transfer delivers an AI/ML model over the air interface, either parameters of a model structure known at the receiving end or a new model with parameters. Delivery may contain a full model or a partial model. In one embodiment, the network is responsible for model control, i.e., model activation/deactivation/switching.
The UE also includes a set of control modules that carry out functional tasks. These control modules can be implemented by circuits, software, firmware, or a combination of them. An AI-ML model receiver 191 receives an AI-ML model from an AI server in the wireless network, wherein the UE is connected with a radio access network (RAN) node of the wireless network. An information module 192 obtains related model information of the AI-ML model. An identification module 193 provides the related model information of the received AI-ML model to the wireless network.
The AI server, such as 211 and 212 integrates with wireless network, which includes exemplary CN/OAM 221, RAN node 231 and the UEs. In one embodiment, UE obtains an AI-ML model from the OTT server by Model delivery through NF/network server. For example, at step 251, AI server 211 delivers AI-ML model with related model information to UE through the network server, such as CN or OAM 221. At step 252, AI server 212 delivers AI-ML model with related model information to UE through the network server, such as CN or OAM 221. In one embodiment, the AI-ML model is transferred through RRC messages. In another embodiment, the AI-ML model is transferred through NAS messages. In one embodiment, the AI-ML model and model related information are delivered from the OTT server to NF/network server at step 261. In one embodiment, the NF/network server is an entity in the core network (CN). In another embodiment, the NF/network server is Operation Administration and Maintenance (OAM). In one embodiment, the NF/network server delivers the AI-ML model to an entity on the next-generation RAN (NG-RAN) side, such as RAN 231. In one embodiment, at step 262, the CN delivers model list and mapping to RAN node.
In one novel aspect, The AI-ML model is identified and managed for LCM by model ID. In one embodiment 292, Model ID identifies an AI-ML model with related model information. In one embodiment, UE vendor/OTT server manages its own Model IDs. For example, UE vendor/OTT server 211 manages its own Model ID list-1 281; UE vendor/OTT server 212 manages its own Model ID list-1 282. In one embodiment, a network entity manages its model list, such a Model ID list-2 285. In one embodiment, the network reuses the model IDs provided by different UE vendor/OTT servers. In one embodiment, the network re-structures the model IDs for models from different UE vendor/OTT servers. In one embodiment, RAN node 231 maintains a mapping 286 between Model ID list-1x, such as Mode ID list-1 281 and Model ID list-1 282, from the UE vendor/OTT server side and Model ID list-2, such as Model ID list-2 285, from the network side. In one embodiment, the RAN node assigns model index for each configured model during RRC connection for model control (activation/deactivation, switch, fallback). The RAN node with model index maintains a linkage among the model index, the Model ID list-1x and Model ID list-2.
At step 311, AI-ML model is delivered from the AI server, such the UE vendor/OTT server 302 together with related model information. At step 312, UE 301 receives AI-ML model with model related information. In one embodiment, the model related information is also identified by the model ID, which is consistent with model ID in model list-1 of the AI server 302. During model training and model development, each model should be identified by an identifier and even a version number considering the model can be updated. Considering the AI/ML model is kept proprietary and left to implementation, it's very likely that different vendors may have different sets of AI/ML models even for the same functionality. Therefore, model IDs in each vendor's repository are used for model development and management, which is implementation and vendor specific. We call it original model ID or Model ID list-1.
At step 321, CN 304 receives the related model information identified by model ID of Model ID list-1 managed/assigned by AI server 302. The network either reuses the model ID for its own Model ID list-2 or re-structures a new model ID for Model ID list-2 owned by the network. The model ID used for model development and management in repository for each vendor is not suitable to be used directly in model control over the air interface. If the network is responsible for model control, i.e., model activation/deactivation/switching, it needs to re-structure the UE vendor specific model ID lists to another form of model IDs, which are manageable by the network. We call it global model ID, or Model ID list-2. Each global model ID of this type should be able to uniquely identify each AI model known to the network. Certain network functions, e.g., management data analytics function (MDAF) may be responsible for model ID management, where the linkage between the original vendor-specific model IDs of the Model ID list-1 and the global model IDs of the Model ID list-2 are maintained. In one embodiment, at step 322, the network sends the linkage/mapping of the Model ID list-1 and Model ID list-2 to RAN 303.
At step 331, UE 301 sends UE capability report with model ID to RAN 303. According to the reported UE capability, RAN, at step 351, configures the models to be used during RRC connection to UE. In one embodiment, RAN reuses the Model IDs provided by different UE vendor/OTT servers.
In one embodiment 340, RAN assigns a Model Index to facilitate model control during the RRC lifetime. During model control in model-ID-based LCM, an appropriate AI-ML model among a set of AI/ML models is chosen for usage, which can fit one or more factors including the scenario, the configuration, and the site. Considering model activation, deactivation, switch, and/or selection only occurs when UE is in CONNECTED state, a temporary model index can be assigned to each model through model configuration. Model activation/deactivation/switch/selection and fallback can rely on the model index, similar as SCell index for model activation, deactivation, switching and fallback. Based on the mapping between Model ID list-1 and Model ID list-2, RAN configures the models to be used during the RRC connection. RAN assigns Model Indexes to facilitate model control during the RRC life time for different purposes, e.g. beam management (BM), channel state information (CSI). In one embodiment, the linkage between the Model Index configured by RAN and the Model ID configured by NF/network server is provided. Model Index list of RAN is mapped to model ID list-2 of NF/network server, thus the mapping relationship has been established with Model ID list-1, Model ID list-2, and Model Index list. For AI-ML over air interface, a set of specific models, e.g., scenario-specific, configuration-specific and/or site-specific models may be developed. Therefore, multiple AI-ML models for the same functionality will be supported. For model-ID-based LCM, the multiple models are visible to the peer entity, e.g., the UE-sided models are visible to the network side. For functionality-based LCM, even if there may be multiple AI-ML models at the UE side, the model switching operation at the UE side is transparent to the network.
In one embodiment, the AI-ML models are delivered from the UE vendor/OTT server to UE through UP traffic. In one embodiment, Model identification is performed based on UE-vendor specific procedure. In one embodiment, Model identification is performed based on Chipset specific procedure. UE parses the AI/ML model from UP traffic from the application layer and delivers the AI-ML model to the modem, the AI/ML model delivery over air interface is transparent to NF and RAN/CN. In one embodiment, the Model Identification is performed through RRC procedure between UE and RAN. In one embodiment, the Model Identification is performed through NAS procedure between UE and CN. In one embodiment, UE reports model identification information when a new AI-ML model or a new version of an AI-ML model is available on the UE side.
At step 510, AI-ML model and related mode information is delivered to the UE. At step 521, UE 501 receives AI-ML model. In one embodiment, the AI-ML model is delivered through UP traffic from the wireless network. At step 522, UE 501 obtains related model information to perform model identification procedure. In one embodiment, the model identification is performed based on UE-vendor specific procedure. In one embodiment, the model identification is performed based on Chipset specific procedure. UE parses the AI-ML model from UP traffic from the application layer and delivers the AI-ML model to the modem. In one embodiment, the AI-ML model delivery over air interface is transparent to NF and RAN/CN.
At step 530, UE performs model identification using NAS procedure. At step 551, UE sends UE capability with model ID to RAN 503. In one embodiment, the model and/or functionality information is stored at CN. In one embodiment, at step 531, RAN node request model information delivery from CN if it doesn't have the model information for the UE. At step 532, CN 504 sends the linkage/mapping between Model ID list-1 and Model ID list-2 to RAN 503. In another scenario, considering the scenario of mobility, in one embodiment, the model/functionality information stored at CN is delivered to RAN through NG-C interface. At step 552, RAN 503 receive model information identified by the model ID in the UE capability from CN 504. In one embodiment, UE vendor/OTT server manages its own Model IDs. In one embodiment, the network reuses the Model IDs provided by different UE vendor/OTT servers. In one embodiment, the network re-structure the Model IDs for models from different UE vendor/OTT servers. When both the RAN 503 and CN 504 has the model ID mapping, either the model ID of Model ID list-1 or of Model ID list-2 is sent.
In one embodiment, at step 561, RAN node assigns a model Index for each configured AI-ML model. For example, the UE vendor/OTT server manages its own model IDs with Model ID list-1, and the network re-structures the model IDs from different UE vendors with Model ID list-2, Model ID list-1 of UE vendor/OTT server is mapping to Model ID list-2 of the network. RAN configures the models to be used during the RRC connection according to the mapping of Model ID list-1 and Model ID list-2. RAN may assign Model Indexes to facilitate model control during the RRC life time for different purposes, e.g. BM, CSI. At step 570, RAN 503 configures the AI-ML with the model index.
At step 610, AI-ML model is delivered to the UE with the related model information. At step 621, UE 501 receives AI-ML model. In one embodiment, the AI-ML model is delivered through UP traffic from the wireless network. At step 622, UE 601 obtains related model information to perform a model identification procedure. In one embodiment, the model identification is performed based on UE-vendor specific procedure. In one embodiment, the model identification is performed based on Chipset specific procedure. UE parses the AI-ML model from UP traffic from the application layer and delivers the AI-ML model to the modem. In one embodiment, the AI-ML model delivery over air interface is transparent to NF and RAN/CN.
At step 630, UE performs model identification using RRC procedure. In one embodiment, at step 631, RAN 603 sends model information to CN 604. At step 633, CN 604 sends the linkage/mapping between Model ID list-1 and Model ID list-2 to RAN 603. In another scenario, considering the scenario of mobility, in one embodiment, the model/functionality information stored at CN is delivered to RAN through NG-C interface. In one embodiment, UE vendor/OTT server manages its own Model IDs. In one embodiment, the network reuses the Model IDs provided by different UE vendor/OTT servers. In one embodiment, the network re-structure the Model IDs for models from different UE vendor/OTT servers. When both the RAN 603 and CN 604 has the model ID mapping, either the model ID of Model ID list-1 or of Model ID list-2 is sent.
At step 651, UE sends UE capability with model ID to RAN 603. In one embodiment, at step 661, RAN node assigns a model Index for each configured AI-ML model. For example, the UE vendor/OTT server manages its own model IDs with Model ID list-1, and the network re-structures the model IDs from different UE vendors with Model ID list-2, Model ID list-1 of UE vendor/OTT server is mapping to Model ID list-2 of the network. RAN configures the models to be used during the RRC connection according to the mapping of Model ID list-1 and Model ID list-2. RAN may assign Model Indexes to facilitate model control during the RRC life time for different purposes, e.g. BM, CSI. At step 670, RAN 603 configures the AI-ML with the model index.
Although the present invention has been described in connection with certain specific embodiments for instructional purposes, the present invention is not limited thereto. Accordingly, various modifications, adaptations, and combinations of various features of the described embodiments can be practiced without departing from the scope of the invention as set forth in the claims.
Number | Date | Country | Kind |
---|---|---|---|
PCT/CN2023/076840 | Feb 2023 | WO | international |
CN 202410040969.5 | Jan 2024 | CN | national |
This application is filed under 35 U.S.C. § 111 (a) and is based on and hereby claims priority under 35 U.S.C. § 120 and § 365 (c) from International Application No. PCT/CN2023/076840, titled “Methods and apparatus of General Framework for Model/Functionality Identification,” filed on Feb. 17, 2023. This application claims priority under 35 U.S.C. § 119 from Chinese Application Number 202410040969.5, titled “Methods and apparatus of General Framework for Model/Functionality Identification,” filed on Jan. 10, 2024. The disclosure of each of the foregoing documents is incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2023/076840 | Feb 2023 | WO |
Child | 18437086 | US |