The field relates generally to using distributed machine learning models to improve conditional handover (CHO) procedures. More specifically, the field relates to generating connectivity graphs using distributed machine learning (ML) models for use in CHO procedures. Disclosed are aspects related to generating connectivity graphs between device and network nodes for use in CHO procedures.
In the past few years, machine learning has led to major breakthroughs in various areas, such as natural language processing, computer vision, speech recognition, and Internet of Things (IoT), with some breakthroughs related to automation and digitalization tasks. Machine learning can also be applied to the field of telecommunications. Most of this success stems from collecting and processing big data in suitable environments. For some applications of machine learning, this process of collecting data can be incredibly privacy invasive. One potential use case is to improve the results of speech recognition and language translation, while another one is to predict the next word typed on a mobile phone to increase the speed and productivity of the person typing. Another potential use case is predicting when a connected device needs to handover from one network to another before the connected device loses connectivity to its current network. In all cases, it would be beneficial to directly train on the same data instead of using data from other sources. This would allow for training a machine learning (ML) model (referred to herein as “model” also) on the same data distribution (i.i.d.—independent and identically distributed) that is also used for making predictions. However, directly collecting such data might not always be feasible owing to privacy concerns. Users may not prefer nor have any interest in sending everything they type to a remote server/cloud.
One recent solution to address this is the introduction of federated learning, a new distributed machine learning approach where the training data does not leave the users' computing device at all. Instead of sharing their data directly, the client computing devices themselves compute weight updates using their locally available data. It is a way of training a model without directly inspecting clients' or users' data on a server node or computing device. Federated learning is a collaborative form of machine learning where the training process is distributed among many users. A server node or computing device has the role of coordinating between models, but most of the work is not performed by a central entity anymore but by a federation of users or clients.
After the model is initialized in every user or client computing device, a certain number of devices are randomly selected to improve the model. Each sampled user or client computing device receives the current model from the server node or computing device and uses its locally available data to compute a model update. All these updates are sent back to the server node or computing device where they are averaged, weighted by the number of training examples that the clients used. The server node or computing device then applies this update to the model, typically by using some form of gradient descent.
Current machine learning approaches require the availability of large datasets, which are usually created by collecting huge amounts of data from user or client computing devices. Federated learning is a more flexible technique that allows training a model without directly seeing the data. Although the machine learning process is used in a distributed way, federated learning is quite different to the way conventional machine learning is used in data centers. The local data used in federated learning may not have the same guarantees about data distributions as in traditional machine learning processes, and communication is oftentimes slow and unstable between the local users or client computing devices and the server node or computing device. To be able to perform federated learning efficiently, proper optimization processes need to be adapted within each user machine or computing device. For instance, different telecommunications operators will each generate huge alarm datasets and relevant features. In this situation, there may be a good list of false alarms compared to the list of true alarms. For such a machine learning classification task, typically, the dataset of all operators in a central hub/repository would be required beforehand. This is required since different operators will encompass a variety of features, and the resultant model will learn their characteristics. However, this scenario is extremely impractical in real-time since it requires multiple regulatory and geographical permissions; and, moreover, it is extremely privacy-invasive for the operators. The operators often will not want to share their customers' data out of their premises. Hence, distribute machine learning, such as federated learning, may provide a suitable alternative that can be leveraged to greater benefit in such circumstances.
Handover occurs when a connected device (such as a cell phone) or user equipment (UE) switches its connection from one network such as a cell tower or other terrestrial network (TN) to another such as a non-terrestrial network (NTN) such as a satellite. In a stationary context, the HO decision might not be tricky. However, if both the serving cell and the UE is non-stationary and moving at high speeds (such as a satellite), forecasting and then optimizing a set of HO parameters in advance is critical to the success of HO.
There is a tradeoff for UE being connected to non-terrestrial network (NTN) or terrestrial network (TN). On one hand, the cost of UE being connected to TN is less, for example less UE energy consumption, as compared to being connected to NTN. On the other hand, once hand-in (NTN->TN HO) happens due to short sighted decisions especially in rural areas, there is a risk for UE losing connectivity to TN after some time in the case of high mobility due to TN's limited coverage area as compared to NTN. In addition, there is a risk that the TN network gets overloaded or might serve UE with poor quality of experience (QoE) given a particular service (such as broadband video streaming) at particular time of the day (e.g., busy hour). This then yields a tradeoff between the NTN's serving cost and end-user QoE.
Therefore, there exists a need for a UE being connected to NTN, where the UE should have a smart and pro-active HO control mechanism that decides whether or not to handover to TN based on its forecasted trajectory and UE context in the future, given that connecting back from TN to NTN would be hard as the connection of UE to the satellite was lost before. When to hand-in (NTN to TN) should be decided carefully with good forecasting of the future UE and network state in advance.
In earlier releases than 5G, the handover (HO) command was triggered after a user equipment (UE) reports an A3 event indicating that a UE reference signal received power (RSRP) on the serving cell is getting worse than the UE RSRP at a neighboring cell. However, it is sometimes too late to HO after sending the report since the quality can be too bad to send the report or the HO commands. In 5G new radio (NR), there are ways to cope with late HO that eventually causes radio link failure (RLF) and HO failure such as conditional HO (TS 38.331, TS 38.423). In conditional HO, the handover command is sent from the network to the UE in advance (based on previous measurement reports). Along with the early HO command, a list of candidate neighboring cell names are also sent. Both HO command and the list of candidate neighboring cells are stored in the UE. The UE does not perform HO immediately, but triggers the HO command only when the certain criteria is met (such as threshold delta between source and any target cell). The candidate neighboring cells received from the network should also be neither too large (otherwise might unnecessarily allocate resources in other cells, especially high load ones, during HO preparation), nor too small (not to reduce the risk of coverage and RLF).
Therefore, there exists a need to combine all the features in a machine-learning (ML) model so that a right planning for the HO control can be given in advance since there are massive number of input parameters and conditions (i.e., priors) to consider in the evaluation which might be hard to accomplish with rule based systems. There further exists a need to incorporate federated or distributed learning into this machine-learning model.
The method and device disclosed herein is not a replacement for existing conditional HO mechanism, but rather to improve it with more holistic view (obtained via collaborative learning with considerations of much more attributes and dimensions) and help CHO (conditional handover) to associate with optimum decision thresholds. This is especially important since in the CHO, there is a tradeoff between energy consumption and HO failure. The current implementation of CHO is mainly to reduce the risk of HO failure with the cost of allocating and preparing more target cells for the handover, and it might be that the suboptimal preparations in advance might allocate unnecessary resources such as increased energy consumption and increased QoS resource allocation on the target cells.
The HO alternatives are sent from TN to UE in the form of a graph representation, where the nodes in the graph are alternative network nodes where the HO can be made to, and the link between the nodes indicate the ranking with respect to the estimated overall quality if the HO is made to the corresponding target network.
The ML based collaborative method disclosed herein reduces the NTN HO mechanism decision with long and short term predictions. The goal is to minimize the UE's handover failure that occurs during HO from/to NTN.
The method disclosed herein includes cascaded/hierarchically structured KPI and UE context prediction models. The output of the final prediction model is fed into a Graph Neural Network to generate and recommend the graph structure given the predicted state of the network. The generated graph represents the connectivity of UE to the network which would yield minimum HO failure to/from NTN network and reduce cost by extending the connectivity time of the UE to TN if the UE is not expected to need NTN. As the structure of the generated graph can be similar in different time and location, the method disclosed herein includes a collaborative machine learning method on top of it to enable the generative and the prediction models to share weights in between each other. This necessitates a grouping of similar models before each model joins the training.
In some aspects disclosed herein, the goal is to generate the best connectivity graph between the UE and the network provider nodes (NTN and TN) to minimize the energy consumption and handovers and to maximize the objective QoE (derived from pre-trained model). The method and device disclosed herein considers the existing HO mechanisms such as conditional HO(CHO), and enhances it with a data-driven decision.
The method and device disclosed herein provides several advantages, including but not limited to: preventing the UE from going into a disconnected state due to handover failure or radio link failure; preventing the UE from being out of coverage by a proactive handover decision; reducing the cost of serving by minimizing the serving time from NTN; and reducing the energy consumption by reducing the number of handovers.
Internet of Things (IoT) Connectivity Management Use Case
In some cases an IoT device drops connectivity from a source cell before reestablishing on a new base station target cell; followed by the fact the device cannot establish a connectivity with the target cell. In the worst case, the scenario the IoT device completely loses connectivity and cannot connect, and is pushed into a dead-lock. In UE's this is not a problem since the human user can initiate a restart or can move close to one base station when the phenomena is experienced to trigger reconnection. But for remote stationary, IoT devices it is harder to do unless there is an auto-reboot mechanism (watch-dog timer) implementation in place.
The factors such as environmental, the traffic load of the candidate base stations are critical to the HO decision as they cause high interference. One example is the foliage, e.g., IoT sensors installed within a large forest area for early fire detection. Another example could be the geographical areas with river and large water damps, as well as heavy rainfall scenarios in a monsoon season climate. These factors can create additional interference that impacts measurement reports observed in the UE's in the area that might eventually cause HO failures.
In a first aspect, a method for a distributed machine learning (ML) assisted conditional handover (CHO) procedure for a connected device is provided. The method comprises receiving, by the connected device, one or more measurement configurations from a source station; transmitting, by the connected device, one or more measurement reports to the source station; receiving and storing, by the connected device, one or more CHO commands from the source station, the one or more CHO commands having at least one triggering condition for CHO to one or more candidate target cells, wherein the one or more CHO commands are determined by inputting the one or more measurement reports into a distributed ML-generated reference model; and evaluating, by the connected device, whether the at least one triggering condition in any of the one or more CHO commands is fulfilled.
In some aspects, the method further comprises executing, by the connected device, a decision for handover based on the evaluation, wherein the connected device executes the handover if the one or more CHO commands is fulfilled and does not execute the handover if the one or more CHO commands is not fulfilled.
In some aspects, the distributed ML-generated reference model is generated by: collecting data for a set of device and network node agents, wherein the collected data include attribute information for at least one connected device used for training, attribute information for at least one non-terrestrial network (NTN), and attribute information for at least one terrestrial network (TN); building, using the attribute information from the collected data, dependency graphs representing interaction relationships between the set of device and network node agents; constructing, using the dependency graphs, graph-based ML models representing profiles for the set of device and network node agents, wherein constructing the graph-based ML models includes generating connectivity between the set of device and network node agents; and training and validating the graph-based ML models using the generated connectivity for the set of device and network node agents.
In some aspects, the data for the set of device and network node agents comprises one or more of: the cost of the at least one training device being connected to the at least one NTN; the cost of the at least one training device being connected to the at least one TN; the service and application usage time by the at least one training device within a next time interval (T); the throughput and latency requirements of a predicted service in the at least one training device; and the conditions of at least one of the networks available to the at least one training device.
In some aspects, the attribute information for the cost of the at least one connected device for training being connected to the at least one NTN comprises one or more of: the previous cost of service for all available NTNs; the time when the previous service occurred; and the total power consumption of the at least one connected device for training and the at least one NTN during service.
In some aspects, the attribute information for the cost of the at least one connected device for training being connected to the at least one TN comprises one or more of: the previous cost of service for all available TNs; the time when the previous service occurred; and the total power consumption of the at least one connected device for training and the at least one TN during service.
In some aspects, the attribute information for the service and application usage time by the at least connected device for training within the next time interval (T) comprises one or more of: the type of service being used in the at least one training device; the time of usage; and the historical behavior of users of the at least one training device.
In some aspects, the attribute information for the throughput and latency requirements of the predicted service in the at least one connected device for training comprises one or more of: the location of the at least one connected device for training; the speed at which the at least connected device for training is moving; and the historical mobility trajectory of the at least connected device for training.
In some aspects, the attribute information for the conditions of the at least one of the networks available to the at least connected device for training comprises one or more of: the proximity of all potential satellites in the at least one network; the available radio network deployment topology from the at least one network's knowledge base; the propagation conditions of the at least one network; the downlink signal-to-interference-plus-noise ratio as a function of time for the at least one network; the system load on the at least one network; the number of radio link failure (RLF) events in the at least one network; the historical and current handover failure (HOF) events in the at least one network; the historical and current ping-pong rate in the at least one network; the historical average time in outage of the at least one network; the predicted coverage of the at least one network's TN's per time slot; the historical availability of TNs and NTNs in the at least one network; the availability of TNs and NTNs in the at least one network in the past per timeslot at a given geo-location and predicted availability in a plurality of future time intervals; and the current and previous TN names and NTN names and elevations in the at least one network.
In some aspects, the collected data further includes data collected by a collaborative machine learning method. In some aspects, the collaborative machine learning method is a federated learning method.
In some aspects, the graph-based ML models are generated according to:
In some aspects, the source station is a non-terrestrial network (NTN).
In some aspects, the one or more candidate target cells are terrestrial networks (TNs).
In some aspects, the connected device is a user equipment (UE).
In some aspects, the at least one or more CHO commands includes at least one of: a leaving condition, a target cell identity (ID), a life timer, a CHO command priority, and a CHO command ID.
According to a second aspect, a connected device is provided, the device comprising processing circuitry; and a memory, said memory containing instructions executable by said processing circuitry, whereby said connected device is operative to perform the method of any one of the previously disclosed aspects.
According to a third aspect, a connected device is provided, wherein the connected device is adapted to perform the method of any one of the previously disclosed aspects.
In some aspects, the connected device is a user equipment (UE).
In some aspects, the connected device is any device selected from a group consisting of: a sensor, a ground vehicle, an aircraft, or a watercraft, wherein the selected device has connectivity capabilities to both terrestrial and non-terrestrial networks.
According to a third aspect, a computer program comprising instructions for adapting an apparatus to perform the method of any one of the previously disclosed aspects is provided.
According to a fourth aspect, a carrier containing the computer program previously disclosed, wherein the carrier is one of an electronic signal, optical signal, radio signal, or compute readable storage medium, is provided.
In some aspects, t the graph-based ML models are graph neural networks.
The accompanying drawings, which are incorporated herein and form part of the specification, illustrate various aspects.
In some aspects, the following relationships exist between the various sequence graphs and features shown in
In some aspects, an adjacency matrix may be represented by:
Depending on the results of the predicted models and also the UE requirements in time and space, scheduling of UE's to the right network nodes (TN or NTN) can be made using the graph neural network (GNN) algorithm. The ML model needs to be formulated such that the output of the model is the decision to handover or not rather while the intermediate results from predictions are used along the final decision.
In some aspects, a topology (connectivity) graph is generated given the conditions. The conditions are a function of multiple models that each consists of various application, radio, network and energy metrics as mentioned previously in this disclosure. The goal here is to generate realistic connectivity graphs between UE and the candidate network nodes that the UE will be connected to. One candidate technique is GraphRNN which is used in drug discovery use cases, for example. The same technique can either build a generation from scratch or complete an incomplete graph (e.g., in the cases the UE is aware of only a subset of nodes but would like to have alternative nodes as well, in that case graph completion can be performed).
In some aspects, the graph is represented by the adjacency matrix previously disclosed, and the space complexity of adjacency matrix is O(n2). Since the candidate network nodes that the graph would involve/generate is not expected to be high (e.g., 4-5 nodes max.), space complexity in our use case might not be a big issue. However, the order of the graph is important which will be constructed with the ranking information, hence the problem might rise with a O(N!) time complexity. Again, due to not large N, this is not foreseen as a major issue. Graph generation includes two main steps: i) obtain the distribution of the dataset; ii) draw samples from the obtained distribution so that the drawn samples will be as similar as the generated samples by a learned model (parameterized by theta). So the goal is to learn the parameters theta of the generative model (e.g., GraphRNN). In some aspects, the goal is to maximize the likelihood that the aspects disclosed herein can take the average of argmax, hence the role of E. In some aspects, this relationship can be expressed by the formula:
Q*=argmaxQ Ex˜pdata log pmodel(X|Q)
The training phase comprises many graphs that the model learns from, and in the inference phase it is asked to generate graphs with given conditions. Since the above equation assumes a IID (identical and independent distribution) of data samples, and since some connectivity graph sets might be very different than others in reality based on geographical location and availability and network conditions, this disclosure proposes group federation where the graphs are separated into groups based on their similarity.
In another aspect, federated learning is utilized (also in the form of federation of graph models) so that the UE's record historical events are shared by means of local (UE) trained updates so that the HO decision uses experience information on previous handovers. As this information is sensitive, a federated learning method can be implemented where only the learned model weights are shared in between rather than the raw GPS data. Just before UE decides, the local model weights are updated, so that the estimation for the TN location is performed via the federated ML model. If the model does not estimate a nearby TN, the measurement should not be performed hence a NTN-to-NTN handover is initiated instead.
While various aspects of the present disclosure are described herein, it should be understood that they have been presented by way of example only, and not limitation. Thus, the breadth and scope of the present disclosure should not be limited by any of the above-described exemplary aspects. Moreover, any combination of the above-described elements in all possible variations thereof is encompassed by the disclosure unless otherwise indicated herein or otherwise clearly contradicted by context.
Additionally, while the processes described above and illustrated in the drawings are shown as a sequence of steps, this was done solely for the sake of illustration. Accordingly, it is contemplated that some steps may be added, some steps may be omitted, the order of the steps may be re-arranged, and some steps may be performed in parallel.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/SE2022/050135 | 2/8/2022 | WO |