The present application claims priority to Indian Provisional Patent Application No. 20/232,1017243 filed on Mar. 15, 2023, the entirety of which is incorporated by reference herein.
The present disclosure is related to Open Radio Access Network (O-RAN) wireless networks, and relates more particularly to optimization for overload control in O-RAN networks.
In the following sections, an overview of Next Generation Radio Access Network (NG-RAN) architecture and 5G New Radio (NR) stacks will be discussed. 5G NR (New Radio) user and control plane functions with monolithic gNB (gNodeB) are shown in
NG-Radio Access Network (NG-RAN) architecture from 3GPP TS 38.401 is shown in
In this section, an overview of Layer 2 (L2) of 5G NR will be provided in connection with
1) Medium Access Control (MAC) 501 in
2) Radio Link Control (RLC) 502 in
3) Packet Data Convergence Protocol (PDCP) 503 in
4) Service Data Adaptation Protocol (SDAP) 504 in
Open Radio Access Network (O-RAN) is based on disaggregated components which are connected through open and standardized interfaces based on 3GPP NG-RAN. An overview of O-RAN with disaggregated RAN CU (Centralized Unit), DU (Distributed Unit), and RU (Radio Unit), near-real-time Radio Intelligent Controller (RIC) and non-real-time RIC is illustrated in
As shown in
A cell site can comprise multiple sectors, and each sector can support multiple cells. For example, one site could comprise three sectors and each sector could support eight cells (with eight cells in each sector on different frequency bands). One CU-CP (CU-Control Plane) could support multiple DUs and thus multiple cells. For example, a CU-CP could support 1,000 cells and around 100,000 User Equipments (UEs). Each UE could support multiple Data Radio Bearers (DRB) and there could be multiple instances of CU-UP (CU-User Plane) to serve these DRBs. For example, each UE could support 4 DRBs, and 400,000 DRBs (corresponding to 100,000 UEs) may be served by five CU-UP instances (and one CU-CP instance).
The DU could be located in a private data center, or it could be located at a cell-site. The CU could also be in a private data center or even hosted on a public cloud system. The DU and CU, which are typically located at different physical locations, could be tens of kilometers apart. The CU communicates with a 5G core system, which could also be hosted in the same public cloud system (or could be hosted by a different cloud provider). A RU (Radio Unit) (shown as O-RU 803 in
The E2 nodes (CU and DU) are connected to the near-real-time RIC 132 using the E2 interface. The E2 interface is used to send data (e.g., user, cell, slice KPMs) from the RAN, and deploy control actions and policies to the RAN at near-real-time RIC 132. The application or service at the near-real-time RIC 132 that deploys the control actions and policies to the RAN are called xApps. The near-real-time RIC 132 is connected to the non-real-time RIC 133 (which is shown as part of Service Management and Orchestration (SMO) Framework 805 in
In this section, PDU sessions, DRBs, and quality of service (QOS) flows will be discussed. In 5G networks, PDU connectivity service is a service that provides exchange of PDUs between a UE and a data network identified by a Data Network Name (DNN). The PDU Connecitivity service is supported via PDU sessions that are established upon request from the UE. The DNN defines the interface to a specific external data network. One or more QoS flows can be supported in a PDU session. All the packets belonging to a specific QoS flow have the same 5QI (5G QOS Identifier). A PDU session consists of the following: Data Radio Bearer which is between UE and CU in RAN; and an NG-U GTP tunnel which is between CU and UPF (User Plane Function) in the core network.
The following should be noted for 3GPP 5G network architecture, which is illustrated in
2) The transport connection between the DU 305 and the CU-UP 304b uses a single GTP-U tunnel per DRB.
3) SDAP:
4) User plane (UP) protocol includes a field to identify the QoS flow and is present between CU and UPF 903 (in the core network).
5) One (logical) RLC queue is implemented per DRB (or per logical channel), as shown in
In this section, standardized 5QI to QoS characteristics mapping will be discussed. As per 3GPP TS 23.501, the one-to-one mapping of standardized 5QI values to 5G QoS characteristics is specified in Table 1 shown below. The first column represents the 5QI value. The second column lists the different resource types, i.e., as one of Non-GBR, GBR, Delay-critical GBR. The third column (“Default Priority Level”) represents the priority level Priority5QI, for which lower the value the higher the priority of the corresponding QoS flow. The fourth column represents the Packet Delay Budget (PDB), which defines an upper bound for the time that a packet may be delayed between the UE and the N6 termination point at the UPF. The fifth column represents the Packet Error Rate (PER). The sixth column represents the maximimum data burst volume for delay-critical GBR types. The seventh column represents averaging window for GBR, delay critical GBR types.
For example, as shown in Table 1, 5QI value 1 is of resource type GBR with the default priority value of 20, PDB of 100 ms, PER of 0.01, and averaging window of 2000 ms. Conversational voice falls under this catogery. Similarly, as shown in Table 1, 5QI value 7 is of resource type Non-GBR with the default priority value of 70, PDB of 100 ms and PER of 0.001. Voice, video (live streaming), and interactive gaming fall under this catogery.
In this section, Radio Resource Management (RRM) will be discussed (a block diagram for an example RRM with a MAC Scheduler is shown in
In the above expression, the parameters are defined as follows:
a) P5QI is the priority metric corresponding to the QoS class (5QI) of the logical channel. Incoming traffic from a DRB is mapped to Logical Channel (LC) at RLC level. P5QI is a function of the default 5QI priority value, Priority5QI, of a QoS flow that is mapped to the current LC. The lower the value of Priority5QI the higher the priority of the corresponding QoS flow. For example, Voice over New Radio (VoNR) (with 5Q1 of 1) will have a much higher P5QI compared to web browsing (with 5QI of 9).
b) PGBR is the priority metric corresponding to the target bit rate of the corresponding logical channel. The GBR metric PGBR represents the fraction of data that must be delivered to the UE within the time left in the current averaging window Tavg_win (as per 5QI table, default is 2000 msec.) to meet the UE's GBR requirement. PGBR is calculated as follows:
P
GBR
=remData/targetData
where
c) PPDBDU is the priority metric corresponding to the packet delay budget at DU (i.e., PDBDU). PPDBDU=1 if PDBDU<=QDelayRLC and PPDBDU=1/(PPDBDU−QDelayRLC) if PDBDU>QDelayRLC where both PDBDU (Packet Delay Budget at DU) and RLC Queuing delay, QDelayRLC, are measured in terms of slots. QDelayRLC=(t−TRLC) is the delay of the oldest RLC packet in the QoS flow that has not been scheduled yet, and it is calculated as the difference in time between the service data unit (SDU) insertion in RLC queue to current time where t:=current time instant, TRLC:=time instant when oldest SDU was inserted in RLC.
d) PPF is the priority metric corresponding to proportional fair (PF) metric of the UE. PPF is the classical PF Metric, calculated on a per UE basis as PPF=r/Ravg,
where
e) In addition, the following weights are defined: 1) W5QI is the weight of P5QI; 2) WGBR is the weight of PGBR; 3) WPDB is the weight of PPDB; and 4) WPF is the weight of PPF. Each of the above weights is set to a value between 0 and 1.
In another example variant, the scheduling priority of a UE is determined as follows: PUE=W5QI*P5QI+maximum (WGBR*PGBR+WPDB*PPDBDU)+WPF*PPF.
In yet another example variant, the scheduling priority of a UE is determined as follows: PUE=(W5QI*P5QI+WPF*PPF)*maximum (WGBR*PGBR+WPDB*PPDBDU).
The scheduling priority of a UE is based on the maximum logical channel priority value across the logical channels (LCs) of the UE, and the resources allocated to a UE are based on this maximum logical channel priority.
In this section, flow control between the CU-UP and the DU will be discussed. The F1-U interface, for which the associated procedures and functionality are defined in 3GPP TS 38.425, supports NR User Plane (NR UP) protocol which provides support for flow control and reliability between CU-UP and DU for each DRB. Three types of PDU are supported by NR-UP:
1) Downlink User Data (DUD) PDU from CU-UP to DU (see, e.g.,
2) Downlink Data Delivery Status (DDDS) PDU from DU to CU-UP (as shown, e.g., in
3) Transfer of (Radio) Assistance Information (TAI) PDU from DU to CU-UP (see, e.g.,
In this section, non-stand-alone (NSA) architecture will be discussed in conjunction with
i) 4G eNB as well as 5G gNB use 4G EPC 1801, and 5G Core is not used.
ii) In the architecture shown in
In a variant of the NSA architecture, DL data from 4G EPC is first sent to 4G CU-UP, where the DL data is split across two transmission paths (or network legs): 1) 4G CU-UP to 4G DU, then to NSA UE; and 2) 4G CU-UP to 5G DU, then to NSA UE, and the DL data of the two transmission paths are combined at the NSA UE.
DU and CU can be overloaded in stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures, and this can result in undesirable system behavior. For example, the MAC scheduler at the DU needs to compute scheduling metric for active users and select certain number of active users to serve every TTI (Transmission Time Interval) for the downlink as well as the uplink direction for each cell in a DU. Each TTI is of short duration such as 1 ms for 15 KHz SCS (Sub-Carrier Spacing) 5G systems, 0.5 ms for 30 KHz SCS and it keeps reducing as SCS increases to higher values. As the number of Data Radio Bearers (DRBs) increases in the system, the processing overhead also goes up, and it is not always feasible to do resource allocation using typical methods (such as the above enhanced proportional fair way of resource allocation) for a given base station system. If DU L2 (Layer 2) is not able to do its required processing every TTI, it results in an overload condition, called TTI-overstretch, and this can result in undesirable system behavior.
Similarly, CU-UP can be overloaded in the O-RAN architecture. For example, some of the PDCP queues can become full and start dropping packets. This can also result in degradation of end-to-end throughput for applications running over TCP type of protocols.
Accordingly, there is a need for an enhanced system and method to effectively control overload situations in the O-RAN networks.
Accordingly, what is desired is a system and method to effectively control overload situations in the O-RAN networks.
According to an example embodiment, DL and UL overload control optimizations are implemented at the DU for stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures.
According to an example embodiment, instead of computing the scheduling metric for every active UE (or logical channel) in each slot (or TTI), i) scheduling metric is computed for a subset of UEs in each slot, and ii) at the same time, scheduling metric is computed for each active UE once in a window of length W slots.
According to an example embodiment, the DU implements a DL overload control action including a change in at least one of a desired buffer size (DBS) and a desired data rate (DDR) which is sent as part of DL Data Delivery Status (DDDS) message from the DU to a centralized unit user plane (CU-UP) for each selected DRB.
According to an example embodiment, DL and UL overload control optimizations for stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures are implemented at the DU based on the overload analysis performed at a near-RT RIC.
According to an example embodiment, various network parameters including processing load, number of active UEs, PRB utilization, etc., are communicated from the DU to the near-RT RIC, which analyzes the received parameters to detect an overload condition.
According to an example embodiment, the overload control action includes computing scheduling metrics for only a subset of UEs among active UEs in each TTI within a time window of W number of TTIs, wherein W>1, and wherein the scheduling metric for each active UE is computed only once within the time window of W number of TTIs.
According to an example embodiment, DL and UL overload control optimizations for stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures are implemented at the CU-UP based on the overload analysis performed at a near-RT RIC.
According to an example embodiment, various network parameters including processing load, number of active UEs, PRB utilization, etc., are communicated from the CU-UP to the near-RT RIC, which analyzes the received parameters to detect an overload condition.
According to an example embodiment, the overload control action includes instructing, by the near-RT RIC to the DU, to one of i) dynamically allocate more buffer space to the corresponding RLC queues in the DU, or ii) adjust scheduler weights to give higher preference to the corresponding DRBs.
According to an example embodiment, DL and UL overload control optimizations are implemented at the CU-UP for stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures.
According to an example embodiment, the overload control action includes instructing, by the CU-UP to the DU, to one of i) dynamically allocate more buffer space to the corresponding RLC queues in the DU, or ii) adjust scheduler weights to give higher preference to the corresponding DRBs.
According to an example method, overload control optimizations are implemented for non-stand-alone (NSA) architectures using the DU.
According to an example method, overload control optimizations are implemented for non-stand-alone (NSA) architectures at the DU based on the overload analysis performed at a near-RT RIC.
According to an example embodiment, in the case of the NSA architecture wireless network, the overload control action includes adjusting relative data transmission rates between a 4G leg of data transmission and a 5G leg of data transmission to a served UE. For this application the following terms and definitions shall apply:
The term “network” as used herein includes both networks and internetworks of all kinds, including the Internet, and is not limited to any particular type of network or inter-network.
The terms “first” and “second” are used to distinguish one element, set, data, object or thing from another, and are not used to designate relative position or arrangement in time.
The terms “coupled”, “coupled to”, “coupled with”, “connected”, “connected to”, and “connected with” as used herein each mean a relationship between or among two or more devices, apparatus, files, programs, applications, media, components, networks, systems, subsystems, and/or means, constituting any one or more of (a) a connection, whether direct or through one or more other devices, apparatus, files, programs, applications, media, components, networks, systems, subsystems, or means, (b) a communications relationship, whether direct or through one or more other devices, apparatus, files, programs, applications, media, components, networks, systems, subsystems, or means, and/or (c) a functional relationship in which the operation of any one or more devices, apparatus, files, programs, applications, media, components, networks, systems, subsystems, or means depends, in whole or in part, on the operation of any one or more others thereof.
The above-described and other features and advantages of the present disclosure will be appreciated and understood by those skilled in the art from the following detailed description, drawings, and appended claims.
According to an example embodiment of the method according to the present disclosure, DL and UL overload control optimizations are implemented at the DU for stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures. An L2 scheduler needs to compute the scheduling metric and select certain number of active users to serve every Transmission Time Interval (TTI) for the downlink as well as the uplink direction for each cell in a DU. As the number of Data Radio Bearers (DRBs) or the associated Logical Channels (LCs) increases in a given base station system, the processing overhead increases (e.g., reaches an overload condition) and it is not always feasible to do resource allocation using conventional methods for the given base station system.
According to an example embodiment, in the case of an overload condition (or phase), instead of computing the scheduling metric for every active UE (or logical channel) in each slot (or TTI), i) scheduling metric is computed for a subset of UEs in each slot, and ii) at the same time, scheduling metric is computed for each active UE once in a window of length having W slots. For example, if we have two hundred active UEs, we can use a window length of four slots and compute the scheduling metric for fifty active UEs in every slot. This subset of users for each slot is selected using selected policies. In one such policy, we consider UEs running delay-sensitive applications (e.g., video conferencing) and divide these UEs substantially equally in each slot in the above-described window of W slots, W>1 (e.g., 4 in the example), for the purpose of updating scheduling metric during overload conditions. In the slot (TTI) m of this window, we can update the scheduling metric for those delay-sensitive applications whose RLC packets have been in the DU's RLC queues for longer (or at least equal) duration compared to the UEs in the subsequent slot (m+1). In this manner, in addition to reducing the number of UEs for which we compute the scheduling metric in every slot, we also reduce the number of UEs that are selected to schedule in each slot, thereby helping to reduce the processing overhead during the overload phase.
In the above example of a window having a length W of 4 slots and 200 active UEs, let's assume the MAC scheduler serves the UEs in the following manner: UE u5 is served in slot 1 of the window; two UEs, u65 and u67, are served in slot 2; three UEs, u129, u130 and u131, are served in slot 3; and UE u178 is served in slot 4. Each active UE will be served once or not served at all in each window having a length of W slots, e.g., as u5 was served in the first slot of the window, u5 can't be served in the slots 2,3 and 4 of the window, but u5 can be served in the next window if selected by the scheduler.
In addition to the already active UEs, new UEs may also become active during the above window of length W. If the window of length W is smaller than a specified minimum threshold, W_min_thresh, we start computing the scheduling metric of these newly added UEs in the next window, but otherwise, the following can be performed:
i) As new UEs becomes active during a given window, the workload of servicing these newly added active UEs is uniformly distributed in the remaining slots in that given window, if the number of newly added active UEs is below a specified threshold. In this case, the scheduling metric of a newly added UE is computed in one of the remaining slots in that window, and the scheduler subsequently decides whether or not to select this UE for scheduling in that particular slot.
ii) In an alternative technique, a specified metric, M, is continually updated until a specified threshold, M_max, is reached. The metric M is equal to the sum of the products of (number of newly added active UEs in a TTI slot in a given window of length W) and (remaining TTI slots in that window) for all slots in the given window up to the particular slot for which M is being calculated. As an example, let's assume a window length W=4 slots and nk number of new active UEs are added in the kth slot of this window (with k=1,2,3,4). In this case, the metric M is equal to (n1)·(3) in the first slot; (n1)·(3)+(n2)(2) in the second slot; (n1)·(3)+(n2)·(2)+(n3)·(1) in the third slot; and (n1)·(3)+(n2)·(2)+(n3)·(1)+(n4)·(0) in the fourth slot. The number of newly added active UEs in this window of length W=4 is n1+n2+n3+n4. We keep computing the scheduling metric of newly added active UEs as long as the value of the metric M stays below the specified threshold, M_max. For example, we compute the scheduling metric of n1 newly added active UEs in the remaining slots if (n1)·(3)<M_max. If (n1)·(3)+(n2)·(2) is greater than M_max but (n1)·(3)<M_max, we compute the scheduling metric of n1 newly added active UEs in that window and not for n2 UEs which became active in the second slot in that window.
In some scenarios, UEs can send a large number of small size packets at a high packets-per-second rate, and it may not be always feasible for DU to process these packets, thereby necessitating UL overload control optimization, e.g., at the DU. In the present example embodiment, the CPU load for DU processors for handling layer 2 (and related) mechanisms is monitored to determine whether the CPU load exceeds a specified threshold, in which case i) the number of UEs for which UL grant is given is reduced, and ii) the size of the uplink grant given to each UE is reduced.
According to an alternative example embodiment, the number of UL packets received at the DU from each active UE in every slot is counted to determine whether this aggregate number (i.e., the sum across all UEs which are sending UL data to this DU) exceeds a specified threshold and stays above the threshold for a specified period of time, in which case i) the number of UEs for which UL grant is given is reduced, and ii) the size of the uplink grant given to each UE is reduced.
In some scenarios, application servers can keep transmitting a large number of small-size packets at a very high packet sending rate. At the User Plane Function (UPF) in the core network, traffic policing (or shaping) can be performed, but that may be happening on the basis of data rate (e.g., in Mbps or Gbps) but not in terms of number of packets per second. If these large number of small size packets reach the DU at a very high input packet rate, the DU may not have the hardware and/or software resources to process these packets, thereby necessitating DL overload control optimization, e.g., at the DU. In the present example embodiment, the processing load (or DL packet rate in terms of packets per second) at DU can be monitored to determine whether the processing load (or DL packet rate) exceeds a specified threshold, in which case the Desired Buffer Size (DBS) and/or Desired Data Rate (DDR) which is sent as part of DDDS (DL Data Delivery Status) message from DU to CU-UP for each selected DRB is changed. In one variant, a reduced value of DBS is sent (as part of DDDS message) for selected DRBs from DU to CU-UP. In another variant, a reduced Desired Data Rate which is sent from DU to CU-UP as part of DDDS for selected DRBs is implemented. These reductions help to keep the packets at CU-UP somewhat longer (or to drop some such excess packets at CU-UP if needed), thereby helping to manage the overload situation at DU.
According to an example embodiment, DL and UL overload control optimizations for stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures are implemented at the DU based on the overload analysis performed at a near-RT RIC. In this example embodiment, the near-RT RIC subscribes to the following parameters (e.g., performance measures) from the DU (as explained below in further detail in connection with
1) DU processing load for downlink (DL) traffic (considering multiple processors available in the DU server).
2) DU processing load for uplink (UL) traffic (considering multiple processors available in the DU server).
3) Number of active users.
4) DL PRB utilization.
5) UL PRB utilization.
6) Percentage of DRBs associated with each 5QI.
7) Channel condition measures for UEs in each cell (supported by that DU).
8) Number of DL packets (RLC SDUs) input to DU from CU-UP.
9) Number of UL packets input to DU from all the UEs supported by that DU.
10) Activity factor for each DRB.
11) Statistics related to QoS performance of different DRBs from the DU, e.g., waiting time in RLC queues along with 5QI of that DRB, throughput violations for GBR (Guaranteed Bit Rate) DRBs, and packet error rate.
The above-listed parameters are measured or computed every measurement interval T and communicated from the E2 Node 131 (DU in this example) to the near-RT-RIC 132 (e.g., as shown in
As shown in
We now turn to
As an example of an action to handle the overload control gracefully, the near-RT RIC 132 can ask the DU 131b to start computing the scheduling metric for only a subset of active UEs in every slot (TTI). In this case, the near-RT RIC 132 can also provide the transmission window length and the subset of active UEs for which scheduling metric needs to be computed in a given TTI within the transmission window. Alternatively, the near-RT RIC 132 can provide one or more policies to the DU 131b to enable the DU to identify a subset of UEs for which the scheduling metric needs to be computed in a given TTI within the transmission window. Along with this, the near-RT RIC 132 can also consider the 5QIs of DRBs that are being supported by active UEs in each cell, and the near-RT RIC 132 can also ask the DU 131b to reduce the amount of resources that the DU 131b provides to each cell given the 5QI DRB mix for that cell (e.g., the number of UEs of different 5QI combinations which can be scheduled per slot, or the amount of PRBs which can be allocated in each cell given the 5QI mix of UEs in that cell). Suitable objects are added in the E2 protocol for this purpose. Finally, as shown by the process arrow 1405, when the overload condition no longer exists, the near-RT RIC 132 instructs the DU 131b to switch over to the normal mode of operation.
According to an example embodiment, DL and UL overload control optimizations for stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures are implemented at the CU-UP based on the overload analysis performed at a near-RT RIC. In this example embodiment, the near-RT RIC subscribes to the following parameters (e.g., performance measures) from a particular CU-UP (as explained below in further detail in connection with
1) The number of DRBs for which the particular CU-UP is carrying traffic (for a given DU).
2) DL and UL processing load at the particular CU-UP (across multiple processors available in the CU-UP server).
3) PDCP service data unit (SDU) or PDU drops in the PDCP queues at the CU-UP due to non-availability of buffer space in the corresponding RLC queues in the DU (as reported by the DU to CU-UP, as part of flow control feedback message DDDS).
4) Identity of DRBs for which PDCP SDUs or PDUs are being dropped at the CU-UP due to non-availability of space in CU-UP queues.
5) Available buffer space in PDCP queues at the CU-UP.
The above-listed parameters are measured or computed every measurement interval T and communicated from the E2 Node 231 (CU-UP in this example) to the near-RT-RIC 132 (e.g., as shown in
As shown in
We now turn to
As an example of an action to handle the overload control gracefully (e.g., as shown by the process arrow 2404), the near-RT RIC 132 can identify the DRBs which are dropping packets in the PDCP queues at the CU-UP 131a during overload conditions (e.g., when available buffer space at the CU-UP is below a specified threshold). Under normal conditions, the CU-UP can potentially allocate a higher amount of buffer space to such DRBs to reduce the packet drops from the PDCP queues. If the CU-UP 131a is overloaded, it may not be possible to allocate more buffer space to corresponding PDCP queues at the CU-UP 131a. If the near-RT-RIC 132 detects that CU-UP 131a is overloaded and the DU 131b is not overloaded, the near-RT RIC 132 sends message(s) to the corresponding DU 131b to dynamically allocate more buffer space to the corresponding RLC queues or adjust MAC scheduler weights to give higher preference to the corresponding DRBs (which actions are referenced by 2405; process arrow 2406 shows DU 131b sending to the CU-UP 131a a higher desired buffer size (DBS) as part of DDDS feedback for select DRBs), thereby helping to create space in the RLC queues for such DRBs, which in turn helps the CU-UP in sending packets faster from the PDCP queues to the RLC queues at the DU for such DRBs (leading to PDCP queues clearing up, as shown at 2407a, and a higher rate of DL data for select DRBs, as shown by the process arrow 2407b). Suitable objects are added in the E2 protocol for this purpose.
According to an alternative example embodiment, DL and UL overload control optimizations for stand-alone (SA) 4G/5G and non-stand-alone (NSA) architectures are implemented at the CU-UP without the assistance of a near-RT RIC. This alternative example embodiment is substantially similar to the above-described embodiment utilizing the near-RT RIC in conjunction with the CU-UP, except for the following differences:
1) The overload-control-related processing occurs at the CU-UP, including the detection of an overload condition (e.g., based on at least available buffer space in PDCP queues at the CU-UP, the processing overload at CU-UP, etc., as listed above).
2) The CU-UP communicates the overload-control-related actions, e.g., asking the DU to provide more buffer space to the corresponding RLC queues in the DU, or adjust scheduler weights to give higher preference to these DRBs during the CU-UP overload phase. These overload-control-related actions can be communicated by using reserved bits in the downlink NR-U data plane PDU which is sent from the CU-UP to the DU (as per 3GPP TS38.425).
According to an example embodiment of a method, overload control optimization for NSA architecture is provided. As previously discussed in connection with
When a DU gets overloaded, the enhanced DDDS message with the “DU overload extension” field is used to communicate a DU overload situation to the 5G CU-UP 1802a. As an example, if a DU (e.g., 5G DU 1805 or 4G DU 1803) gets overloaded (e.g., as indicated by the DU's CPU utilization exceeding a specified threshold), this overload situation can be indicated (e.g., in a DDDS with DU overload extension 2801a from 5G DU 1805, or in a DDDS with DU overload extension 2801b from 4G DU 1803 in
In an alternative embodiment, a combination of the above-described techniques can be implemented. For example, an initial phase of overload control optimization can utilize the above-described “DU overload extension” technique, and if the overload condition persists, a subsequent phase of overload control optimization can additionally utilize the previously described overload control technique implemented by the DU. In yet another example, an initial phase of overload control optimization can utilize the previously described overload control technique implemented by the DU, and if the overload condition persists, a subsequent phase of overload control optimization can additionally utilize the above-described “DU overload extension” technique. A particular combination (e.g., one of the above) can be chosen depending on the network operator policies.
In another alternative embodiment, the above-described “DU overload extension” technique can be distributed across DUs, CU and RIC. In this example embodiment, 5G and 4G DUs communicate various parameters to a near-RT-RIC, e.g., parameters which help the near-RT RIC determine whether or not a DU is getting overloaded, and activity factor for various DRBs. At the same time, the 5G CU-UP continues communicating packet-splitting-related parameters to the near-RT-RIC. Once an overload condition for the 5G DU is detected by the near-RT RIC, the near-RT RIC changes the splitting ratio for some DRBs (e.g., DRBs with high activity factor) and informs the CU-UP of the change in the splitting ratio. In response, the CU-UP can reduce the data rate across the 5G leg for some DRBs to decrease their activity factor for the 5G DU, while the data rate across the 4G leg for these DRBs can be increased. A similar solution can be implemented when an overload condition for the 4G DU is detected, i.e., the data rate across the 4G leg is reduced, while the data rate across the 5G leg can be increased.
While the present disclosure has been described with reference to one or more exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from the scope of the present disclosure. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the disclosure without departing from the scope thereof. Therefore, it is intended that the present disclosure not be limited to the particular embodiment(s) disclosed as the best mode contemplated, but that the disclosure will include all embodiments falling within the scope of the appended claims.
For the sake of completeness, a list of abbreviations used in the present specification is provided below:
Number | Date | Country | Kind |
---|---|---|---|
202321017243 | Mar 2023 | IN | national |