RADIO LINK FAILURE RECOVERY WITH SRB3 IN MR-DC

Information

  • Patent Application
  • 20210014923
  • Publication Number
    20210014923
  • Date Filed
    September 30, 2020
    4 years ago
  • Date Published
    January 14, 2021
    3 years ago
Abstract
Embodiments herein provide a transfer framework and procedure for a failure indication and subsequent recover message for a user equipment (UE) in multi-radio access technology (MR)-dual connectivity (DC). For example, in embodiments, a UE may send an indication of a radio link failure (RLF) on a master node (MN) to a secondary node (SN) via a signalling radio bearer type 3 (SRB3). The SN may forward the indication to the MN (e.g., in a transparent manner). In some embodiments, the indication may be formatted in the same format as messages on signalling radio bearer type 1 (SRB1). Other embodiments may be described and claimed.
Description
FIELD

Embodiments relate generally to the technical field of wireless communications.


BACKGROUND

A user equipment (UE) configured with multi-radio access technology (MR)-dual connectivity (DC) has a group of cells that belong to a Master cell group (MCG) that are connected to the Master Node (MN) and another group of cells that belong to a Secondary Cell Group (SCG) connected to a Secondary Node (SN). When there is a failure of the radio link to the Primary cell (Pcell) of the MCG, normally the UE does a re-establishment procedure that is disruptive and can result in loss of data.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments will be readily understood by the following detailed description in conjunction with the accompanying drawings. To facilitate this description, like reference numerals designate like structural elements. Embodiments are illustrated by way of example and not by way of limitation in the figures of the accompanying drawings.



FIG. 1 illustrates a network environment including a user equipment (UE), a master node (MN), and a secondary node (SN), in accordance with various embodiments.



FIG. 2 illustrates an example procedure in accordance with various embodiments.



FIG. 3 illustrates a process of a UE in accordance with various embodiments.



FIG. 4 illustrates a process of a SN in accordance with various embodiments.



FIG. 5 illustrates an example architecture of a system of a network, in accordance with various embodiments.



FIG. 6 illustrates an example next generation radio access network (NG-RAN) in accordance with various embodiments.



FIG. 7 illustrates an example architecture of a system including a first core network, in accordance with various embodiments.



FIGS. 8A and 8B illustrate example architectures of a system including a second core network, in accordance with various embodiments.



FIG. 9 illustrates an example of infrastructure equipment in accordance with various embodiments.



FIG. 10 illustrates an example of a computer platform in accordance with various embodiments.



FIG. 11 illustrates example components of baseband circuitry and radio front end modules in accordance with various embodiments.



FIG. 12 is a block diagram illustrating components, according to some example embodiments, able to read instructions from a machine-readable or computer-readable medium (e.g., a non-transitory machine-readable storage medium) and perform any one or more of the methodologies discussed herein.





DETAILED DESCRIPTION

The following detailed description refers to the accompanying drawings. The same reference numbers may be used in different drawings to identify the same or similar elements. In the following description, for purposes of explanation and not limitation, specific details are set forth such as particular structures, architectures, interfaces, techniques, etc. in order to provide a thorough understanding of the various aspects of various embodiments. However, it will be apparent to those skilled in the art having the benefit of the present disclosure that the various aspects of the various embodiments may be practiced in other examples that depart from these specific details. In certain instances, descriptions of well-known devices, circuits, and methods are omitted so as not to obscure the description of the various embodiments with unnecessary detail. For the purposes of the present document, the phrase “A or B” means (A), (B), or (A and B).


In accordance with various embodiments, when there is a failure of the radio interface towards a primary cell in a multi-radio access technology (MR)-dual connectivity (DC) configuration, a UE can indicate the failure through a secondary group. Currently, there is no specific mechanism for sending the failure message and subsequent recovery handling.


Embodiments herein provide a transfer framework and procedure for the failure indication and subsequent recover message that can re-use much of the existing systems and implementation of the UE handling of these messages. For example, in embodiments, a user equipment (UE) may send an indication of a radio link failure (RLF) on a master node (MN) to a secondary node (SN) via a signalling radio bearer type 3 (SRB3). The SN may forward the indication to the MN (e.g., in a transparent manner). In some embodiments, the indication may be formatted in the same format as messages on signalling radio bearer type 1 (SRB1).


By maximizing the re-use of existing systems and implementation, the embodiments herein introduce less complexity than other solutions.


A UE (e.g., UE 501 of FIG. 5) configured with MR-DC has a group of cells that belong to a Master cell group (MCG) that are connected to the Master Node (MN) and another group of cells that belong to a Secondary Cell Group (SCG) connected to a Secondary Node (SN). When the UE is configured with MR-DC, the RRC signalling link between the UE and the SN is called SRB3.


When there is a failure of the radio link to the Primary cell (Pcell) of the MCG, normally the UE does a re-establishment procedure that is disruptive and can result in loss of data. The UE in MR-DC can be configured with a split SRB connecting the UE and the MN. A split SRB has two paths between the MN and UE—one path over the MN and other path over the SN. These paths are L2 based and the radio resource control (RRC) is still between the MN and UE. With a split SRB, when there is an RLF towards the Pcell, the UE and MN can still exchange the failure indication and recovery messages over the split SRB leg that is through the SN.


Another configuration is where the UE is configured with SRB3. When the link to the SN is still available, the UE can instead indicate the failure of the Pcell in a message sent to the SN over SRB3. The network can then reconfigure the UE to recover the radio link to the PCell, for example using a Handover (HO) procedure to another PCell.


It has already been agreed to support MN failure indication and recovery procedures over split SRB1. When split SRB is not used or a direct path to MCG RRC is not available from the UE, this may be extended to support similar failure indication and recovery procedure over SRB3.


In general, it is beneficial to maximise re-use of the procedures between split SRB1 and SRB3 as this minmizes additional implementation effort. Various embodiments attempt to make the split SRB1 procedures directly applicable also over SRB3. One embodiment includes a transparent model where SRB3 simply acts as a transport mechanism for the messages defined over SRB1, with the corresponding transparent transfer over X2/XN. An example of this transparent model is shown by FIG. 1.


Aspects of the transparent model with respect to MN failure handling over SRB3 include (1) transfer of the failure indication; (2) MN originated RRC message—contents and transfer; and (3) subsequent handling MN RRC message in the UE—transfer of “response” message, failure handling etc. Each of these aspects are discussed in more detail below.


Failure Indication


To maximise re-use of existing messages and procedures, the same failure message and contents defined for split SRB1 may be re-used over SRB3.


In some embodiments, the failure indication is carried transparently over SRB3 and Xn/X2 to MCG. ULInformationTransferMRDC is well suited for this use over SRB3. In these embodiments, the ULInformationTransferMRDC is used for the transfer of the same failure message over SRB3.


In some embodiments, to simplify and minimise further message interactions, no additional UL message is sent over SRB3 or SRB1 subsequent to the failure indication. In these embodiments, the transfer of UL messages over SRB3 and SRB1 should be suspended after transfer of the failure indication using ULInformationTransferMRDC over SRB3.


MN Originated RRC Recovery Procedure


On receipt of the failure indication over SN, MN initiates recovery procedure. The following MN originated RRC messages may be used: a reconfiguration with sync or a release message. That is, the MN must perform a HO or release the connection.


Since there is no direct communication path between the MN and the UE, these RRC messages also need to be transported over the SN. HO messages can include SCG configuration in an encapsulated SN originated RRC message. This means that there are at least two possible options: Option 1 involves the SN encapsulating the MN originated RRC message in the SN originated RRC message along with the SN configuration and send to the UE; and option 2 involves the SN configuration being sent to the MN and encapsulated in the MN originated RRC message as today and then further encapsulated in an SN RRC message to be sent to the UE over SRB3.


While option 1 is more efficient and faster, it will introduce another RRC architecture model and corresponding handling where MN originated RRC messages is encapsulated by SN RRC message along with SN configuration. To minimise additional complexity, and as discussed above, to minimise implementation changes by re-using existing messages and procedures, embodiments herein adopt option 2. In these embodiments, MN RRC messages, which may include SN originated RRC messages, are encapsulated in an SN RRC message and transparently transported over SRB3. See e.g., FIG. 2.


In these embodiments, any SN configuration information can be already included in the SN originated RRC message encapsulated in the MN originated RRC message. Additionally, in some embodiments no additional SN originated information is included in the SN originated encapsulating RRC message sent over SRB3. Such embodiments minimise additional complexity of UE processing as discussed further below.


UE Processing the RRC Message Received Over SRB3


As discussed above, the encapsulating SN originated RRC message does not contain any SCG configuration. It is simply a transport of the MN RRC message. Therefore, it is not necessary and it is simpler not to have a joint success failure for the SN originated encapsulating message. In embodiments, on successful receipt and processing of the SN originated encapsulating message, the UE “SN RRC” simply “transfers” the encapsulated MN originated RRC message to the “MN RRC” for processing. The procedure for the SN originated encapsulating message completes with this UE internal “transfer” of the encapsulated MN RRC message. In various embodiments, the SN originated encapsulating message is handled transparently by the “SN RRC” in the UE and the MN RRC message is transferred to the “MN RRC” in the UE. There is no joint success/failure of the message. That is, the SRB3 transfer message can succeed independent of the contained MN RRC message.


In embodiments, on receipt of the encapsulated MN RRC message, the “MN RRC” processes this message as though it was received over SRB1. That is, the procedural section relevant for the corresponding RRC message is re-used as is for the processing of this MN RRC messages irrespective of whether it was received over SRB1 or SRB3. This is also applicable for the failure handling behavior and generation and transfer of the “complete” messages at the UE.


One special handling needed for the aforementioned embodiments is to resume the suspended transmission over SRB1 is resumed as part of the processing of the received MN RRC message. Another consequence of the aforementioned embodiments is that the UE processes the MN RRC message and the encapsulating SN RRC message independently. This implies, for example, if as a consequence of processing the RRC Reconfiguration message, the SCG/SRB3 is released, there is no guarantee that the radio link control (RLC) acknowledgement (ACK) for the encapsulating SN RRC message will be sent to SN.


As with such failure scenarios, also as discussed previously in different contexts, the MN may not be sure which was the last received MN RRC message at the UE on receipt of the failure indication. Different solutions such as indicating the transaction id of the last received DL message in the failure indication, have been discussed before for this. However, as in the previous scenarios, it can be left to network implementation, using Full configuration for example, to handling any abnormalities that may arise.


MR-DC Aspects

MR-DC involves a multiple Rx/Tx UE (e.g., UE 501, UE 701, and/or UE 801 of FIGS. 5, 7, and 8A,8B, respectively) configured to utilize radio resources provided by two distinct schedulers in two different nodes (e.g., RAN nodes 511 in FIG. 5 infra) connected via a non-ideal backhaul, where one of these nodes provides E-UTRA access and the other node provides NR access. One scheduler is located in a Master Node (MN) and the other scheduler is located in a Secondary Node (SN). The MN and SN are connected via a network interface and at least the MN is connected to the core network (e.g., CN 520, CN720, and/or CN 820 of FIGS. 5, 7, and 8A,8B infra). MR-DC includes different types of dual connectivity scenarios, such as EN-DC (E-UTRA-NR Dual Connectivity), NGEN-DC (NG-RAN E-UTRA-NR Dual Connectivity), and NE-DC (NR-E-UTRA Dual Connectivity).


In EN-DC, the UE is connected to one eNB that acts as an MN and one en-gNB that acts as an SN. The eNB is connected to an EPC (see e.g., CN 720 of FIG. 7) and the en-gNB is connected to the eNB via an X2 interface. The en-gNB is a node that provides NR user plane and control plane protocol terminations towards the UE, and acts as the SN in EN-DC. In EN-DC, the involved core network entity is the MME (e.g., MME 721 of FIG. 7), wherein the S1-MME (or S1 control plane) interface is terminated in or at the MN, and the MN and the SN are interconnected via an X2 control place interface (X2-C).


In NR-EN, the UE is connected to one gNB that acts as the MN and one ng-eNB that acts as the SN. The gNB is connected to a 5GC (see e.g., CN 820 of FIG. 8) and the ng-eNB (Master Node eNB) is connected to the gNB via an Xn interface. In NE-DC, the UE is connected to one gNB that acts as an MN and one ng-eNB that acts as a SN, wherein the gNB is connected to 5GC and the ng-eNB is connected to the gNB via the Xn interface. In MR-DC with 5GC (NGEN-DC and/or NE-DC), the involved core network entity is the AMF (e.g., AMF 821 of FIG. 8), wherein the NG control place (NG-C) interface is terminated in or at the MN, and the MN and the SN are interconnected via an Xn control place interface (Xn-C).


The UE considers itself to be in EN-DC, if and only if it is configured with nr-SecondaryCellGroupConfig according to 3GPP TS 36.331, and it is connected to EPC. The UE considers itself to be in NGEN-DC, if and only if it is configured with nr-SecondaryCellGroupConfig according to 3GPP TS 36.331, and it is connected to SGC. The UE considers itself to be in NE-DC, if and only if it is configured with mrdc-SecondaryCellGroup set to eutra-SCG. The UE considers itself to be in NR-DC, if and only if it is configured with mrdc-SecondaryCellGroup set to nr-SCG. The UE considers itself to be in MR-DC, if and only if it is in (NG)EN-DC, NE-DC or NR-DC.


When the UE is in NR-DC, the network may provide a UE configured with an SCG with an sk-Counter even when no DRB is setup using the secondary key (S-KgNB) in order to allow the configuration of SRB3.


In (NG)EN-DC and NR-DC, SRB3 can be used for measurement configuration and reporting, to (re)configure MAC, RLC, PHY, and RLF timers and constants of the SCG configuration, and to reconfigure PDCP for DRBs associated with the S-KgNB or SRB3, and to reconfigure SDAP for DRBs associated with S-KgNB in (NG)EN-DC and NR-DC, provided that the (re-)configuration does not require any MN involvement. In EN-DC, only measConfig, radioBearerConfig and/or secondaryCellGroup are included in RRCReconfiguration received via SRB3.


MR-DC Control Plane


In MR-DC, the UE has a single RRC state, based on the MN RRC and a single C-plane connection towards the CN. Each RAN node has its own RRC entity (e.g., an E-UTRA RRC entity if the node is an eNB or an NR RRC entity if the node is a gNB), which generate RRC PDUs to be sent to the UE. RRC PDUs generated by the SN can be transported via the MN to the UE. The MN always sends the initial SN RRC configuration via MCG SRB (SRB1), but subsequent reconfigurations may be transported via MN or SN. When transporting RRC PDU from the SN, the MN does not modify the UE configuration provided by the SN.


In EN-DC, SRB1 uses E-UTRA PDCP at initial connection establishment. After the initial connection establishment, MCG SRBs (SRB1 and SRB2) can be configured by the network to use either E-UTRA PDCP or NR PDCP. A PDCP version change (release of old PDCP and establish of new PDCP) of SRBs can be supported in either direction, from E-UTRA PDCP to NR PDCP or viceversa, via a handover procedure (reconfiguration with mobility). Alternatively, for the initial change from E-UTRA PDCP to NR PDCP, with a reconfiguration without mobility, when the network knows there is no UL data in buffer and before the initial security activation. For EN-DC capable UEs, NR PDCP can be configured for DRBs and SRBs also before EN-DC is configured.


If the SN is a gNB (e.g., in EN-DC and NGEN-DC), the UE can be configured to establish an SRB with the SN (SRB3) to enable RRC PDUs for the SN to be sent directly between the UE and the SN. RRC PDUs for the SN can only be transported directly to the UE for SN RRC reconfiguration not requiring any coordination with the MN. Measurement reporting for mobility within the SN can be done directly from the UE to the SN if SRB3 is configured.


Split SRB is supported for all MR-DC options, allowing duplication of RRC PDUs generated by the MN, via the direct path and via the SN. Split SRB uses NR PDCP. This version of the specification does not support the duplication of RRC PDUs generated by the SN via the MN and SN paths.


In EN-DC, the SCG configuration is kept in the UE during suspension. The UE releases the SCG configuration, except for the radio bearer configuration, during resumption initiation.


MR-DC User Plane


Three bearer types can be used in MR-DC from the UE perspective. These three bearers include MCG bearer, SCG bearer, and split bearer. For EN-DC, the network can configure either E-UTRA PDCP or NR PDCP for MN terminated MCG bearers while NR PDCP is always used for all other bearers. In MR-DC with 5GC (NGEN-DC, NE-DC), NR PDCP is used for all bearer types. In NGEN-DC, E-UTRA RLC/MAC is used in the MN while NR RLC/MAC is used in the SN. In NE-DC, NR RLC/MAC is used in the MN while E-UTRA RLC/MAC is used in the SN.


From the network perspective, each bearer (MCG, SCG, and split bearer) can be terminated either in MN or in SN. Even if only SCG bearers are configured for a UE, for SRB1 and SRB2 the logical channels are always configured at least in the MCG—this is still an MR-DC configuration and a Pcell always exists. If only MCG bearers are configured for a UE, there is no SCG, this is still considered an MR-DC configuration, as long as at least one of the bearers is terminated in the SN.


There are different user plane connectivity options of the MN and SN involved in MR-DC for the UE, which depends on the bearer option configured. For MN terminated bearers, the user plane connection to the CN entity is terminated in the MN. For SN terminated bearers, the user plane connection to the CN entity is terminated in the SN. For split bearers, MN terminated SCG bearers and SN terminated MCG bearers, PDCP data is transferred between the MN and the SN via the MN-SN user plane interface


The transport of user plane data over the Uu interface either involves MCG or SCG radio resources or both. For MCG bearers, only MCG radio resources are involved. For SCG bearers, only SCG radio resources are involved. For split bearers, both MCG and SCG radio resources are involved.


For EN-DC, the X2-U interface is the user plane interface between MN and SN, and S1-U is the user plane interface between the MN, the SN or both and the S-GW. For NGEN-DC and NE-DC, the Xn-U interface is the user plane interface between MN and SN, and NG-U is the user plane interface between the MN, the SN, or both and the UPF.


Layer 1 Aspects


In MR-DC, two or more Component Carriers (CCs) may be aggregated over two cell groups. The UE may simultaneously receive or transmit on multiple CCs depending on its capabilities. The maximum number of configured CCs for a UE is 32 for DL and UL. Depending on UE capabilities, up to 31 CCs can be configured for an E-UTRA cell group when the NR cell group is configured. In EN-DC, the SN may configure the same Physical Cell ID (PCI) to more than one NR cell it serves. To avoid PCI confusion for EN-DC, NR PCIs should be allocated in a way that an NR cell is uniquely identifyable by E-UTRA cell information. This E-UTRA cell is in the coverage area of an NR cell included in the EN-DC operation. In addition, NR PCIs should only be re-used in NR cells on the same SSB frequency sufficiently distant from each other. X2-C signalling supports disambiguation of NR PCIs by including the ECGI in respective X2AP EN-DC messages (e.g., SGNB ADDITION REQUEST).


Layer 2 Aspects


At the MAC layer for MR-DC, the UE is configured with two MAC entities one MAC entity for the MCG and one MAC entity for the SCG. Semi-persistent scheduling (SPS) resources can be configured on both the Pcell and the PSCell. Additionally, the buffer status report (BSR) configuration, triggering, and reporting are independently performed per cell group. For split bearers, the PDCP data is considered in BSR in the cell group(s) configured by RRC. Additionally, separate DRX configurations are provided for MCG and SCG.


At the RLC layer, both RLC AM and UM can be configured for MR-DC, for all bearer types (MCG, SCG and split bearers). At the PDCP layer for EN-DC, CA packet duplication (see [3]) is not applied to CA in the MN. In EN-DC and NGEN-DC, CA packet duplication can only be configured for SCG bearer. In NE-DC, CA packet duplication can only be configured for MCG bearer. In EN-DC, RoHC can be configured for all the bearer types. At the SDAP layer for NGEN-DC and NE-DC, the UE can be configured with two SDAP protocol entities for each individual PDU session, one for MN and another one for SN.


Cell Measurement and Cell Reselection Aspects


Radio Resource Control (RRC) includes a measurement configuration and reporting function that is used for establishment/modification/release of measurements (e.g., intra-frequency, inter-frequency, and inter-RAT measurements).


The network may configure an RRC_CONNECTED UE to perform measurements and report them in accordance with the measurement configuration. The measurement configuration is provided by means of dedicated signalling using the RRCReconfiguration. The network may configure the UE to perform NR measurements, and/or Inter-RAT measurements of E-UTRA frequencies. The network may configure the UE to report measurement information based on SS/PBCH block(s), including measurement results per SS/PBCH block; measurement results per cell based on SS/PBCH block(s); and/or SS/PBCH block(s) indexes. The network may configure the UE to report measurement information based on CSI-RS resources, including measurement results per CSI-RS resource; measurement results per cell based on CSI-RS resource(s); and/or CSI-RS resource measurement identifiers.


The measurement configuration includes the following parameters: measurement objects, reporting configurations, measurement identities, quantity configurations, and measurement gaps. The reporting configurations include lists of reporting configurations where there can be one or multiple reporting configurations per measurement object. Each reporting configuration includes a reporting criterion, reference signal (RS) type, and reporting format.


The measurement identities include list of measurement identities where each measurement identity links one measurement object with one reporting configuration. By configuring multiple measurement identities, it is possible to link more than one measurement object to the same reporting configuration, as well as to link more than one reporting configuration to the same measurement object. The measurement identity is also included in the measurement report that triggered the reporting, serving as a reference to the network.


The quantity configuration defines the measurement filtering configuration used for all event evaluation and related reporting of that measurement type. For NR measurements, the network may configure up to 2 quantity configurations with a reference in the NR measurement object to the configuration that is to be used. In each configuration, different filter coefficients can be configured for different measurement quantities, for different RS types, and for measurements per cell and per beam. Measurement gaps are periods that the UE may use to perform measurements (e.g., no (UL, DL) transmissions are scheduled).


A measurement object is a list of objects on which the UE is to perform the measurements. For intra-frequency and inter-frequency measurements, a measurement object indicates the frequency/time location and subcarrier spacing of reference signals to be measured. Associated with this measurement object, the network may configure a list of cell specific offsets, a list of ‘blacklisted’ cells and a list of ‘whitelisted’ cells. Blacklisted cells are not applicable in event evaluation or measurement reporting. Whitelisted cells are the only ones applicable in event evaluation or measurement reporting. The UE determines which MO corresponds to each serving cell frequency from the frequencyInfoDL in ServingCellConfigCommon within the serving cell configuration.


A UE in RRC_CONNECTED maintains a measurement object list, a reporting configuration list, and a measurement identities list according to signalling and procedures in this specification. The measurement object list possibly includes NR intra-frequency object(s), NR inter-frequency object(s) and inter-RAT objects. Similarly, the reporting configuration list includes NR and inter-RAT reporting configurations. Any measurement object can be linked to any reporting configuration of the same RAT type. Some reporting configurations may not be linked to a measurement object. Likewise, some measurement objects may not be linked to a reporting configuration.


The measurement procedures distinguish the following types of cells: (1) The NR serving cell(s)—these are the SpCell and one or more SCells; (2) Listed cells—these are cells listed within the measurement object(s); and (3) Detected cells—these are cells that are not listed within the measurement object(s) but are detected by the UE on the SSB frequency(ies) and subcarrier spacing(s) indicated by the measurement object(s). For NR measurement object(s), the UE measures and reports on the serving cell(s), listed cells and/or detected cells. For inter-RAT measurements object(s) of E-UTRA, the UE measures and reports on listed cells and detected cells. Whenever the procedural specification, other than contained in sub-clause 5.5.2 of 3GPP TS 38.331, refers to a field it concerns a field included in the VarMeasConfig unless explicitly stated otherwise (e.g., only the measurement configuration procedure covers the direct UE action related to the received measConfig).


In NR-DC, the UE may receive two independent measConfig: a measConfig, associated with MCG, that is included in the RRCReconfiguration message received via SRB1; and a measConfig, associated with SCG, that is included in the RRCReconfiguration message received via SRB3, or, alternatively, included within a RRCReconfiguration message embedded in a RRCReconfiguration message received via SRB1. In this case, the UE maintains two independent VarMeasConfig and VarMeasReportList, one associated with each measConfig, and independently performs all the procedures in clause 5.5 of 3GPP TS 38.331 for each measConfig and the associated VarMeasConfig and VarMeasReportList, unless explicitly stated otherwise.


RRC also includes an RRC connection control function that is used for connection mobility including, e.g., intra-frequency and inter-frequency handover, associated security handling (e.g., key/algorithm change), and specification of RRC context information transferred between network nodes. The RRC connection control function may control or instruct a network node and/or UE to perform an RRC reconfiguration procedure. The purpose of the RRC reconfiguration procedure is to modify an RRC connection to establish/modify/release radio bearers (RBs), to perform reconfiguration with synchronization (sync), to setup/modify/release measurements, to add/modify/release SCells and cell groups. As part of the RRC reconfiguration procedure, NAS dedicated information may be transferred from the Network to the UE.


The Network may initiate the RRC reconfiguration procedure to a UE in RRC_CONNECTED mode. The Network applies the procedure as follows: the establishment of RBs (other than SRB1, that is established during RRC connection establishment) is performed only when AS security has been activated; the addition of Secondary Cell Group and SCells is performed only when AS security has been activated; and the reconfigurationWithSync is included in secondaryCellGroup only when at least one DRB is setup in SCG.


MR-DC involves a multiple Rx/Tx UE configured to utilize radio resources provided by two distinct schedulers in two different nodes connected via non-ideal backhaul, one providing Evolved Universal Terrestrial Radio Access (E-UTRA) access and the other one providing NR access. One scheduler is located in a MN and the other in the Secondary Node (SN). The MN and SN are connected via a network interface and at least the MN is connected to the core network.


MR-DC includes EN-DC or NGEN-DC. In EN-DC, a UE may be connected to one eNB that acts as an MN and one gNB that acts as an SN. The eNB is connected to an EPC and the en-gNB is connected to the eNB via the X2 interface. The en-gNB is a node that provides NR user plane and control plane protocol terminations towards the UE, and acts as the SN in EN-DC. In NR-EN, a UE may be connected to one gNB that acts as the MN and one ng-eNB that acts as a SN. The gNB is connected to 5GC and the ng-eNB (Master Node eNB) is connected to the gNB via the Xn interface.


In the RRC_IDLE state, the cell reselection procedure allows the UE to select a more suitable cell and to camp on that cell. When the UE is in either Camped Normally state or Camped on Any Cell state on a cell, the UE shall attempt to detect, synchronise, and monitor intra-frequency, inter-frequency and inter-RAT cells indicated by the serving cell. For intra-frequency and inter-frequency cells the serving cell may not provide explicit neighbour list but carrier frequency information and bandwidth information only. UE measurement activity is also controlled by measurement rules defined in 3GPP TS 36.304, allowing the UE to limit its measurement activity. For idle mode cell re-selection purposes, the UE shall be capable of monitoring at least: Intra-frequency carrier, and depending on UE capability, [7] NR inter-frequency carriers, depending on UE capability, [7] FDD E-UTRA inter-RAT carriers, and depending on UE capability, [7] TDD E-UTRA inter-RAT carriers. In addition to theses requirements, a UE supporting E-UTRA measurements in RRC_IDLE state shall be capable of monitoring a total of at least [14] carrier frequency layers, which includes serving layer, comprising of any above defined combination of E-UTRA FDD, E-UTRA TDD and NR layers.


In the RRC_IDLE state, the UE measures the SS-RSRP and SS-RSRQ level of the serving cell and evaluates the cell selection criterion S for the serving cell at least every DRX cycle. The UE filters the SS-RSRP and SS-RSRQ measurements of the serving cell using at least 2 measurements. Within the set of measurements used for the filtering, at least two measurements shall be spaced by, at least [DRX cycle/2]. If the UE has evaluated a number of DRX cycles Nserv consecutive DRX cycles that the serving cell does not fulfil the cell selection criterion S, the UE initiates the measurements of all neighbour cells indicated by the serving cell, regardless of the measurement rules currently limiting UE measurement activities. If the UE in RRC_IDLE has not found any new suitable cell based on searches and measurements using the intra-frequency, inter-frequency and inter-RAT information indicated in the system information for [10] s, the UE initiates cell selection procedures for the selected PLMN.


Systems and Implementations


FIG. 5 illustrates an example architecture of a system 500 of a network, in accordance with various embodiments. The following description is provided for an example system 500 that operates in conjunction with the LTE system standards and 5G or NR system standards as provided by 3GPP technical specifications. However, the example embodiments are not limited in this regard and the described embodiments may apply to other networks that benefit from the principles described herein, such as future 3GPP systems (e.g., Sixth Generation (6G)) systems, IEEE 802.16 protocols (e.g., WMAN, WiMAX, etc.), or the like.


As shown by FIG. 5, the system 500 includes UE 501a and UE 501b (collectively referred to as “UEs 501” or “UE 501”). In this example, UEs 501 are illustrated as smartphones (e.g., handheld touchscreen mobile computing devices connectable to one or more cellular networks), but may also comprise any mobile or non-mobile computing device, such as consumer electronics devices, cellular phones, smartphones, feature phones, tablet computers, wearable computer devices, personal digital assistants (PDAs), pagers, wireless handsets, desktop computers, laptop computers, in-vehicle infotainment (IVI), in-car entertainment (ICE) devices, an Instrument Cluster (IC), head-up display (HUD) devices, onboard diagnostic (OBD) devices, dashtop mobile equipment (DME), mobile data terminals (MDTs), Electronic Engine Management System (EEMS), electronic/engine control units (ECUs), electronic/engine control modules (ECMs), embedded systems, microcontrollers, control modules, engine management systems (EMS), networked or “smart” appliances, MTC devices, M2M, IoT devices, and/or the like.


In some embodiments, any of the UEs 501 may be IoT UEs, which may comprise a network access layer designed for low-power IoT applications utilizing short-lived UE connections. An IoT UE can utilize technologies such as M2M or MTC for exchanging data with an MTC server or device via a PLMN, ProSe or D2D communication, sensor networks, or IoT networks. The M2M or MTC exchange of data may be a machine-initiated exchange of data. An IoT network describes interconnecting IoT UEs, which may include uniquely identifiable embedded computing devices (within the Internet infrastructure), with short-lived connections. The IoT UEs may execute background applications (e.g., keep-alive messages, status updates, etc.) to facilitate the connections of the IoT network.


The UEs 501 may be configured to connect, for example, communicatively couple, with an or RAN 510. In embodiments, the RAN 510 may be an NG RAN or a 5G RAN, an E-UTRAN, or a legacy RAN, such as a UTRAN or GERAN. As used herein, the term “NG RAN” or the like may refer to a RAN 510 that operates in an NR or 5G system 500, and the term “E-UTRAN” or the like may refer to a RAN 510 that operates in an LTE or 4G system 500. The UEs 501 utilize connections (or channels) 503 and 504, respectively, each of which comprises a physical communications interface or layer (discussed in further detail below).


In this example, the connections 503 and 504 are illustrated as an air interface to enable communicative coupling, and can be consistent with cellular communications protocols, such as a GSM protocol, a CDMA network protocol, a PTT protocol, a POC protocol, a UMTS protocol, a 3GPP LTE protocol, a 5G protocol, a NR protocol, and/or any of the other communications protocols discussed herein. In embodiments, the UEs 501 may directly exchange communication data via a ProSe interface 505. The ProSe interface 505 may alternatively be referred to as a SL interface 505 and may comprise one or more logical channels, including but not limited to a PSCCH, a PSSCH, a PSDCH, and a PSBCH.


The UE 501b is shown to be configured to access an AP 506 (also referred to as “WLAN node 506,” “WLAN 506,” “WLAN Termination 506,” “WT 506” or the like) via connection 507. The connection 507 can comprise a local wireless connection, such as a connection consistent with any IEEE 802.11 protocol, wherein the AP 506 would comprise a wireless fidelity (Wi-Fi®) router. In this example, the AP 506 is shown to be connected to the Internet without connecting to the core network of the wireless system (described in further detail below). In various embodiments, the UE 501b, RAN 510, and AP 506 may be configured to utilize LWA operation and/or LWIP operation. The LWA operation may involve the UE 501b in RRC_CONNECTED being configured by a RAN node 511a-b to utilize radio resources of LTE and WLAN. LWIP operation may involve the UE 501b using WLAN radio resources (e.g., connection 507) via IPsec protocol tunneling to authenticate and encrypt packets (e.g., IP packets) sent over the connection 507. IPsec tunneling may include encapsulating the entirety of original IP packets and adding a new packet header, thereby protecting the original header of the IP packets.


The RAN 510 can include one or more AN nodes or RAN nodes 511a and 511b (collectively referred to as “RAN nodes 511” or “RAN node 511”) that enable the connections 503 and 504. As used herein, the terms “access node,” “access point,” or the like may describe equipment that provides the radio baseband functions for data and/or voice connectivity between a network and one or more users. These access nodes can be referred to as BS, gNBs, RAN nodes, eNBs, NodeBs, RSUs, TRxPs or TRPs, and so forth, and can comprise ground stations (e.g., terrestrial access points) or satellite stations providing coverage within a geographic area (e.g., a cell). As used herein, the term “NG RAN node” or the like may refer to a RAN node 511 that operates in an NR or 5G system 500 (e.g., a gNB), and the term “E-UTRAN node” or the like may refer to a RAN node 511 that operates in an LTE or 4G system 500 (e.g., an eNB). According to various embodiments, the RAN nodes 511 may be implemented as one or more of a dedicated physical device such as a macrocell base station, and/or a low power (LP) base station for providing femtocells, picocells or other like cells having smaller coverage areas, smaller user capacity, or higher bandwidth compared to macrocells.


In some embodiments, all or parts of the RAN nodes 511 may be implemented as one or more software entities running on server computers as part of a virtual network, which may be referred to as a CRAN and/or a virtual baseband unit pool (vBBUP). In these embodiments, the CRAN or vBBUP may implement a RAN function split, such as a PDCP split wherein RRC and PDCP layers are operated by the CRAN/vBBUP and other L2 protocol entities are operated by individual RAN nodes 511; a MAC/PHY split wherein RRC, PDCP, RLC, and MAC layers are operated by the CRAN/vBBUP and the PHY layer is operated by individual RAN nodes 511; or a “lower PHY” split wherein RRC, PDCP, RLC, MAC layers and upper portions of the PHY layer are operated by the CRAN/vBBUP and lower portions of the PHY layer are operated by individual RAN nodes 511. This virtualized framework allows the freed-up processor cores of the RAN nodes 511 to perform other virtualized applications. In some implementations, an individual RAN node 511 may represent individual gNB-DUs that are connected to a gNB-CU via individual F1 interfaces (not shown by FIG. 5). In these implementations, the gNB-DUs may include one or more remote radio heads or RFEMs (see, e.g., FIG. 9), and the gNB-CU may be operated by a server that is located in the RAN 510 (not shown) or by a server pool in a similar manner as the CRAN/vBBUP. Additionally or alternatively, one or more of the RAN nodes 511 may be next generation eNBs (ng-eNBs), which are RAN nodes that provide E-UTRA user plane and control plane protocol terminations toward the UEs 501, and are connected to a 5GC (e.g., CN 820 of FIG. 8) via an NG interface (discussed infra).


In V2X scenarios one or more of the RAN nodes 511 may be or act as RSUs. The term “Road Side Unit” or “RSU” may refer to any transportation infrastructure entity used for V2X communications. An RSU may be implemented in or by a suitable RAN node or a stationary (or relatively stationary) UE, where an RSU implemented in or by a UE may be referred to as a “UE-type RSU,” an RSU implemented in or by an eNB may be referred to as an “eNB-type RSU,” an RSU implemented in or by a gNB may be referred to as a “gNB-type RSU,” and the like. In one example, an RSU is a computing device coupled with radio frequency circuitry located on a roadside that provides connectivity support to passing vehicle UEs 501 (vUEs 501). The RSU may also include internal data storage circuitry to store intersection map geometry, traffic statistics, media, as well as applications/software to sense and control ongoing vehicular and pedestrian traffic. The RSU may operate on the 5.9 GHz Direct Short Range Communications (DSRC) band to provide very low latency communications required for high speed events, such as crash avoidance, traffic warnings, and the like. Additionally or alternatively, the RSU may operate on the cellular V2X band to provide the aforementioned low latency communications, as well as other cellular communications services. Additionally or alternatively, the RSU may operate as a Wi-Fi hotspot (2.4 GHz band) and/or provide connectivity to one or more cellular networks to provide uplink and downlink communications. The computing device(s) and some or all of the radiofrequency circuitry of the RSU may be packaged in a weatherproof enclosure suitable for outdoor installation, and may include a network interface controller to provide a wired connection (e.g., Ethernet) to a traffic signal controller and/or a backhaul network.


Any of the RAN nodes 511 can terminate the air interface protocol and can be the first point of contact for the UEs 501. In some embodiments, any of the RAN nodes 511 can fulfill various logical functions for the RAN 510 including, but not limited to, radio network controller (RNC) functions such as radio bearer management, uplink and downlink dynamic radio resource management and data packet scheduling, and mobility management.


In embodiments, the UEs 501 can be configured to communicate using OFDM communication signals with each other or with any of the RAN nodes 511 over a multicarrier communication channel in accordance with various communication techniques, such as, but not limited to, an OFDMA communication technique (e.g., for downlink communications) or a SC-FDMA communication technique (e.g., for uplink and ProSe or sidelink communications), although the scope of the embodiments is not limited in this respect. The OFDM signals can comprise a plurality of orthogonal subcarriers.


In some embodiments, a downlink resource grid can be used for downlink transmissions from any of the RAN nodes 511 to the UEs 501, while uplink transmissions can utilize similar techniques. The grid can be a time-frequency grid, called a resource grid or time-frequency resource grid, which is the physical resource in the downlink in each slot. Such a time-frequency plane representation is a common practice for OFDM systems, which makes it intuitive for radio resource allocation. Each column and each row of the resource grid corresponds to one OFDM symbol and one OFDM subcarrier, respectively. The duration of the resource grid in the time domain corresponds to one slot in a radio frame. The smallest time-frequency unit in a resource grid is denoted as a resource element. Each resource grid comprises a number of resource blocks, which describe the mapping of certain physical channels to resource elements. Each resource block comprises a collection of resource elements; in the frequency domain, this may represent the smallest quantity of resources that currently can be allocated. There are several different physical downlink channels that are conveyed using such resource blocks.


The UE 501 is also capable of detecting radio link failures (RLFs). Detection and recovery of physical layer problems in RRC_CONNECTED is based on out-of-sync and in-sync indications. Upon receiving N310 consecutive out-of-sync indications for the SpCell from lower layers while neither T300, T301, T304, T311 nor T319 are running, the UE 501 starts timer T310 for the corresponding SpCell. Upon receiving N311 consecutive in-sync indications for the SpCell from lower layers while T310 is running, the UE 501 stops the timer T310 for the corresponding SpCell. In this case, the UE 501 maintains the RRC connection without explicit signalling. In other words, the UE 501 maintains the entire radio resource configuration. Periods in time where neither in-sync nor out-of-sync is reported by L1 do not affect the evaluation of the number of consecutive in-sync or out-of-sync indications.


To detect RLFs, the UE 501, detects or determines expiration of timer T310 in Pcell, a random access problem indication from MCG MAC while neither T300, T301, T304, T311 nor T319 are running, or an indication from MCG RLC that the maximum number of retransmissions has been reached. In either of these cases, the UE 501 initiates the failure information procedure as specified in 5.7.5 of 3GPP TS 38.331 to report RLC failure if the indication is from MCG RLC and CA duplication is configured and activated, and for the corresponding logical channel allowedServingCells only includes SCell(s). Otherwise, the UE 501 considers an RLF to be detected for the MCG (e.g., RLF); if AS security has not been activated, the UE 501 performs the actions upon going to RRC_IDLE as specified in 5.3.11 of 3GPP TS 38.331, with release cause ‘other’; else if AS security has been activated but SRB2 and at least one DRB have not been setup, the UE 501 performs the actions upon going to RRC_IDLE as specified in 5.3.11 of 3GPP 38.331, with release cause ‘RRC connection failure’; otherwise, the UE 501 initiates the connection re-establishment procedure as specified in 5.3.7 of 3GPP TS 38.331.


Additionally or alternatively, when the UE 501 detects or determines expiration of timer T310 in PSCell, a random access problem indication from SCG MAC, or an indication from SCG RLC that the maximum number of retransmissions has been reached. In either of these cases, the UE 501 initiates the failure information procedure as specified in 5.7.5 of 3GPP 38.331 to report RLC failure if the indication is from SCG RLC and CA duplication is configured and activated; and for the corresponding logical channel allowedServingCells only includes SCell(s). Otherwise, the UE 501 considers an RLF to be detected for the SCG (e.g., SCG RLF), and initiates the SCG failure information procedure as specified in 5.7.3 of 3GPP 38.331 to report SCG RLF.


Additionally, the UE 501 may initiate a failure information procedure when there is a need inform the network about a failure detected by the UE 501, such as a detected RLF. In particular, the UE 501 initiates the procedure when the UE 501 detects failure for an RLC bearer in accordance with 5.3.10.3 of 3GPP TS 38.331. Upon initiating the failure information procedure, the UE 501 initiates transmission of the FailureInformation message as specified in 5.7.5.3 of 3GPP TS 38.331. If the failure information is used to inform the network about a failure for an MCG RLC bearer, the UE 501 submits the FailureInformation message to lower layers for transmission via SRB1. If the failure information is used to inform the network about a failure for an SCG RLC bearer, and if SRB3 is configured, the UE 501 submits the FailureInformation message to lower layers for transmission via SRB3. If the failure information is used to inform the network about a failure for an SCG RLC bearer, and if the UE 501 is in (NG)EN-DC, the UE 501 submits the FailureInformation message via the E-UTRA MCG embedded in E-UTRA RRC message ULInformationTransferMRDC as specified in 3GPP TS 36.331. If the failure information is used to inform the network about a failure for an SCG RLC bearer, and if the UE 501 is in NR-DC, the UE 501 submits the FailureInformation message via the NR MCG embedded in NR RRC message ULInformationTransferMRDC as specified in section 5.7.2a.3 of 3GPP TS 38.331.


According to various embodiments, the UEs 501 and the RAN nodes 511 communicate data (e.g., transmit and receive) data over a licensed medium (also referred to as the “licensed spectrum” and/or the “licensed band”) and an unlicensed shared medium (also referred to as the “unlicensed spectrum” and/or the “unlicensed band”). The licensed spectrum may include channels that operate in the frequency range of approximately 400 MHz to approximately 3.8 GHz, whereas the unlicensed spectrum may include the 5 GHz band.


To operate in the unlicensed spectrum, the UEs 501 and the RAN nodes 511 may operate using LAA, eLAA, and/or feLAA mechanisms. In these implementations, the UEs 501 and the RAN nodes 511 may perform one or more known medium-sensing operations and/or carrier-sensing operations in order to determine whether one or more channels in the unlicensed spectrum is unavailable or otherwise occupied prior to transmitting in the unlicensed spectrum. The medium/carrier sensing operations may be performed according to a listen-before-talk (LBT) protocol.


LBT is a mechanism whereby equipment (e.g., UEs 501 RAN nodes 511, etc.) senses a medium (e.g., a channel or carrier frequency) and transmits when the medium is sensed to be idle (or when a specific channel in the medium is sensed to be unoccupied). The medium sensing operation may include CCA, which utilizes at least ED to determine the presence or absence of other signals on a channel in order to determine if a channel is occupied or clear. This LBT mechanism allows cellular/LAA networks to coexist with incumbent systems in the unlicensed spectrum and with other LAA networks. ED may include sensing RF energy across an intended transmission band for a period of time and comparing the sensed RF energy to a predefined or configured threshold.


Typically, the incumbent systems in the 5 GHz band are WLANs based on IEEE 802.11 technologies. WLAN employs a contention-based channel access mechanism, called CSMA/CA. Here, when a WLAN node (e.g., a mobile station (MS) such as UE 501, AP 506, or the like) intends to transmit, the WLAN node may first perform CCA before transmission. Additionally, a backoff mechanism is used to avoid collisions in situations where more than one WLAN node senses the channel as idle and transmits at the same time. The backoff mechanism may be a counter that is drawn randomly within the CWS, which is increased exponentially upon the occurrence of collision and reset to a minimum value when the transmission succeeds. The LBT mechanism designed for LAA is somewhat similar to the CSMA/CA of WLAN. In some implementations, the LBT procedure for DL or UL transmission bursts including PDSCH or PUSCH transmissions, respectively, may have an LAA contention window that is variable in length between X and Y ECCA slots, where X and Y are minimum and maximum values for the CWSs for LAA. In one example, the minimum CWS for an LAA transmission may be 9 microseconds (μs); however, the size of the CWS and a MCOT (e.g., a transmission burst) may be based on governmental regulatory requirements.


The LAA mechanisms are built upon CA technologies of LTE-Advanced systems. In CA, each aggregated carrier is referred to as a CC. A CC may have a bandwidth of 1.4, 3, 5, 10, 15 or 20 MHz and a maximum of five CCs can be aggregated, and therefore, a maximum aggregated bandwidth is 100 MHz. In FDD systems, the number of aggregated carriers can be different for DL and UL, where the number of UL CCs is equal to or lower than the number of DL component carriers. In some cases, individual CCs can have a different bandwidth than other CCs. In TDD systems, the number of CCs as well as the bandwidths of each CC is usually the same for DL and UL.


CA also comprises individual serving cells to provide individual CCs. The coverage of the serving cells may differ, for example, because CCs on different frequency bands will experience different pathloss. A primary service cell or PCell may provide a PCC for both UL and DL, and may handle RRC and NAS related activities. The other serving cells are referred to as SCells, and each SCell may provide an individual SCC for both UL and DL. The SCCs may be added and removed as required, while changing the PCC may require the UE 501 to undergo a handover. In LAA, eLAA, and feLAA, some or all of the SCells may operate in the unlicensed spectrum (referred to as “LAA SCells”), and the LAA SCells are assisted by a PCell operating in the licensed spectrum. When a UE is configured with more than one LAA SCell, the UE may receive UL grants on the configured LAA SCells indicating different PUSCH starting positions within a same subframe.


The PDSCH carries user data and higher-layer signaling to the UEs 501. The PDCCH carries information about the transport format and resource allocations related to the PDSCH channel, among other things. It may also inform the UEs 501 about the transport format, resource allocation, and HARQ information related to the uplink shared channel. Typically, downlink scheduling (assigning control and shared channel resource blocks to the UE 501b within a cell) may be performed at any of the RAN nodes 511 based on channel quality information fed back from any of the UEs 501. The downlink resource assignment information may be sent on the PDCCH used for (e.g., assigned to) each of the UEs 501.


The PDCCH uses CCEs to convey the control information. Before being mapped to resource elements, the PDCCH complex-valued symbols may first be organized into quadruplets, which may then be permuted using a sub-block interleaver for rate matching. Each PDCCH may be transmitted using one or more of these CCEs, where each CCE may correspond to nine sets of four physical resource elements known as REGs. Four Quadrature Phase Shift Keying (QPSK) symbols may be mapped to each REG. The PDCCH can be transmitted using one or more CCEs, depending on the size of the DCI and the channel condition. There can be four or more different PDCCH formats defined in LTE with different numbers of CCEs (e.g., aggregation level, L=1, 2, 4, or 8).


Some embodiments may use concepts for resource allocation for control channel information that are an extension of the above-described concepts. For example, some embodiments may utilize an EPDCCH that uses PDSCH resources for control information transmission. The EPDCCH may be transmitted using one or more ECCEs. Similar to above, each ECCE may correspond to nine sets of four physical resource elements known as an EREGs. An ECCE may have other numbers of EREGs in some situations.


The RAN nodes 511 may be configured to communicate with one another via interface 512. In embodiments where the system 500 is an LTE system (e.g., when CN 520 is an EPC), the interface 512 may be an X2 interface 512. The X2 interface may be defined between two or more RAN nodes 511 (e.g., two or more eNBs and the like) that connect to EPC 520, and/or between two eNBs connecting to EPC 520. In some implementations, the X2 interface may include an X2 user plane interface (X2-U) and an X2 control plane interface (X2-C). The X2-U may provide flow control mechanisms for user data packets transferred over the X2 interface, and may be used to communicate information about the delivery of user data between eNBs. For example, the X2-U may provide specific sequence number information for user data transferred from a MeNB to an SeNB; information about successful in sequence delivery of PDCP PDUs to a UE 501 from an SeNB for user data; information of PDCP PDUs that were not delivered to a UE 501; information about a current minimum desired buffer size at the SeNB for transmitting to the UE user data; and the like. The X2-C may provide intra-LTE access mobility functionality, including context transfers from source to target eNBs, user plane transport control, etc.; load management functionality; as well as inter-cell interference coordination functionality.


In embodiments where the system 500 is a 5G or NR system (e.g., when CN 520 is an 5GC), the interface 512 may be an Xn interface 512. The Xn interface is defined between two or more RAN nodes 511 (e.g., two or more gNBs and the like) that connect to 5GC 520, between a RAN node 511 (e.g., a gNB) connecting to 5GC 520 and an eNB, and/or between two eNBs connecting to 5GC 520. In some implementations, the Xn interface may include an Xn user plane (Xn-U) interface and an Xn control plane (Xn-C) interface. The Xn-U may provide non-guaranteed delivery of user plane PDUs and support/provide data forwarding and flow control functionality. The Xn-C may provide management and error handling functionality, functionality to manage the Xn-C interface; mobility support for UE 501 in a connected mode (e.g., CM-CONNECTED) including functionality to manage the UE mobility for connected mode between one or more RAN nodes 511. The mobility support may include context transfer from an old (source) serving RAN node 511 to new (target) serving RAN node 511; and control of user plane tunnels between old (source) serving RAN node 511 to new (target) serving RAN node 511. A protocol stack of the Xn-U may include a transport network layer built on Internet Protocol (IP) transport layer, and a GTP-U layer on top of a UDP and/or IP layer(s) to carry user plane PDUs. The Xn-C protocol stack may include an application layer signaling protocol (referred to as Xn Application Protocol (Xn-AP)) and a transport network layer that is built on SCTP. The SCTP may be on top of an IP layer, and may provide the guaranteed delivery of application layer messages. In the transport IP layer, point-to-point transmission is used to deliver the signaling PDUs. In other implementations, the Xn-U protocol stack and/or the Xn-C protocol stack may be same or similar to the user plane and/or control plane protocol stack(s) shown and described herein.


The RAN 510 is shown to be communicatively coupled to a core network—in this embodiment, core network (CN) 520. The CN 520 may comprise a plurality of network elements 522, which are configured to offer various data and telecommunications services to customers/subscribers (e.g., users of UEs 501) who are connected to the CN 520 via the RAN 510. The components of the CN 520 may be implemented in one physical node or separate physical nodes including components to read and execute instructions from a machine-readable or computer-readable medium (e.g., a non-transitory machine-readable storage medium). In some embodiments, NFV may be utilized to virtualize any or all of the above-described network node functions via executable instructions stored in one or more computer-readable storage mediums (described in further detail below). A logical instantiation of the CN 520 may be referred to as a network slice, and a logical instantiation of a portion of the CN 520 may be referred to as a network sub-slice. NFV architectures and infrastructures may be used to virtualize one or more network functions, alternatively performed by proprietary hardware, onto physical resources comprising a combination of industry-standard server hardware, storage hardware, or switches. In other words, NFV systems can be used to execute virtual or reconfigurable implementations of one or more EPC components/functions.


Generally, the application server 530 may be an element offering applications that use IP bearer resources with the core network (e.g., UMTS PS domain, LTE PS data services, etc.). The application server 530 can also be configured to support one or more communication services (e.g., VoIP sessions, PTT sessions, group communication sessions, social networking services, etc.) for the UEs 501 via the EPC 520.


In embodiments, the CN 520 may be a 5GC (referred to as “5GC 520” or the like), and the RAN 510 may be connected with the CN 520 via an NG interface 513. In embodiments, the NG interface 513 may be split into two parts, an NG user plane (NG-U) interface 514, which carries traffic data between the RAN nodes 511 and a UPF, and the S1 control plane (NG-C) interface 515, which is a signaling interface between the RAN nodes 511 and AMFs.


In embodiments, the CN 520 may be a 5G CN (referred to as “5GC 520” or the like), while in other embodiments, the CN 520 may be an EPC). Where CN 520 is an EPC (referred to as “EPC 520” or the like), the RAN 510 may be connected with the CN 520 via an S1 interface 513. In embodiments, the S1 interface 513 may be split into two parts, an S1 user plane (S1-U) interface 514, which carries traffic data between the RAN nodes 511 and the S-GW, and the S1-MME interface 515, which is a signaling interface between the RAN nodes 511 and MMES.



FIG. 6 shows an example NG-RAN architecture in accordance with various embodiments. The NG-RAN comprises a set of gNBs connected to the 5GC (e.g., CN 820 of FIGS. 8A and 8B discussed infra) through the NG interface. The NG-RAN of FIG. 6 may correspond to RAN 510 of FIG. 5 and/or RAN 811 of FIG. 8, and the gNBs in FIG. 6 may correspond to the RAN nodes 511 of FIG. 5.


A gNB can support FDD mode, TDD mode, or dual mode operation. One or more gNBs can be interconnected to one another through the Xn interface. A gNB may include a gNB-CU and one or more gNB-DU(s). A gNB-CU is a logical node that hosts RRC, SDAP, and PDCP protocols of the gNB or RRC and PDCP protocols of an en-gNB that controls the operation of one or more gNB-DUs. The gNB-CU terminates the F1 interface connected with an individual gNB-DU. A gnB-DU is a logical node that hosts the RLC, MAC, and PHY layers of the gNB or en-gNB, and its operation is partly controlled by a gNB-CU. An individual gNB-DU supports one or more cells and an individual cell is supported by only one gNB-DU. The gNB-DU terminates the F1 interface connected with the gNB-CU.


In some implementations, the gNB-CU may be divided into a gNB-CU-Control Plane (gNB-CU-CP) and a gNB-CU-User Plane (gNB-CU-UP). The gNB-CU-CP is a logical node that hosts the RRC and the control plane part(s) of the PDCP protocol of the gNB-CU for an en-gNB or a gNB. In these implementations, the gNB-CU-CP terminates an E1 interface that connects the gNB-CU-CP with the gNB-CU-UP (not shown) and an F1-C interface connected with the gNB-DU. The gNB-CU-UP is a logical node that hosts the user plane part(s) of the PDCP protocol of the gNB-CU for an en-gNB, and the user plane part(s) of the PDCP protocol and the SDAP protocol of the gNB-CU for a gNB. In these implementations, the gNB-CU-UP terminates the E1 interface connected with the gNB-CU-CP and the F1-U interface connected with the gNB-DU.


A gNB-CU is connected with one or more gNB-DUs via respective F1 interfaces. In most implementations, one gNB-DU is connected to only one gNB-CU. For resiliency, a gNB-DU may be connected to multiple gNB-CUs by appropriate implementation. The NG, Xn and F1 are logical interfaces.


For the NG-RAN, the NG and Xn-C interfaces terminate in the gNB-CU for a gNB comprising a gNB-CU and gNB-DUs. For EN-DC, the S1-U and X2-C interfaces for a gNB consisting of a gNB-CU and gNB-DUs, terminate in the gNB-CU. The gNB-CU and connected gNB-DUs are only visible to other gNBs and the 5GC as a gNB.


The node hosting user plane part of NR PDCP (e.g., gNB-CU, gNB-CU-UP, and for EN-DC, MeNB or SgNB depending on the bearer split) shall perform user inactivity monitoring and further informs its inactivity or (re)activation to the node having C-plane connection towards the core network (e.g., over E1, X2). The node hosting NR RLC (e.g., gNB-DU) may perform user inactivity monitoring and further inform its inactivity or (re)activation to the node hosting control plane, e.g., gNB-CU or gNB-CU-CP.


The UL PDCP configuration (e.g., how the UE uses the UL at the assisting node) is indicated via the X2 control place interface (X2-C) for EN-DC, Xn control place interface (Xn-C) for the NG-RAN and F1 control place interface (F1-C). Radio Link Outage/Resume for DL and/or UL is indicated via the X2 user place interface (X2-U) for EN-DC, Xn user place interface (Xn-U) for NG-RAN, and the F1 user place interface (F1-U).


The NG-RAN is layered into a Radio Network Layer (RNL) and a Transport Network Layer (TNL). The NG-RAN architecture of FIG. 6, e.g., the NG-RAN logical nodes and interfaces between them, is defined as part of the RNL. For each NG-RAN interface (e.g., NG, Xn, F1) the related TNL protocol and the functionality are specified. The TNL provides services for user plane transport, signalling transport. In an NG-Flex configuration, each gNB is connected to all AMFs within an AMF Region or AMF pool. In some implementations, the NDS/IP security protection may be be applied for control plane and user plane data on the TNL of the various NG-RAN interfaces has to be supported.


The F1 interface is a 5G radio network layer signaling protocol for signaling service between the gNB-CU and a gNB-DU. The F1 interface is an open interface that supports the exchange of signaling information between respective endpoints at the gNB-CU and gNB-DU. In addition, the F1 interface supports data transmission to the respective endpoints, and from a logical standpoint, the F1 interface is a point-to-point interface between the endpoints. In some implementations, the point-to-point logical interface may exist in the absence of a physical direct connection between the endpoints.


The F1 interface also supports support control plane and user plane separation. In this regard, the CU is separated in an F1-C and an F1-U. The F1-U includes a user data transfer function, which allows the gNB-DU and gNB-CU to transfer user data between the gNB-CU and gNB-DU. The F1-U also includes a flow control function that allows downlink user data flow to the gNB-DU to be controlled.


The F1-C comprises interface management functions/procedures, UE context management functions/procedures, RRC message transfer functions/procedures, system information delivery functions/procedures, paging functions/procedures, and warning message transmission functions/procedures


The interface management functions/procedures include error indication, reset, F1 setup, gNB-DU configuration update, gNB-CU configuration update, and gNB-DU resource coordination functions/procedures. The error indication function is used by the gNB-DU or gNB-CU to indicate to the gNB-CU or gNB-DU that an error has occurred. The reset function is used by both the gNB-DU and the gNB-CU to initialize a peer entity after node setup and after a failure event has occurred.


The F1 setup function allows the gNB-DU and the gNB-CU to exchange application level data needed for the gNB-DU and gNB-CU to interoperate correctly on the F1 interface. The purpose of the F1 Setup procedure is to exchange application level data needed for the gNB-DU and the gNB-CU to correctly interoperate on the F1 interface. The F1 Setup procedure may be the first F1AP procedure triggered after a TNL association has become operational. The F1 Setup procedure uses non-UE associated signaling. In embodiments, the gNB-DU initiates the F1 Setup procedure by sending an F1 SETUP REQUEST message including appropriate data to the gNB-CU, and the gNB-CU responds with an F1 SETUP RESPONSE message including appropriate data. If the gNB-CU cannot accept the setup, it should respond with a F1 SETUP FAILURE and appropriate cause value.


The gNB-CU Configuration Update and gNB-DU Configuration Update functions allow to update application level configuration data needed between gNB-CU and gNB-DU to interoperate correctly over the F1 interface, and activates or deactivates cells. The F1 setup and gNB-DU Configuration Update functions allow a gNB-DU to inform the gNB-CU about the S-NSSAI(s) supported by the gNB-DU.


The resource coordination function is used to transfer information about frequency resource sharing between a gNB-CU and a gNB-DU.


The RRC message transfer functions allow RRC messages to be transferred between gNB-CU and gNB-DU. RRC messages are transferred over F1-C. The gNB-CU is responsible for the encoding of the dedicated RRC message with assistance information provided by gNB-DU.


The F1 UE context management functions support the establishment, modification, and release of the necessary overall UE context. The establishment of the F1 UE context is initiated by the gNB-CU and accepted or rejected by the gNB-DU based on admission control criteria (e.g., resource not available). The modification of the F1 UE context can be initiated by either the gNB-CU or the gNB-DU, where the receiving node can accept or reject the modification request. The F1 UE context management function also supports the release of the context previously established in the gNB-DU. The release of the context is triggered by the gNB-CU either directly or following a request received from the gNB-DU. The gNB-CU request the gNB-DU to release the UE Context when the UE enters RRC_IDLE or RRC_INACTIVE. This procedure is also used to command the gNB-DU to stop data transmission for the UE.


The F1 UE context management functions can be also used to manage DRBs and SRBs, e.g., establishing, modifying and releasing DRB and SRB resources. The establishment and modification of DRB resources are triggered by the gNB-CU and accepted/rejected by the gNB-DU based on resource reservation information and QoS information to be provided to the gNB-DU. For each DRB to be setup or modified, the S-NSSAI may be provided by gNB-CU to the gNB-DU in the UE Context Setup procedure and the UE Context Modification procedure. The mapping between QoS flows and radio bearers is performed by gNB-CU and the granularity of bearer related management over F1 is radio bearer level. For NG-RAN, the gNB-CU provides an aggregated DRB QoS profile and QoS flow profile to the gNB-DU, and the gNB-DU either accepts the request or rejects it with appropriate cause value. To support packet duplication for intra-gNB-DU CA one data radio bearer should be configured with two GTP-U tunnels between gNB-CU and a gNB-DU.


F1-C signalling bearers provide provisioning of reliable transfer of F1 application protocol (AP) messages over the F1-C interface; provisioning of networking and routing functions; provisioning of redundancy in the signalling network; and provide support for flow control and congestion control.


The F1 interface provide separation between the radio network layer and transport network layer separation. The radio network layer comprises an Application Protocol (F1AP) layer. The F1AP is an application protocol that supports the functions and signaling services of the F1 interface. F1AP services are divided into two groups: Non-UE-associated services and UE-associated services. Non-UE-associated services are services related to the whole F1 interface instance between the gNB-DU and gNB-CU utilizing a non UE-associated signaling connection. UE-associated services are services related to one UE. F1AP functions that provide these services are associated with a UE-associated signaling connection that is maintained for the UE in question.


The Transport Network Layer comprises Stream Control Transmission Protocol (SCTP) layer on top of an internet protocol (IP) layer, where the IP layer is on top of a data link layer and a physical layer under the data link layer. The IP layer supports IPv6 as described by IETF RFC 2460 and/or IPv4 as described by IETF RFC 791, as well as Diffserv Code Point marking as described in IETF RFC 2474. Additionally, point-to-point transmission may be used to deliver F1AP messages. The data link layer may be any suitable data link layer protocol such as PPP, Ethernet, or the like.


The SCTP layer is a transport layer as described by IETF RFC 4960. There may be one or more SCTP associations established between individual gNB-CU and gNB-DU pairs, which may be established by the gNB-DU. A single SCTP association is used for F1AP elementary procedures within a set of SCTP associations established between a gNB-CU and gNB-DU pair. Additionally, the SCTP layer may implement SCTP congestion control mechanisms that initiate higher layer protocols to reduce signalling traffic at a source and prioritize certain messages.


The E1 interface provides means for interconnecting a gNB-CU-CP and a gNB-CU-UP of a gNB-CU within an NG-RAN, or for interconnecting a gNB-CU-CP and a gNB-CU-UP of an en-gNB within an E-UTRAN. The E1 interface is an open interface that supports the exchange of signaling information between respective endpoints at the gNB-CU-CP and the gNB-CU-UP. In addition, the E1 interface supports data transmission to the respective endpoints, and from a logical standpoint, the E1 interface is a point-to-point interface between the endpoints. In some implementations, the point-to-point logical interface may exist in the absence of a physical direct connection between the endpoints. The E1 interface enables the exchange of UE associated information and non-UE associated information, and is a control interface that is not used for user data forwarding.


The E1 interface comprises E1 interface management functions, E1 bearer context management function, and TEID allocation functions. The E1 interface management functions include an error indication function, a reset function, an E1 setup function, and gNB-CU-UP Configuration Update and gNB-CU-CP Configuration Update functions. The error indication function is used by the gNB-CU-UP or gNB-CU-CP to indicate to the gNB-CU-CP or gNB-CU-UP that an error has occurred. The reset function is used to initialize the peer entity after node setup and after a failure event occurred. This procedure can be used by both the gNB-CU-UP and the gNB-CU-CP. The E1 setup function allows to exchange application level data needed for the gNB-CU-UP and gNB-CU-CP to interoperate correctly on the E1 interface. The E1 setup is initiated by both the gNB-CU-UP and gNB-CU-CP. The gNB-CU-UP Configuration Update and gNB-CU-CP Configuration Update functions allow to update application level configuration data needed between the gNB-CU-CP and the gNB-CU-UP to interoperate correctly over the E1 interface. The E1 setup and gNB-CU-UP Configuration Update functions allow to inform S-NSSAI(s) and PLMN-ID(s) supported by the gNB-CU-UP.


The E1 bearer context management function establishment is initiated by the gNB-CU-CP and accepted or rejected by the gNB-CU-UP based on admission control criteria (e.g., resource not available). The purpose of the Bearer Context Setup procedure is to allow the gNB-CU-CP to establish a bearer context in the gNB-CU-UP. The procedure uses UE-associated signalling. The modification of the E1 bearer context can be initiated by either gNB-CU-CP or gNB-CU-UP. The receiving node can accept or reject the modification. The E1 bearer context management function also supports the release of the bearer context previously established in the gNB-CU-UP. The release of the bearer context is triggered by the gNB-CU-CP either directly or following a request received from the gNB-CU-UP. This function is used to setup and modify the QoS-flow to DRB mapping configuration. The gNB-CU-CP decides flow-to-DRB mapping and provides the generated SDAP and PDCP configuration to the gNB-CU-UP.


The TEID allocation functions are used to allocated TEIDs. The gNB-CU-UP is responsible for the allocation of the F1-U UL GTP TEID for each data radio bearer. The gNB-CU-UP is responsible for the allocation of the S1-U DL GTP TEID for each E-RAB and the NG-U DL GTP TEID for each PDU Session. The gNB-CU-UP is responsible for the allocation of the X2-U DL/UL GTP TEID or the Xn-U DL/UL GTP TEID for each data radio bearer.



FIG. 7 illustrates an example architecture of a system 700 including a first CN 720, in accordance with various embodiments. In this example, system 700 may implement the LTE standard wherein the CN 720 is an EPC 720 that corresponds with CN 520 of FIG. 5. Additionally, the UE 701 may be the same or similar as the UEs 501 of FIG. 5, and the E-UTRAN 710 may be a RAN that is the same or similar to the RAN 510 of FIG. 5, and which may include RAN nodes 511 discussed previously. The CN 720 may comprise MMEs 721, an S-GW 722, a P-GW 723, a HSS 724, and a SGSN 725.


The MMEs 721 may be similar in function to the control plane of legacy SGSN, and may implement MM functions to keep track of the current location of a UE 701. The MMEs 721 may perform various MM procedures to manage mobility aspects in access such as gateway selection and tracking area list management. MM (also referred to as “EPS MM” or “EMM” in E-UTRAN systems) may refer to all applicable procedures, methods, data storage, etc. that are used to maintain knowledge about a present location of the UE 701, provide user identity confidentiality, and/or perform other like services to users/subscribers. Each UE 701 and the MME 721 may include an MM or EMM sublayer, and an MM context may be established in the UE 701 and the MME 721 when an attach procedure is successfully completed. The MM context may be a data structure or database object that stores MM-related information of the UE 701. The MMEs 721 may be coupled with the HSS 724 via an S6a reference point, coupled with the SGSN 725 via an S3 reference point, and coupled with the S-GW 722 via an S11 reference point.


The SGSN 725 may be a node that serves the UE 701 by tracking the location of an individual UE 701 and performing security functions. In addition, the SGSN 725 may perform Inter-EPC node signaling for mobility between 2G/3G and E-UTRAN 3GPP access networks; PDN and S-GW selection as specified by the MMEs 721; handling of UE 701 time zone functions as specified by the MMEs 721; and MME selection for handovers to E-UTRAN 3GPP access network. The S3 reference point between the MMEs 721 and the SGSN 725 may enable user and bearer information exchange for inter-3GPP access network mobility in idle and/or active states.


The HSS 724 may comprise a database for network users, including subscription-related information to support the network entities' handling of communication sessions. The EPC 720 may comprise one or several HSSs 724, depending on the number of mobile subscribers, on the capacity of the equipment, on the organization of the network, etc. For example, the HSS 724 can provide support for routing/roaming, authentication, authorization, naming/addressing resolution, location dependencies, etc. An S6a reference point between the HSS 724 and the MMEs 721 may enable transfer of subscription and authentication data for authenticating/authorizing user access to the EPC 720 between HSS 724 and the MMEs 721.


The S-GW 722 may terminate the S1 interface 513 (“S1-U” in FIG. 7) toward the RAN 710, and routes data packets between the RAN 710 and the EPC 720. In addition, the S-GW 722 may be a local mobility anchor point for inter-RAN node handovers and also may provide an anchor for inter-3GPP mobility. Other responsibilities may include lawful intercept, charging, and some policy enforcement. The S11 reference point between the S-GW 722 and the MMEs 721 may provide a control plane between the MMEs 721 and the S-GW 722. The S-GW 722 may be coupled with the P-GW 723 via an S5 reference point.


The P-GW 723 may terminate an SGi interface toward a PDN 730. The P-GW 723 may route data packets between the EPC 720 and external networks such as a network including the application server 530 (alternatively referred to as an “AF”) via an IP interface 525 (see e.g., FIG. 5). In embodiments, the P-GW 723 may be communicatively coupled to an application server (application server 530 of FIG. 5 or PDN 730 in FIG. 7) via an IP communications interface 525 (see, e.g., FIG. 5). The S5 reference point between the P-GW 723 and the S-GW 722 may provide user plane tunneling and tunnel management between the P-GW 723 and the S-GW 722. The S5 reference point may also be used for S-GW 722 relocation due to UE 701 mobility and if the S-GW 722 needs to connect to a non-collocated P-GW 723 for the required PDN connectivity. The P-GW 723 may further include a node for policy enforcement and charging data collection (e.g., PCEF (not shown)). Additionally, the SGi reference point between the P-GW 723 and the packet data network (PDN) 730 may be an operator external public, a private PDN, or an intra operator packet data network, for example, for provision of IMS services. The P-GW 723 may be coupled with a PCRF 726 via a Gx reference point.


PCRF 726 is the policy and charging control element of the EPC 720. In a non-roaming scenario, there may be a single PCRF 726 in the Home Public Land Mobile Network (HPLMN) associated with a UE 701's Internet Protocol Connectivity Access Network (IP-CAN) session. In a roaming scenario with local breakout of traffic, there may be two PCRFs associated with a UE 701's IP-CAN session, a Home PCRF (H-PCRF) within an HPLMN and a Visited PCRF (V-PCRF) within a Visited Public Land Mobile Network (VPLMN). The PCRF 726 may be communicatively coupled to the application server 730 via the P-GW 723. The application server 730 may signal the PCRF 726 to indicate a new service flow and select the appropriate QoS and charging parameters. The PCRF 726 may provision this rule into a PCEF (not shown) with the appropriate TFT and QCI, which commences the QoS and charging as specified by the application server 730. The Gx reference point between the PCRF 726 and the P-GW 723 may allow for the transfer of QoS policy and charging rules from the PCRF 726 to PCEF in the P-GW 723. An Rx reference point may reside between the PDN 730 (or “AF 730”) and the PCRF 726.



FIGS. 8A and 8B illustrate example system architectures 800-1 and 800-2 (collectively, system 800) of a 5GC, in accordance with various embodiments. In particular, FIG. 8A shows an exemplary 5G system architecture 800-1 in a reference point representation where interactions between NFs are represented by corresponding point-to-point reference points Ni, and FIG. 8B illustrates an exemplary 5G system architecture 300D in a service-based representation where interactions between NFs are represented by corresponding service-based interfaces. The system 800 is shown to include a UE 801, which may be the same or similar to the UEs 501, 701 discussed previously; a (R)AN 810, which may be the same or similar to the RAN 510 and RAN 730 discussed previously, and which may include RAN nodes 511 discussed previously; and a DN 803, which may be, for example, operator services, Internet access or 3rd party services, and may correspond with the PDN of FIG. 7; and a 5GC 820. The 5GC 820 may include an AUSF 822; an AMF 821; a SMF 824; a NEF 823; a PCF 826; a NRF 825; a UDM 827; an AF 828; a UPF 802; a NSSF 829; and SCP 830.


The reference point representation of FIG. 8A shows various interactions between corresponding NFs. For example, FIG. 8A illustrates the following reference points: N1 (between the UE 801 and the AMF 821), N2 (between the RAN 810 and the AMF 821), N3 (between the RAN 810 and the UPF 802), N4 (between the SMF 824 and the UPF 802), N5 (between the PCF 826 and the AF 828), N6 (between the UPF 802 and the DN 803), N7 (between the SMF 824 and the PCF 826), N8 (between the UDM 827 and the AMF 821), N9 (between two UPFs 802), N10 (between the UDM 827 and the SMF 824), N11 (between the AMF 821 and the SMF 824), N12 (between the AUSF 822 and the AMF 821), N13 (between the AUSF 822 and the UDM 827), N14 (between two AMFs 821), N15 (between the PCF 826 and the AMF 821 in case of a non-roaming scenario, or between the PCF 826 and a visited network and AMF 821 in case of a roaming scenario), N16 (between two SMFs; not shown), and N22 (between AMF 821 and NSSF 825). Other reference point representations not shown in FIG. 8A can also be used.


The service-based representation of FIG. 8B represents NFs within the control plane that enable other authorized NFs to access their services. In this regard, 5G system architecture 800-2 can include the following service-based interfaces: Namf (a service-based interface exhibited by the AMF 821), Nsmf (a service-based interface exhibited by the SMF 824), Nnef 364C (a service-based interface exhibited by the NEF 823), Npcf (a service-based interface exhibited by the PCF 826), Nudm (a service-based interface exhibited by the UDM 827), Naf (a service-based interface exhibited by the AF 828), Nnrf (a service-based interface exhibited by the NRF 825), Nnssf (a service-based interface exhibited by the NSSF 829), Nausf (a service-based interface exhibited by the AUSF 822). Other service-based interfaces (e.g., Nudr, N5g-eir, and Nudsf) not shown in FIG. 8B can also be used. In embodiments, the NEF 823 can provide an interface to MEC host 836, which can be used to process wireless connections with the RAN 810.


The UPF 802 acts as an anchor point for intra-RAT and inter-RAT mobility, an external PDU session point of interconnect to DN 803, which can include, for example, operator services, Internet access, or third-party services; and a branching point to support multi-homed PDU session. The UPF 802 may also perform packet routing and forwarding, perform packet inspection, enforce the user plane part of policy rules, lawfully intercept packets (UP collection), perform traffic usage reporting, perform QoS handling for a user plane (e.g., packet filtering, gating, UL/DL rate enforcement), perform Uplink Traffic verification (e.g., SDF to QoS flow mapping), transport level packet marking in the uplink and downlink, and perform downlink packet buffering and downlink data notification triggering. UPF 802 may include an uplink classifier to support routing traffic flows to a data network. The DN 803 may represent various network operator services, Internet access, or third party services. DN 803 may include, or be similar to, application server(s) 530 discussed previously. The UPF 802 may interact with the SMF 824 via an N4 reference point between the SMF 824 and the UPF 802.


The AUSF 822 stores data for authentication of UE 801 and handle authentication-related functionality. The AUSF 822 may facilitate a common authentication framework for various access types. The AUSF 822 may communicate with the AMF 821 via an N12 reference point between the AMF 821 and the AUSF 822; and may communicate with the UDM 827 via an N13 reference point between the UDM 827 and the AUSF 822. Additionally, the AUSF 822 may exhibit an Nausf service-based interface.


The UDM 827 may handle subscription-related information to support the network entities' handling of communication sessions, and may store subscription data of UE 801. For example, subscription data may be communicated between the UDM 827 and the AMF 821 via an N8 reference point between the UDM 827 and the AMF. The UDM 827 may include two parts, an application FE and a UDR (not shown). The UDR may store subscription data and policy data for the UDM 827 and the PCF 826, and/or structured data for exposure and application data (including PFDs for application detection, application request information for multiple UEs 801) for the NEF 823. The Nudr service-based interface may be exhibited by the UDR 221 to allow the UDM 827, PCF 826, and NEF 823 to access a particular set of the stored data, as well as to read, update (e.g., add, modify), delete, and subscribe to notification of relevant data changes in the UDR. The UDM may include a UDM-FE, which is in charge of processing credentials, location management, subscription management and so on. Several different front ends may serve the same user in different transactions. The UDM-FE accesses subscription information stored in the UDR and performs authentication credential processing, user identification handling, access authorization, registration/mobility management, and subscription management. The UDR may interact with the SMF 824 via an N10 reference point between the UDM 827 and the SMF 824. UDM 827 may also support SMS management, wherein an SMS-FE implements the similar application logic as discussed previously. Additionally, the UDM 827 may exhibit the Nudm service-based interface.


In some aspects, the UDM 827 can be coupled to an application server 840, which can include a telephony application server (TAS), another application server (AS), an edge compute node (e.g., Multi-access Edge Computing (MEC) host or the like) a Content Delivery Network (CDN) node, or some other system.


The AMF 821 may be responsible for registration management (e.g., for registering UE 801, etc.), connection management, reachability management, mobility management, and lawful interception of AMF-related events, and access authentication and authorization. The AMF 821 may be a termination point for the an N11 reference point between the AMF 821 and the SMF 824. The AMF 821 may provide transport for SM messages between the UE 801 and the SMF 824, and act as a transparent proxy for routing SM messages. AMF 821 may also provide transport for SMS messages between UE 801 and an SMSF (not shown). AMF 821 may act as SEAF, which may include interaction with the AUSF 822 and the UE 801, receipt of an intermediate key that was established as a result of the UE 801 authentication process. Where USIM based authentication is used, the AMF 821 may retrieve the security material from the AUSF 822. AMF 821 may also include a SCM function, which receives a key from the SEA that it uses to derive access-network specific keys. Furthermore, AMF 821 may be a termination point of a RAN CP interface, which may include or be an N2 reference point between the (R)AN 810 and the AMF 821; and the AMF 821 may be a termination point of NAS (N1) signalling, and perform NAS ciphering and integrity protection.


AMF 821 may also support NAS signalling with a UE 801 over an N3 IWF interface. The N3IWF may be used to provide access to untrusted entities. N3IWF may be a termination point for the N2 interface between the (R)AN 810 and the AMF 821 for the control plane, and may be a termination point for the N3 reference point between the (R)AN 810 and the UPF 802 for the user plane. As such, the AMF 821 may handle N2 signalling from the SMF 824 and the AMF 821 for PDU sessions and QoS, encapsulate/de-encapsulate packets for IPSec and N3 tunnelling, mark N3 user-plane packets in the uplink, and enforce QoS corresponding to N3 packet marking taking into account QoS requirements associated with such marking received over N2. N3IWF may also relay uplink and downlink control-plane NAS signalling between the UE 801 and AMF 821 via an N1 reference point between the UE 801 and the AMF 821, and relay uplink and downlink user-plane packets between the UE 801 and UPF 802. The N3IWF also provides mechanisms for IPsec tunnel establishment with the UE 801. The AMF 821 may exhibit an Namf service-based interface, and may be a termination point for an N14 reference point between two AMFs 821 and an N17 reference point between the AMF 821 and a 5G-EIR (not shown).


The UE 801 may need to register with the AMF 821 in order to receive network services. RM is used to register or deregister the UE 801 with the network (e.g., AMF 821), and establish a UE context in the network (e.g., AMF 821). The UE 801 may operate in an RM-REGISTERED state or an RM-DEREGISTERED state. In the RM-DEREGISTERED state, the UE 801 is not registered with the network, and the UE context in AMF 821 holds no valid location or routing information for the UE 801 so the UE 801 is not reachable by the AMF 821. In the RM-REGISTERED state, the UE 801 is registered with the network, and the UE context in AMF 821 may hold a valid location or routing information for the UE 801 so the UE 801 is reachable by the AMF 821. In the RM-REGISTERED state, the UE 801 may perform mobility Registration Update procedures, perform periodic Registration Update procedures triggered by expiration of the periodic update timer (e.g., to notify the network that the UE 801 is still active), and perform a Registration Update procedure to update UE capability information or to re-negotiate protocol parameters with the network, among others.


The AMF 821 may store one or more RM contexts for the UE 801, where each RM context is associated with a specific access to the network. The RM context may be a data structure, database object, etc. that indicates or stores, inter alia, a registration state per access type and the periodic update timer. The AMF 821 may also store a 5GC MM context that may be the same or similar to the (E)MM context discussed previously. In various embodiments, the AMF 821 may store a CE mode B Restriction parameter of the UE 801 in an associated MM context or RM context. The AMF 821 may also derive the value, when needed, from the UE's usage setting parameter already stored in the UE context (and/or MM/RM context).


CM may be used to establish and release a signaling connection between the UE 801 and the AMF 821 over the N1 interface. The signaling connection is used to enable NAS signaling exchange between the UE 801 and the CN 820, and comprises both the signaling connection between the UE and the AN (e.g., RRC connection or UE-N3IWF connection for non-3GPP access) and the N2 connection for the UE 801 between the AN (e.g., RAN 810) and the AMF 821. The UE 801 may operate in one of two CM states, CM-IDLE mode or CM-CONNECTED mode. When the UE 801 is operating in the CM-IDLE state/mode, the UE 801 may have no NAS signaling connection established with the AMF 821 over the N1 interface, and there may be (R)AN 810 signaling connection (e.g., N2 and/or N3 connections) for the UE 801. When the UE 801 is operating in the CM-CONNECTED state/mode, the UE 801 may have an established NAS signaling connection with the AMF 821 over the N1 interface, and there may be a (R)AN 810 signaling connection (e.g., N2 and/or N3 connections) for the UE 801. Establishment of an N2 connection between the (R)AN 810 and the AMF 821 may cause the UE 801 to transition from CM-IDLE mode to CM-CONNECTED mode, and the UE 801 may transition from the CM-CONNECTED mode to the CM-IDLE mode when N2 signaling between the (R)AN 810 and the AMF 821 is released.


The SMF 824 may be responsible for SM (e.g., session establishment, modify and release, including tunnel maintain between UPF and AN node); UE IP address allocation and management (including optional authorization); selection and control of UP function; configuring traffic steering at UPF to route traffic to proper destination; termination of interfaces toward policy control functions; controlling part of policy enforcement and QoS; lawful intercept (for SM events and interface to LI system); termination of SM parts of NAS messages; downlink data notification; initiating AN specific SM information, sent via AMF over N2 to AN; and determining SSC mode of a session. SM may refer to management of a PDU session, and a PDU session or “session” may refer to a PDU connectivity service that provides or enables the exchange of PDUs between a UE 801 and a data network (DN) 803 identified by a Data Network Name (DNN). PDU sessions may be established upon UE 801 request, modified upon UE 801 and 5GC 820 request, and released upon UE 801 and 5GC 820 request using NAS SM signaling exchanged over the N1 reference point between the UE 801 and the SMF 824. Upon request from an application server, the 5GC 820 may trigger a specific application in the UE 801. In response to receipt of the trigger message, the UE 801 may pass the trigger message (or relevant parts/information of the trigger message) to one or more identified applications in the UE 801. The identified application(s) in the UE 801 may establish a PDU session to a specific DNN. The SMF 824 may check whether the UE 801 requests are compliant with user subscription information associated with the UE 801. In this regard, the SMF 824 may retrieve and/or request to receive update notifications on SMF 824 level subscription data from the UDM 827.


The SMF 824 may include the following roaming functionality: handling local enforcement to apply QoS SLAB (VPLMN); charging data collection and charging interface (VPLMN); lawful intercept (in VPLMN for SM events and interface to LI system); and support for interaction with external DN for transport of signalling for PDU session authorization/authentication by external DN. An N16 reference point between two SMFs 824 may be included in the system 800, which may be between another SMF 824 in a visited network and the SMF 824 in the home network in roaming scenarios. Additionally, the SMF 824 may exhibit the Nsmf service-based interface.


The NEF 823 may provide means for securely exposing the services and capabilities provided by 3GPP network functions for third party, internal exposure/re-exposure, Application Functions (e.g., AF 828), edge computing or fog computing systems, etc. In such embodiments, the NEF 823 may authenticate, authorize, and/or throttle the AFs. NEF 823 may also translate information exchanged with the AF 828 and information exchanged with internal network functions. For example, the NEF 823 may translate between an AF-Service-Identifier and an internal 5GC information. NEF 823 may also receive information from other network functions (NFs) based on exposed capabilities of other network functions. This information may be stored at the NEF 823 as structured data, or at a data storage NF using standardized interfaces. The stored information can then be re-exposed by the NEF 823 to other NFs and AFs, and/or used for other purposes such as analytics. Additionally, the NEF 823 may exhibit an Nnef service-based interface.


The NRF 825 may support service discovery functions, receive NF discovery requests from NF instances, and provide the information of the discovered NF instances to the NF instances. NRF 825 also maintains information of available NF instances and their supported services. As used herein, the terms “instantiate,” “instantiation,” and the like may refer to the creation of an instance, and an “instance” may refer to a concrete occurrence of an object, which may occur, for example, during execution of program code. Additionally, the NRF 825 may exhibit the Nnrf service-based interface. The NRF 825 also supports service discovery functions, wherein the NRF 825 receives NF Discovery Request from NF instance or the SCP 830, and provides the information of the discovered NF instances (be discovered) to the NF instance or SCP 830.


The PCF 826 may provide policy rules to control plane function(s) to enforce them, and may also support unified policy framework to govern network behaviour. The PCF 826 may also implement an FE to access subscription information relevant for policy decisions in a UDR of the UDM 827. The PCF 826 may communicate with the AMF 821 via an N15 reference point between the PCF 826 and the AMF 821, which may include a PCF 826 in a visited network and the AMF 821 in case of roaming scenarios. The PCF 826 may communicate with the AF 828 via an N5 reference point between the PCF 826 and the AF 828; and with the SMF 824 via an N7 reference point between the PCF 826 and the SMF 824. The system 800 and/or CN 820 may also include an N24 reference point between the PCF 826 (in the home network) and a PCF 826 in a visited network. Additionally, the PCF 826 may exhibit an Npcf service-based interface.


The AF 828 may provide application influence on traffic routing, provide access to the NCE, and interact with the policy framework for policy control. The NCE may be a mechanism that allows the 5GC 820 and AF 828 to provide information to each other via NEF 823, which may be used for edge computing implementations. In such implementations, the network operator and third party services may be hosted close to the UE 801 access point of attachment to achieve an efficient service delivery through the reduced end-to-end latency and load on the transport network. For edge computing implementations, the 5GC may select a UPF 802 close to the UE 801 and execute traffic steering from the UPF 802 to DN 803 via the N6 interface. This may be based on the UE subscription data, UE location, and information provided by the AF 828. In this way, the AF 828 may influence UPF (re)selection and traffic routing. Based on operator deployment, when AF 828 is considered to be a trusted entity, the network operator may permit AF 828 to interact directly with relevant NFs. Additionally, the AF 828 may exhibit an Naf service-based interface.


The NSSF 829 may select a set of network slice instances serving the UE 801. The NSSF 829 may also determine allowed NSSAI and the mapping to the subscribed S-NSSAIs, if needed. The NSSF 829 may also determine the AMF set to be used to serve the UE 801, or a list of candidate AMF(s) 821 based on a suitable configuration and possibly by querying the NRF 825. The selection of a set of network slice instances for the UE 801 may be triggered by the AMF 821 with which the UE 801 is registered by interacting with the NSSF 829, which may lead to a change of AMF 821. The NSSF 829 may interact with the AMF 821 via an N22 reference point between AMF 821 and NSSF 829; and may communicate with another NSSF 829 in a visited network via an N31 reference point (not shown by FIG. 8). Additionally, the NSSF 829 may exhibit an Nnssf service-based interface.


As discussed previously, the system 800 may include an SMSF, which may be responsible for SMS subscription checking and verification, and relaying SM messages to/from the UE 801 to/from other entities, such as an SMS-GMSC/IWMSC/SMS-router. The SMS may also interact with AMF 821 and UDM 827 for a notification procedure that the UE 801 is available for SMS transfer (e.g., set a UE not reachable flag, and notifying UDM 827 when UE 801 is available for SMS).


The SCP 830 (or individual instances of the SCP 830) supports indirect communication (see e.g., 3GPP TS 23.501 section 7.1.1); delegated discovery (see e.g., 3GPP TS 23.501 section 7.1.1); message forwarding and routing to destination NF/NF service(s), communication security (e.g., authorization of the NF Service Consumer to access the NF Service Producer API) (see e.g., 3GPP TS 33.501), load balancing, monitoring, overload control, etc.; and discovery and selection functionality for UDM(s), AUSF(s), UDR(s), PCF(s) with access to subscription data stored in the UDR based on UE's SUPI, SUCI or GPSI (see e.g., 3GPP TS 23.501 section 6.3). Load balancing, monitoring, overload control functionality provided by the SCP may be implementation specific. The SCP 830 may be deployed in a distributed manner. More than one SCP 830 can be present in the communication path between various NF Services. The SCP 830, although not an NF instance, can also be deployed distributed, redundant, and scalable.


The system architecture 800 may also include other elements that are not shown by FIG. 8A or 8B, such as a Data Storage system/architecture, a 5G-EIR, a SEPP, and the like. The Data Storage system may include a SDSF, an UDSF, and/or the like. Any NF may store and retrieve unstructured data into/from the UDSF (e.g., UE contexts), via N18 reference point between any NF and the UDSF (not shown by FIG. 8). Individual NFs may share a UDSF for storing their respective unstructured data or individual NFs may each have their own UDSF located at or near the individual NFs. Additionally, the UDSF may exhibit an Nudsf service-based interface (not shown by FIG. 8). The 5G-EIR may be an NF that checks the status of PEI for determining whether particular equipment/entities are blacklisted from the network; and the SEPP may be a non-transparent proxy that performs topology hiding, message filtering, and policing on inter-PLMN control plane interfaces.


In another example, the 5G system architecture 800 includes an IP multimedia subsystem (IMS) as well as a plurality of IP multimedia core network subsystem entities, such as call session control functions (CSCFs) (not shown by FIG. 8A or 8B). More specifically, the IMS includes a CSCF, which can act as a proxy CSCF (P-CSCF), a serving CSCF (S-CSCF), an emergency CSCF (E-CSCF), or interrogating CSCF (I-CSCF). The P-CSCF can be configured to be the first contact point for the UE 801 within the IMS. The S-CSCF can be configured to handle the session states in the network, and the E-CSCF can be configured to handle certain aspects of emergency sessions such as routing an emergency request to the correct emergency center or public safety answering point (PSAP). The I-CSCF can be configured to function as the contact point within an operator's network for all IMS connections destined to a subscriber of that network operator, or a roaming subscriber currently located within that network operator's service area. In some aspects, the I-CSCF can be connected to another IP multimedia network, for example, an IMS operated by a different network operator.


Additionally, there may be many more reference points and/or service-based interfaces between the NF services in the NFs; however, these interfaces and reference points have been omitted from FIGS. 8A and 8B for clarity. In one example, the CN 820 may include an Nx interface, which is an inter-CN interface between an MME and the AMF 821 in order to enable interworking between system 800 and an EPC. Other example interfaces/reference points may include an N5g-EIR service-based interface exhibited by a 5G-EIR, an N27 reference point between the NRF in the visited network and the NRF in the home network; and an N31 reference point between the NSSF in the visited network and the NSSF in the home network.



FIG. 9 illustrates an example of infrastructure equipment 900 in accordance with various embodiments. The infrastructure equipment 900 (or “system 900”) may be implemented as a base station, radio head, RAN node such as the RAN nodes 511 and/or AP 506 shown and described previously, application server(s) 530, and/or any other element/device discussed herein. In other examples, the system 900 could be implemented in or by a UE.


The system 900 includes application circuitry 905, baseband circuitry 910, one or more radio front end modules (RFEMs) 915, memory circuitry 920, power management integrated circuitry (PMIC) 925, power tee circuitry 930, network controller circuitry 935, network interface connector 940, satellite positioning circuitry 945, and user interface 950. In some embodiments, the device 900 may include additional elements such as, for example, memory/storage, display, camera, sensor, or input/output (I/O) interface. In other embodiments, the components described below may be included in more than one device. For example, said circuitries may be separately included in more than one device for CRAN, vBBU, or other like implementations.


Application circuitry 905 includes circuitry such as, but not limited to one or more processors (or processor cores), cache memory, and one or more of low drop-out voltage regulators (LDOs), interrupt controllers, serial interfaces such as SPI, I2C or universal programmable serial interface module, real time clock (RTC), timer-counters including interval and watchdog timers, general purpose input/output (I/O or IO), memory card controllers such as Secure Digital (SD) MultiMediaCard (MMC) or similar, Universal Serial Bus (USB) interfaces, Mobile Industry Processor Interface (MIPI) interfaces and Joint Test Access Group (JTAG) test access ports. The processors (or cores) of the application circuitry 905 may be coupled with or may include memory/storage elements and may be configured to execute instructions stored in the memory/storage to enable various applications or operating systems to run on the system 900. In some implementations, the memory/storage elements may be on-chip memory circuitry, which may include any suitable volatile and/or non-volatile memory, such as DRAM, SRAM, EPROM, EEPROM, Flash memory, solid-state memory, and/or any other type of memory device technology, such as those discussed herein.


The processor(s) of application circuitry 905 may include, for example, one or more processor cores (CPUs), one or more application processors, one or more graphics processing units (GPUs), one or more reduced instruction set computing (RISC) processors, one or more Acorn RISC Machine (ARM) processors, one or more complex instruction set computing (CISC) processors, one or more digital signal processors (DSP), one or more FPGAs, one or more PLDs, one or more ASICs, one or more microprocessors or controllers, or any suitable combination thereof. In some embodiments, the application circuitry 905 may comprise, or may be, a special-purpose processor/controller to operate according to the various embodiments herein. As examples, the processor(s) of application circuitry 905 may include one or more Intel Pentium®, Core®, or Xeon® processor(s); Advanced Micro Devices (AMD) Ryzen® processor(s), Accelerated Processing Units (APUs), or Epyc® processors; ARM-based processor(s) licensed from ARM Holdings, Ltd. such as the ARM Cortex-A family of processors and the ThunderX2® provided by Cavium™, Inc.; a MIPS-based design from MIPS Technologies, Inc. such as MIPS Warrior P-class processors; and/or the like. In some embodiments, the system 900 may not utilize application circuitry 905, and instead may include a special-purpose processor/controller to process IP data received from an EPC or SGC, for example.


In some implementations, the application circuitry 905 may include one or more hardware accelerators, which may be microprocessors, programmable processing devices, or the like. The one or more hardware accelerators may include, for example, computer vision (CV) and/or deep learning (DL) accelerators. As examples, the programmable processing devices may be one or more a field-programmable devices (FPDs) such as field-programmable gate arrays (FPGAs) and the like; programmable logic devices (PLDs) such as complex PLDs (CPLDs), high-capacity PLDs (HCPLDs), and the like; ASICs such as structured ASICs and the like; programmable SoCs (PSoCs); and the like. In such implementations, the circuitry of application circuitry 905 may comprise logic blocks or logic fabric, and other interconnected resources that may be programmed to perform various functions, such as the procedures, methods, functions, etc. of the various embodiments discussed herein. In such embodiments, the circuitry of application circuitry 905 may include memory cells (e.g., erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), flash memory, static memory (e.g., static random access memory (SRAM), anti-fuses, etc.)) used to store logic blocks, logic fabric, data, etc. in look-up-tables (LUTs) and the like.


The baseband circuitry 910 may be implemented, for example, as a solder-down substrate including one or more integrated circuits, a single packaged integrated circuit soldered to a main circuit board or a multi-chip module containing two or more integrated circuits. The various hardware electronic elements of baseband circuitry 910 are discussed infra with regard to FIG. 11.


User interface circuitry 950 may include one or more user interfaces designed to enable user interaction with the system 900 or peripheral component interfaces designed to enable peripheral component interaction with the system 900. User interfaces may include, but are not limited to, one or more physical or virtual buttons (e.g., a reset button), one or more indicators (e.g., light emitting diodes (LEDs)), a physical keyboard or keypad, a mouse, a touchpad, a touchscreen, speakers or other audio emitting devices, microphones, a printer, a scanner, a headset, a display screen or display device, etc. Peripheral component interfaces may include, but are not limited to, a nonvolatile memory port, a universal serial bus (USB) port, an audio jack, a power supply interface, etc.


The radio front end modules (RFEMs) 915 may comprise a millimeter wave (mmWave) RFEM and one or more sub-mmWave radio frequency integrated circuits (RFICs). In some implementations, the one or more sub-mmWave RFICs may be physically separated from the mmWave RFEM. The RFICs may include connections to one or more antennas or antenna arrays (see e.g., antenna array 1111 of FIG. 11 infra), and the RFEM may be connected to multiple antennas. In alternative implementations, both mmWave and sub-mmWave radio functions may be implemented in the same physical RFEM 915, which incorporates both mmWave antennas and sub-mmWave.


The memory circuitry 920 may include one or more of volatile memory including dynamic random access memory (DRAM) and/or synchronous dynamic random access memory (SDRAM), and nonvolatile memory (NVM) including high-speed electrically erasable memory (commonly referred to as Flash memory), phase change random access memory (PRAM), magnetoresistive random access memory (MRAM), etc., and may incorporate the three-dimensional (3D) cross-point (XPOINT) memories from Intel® and Micron®. Memory circuitry 920 may be implemented as one or more of solder down packaged integrated circuits, socketed memory modules and plug-in memory cards.


The PMIC 925 may include voltage regulators, surge protectors, power alarm detection circuitry, and one or more backup power sources such as a battery or capacitor. The power alarm detection circuitry may detect one or more of brown out (under-voltage) and surge (over-voltage) conditions. The power tee circuitry 930 may provide for electrical power drawn from a network cable to provide both power supply and data connectivity to the infrastructure equipment 900 using a single cable.


The network controller circuitry 935 may provide connectivity to a network using a standard network interface protocol such as Ethernet, Ethernet over GRE Tunnels, Ethernet over Multiprotocol Label Switching (MPLS), or some other suitable protocol. Network connectivity may be provided to/from the infrastructure equipment 900 via network interface connector 940 using a physical connection, which may be electrical (commonly referred to as a “copper interconnect”), optical, or wireless. The network controller circuitry 935 may include one or more dedicated processors and/or FPGAs to communicate using one or more of the aforementioned protocols. In some implementations, the network controller circuitry 935 may include multiple controllers to provide connectivity to other networks using the same or different protocols.


The positioning circuitry 945 includes circuitry to receive and decode signals transmitted/broadcasted by a positioning network of a global navigation satellite system (GNSS). Examples of navigation satellite constellations (or GNSS) include United States' Global Positioning System (GPS), Russia's Global Navigation System (GLONASS), the European Union's Galileo system, China's BeiDou Navigation Satellite System, a regional navigation system or GNSS augmentation system (e.g., Navigation with Indian Constellation (NAVIC), Japan's Quasi-Zenith Satellite System (QZSS), France's Doppler Orbitography and Radio-positioning Integrated by Satellite (DORIS), etc.), or the like. The positioning circuitry 945 comprises various hardware elements (e.g., including hardware devices such as switches, filters, amplifiers, antenna elements, and the like to facilitate OTA communications) to communicate with components of a positioning network, such as navigation satellite constellation nodes. In some embodiments, the positioning circuitry 945 may include a Micro-Technology for Positioning, Navigation, and Timing (Micro-PNT) IC that uses a master timing clock to perform position tracking/estimation without GNSS assistance. The positioning circuitry 945 may also be part of, or interact with, the baseband circuitry 910 and/or RFEMs 915 to communicate with the nodes and components of the positioning network. The positioning circuitry 945 may also provide position data and/or time data to the application circuitry 905, which may use the data to synchronize operations with various infrastructure (e.g., RAN nodes 511, etc.), or the like.


The components shown by FIG. 9 may communicate with one another using interface circuitry, which may include any number of bus and/or IX technologies such as ISA, extended ISA, I2C, SPI, point-to-point interfaces, power management bus (PMBus), PCI, PCIe, PCIx, Intel® UPI, Intel® Accelerator Link, Intel® CXL, CAPI, OpenCAPI, Intel® QPI, UPI, Intel® OPA IX, RapidIO™ system IXs, CCIX, Gen-Z Consortium IXs, a HyperTransport interconnect, NVLink provided by NVIDIA®, and/or any number of other IX technologies. The IX technology may be a proprietary bus, for example, used in an SoC based system.



FIG. 10 illustrates an example of a platform 1000 (or “device 1000”) in accordance with various embodiments. In embodiments, the computer platform 1000 may be suitable for use as UEs 501, 701, 801, application servers 530, and/or any other element/device discussed herein. The platform 1000 may include any combinations of the components shown in the example. The components of platform 1000 may be implemented as integrated circuits (ICs), portions thereof, discrete electronic devices, or other modules, logic, hardware, software, firmware, or a combination thereof adapted in the computer platform 1000, or as components otherwise incorporated within a chassis of a larger system. The block diagram of FIG. 10 is intended to show a high level view of components of the computer platform 1000. However, some of the components shown may be omitted, additional components may be present, and different arrangement of the components shown may occur in other implementations.


Application circuitry 1005 includes circuitry such as, but not limited to one or more processors (or processor cores), cache memory, and one or more of LDOs, interrupt controllers, serial interfaces such as SPI, I2C or universal programmable serial interface module, RTC, timer-counters including interval and watchdog timers, general purpose I/O, memory card controllers such as SD MMC or similar, USB interfaces, MIPI interfaces, and JTAG test access ports. The processors (or cores) of the application circuitry 1005 may be coupled with or may include memory/storage elements and may be configured to execute instructions stored in the memory/storage to enable various applications or operating systems to run on the system 1000. In some implementations, the memory/storage elements may be on-chip memory circuitry, which may include any suitable volatile and/or non-volatile memory, such as DRAM, SRAM, EPROM, EEPROM, Flash memory, solid-state memory, and/or any other type of memory device technology, such as those discussed herein.


The processor(s) of application circuitry 905 may include, for example, one or more processor cores, one or more application processors, one or more GPUs, one or more RISC processors, one or more ARM processors, one or more CISC processors, one or more DSP, one or more FPGAs, one or more PLDs, one or more ASICs, one or more microprocessors or controllers, a multithreaded processor, an ultra-low voltage processor, an embedded processor, some other known processing element, or any suitable combination thereof. In some embodiments, the application circuitry 905 may comprise, or may be, a special-purpose processor/controller to operate according to the various embodiments herein.


As examples, the processor(s) of application circuitry 1005 may include an Intel® Architecture Core™ based processor, such as a Quark™, an Atom™, an i3, an i5, an i7, or an MCU-class processor, or another such processor available from Intel® Corporation, Santa Clara, Calif. The processors of the application circuitry 1005 may also be one or more of Advanced Micro Devices (AMD) Ryzen® processor(s) or Accelerated Processing Units (APUs); A5-A9 processor(s) from Apple® Inc., Snapdragon™ processor(s) from Qualcomm® Technologies, Inc., Texas Instruments, Inc.® Open Multimedia Applications Platform (OMAP)™ processor(s); a MIPS-based design from MIPS Technologies, Inc. such as MIPS Warrior M-class, Warrior I-class, and Warrior P-class processors; an ARM-based design licensed from ARM Holdings, Ltd., such as the ARM Cortex-A, Cortex-R, and Cortex-M family of processors; or the like. In some implementations, the application circuitry 1005 may be a part of a system on a chip (SoC) in which the application circuitry 1005 and other components are formed into a single integrated circuit, or a single package, such as the Edison™ or Galileo™ SoC boards from Intel® Corporation.


Additionally or alternatively, application circuitry 1005 may include circuitry such as, but not limited to, one or more a field-programmable devices (FPDs) such as FPGAs and the like; programmable logic devices (PLDs) such as complex PLDs (CPLDs), high-capacity PLDs (HCPLDs), and the like; ASICs such as structured ASICs and the like; programmable SoCs (PSoCs); and the like. In such embodiments, the circuitry of application circuitry 1005 may comprise logic blocks or logic fabric, and other interconnected resources that may be programmed to perform various functions, such as the procedures, methods, functions, etc. of the various embodiments discussed herein. In such embodiments, the circuitry of application circuitry 1005 may include memory cells (e.g., erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), flash memory, static memory (e.g., static random access memory (SRAM), anti-fuses, etc.)) used to store logic blocks, logic fabric, data, etc. in look-up tables (LUTs) and the like.


The baseband circuitry 1010 may be implemented, for example, as a solder-down substrate including one or more integrated circuits, a single packaged integrated circuit soldered to a main circuit board or a multi-chip module containing two or more integrated circuits. The various hardware electronic elements of baseband circuitry 1010 are discussed infra with regard to FIG. 11.


The RFEMs 1015 may comprise a millimeter wave (mmWave) RFEM and one or more sub-mmWave radio frequency integrated circuits (RFICs). In some implementations, the one or more sub-mmWave RFICs may be physically separated from the mmWave RFEM. The RFICs may include connections to one or more antennas or antenna arrays (see e.g., antenna array 1111 of FIG. 11 infra), and the RFEM may be connected to multiple antennas. In alternative implementations, both mmWave and sub-mmWave radio functions may be implemented in the same physical RFEM 1015, which incorporates both mmWave antennas and sub-mmWave.


The memory circuitry 1020 may include any number and type of memory devices used to provide for a given amount of system memory. As examples, the memory circuitry 1020 may include one or more of volatile memory including random access memory (RAM), dynamic RAM (DRAM) and/or synchronous dynamic RAM (SDRAM), and nonvolatile memory (NVM) including high-speed electrically erasable memory (commonly referred to as Flash memory), phase change random access memory (PRAM), magnetoresistive random access memory (MRAM), etc. The memory circuitry 1020 may be developed in accordance with a Joint Electron Devices Engineering Council (JEDEC) low power double data rate (LPDDR)-based design, such as LPDDR2, LPDDR3, LPDDR4, or the like. Memory circuitry 1020 may be implemented as one or more of solder down packaged integrated circuits, single die package (SDP), dual die package (DDP) or quad die package (Q17P), socketed memory modules, dual inline memory modules (DIMMs) including microDIMMs or MiniDIMMs, and/or soldered onto a motherboard via a ball grid array (BGA). In low power implementations, the memory circuitry 1020 may be on-die memory or registers associated with the application circuitry 1005. To provide for persistent storage of information such as data, applications, operating systems and so forth, memory circuitry 1020 may include one or more mass storage devices, which may include, inter alia, a solid state disk drive (SSDD), hard disk drive (HDD), a micro HDD, resistance change memories, phase change memories, holographic memories, or chemical memories, among others. For example, the computer platform 1000 may incorporate the three-dimensional (3D) cross-point (XPOINT) memories from Intel® and Micron®.


Removable memory circuitry 1023 may include devices, circuitry, enclosures/housings, ports or receptacles, etc. used to couple portable data storage devices with the platform 1000. These portable data storage devices may be used for mass storage purposes, and may include, for example, flash memory cards (e.g., Secure Digital (SD) cards, microSD cards, xD picture cards, and the like), and USB flash drives, optical discs, external HDDs, and the like.


The platform 1000 may also include interface circuitry (not shown) that is used to connect external devices with the platform 1000. The external devices connected to the platform 1000 via the interface circuitry include sensor circuitry 1021 and electro-mechanical components (EMCs) 1022, as well as removable memory devices coupled to removable memory circuitry 1023.


The sensor circuitry 1021 include devices, modules, or subsystems whose purpose is to detect events or changes in its environment and send the information (sensor data) about the detected events to some other a device, module, subsystem, etc. Examples of such sensors include, inter alia, inertia measurement units (IMUS) comprising accelerometers, gyroscopes, and/or magnetometers; microelectromechanical systems (MEMS) or nanoelectromechanical systems (NEMS) comprising 3-axis accelerometers, 3-axis gyroscopes, and/or magnetometers; level sensors; flow sensors; temperature sensors (e.g., thermistors); pressure sensors; barometric pressure sensors; gravimeters; altimeters; image capture devices (e.g., cameras or lensless apertures); light detection and ranging (LiDAR) sensors; proximity sensors (e.g., infrared radiation detector and the like), depth sensors, ambient light sensors, ultrasonic transceivers; microphones or other like audio capture devices; etc.


EMCs 1022 include devices, modules, or subsystems whose purpose is to enable platform 1000 to change its state, position, and/or orientation, or move or control a mechanism or (sub)system. Additionally, EMCs 1022 may be configured to generate and send messages/signalling to other components of the platform 1000 to indicate a current state of the EMCs 1022. Examples of the EMCs 1022 include one or more power switches, relays including electromechanical relays (EMRs) and/or solid state relays (SSRs), actuators (e.g., valve actuators, etc.), an audible sound generator, a visual warning device, motors (e.g., DC motors, stepper motors, etc.), wheels, thrusters, propellers, claws, clamps, hooks, and/or other like electro-mechanical components. In embodiments, platform 1000 is configured to operate one or more EMCs 1022 based on one or more captured events and/or instructions or control signals received from a service provider and/or various clients.


In some implementations, the interface circuitry may connect the platform 1000 with positioning circuitry 1045. The positioning circuitry 1045 includes circuitry to receive and decode signals transmitted/broadcasted by a positioning network of a GNSS. Examples of navigation satellite constellations (or GNSS) include United States' GPS, Russia's GLONASS, the European Union's Galileo system, China's BeiDou Navigation Satellite System, a regional navigation system or GNSS augmentation system (e.g., NAVIC), Japan's QZSS, France's DORIS, etc.), or the like. The positioning circuitry 1045 comprises various hardware elements (e.g., including hardware devices such as switches, filters, amplifiers, antenna elements, and the like to facilitate OTA communications) to communicate with components of a positioning network, such as navigation satellite constellation nodes. In some embodiments, the positioning circuitry 1045 may include a Micro-PNT IC that uses a master timing clock to perform position tracking/estimation without GNSS assistance. The positioning circuitry 1045 may also be part of, or interact with, the baseband circuitry 910 and/or RFEMs 1015 to communicate with the nodes and components of the positioning network. The positioning circuitry 1045 may also provide position data and/or time data to the application circuitry 1005, which may use the data to synchronize operations with various infrastructure (e.g., radio base stations), for turn-by-turn navigation applications, or the like


In some implementations, the interface circuitry may connect the platform 1000 with Near-Field Communication (NFC) circuitry 1040. NFC circuitry 1040 is configured to provide contactless, short-range communications based on radio frequency identification (RFID) standards, wherein magnetic field induction is used to enable communication between NFC circuitry 1040 and NFC-enabled devices external to the platform 1000 (e.g., an “NFC touchpoint”). NFC circuitry 1040 comprises an NFC controller coupled with an antenna element and a processor coupled with the NFC controller. The NFC controller may be a chip/IC providing NFC functionalities to the NFC circuitry 1040 by executing NFC controller firmware and an NFC stack. The NFC stack may be executed by the processor to control the NFC controller, and the NFC controller firmware may be executed by the NFC controller to control the antenna element to emit short-range RF signals. The RF signals may power a passive NFC tag (e.g., a microchip embedded in a sticker or wristband) to transmit stored data to the NFC circuitry 1040, or initiate data transfer between the NFC circuitry 1040 and another active NFC device (e.g., a smartphone or an NFC-enabled POS terminal) that is proximate to the platform 1000.


The driver circuitry 1046 may include software and hardware elements that operate to control particular devices that are embedded in the platform 1000, attached to the platform 1000, or otherwise communicatively coupled with the platform 1000. The driver circuitry 1046 may include individual drivers allowing other components of the platform 1000 to interact with or control various input/output (I/O) devices that may be present within, or connected to, the platform 1000. For example, driver circuitry 1046 may include a display driver to control and allow access to a display device, a touchscreen driver to control and allow access to a touchscreen interface of the platform 1000, sensor drivers to obtain sensor readings of sensor circuitry 1021 and control and allow access to sensor circuitry 1021, EMC drivers to obtain actuator positions of the EMCs 1022 and/or control and allow access to the EMCs 1022, a camera driver to control and allow access to an embedded image capture device, audio drivers to control and allow access to one or more audio devices.


The power management integrated circuitry (PMIC) 1025 (also referred to as “power management circuitry 1025”) may manage power provided to various components of the platform 1000. In particular, with respect to the baseband circuitry 1010, the PMIC 1025 may control power-source selection, voltage scaling, battery charging, or DC-to-DC conversion. The PMIC 1025 may often be included when the platform 1000 is capable of being powered by a battery 1030, for example, when the device is included in a UE 501, 701, 801.


In some embodiments, the PMIC 1025 may control, or otherwise be part of, various power saving mechanisms of the platform 1000. For example, if the platform 1000 is in an RRC_Connected state, where it is still connected to the RAN node as it expects to receive traffic shortly, then it may enter a state known as Discontinuous Reception Mode (DRX) after a period of inactivity. During this state, the platform 1000 may power down for brief intervals of time and thus save power. If there is no data traffic activity for an extended period of time, then the platform 1000 may transition off to an RRC_Idle state, where it disconnects from the network and does not perform operations such as channel quality feedback, handover, etc. The platform 1000 goes into a very low power state and it performs paging where again it periodically wakes up to listen to the network and then powers down again. The platform 1000 may not receive data in this state; in order to receive data, it must transition back to RRC_Connected state. An additional power saving mode may allow a device to be unavailable to the network for periods longer than a paging interval (ranging from seconds to a few hours). During this time, the device is totally unreachable to the network and may power down completely. Any data sent during this time incurs a large delay and it is assumed the delay is acceptable.


A battery 1030 may power the platform 1000, although in some examples the platform 1000 may be mounted deployed in a fixed location, and may have a power supply coupled to an electrical grid. The battery 1030 may be a lithium ion battery, a metal-air battery, such as a zinc-air battery, an aluminum-air battery, a lithium-air battery, and the like. In some implementations, such as in V2X applications, the battery 1030 may be a typical lead-acid automotive battery.


In some implementations, the battery 1030 may be a “smart battery,” which includes or is coupled with a Battery Management System (BMS) or battery monitoring integrated circuitry. The BMS may be included in the platform 1000 to track the state of charge (SoCh) of the battery 1030. The BMS may be used to monitor other parameters of the battery 1030 to provide failure predictions, such as the state of health (SoH) and the state of function (SoF) of the battery 1030. The BMS may communicate the information of the battery 1030 to the application circuitry 1005 or other components of the platform 1000. The BMS may also include an analog-to-digital (ADC) convertor that allows the application circuitry 1005 to directly monitor the voltage of the battery 1030 or the current flow from the battery 1030. The battery parameters may be used to determine actions that the platform 1000 may perform, such as transmission frequency, network operation, sensing frequency, and the like.


A power block, or other power supply coupled to an electrical grid may be coupled with the BMS to charge the battery 1030. In some examples, the power block XS30 may be replaced with a wireless power receiver to obtain the power wirelessly, for example, through a loop antenna in the computer platform 1000. In these examples, a wireless battery charging circuit may be included in the BMS. The specific charging circuits chosen may depend on the size of the battery 1030, and thus, the current required. The charging may be performed using the Airfuel standard promulgated by the Airfuel Alliance, the Qi wireless charging standard promulgated by the Wireless Power Consortium, or the Rezence charging standard promulgated by the Alliance for Wireless Power, among others.


User interface circuitry 1050 includes various input/output (I/O) devices present within, or connected to, the platform 1000, and includes one or more user interfaces designed to enable user interaction with the platform 1000 and/or peripheral component interfaces designed to enable peripheral component interaction with the platform 1000. The user interface circuitry 1050 includes input device circuitry and output device circuitry. Input device circuitry includes any physical or virtual means for accepting an input including, inter alia, one or more physical or virtual buttons (e.g., a reset button), a physical keyboard, keypad, mouse, touchpad, touchscreen, microphones, scanner, headset, and/or the like. The output device circuitry includes any physical or virtual means for showing information or otherwise conveying information, such as sensor readings, actuator position(s), or other like information. Output device circuitry may include any number and/or combinations of audio or visual display, including, inter alia, one or more simple visual outputs/indicators (e.g., binary status indicators (e.g., light emitting diodes (LEDs)) and multi-character visual outputs, or more complex outputs such as display devices or touchscreens (e.g., Liquid Chrystal Displays (LCD), LED displays, quantum dot displays, projectors, etc.), with the output of characters, graphics, multimedia objects, and the like being generated or produced from the operation of the platform 1000. The output device circuitry may also include speakers or other audio emitting devices, printer(s), and/or the like. In some embodiments, the sensor circuitry 1021 may be used as the input device circuitry (e.g., an image capture device, motion capture device, or the like) and one or more EMCs may be used as the output device circuitry (e.g., an actuator to provide haptic feedback or the like). In another example, NFC circuitry comprising an NFC controller coupled with an antenna element and a processing device may be included to read electronic tags and/or connect with another NFC-enabled device. Peripheral component interfaces may include, but are not limited to, a non-volatile memory port, a USB port, an audio jack, a power supply interface, etc.


Although not shown, the components of platform 1000 may communicate with one another using a suitable bus or interconnect (IX) technology, which may include any number of technologies, including ISA, extended ISA, I2C, SPI, point-to-point interfaces, power management bus (PMBus), PCI, PCIe, PCIx, Intel® UPI, Intel® Accelerator Link, Intel® CXL, CAPI, OpenCAPI, Intel® QPI, UPI, Intel® OPA IX, RapidIO™ system IXs, CCIX, Gen-Z Consortium IXs, a HyperTransport interconnect, NVLink provided by NVIDIA®, a Time-Trigger Protocol (TTP) system, a FlexRay system, and/or any number of other IX technologies. The IX 1006 may be a proprietary bus, for example, used in a SoC based system.



FIG. 11 illustrates example components of baseband circuitry 1110 and radio front end modules (RFEM) 1115 in accordance with various embodiments. The baseband circuitry 1110 corresponds to the baseband circuitry 910 and 1010 of FIGS. 9 and 10, respectively. The RFEM 1115 corresponds to the RFEM 915 and 1015 of FIGS. 9 and 10, respectively. As shown, the RFEMs 1115 may include Radio Frequency (RF) circuitry 1106, front-end module (FEM) circuitry 1108, antenna array 1111 coupled together at least as shown.


The baseband circuitry 1110 includes circuitry and/or control logic configured to carry out various radio/network protocol and radio control functions that enable communication with one or more radio networks via the RF circuitry 1106. The radio control functions may include, but are not limited to, signal modulation/demodulation, encoding/decoding, radio frequency shifting, etc. In some embodiments, modulation/demodulation circuitry of the baseband circuitry 1110 may include Fast-Fourier Transform (FFT), precoding, or constellation mapping/demapping functionality. In some embodiments, encoding/decoding circuitry of the baseband circuitry 1110 may include convolution, tail-biting convolution, turbo, Viterbi, or Low Density Parity Check (LDPC) encoder/decoder functionality. Embodiments of modulation/demodulation and encoder/decoder functionality are not limited to these examples and may include other suitable functionality in other embodiments. The baseband circuitry 1110 is configured to process baseband signals received from a receive signal path of the RF circuitry 1106 and to generate baseband signals for a transmit signal path of the RF circuitry 1106. The baseband circuitry 1110 is configured to interface with application circuitry 905/1005 (see FIGS. 9 and 10) for generation and processing of the baseband signals and for controlling operations of the RF circuitry 1106. The baseband circuitry 1110 may handle various radio control functions.


The aforementioned circuitry and/or control logic of the baseband circuitry 1110 may include one or more single or multi-core processors. For example, the one or more processors may include a 3G baseband processor 1104A, a 4G/LTE baseband processor 1104B, a 5G/NR baseband processor 1104C, or some other baseband processor(s) 1104D for other existing generations, generations in development or to be developed in the future (e.g., sixth generation (6G), etc.). In other embodiments, some or all of the functionality of baseband processors 1104A-D may be included in modules stored in the memory 1104G and executed via a Central Processing Unit (CPU) 1104E. In other embodiments, some or all of the functionality of baseband processors 1104A-D may be provided as hardware accelerators (e.g., FPGAs, ASICs, etc.) loaded with the appropriate bit streams or logic blocks stored in respective memory cells. In various embodiments, the memory 1104G may store program code of a real-time OS (RTOS), which when executed by the CPU 1104E (or other baseband processor), is to cause the CPU 1104E (or other baseband processor) to manage resources of the baseband circuitry 1110, schedule tasks, etc. Examples of the RTOS may include Operating System Embedded (OSE)™ provided by Enea®, Nucleus RTOS™ provided by Mentor Graphics®, Versatile Real-Time Executive (VRTX) provided by Mentor Graphics®, ThreadX™ provided by Express Logic®, FreeRTOS, REX OS provided by Qualcomm®, OKL4 provided by Open Kernel (OK) Labs®, or any other suitable RTOS, such as those discussed herein. In addition, the baseband circuitry 1110 includes one or more audio digital signal processor(s) (DSP) 1104F. The audio DSP(s) 1104F include elements for compression/decompression and echo cancellation and may include other suitable processing elements in other embodiments.


In some embodiments, each of the processors 1104A-1104E include respective memory interfaces to send/receive data to/from the memory 1104G. The baseband circuitry 1110 may further include one or more interfaces to communicatively couple to other circuitries/devices, such as an interface to send/receive data to/from memory external to the baseband circuitry 1110; an application circuitry interface to send/receive data to/from the application circuitry 905/1005 of FIG. 9-XT); an RF circuitry interface to send/receive data to/from RF circuitry 1106 of FIG. 11; a wireless hardware connectivity interface to send/receive data to/from one or more wireless hardware elements (e.g., Near Field Communication (NFC) components, Bluetooth®/Bluetooth® Low Energy components, Wi-Fi® components, and/or the like); and a power management interface to send/receive power or control signals to/from the PMIC 1025.


In alternate embodiments (which may be combined with the above described embodiments), baseband circuitry 1110 comprises one or more digital baseband systems, which are coupled with one another via an interconnect subsystem and to a CPU subsystem, an audio subsystem, and an interface subsystem. The digital baseband subsystems may also be coupled to a digital baseband interface and a mixed-signal baseband subsystem via another interconnect subsystem. Each of the interconnect subsystems may include a bus system, point-to-point connections, network-on-chip (NOC) structures, and/or some other suitable bus or interconnect technology, such as those discussed herein. The audio subsystem may include DSP circuitry, buffer memory, program memory, speech processing accelerator circuitry, data converter circuitry such as analog-to-digital and digital-to-analog converter circuitry, analog circuitry including one or more of amplifiers and filters, and/or other like components. In an aspect of the present disclosure, baseband circuitry 1110 may include protocol processing circuitry with one or more instances of control circuitry (not shown) to provide control functions for the digital baseband circuitry and/or radio frequency circuitry (e.g., the radio front end modules 1115).


Although not shown by FIG. 11, in some embodiments, the baseband circuitry 1110 includes individual processing device(s) to operate one or more wireless communication protocols (e.g., a “multi-protocol baseband processor” or “protocol processing circuitry”) and individual processing device(s) to implement PHY layer functions. In these embodiments, the PHY layer functions include the aforementioned radio control functions. In these embodiments, the protocol processing circuitry operates or implements various protocol layers/entities of one or more wireless communication protocols. In a first example, the protocol processing circuitry may operate LTE protocol entities and/or 5G/NR protocol entities when the baseband circuitry 1110 and/or RF circuitry 1106 are part of mmWave communication circuitry or some other suitable cellular communication circuitry. In the first example, the protocol processing circuitry would operate MAC, RLC, PDCP, SDAP, RRC, and NAS functions. In a second example, the protocol processing circuitry may operate one or more IEEE-based protocols when the baseband circuitry 1110 and/or RF circuitry 1106 are part of a Wi-Fi communication system. In the second example, the protocol processing circuitry would operate Wi-Fi MAC and logical link control (LLC) functions. The protocol processing circuitry may include one or more memory structures (e.g., 1104G) to store program code and data for operating the protocol functions, as well as one or more processing cores to execute the program code and perform various operations using the data. The baseband circuitry 1110 may also support radio communications for more than one wireless protocol.


The various hardware elements of the baseband circuitry 1110 discussed herein may be implemented, for example, as a solder-down substrate including one or more integrated circuits (ICs), a single packaged IC soldered to a main circuit board or a multi-chip module containing two or more ICs. In one example, the components of the baseband circuitry 1110 may be suitably combined in a single chip or chipset, or disposed on a same circuit board. In another example, some or all of the constituent components of the baseband circuitry 1110 and RF circuitry 1106 may be implemented together such as, for example, a system on a chip (SoC) or System-in-Package (SiP). In another example, some or all of the constituent components of the baseband circuitry 1110 may be implemented as a separate SoC that is communicatively coupled with and RF circuitry 1106 (or multiple instances of RF circuitry 1106). In yet another example, some or all of the constituent components of the baseband circuitry 1110 and the application circuitry 905/1005 may be implemented together as individual SoCs mounted to a same circuit board (e.g., a “multi-chip package”).


In some embodiments, the baseband circuitry 1110 may provide for communication compatible with one or more radio technologies. For example, in some embodiments, the baseband circuitry 1110 may support communication with an E-UTRAN or other WMAN, a WLAN, a WPAN. Embodiments in which the baseband circuitry 1110 is configured to support radio communications of more than one wireless protocol may be referred to as multi-mode baseband circuitry.


RF circuitry 1106 may enable communication with wireless networks using modulated electromagnetic radiation through a non-solid medium. In various embodiments, the RF circuitry 1106 may include switches, filters, amplifiers, etc. to facilitate the communication with the wireless network. RF circuitry 1106 may include a receive signal path, which may include circuitry to down-convert RF signals received from the FEM circuitry 1108 and provide baseband signals to the baseband circuitry 1110. RF circuitry 1106 may also include a transmit signal path, which may include circuitry to up-convert baseband signals provided by the baseband circuitry 1110 and provide RF output signals to the FEM circuitry 1108 for transmission.


In some embodiments, the receive signal path of the RF circuitry 1106 may include mixer circuitry 1106a, amplifier circuitry 1106b and filter circuitry 1106c. In some embodiments, the transmit signal path of the RF circuitry 1106 may include filter circuitry 1106c and mixer circuitry 1106a. RF circuitry 1106 may also include synthesizer circuitry 1106d for synthesizing a frequency for use by the mixer circuitry 1106a of the receive signal path and the transmit signal path. In some embodiments, the mixer circuitry 1106a of the receive signal path may be configured to down-convert RF signals received from the FEM circuitry 1108 based on the synthesized frequency provided by synthesizer circuitry 1106d. The amplifier circuitry 1106b may be configured to amplify the down-converted signals and the filter circuitry 1106c may be a low-pass filter (LPF) or band-pass filter (BPF) configured to remove unwanted signals from the down-converted signals to generate output baseband signals. Output baseband signals may be provided to the baseband circuitry 1110 for further processing. In some embodiments, the output baseband signals may be zero-frequency baseband signals, although this is not a requirement. In some embodiments, mixer circuitry 1106a of the receive signal path may comprise passive mixers, although the scope of the embodiments is not limited in this respect.


In some embodiments, the mixer circuitry 1106a of the transmit signal path may be configured to up-convert input baseband signals based on the synthesized frequency provided by the synthesizer circuitry 1106d to generate RF output signals for the FEM circuitry 1108. The baseband signals may be provided by the baseband circuitry 1110 and may be filtered by filter circuitry 1106c.


In some embodiments, the mixer circuitry 1106a of the receive signal path and the mixer circuitry 1106a of the transmit signal path may include two or more mixers and may be arranged for quadrature downconversion and upconversion, respectively. In some embodiments, the mixer circuitry 1106a of the receive signal path and the mixer circuitry 1106a of the transmit signal path may include two or more mixers and may be arranged for image rejection (e.g., Hartley image rejection). In some embodiments, the mixer circuitry 1106a of the receive signal path and the mixer circuitry 1106a of the transmit signal path may be arranged for direct downconversion and direct upconversion, respectively. In some embodiments, the mixer circuitry 1106a of the receive signal path and the mixer circuitry 1106a of the transmit signal path may be configured for super-heterodyne operation.


In some embodiments, the output baseband signals and the input baseband signals may be analog baseband signals, although the scope of the embodiments is not limited in this respect. In some alternate embodiments, the output baseband signals and the input baseband signals may be digital baseband signals. In these alternate embodiments, the RF circuitry 1106 may include analog-to-digital converter (ADC) and digital-to-analog converter (DAC) circuitry and the baseband circuitry 1110 may include a digital baseband interface to communicate with the RF circuitry 1106.


In some dual-mode embodiments, a separate radio IC circuitry may be provided for processing signals for each spectrum, although the scope of the embodiments is not limited in this respect.


In some embodiments, the synthesizer circuitry 1106d may be a fractional-N synthesizer or a fractional N/N+1 synthesizer, although the scope of the embodiments is not limited in this respect as other types of frequency synthesizers may be suitable. For example, synthesizer circuitry 1106d may be a delta-sigma synthesizer, a frequency multiplier, or a synthesizer comprising a phase-locked loop with a frequency divider.


The synthesizer circuitry 1106d may be configured to synthesize an output frequency for use by the mixer circuitry 1106a of the RF circuitry 1106 based on a frequency input and a divider control input. In some embodiments, the synthesizer circuitry 1106d may be a fractional N/N+1 synthesizer.


In some embodiments, frequency input may be provided by a voltage controlled oscillator (VCO), although that is not a requirement. Divider control input may be provided by either the baseband circuitry 1110 or the application circuitry 905/1005 depending on the desired output frequency. In some embodiments, a divider control input (e.g., N) may be determined from a look-up table based on a channel indicated by the application circuitry 905/1005.


Synthesizer circuitry 1106d of the RF circuitry 1106 may include a divider, a delay-locked loop (DLL), a multiplexer and a phase accumulator. In some embodiments, the divider may be a dual modulus divider (DMD) and the phase accumulator may be a digital phase accumulator (DPA). In some embodiments, the DMD may be configured to divide the input signal by either N or N+1 (e.g., based on a carry out) to provide a fractional division ratio. In some example embodiments, the DLL may include a set of cascaded, tunable, delay elements, a phase detector, a charge pump and a D-type flip-flop. In these embodiments, the delay elements may be configured to break a VCO period up into Nd equal packets of phase, where Nd is the number of delay elements in the delay line. In this way, the DLL provides negative feedback to help ensure that the total delay through the delay line is one VCO cycle.


In some embodiments, synthesizer circuitry 1106d may be configured to generate a carrier frequency as the output frequency, while in other embodiments, the output frequency may be a multiple of the carrier frequency (e.g., twice the carrier frequency, four times the carrier frequency) and used in conjunction with quadrature generator and divider circuitry to generate multiple signals at the carrier frequency with multiple different phases with respect to each other. In some embodiments, the output frequency may be a LO frequency (fLO). In some embodiments, the RF circuitry 1106 may include an IQ/polar converter.


FEM circuitry 1108 may include a receive signal path, which may include circuitry configured to operate on RF signals received from antenna array 1111, amplify the received signals and provide the amplified versions of the received signals to the RF circuitry 1106 for further processing. FEM circuitry 1108 may also include a transmit signal path, which may include circuitry configured to amplify signals for transmission provided by the RF circuitry 1106 for transmission by one or more of antenna elements of antenna array 1111. In various embodiments, the amplification through the transmit or receive signal paths may be done solely in the RF circuitry 1106, solely in the FEM circuitry 1108, or in both the RF circuitry 1106 and the FEM circuitry 1108.


In some embodiments, the FEM circuitry 1108 may include a TX/RX switch to switch between transmit mode and receive mode operation. The FEM circuitry 1108 may include a receive signal path and a transmit signal path. The receive signal path of the FEM circuitry 1108 may include an LNA to amplify received RF signals and provide the amplified received RF signals as an output (e.g., to the RF circuitry 1106). The transmit signal path of the FEM circuitry 1108 may include a power amplifier (PA) to amplify input RF signals (e.g., provided by RF circuitry 1106), and one or more filters to generate RF signals for subsequent transmission by one or more antenna elements of the antenna array 1111.


The antenna array 1111 comprises one or more antenna elements, each of which is configured convert electrical signals into radio waves to travel through the air and to convert received radio waves into electrical signals. For example, digital baseband signals provided by the baseband circuitry 1110 is converted into analog RF signals (e.g., modulated waveform) that will be amplified and transmitted via the antenna elements of the antenna array 1111 including one or more antenna elements (not shown). The antenna elements may be omnidirectional, direction, or a combination thereof. The antenna elements may be formed in a multitude of arranges as are known and/or discussed herein. The antenna array 1111 may comprise microstrip antennas or printed antennas that are fabricated on the surface of one or more printed circuit boards. The antenna array 1111 may be formed in as a patch of metal foil (e.g., a patch antenna) in a variety of shapes, and may be coupled with the RF circuitry 1106 and/or FEM circuitry 1108 using metal transmission lines or the like.


Processors of the application circuitry 905/1005 and processors of the baseband circuitry 1110 may be used to execute elements of one or more instances of a protocol stack. For example, processors of the baseband circuitry 1110, alone or in combination, may be used execute Layer 3, Layer 2, or Layer 1 functionality, while processors of the application circuitry 905/1005 may utilize data (e.g., packet data) received from these layers and further execute Layer 4 functionality (e.g., TCP and UDP layers). As referred to herein, Layer 3 may comprise a RRC layer, described in further detail below. As referred to herein, Layer 2 may comprise a MAC layer, an RLC layer, and a PDCP layer, described in further detail below. As referred to herein, Layer 1 may comprise a PHY layer of a UE/RAN node, described in further detail below.


Signalling Radio Bearers (SRBs) are Radio Bearers (RBs) that are used only for the transmission of RRC and NAS messages. More specifically, the following SRBs are defined: SRB0 is for RRC messages using the CCCH logical channel; SRB1 is for RRC messages, which may include a piggybacked NAS message, as well as for NAS messages prior to the establishment of SRB2, all using DCCH logical channel; SRB2 is for NAS messages, all using DCCH logical channel—SRB2 has a lower priority than SRB1 and may be configured by the network after AS security activation; and SRB3 is for specific RRC messages when the UE 501 is in (NG)EN-DC or NR-DC, all using DCCH logical channel.


In the DL, piggybacking of NAS messages is used only for one dependant (e.g., with joint success/failure) procedure: bearer establishment/modification/release. In uplink piggybacking of NAS message is used only for transferring the initial NAS message during connection setup and connection resume. The NAS messages transferred via SRB2 are also contained in RRC messages, which do not include any RRC protocol control information. Once AS security is activated, all RRC messages on SRB1, SRB2 and SRB3, including those containing NAS messages, are integrity protected and ciphered by PDCP. NAS independently applies integrity protection and ciphering to the NAS messages (see e.g., 3GPP TS 24.501). Split SRB is supported for all the MR-DC options in both SRB1 and SRB2 (split SRB is not supported for SRB0 and SRB3).



FIG. 12 is a block diagram illustrating components, according to some example embodiments, able to read instructions from a machine-readable or computer-readable medium (e.g., a non-transitory machine-readable storage medium) and perform any one or more of the methodologies discussed herein. Specifically, FIG. 12 shows a diagrammatic representation of hardware resources 1200 including one or more processors (or processor cores) 1210, one or more memory/storage devices 1220, and one or more communication resources 1230, each of which may be communicatively coupled via a bus 1240. For embodiments where node virtualization (e.g., NFV) is utilized, a hypervisor 1202 may be executed to provide an execution environment for one or more network slices/sub-slices to utilize the hardware resources 1200.


The processors 1210 may include, for example, a processor 1212 and a processor 1214. The processor(s) 1210 may be, for example, a central processing unit (CPU), a reduced instruction set computing (RISC) processor, a complex instruction set computing (CISC) processor, a graphics processing unit (GPU), a DSP such as a baseband processor, an ASIC, an FPGA, a radio-frequency integrated circuit (RFIC), another processor (including those discussed herein), or any suitable combination thereof.


The memory/storage devices 1220 may include main memory, disk storage, or any suitable combination thereof. The memory/storage devices 1220 may include, but are not limited to, any type of volatile or nonvolatile memory such as dynamic random access memory (DRAM), static random access memory (SRAM), erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), Flash memory, solid-state storage, etc.


The communication resources 1230 may include interconnection or network interface components or other suitable devices to communicate with one or more peripheral devices 1204 or one or more databases 1206 via a network 1208. For example, the communication resources 1230 may include wired communication components (e.g., for coupling via USB), cellular communication components, NFC components, Bluetooth® (or Bluetooth® Low Energy) components, Wi-Fi® components, and other communication components.


Instructions 1250 may comprise software, a program, an application, an applet, an app, or other executable code for causing at least any of the processors 1210 to perform any one or more of the methodologies discussed herein. The instructions 1250 may reside, completely or partially, within at least one of the processors 1210 (e.g., within the processor's cache memory), the memory/storage devices 1220, or any suitable combination thereof. Furthermore, any portion of the instructions 1250 may be transferred to the hardware resources 1200 from any combination of the peripheral devices 1204 or the databases 1206. Accordingly, the memory of processors 1210, the memory/storage devices 1220, the peripheral devices 1204, and the databases 1206 are examples of computer-readable and machine-readable media.


In some embodiments, the electronic device(s), network(s), system(s), chip(s) or component(s), or portions or implementations thereof, of FIGS. 5-12, or some other figure herein, may be configured to perform one or more processes, techniques, or methods as described herein, or portions thereof. One such process 300 is depicted in FIG. 3. The process 300 may be for a user equipment (UE) capable of communicating using multi-radio access technology (MR)-dual connectivity (DC). At 302, the process 300 may include detecting a radio link failure (RLF). The RLF may be detected, for example, on a master node (MN), e.g. a primary cell of a master cell group. At 304, the process 300 may include encoding an indication of the detected RLF for transmission to a secondary node (SN) over signalling radio bearer type 3 (SRB3). The SN may include, for example, a secondary cell of a secondary cell group. In some embodiments, the SN and MN may be different radio access technologies (RATs), such as LTE and 5G and/or other suitable RAT(s).



FIG. 4 illustrates another process 400 in accordance with various embodiments. In some embodiments, the process 400 may be performed by a SN (e.g., a secondary cell of a SCG) for MR-DC communication with a MN. At 402, the process 400 may include receiving, from the UE via SRB3, signalling radio bearer type 3 (SRB3), an indication of a radio link failure on a master node (MN); and forwarding the indication to the MN.


For one or more embodiments, at least one of the components set forth in one or more of the preceding figures may be configured to perform one or more operations, techniques, processes, and/or methods as set forth in the example section below. For example, the baseband circuitry as described above in connection with one or more of the preceding figures may be configured to operate in accordance with one or more of the examples set forth below. For another example, circuitry associated with a UE, base station, network element, etc. as described above in connection with one or more of the preceding figures may be configured to operate in accordance with one or more of the examples set forth below in the example section.


EXAMPLES

Example 1 includes a UE configured with MR-DC, further configured with SRB3, with radio link failure of the PCell, indicating failure to the SN over SRB3.


Example 2 includes the UE of example 1 and/or some other example(s) herein, wherein the failure indication message and the procedure for generation of it re-use much of that defined for the failure message sent directly over split SRB1.


Example 3 includes the UE of example 1 and/or some other example(s) herein, wherein the recovery message sent over SRB3 is the same as the one sent over split SRB1


Example 4 includes the UE of example 1 and/or some other example(s) herein, wherein the UE handling of the recovery message received over SRB3 re-uses the handling that is defined when the message is received over split SRB1.


Example 5 includes the UE of examples 2-4 and/or some other example(s) herein, wherein these messages are tunneled transparently between the UE and MN over SRB3.


Example 6 includes a method of operating a user equipment (UE) capable of communicating using multi-radio access technology (MR)-dual connectivity (DC), the method comprising: detecting a radio link failure (RLF); and encoding an indication of the detected RLF for transmission to a secondary node (SN) over signalling radio bearer type 3 (SRB3).


Example 7 includes the method of example 6 and/or some other example(s) herein, wherein the encoding comprises: generating a failure indication message used for split signalling radio bearer type 1 (SRB1) to be the failure indication for the SRB3.


Example 8 includes the method of examples 6-7 and/or some other example(s) herein, further comprising: receiving a reconfiguration message or recovery message over SRB3, wherein the reconfiguration message or recovery message is a same reconfiguration message or recovery message sent over a split SRB1.


Example 9 includes the method of example 8 and/or some other example(s) herein, further comprising: processing the reconfiguration message or recovery message in a same manner as the UE is configured to process the reconfiguration message or recovery message of the split SRB1.


Example 10 includes the method of examples 7-9 and/or some other example(s) herein, wherein the failure indication message and the reconfiguration message or recovery message are tunneled transparently between the UE and a master node (MN) over SRB3.


Example 11 includes a method of a secondary node (SN) for multi-radio access technology (MR)-dual connectivity (DC) communication with a user equipment (UE), the method comprising: receiving, from the UE via signalling radio bearer type 3 (SRB3), an indication of a radio link failure on a master node (MN); and forwarding the indication to the MN.


Example 12 may include the method of example 11 and/or some other example(s) herein, wherein the MN is a different radio access technology (RAT) than the SN.


Example 13 includes the method of example 11-12 and/or some other example(s) herein, wherein the indication is received and forwarded in a failure indication message used for split signalling radio bearer type 1 (SRB1).


Example 14 includes the method of example 11-13 and/or some other example(s) herein, further comprising: receiving a reconfiguration message or recovery message from the MN; and forwarding the reconfiguration message or recovery message to the UE over SRB3, wherein the reconfiguration message or recovery message is a same reconfiguration message or recovery message sent over a split SRB1.


Example 15 includes the method of examples 14 and/or some other example(s) herein, wherein the failure indication message and the reconfiguration message or recovery message are tunneled transparently between the MN and the UE over SRB3.


Example 16 may include an apparatus comprising means to perform one or more elements of a method described in or related to any of examples 1-15, or any other method or process described herein.


Example 17 may include one or more non-transitory computer-readable media comprising instructions to cause an electronic device, upon execution of the instructions by one or more processors of the electronic device, to perform one or more elements of a method described in or related to any of examples 1-15, or any other method or process described herein.


Example 18 may include an apparatus comprising logic, modules, or circuitry to perform one or more elements of a method described in or related to any of examples A01-A05, B01-B05, or any other method or process described herein.


Example 19 may include a method, technique, or process as described in or related to any of examples 1-15, or portions or parts thereof.


Example 20 may include an apparatus comprising: one or more processors and one or more computer-readable media comprising instructions that, when executed by the one or more processors, cause the one or more processors to perform the method, techniques, or process as described in or related to any of examples 1-15, or portions thereof.


Example 21 may include a signal as described in or related to any of examples 1-15, or portions or parts thereof.


Example 22 may include a datagram, packet, frame, segment, protocol data unit (PDU), or message as described in or related to any of examples 1-15, or portions or parts thereof, or otherwise described in the present disclosure.


Example 23 may include a signal encoded with data as described in or related to any of examples 1-15, or portions or parts thereof, or otherwise described in the present disclosure.


Example 24 may include a signal encoded with a datagram, packet, frame, segment, protocol data unit (PDU), or message as described in or related to any of examples 1-15, or portions or parts thereof, or otherwise described in the present disclosure.


Example 25 may include an electromagnetic signal carrying computer-readable instructions, wherein execution of the computer-readable instructions by one or more processors is to cause the one or more processors to perform the method, techniques, or process as described in or related to any of examples 1-15, or portions thereof.


Example 26 may include a computer program comprising instructions, wherein execution of the program by a processing element is to cause the processing element to carry out the method, techniques, or process as described in or related to any of examples 1-15, or portions thereof.


Example 27 may include a signal in a wireless network as shown and described herein.


Example 28 may include a method of communicating in a wireless network as shown and described herein.


Example 29 may include a system for providing wireless communication as shown and described herein.


Example 30 may include a device for providing wireless communication as shown and described herein.


Any of the above-described examples may be combined with any other example (or combination of examples), unless explicitly stated otherwise. The foregoing description of one or more implementations provides illustration and description, but is not intended to be exhaustive or to limit the scope of embodiments to the precise form disclosed. Modifications and variations are possible in light of the above teachings or may be acquired from practice of various embodiments.


Terminology

For the purposes of the present document, the following terms and definitions are applicable to the examples and embodiments discussed herein.


The term “circuitry” refers to a circuit or system of multiple circuits configured to perform a particular function in an electronic device. The circuit or system of circuits may be part of, or include one or more hardware components, such as a logic circuit, a processor (shared, dedicated, or group) and/or memory (shared, dedicated, or group), an Application Specific Integrated Circuit (ASIC), a field-programmable gate array (FPGA), programmable logic device (PLD), complex PLD (CPLD), high-capacity PLD (HCPLD), System-on-Chip (SoC), System-in-Package (SiP), Multi-Chip Package (MCP), digital signal processor (DSP), etc., that are configured to provide the described functionality. In addition, the term “circuitry” may also refer to a combination of one or more hardware elements with the program code used to carry out the functionality of that program code. Some types of circuitry may execute one or more software or firmware programs to provide at least some of the described functionality. Such a combination of hardware elements and program code may be referred to as a particular type of circuitry.


The term “processor circuitry” as used herein refers to, is part of, or includes circuitry capable of sequentially and automatically carrying out a sequence of arithmetic or logical operations, or recording, storing, and/or transferring digital data. The term “processor circuitry” may refer to one or more application processors, one or more baseband processors, a physical central processing unit (CPU), a single-core processor, a dual-core processor, a triple-core processor, a quad-core processor, and/or any other device capable of executing or otherwise operating computer-executable instructions, such as program code, software modules, and/or functional processes. The terms “application circuitry” and/or “baseband circuitry” may be considered synonymous to, and may be referred to as, “processor circuitry.”


The term “memory” and/or “memory circuitry” as used herein refers to one or more hardware devices for storing data, including random access memory (RAM), magnetoresistive RAM (MRAM), phase change random access memory (PRAM), dynamic random access memory (DRAM) and/or synchronous dynamic random access memory (SDRAM), core memory, read only memory (ROM), magnetic disk storage mediums, optical storage mediums, flash memory devices or other machine readable mediums for storing data. The term “computer-readable medium” may include, but is not limited to, memory, portable or fixed storage devices, optical storage devices, and various other mediums capable of storing, containing or carrying instructions or data.


The term “interface circuitry” as used herein refers to, is part of, or includes circuitry that enables the exchange of information between two or more components or devices. The term “interface circuitry” may refer to one or more hardware interfaces, for example, buses, I/O interfaces, peripheral component interfaces, network interface cards, and/or the like.


The term “user equipment” or “UE” as used herein refers to a device with radio communication capabilities and may describe a remote user of network resources in a communications network. The term “user equipment” or “UE” may be considered synonymous to, and may be referred to as, client, mobile, mobile device, mobile terminal, user terminal, mobile unit, mobile station, mobile user, subscriber, user, remote station, access agent, user agent, receiver, radio equipment, reconfigurable radio equipment, reconfigurable mobile device, etc. Furthermore, the term “user equipment” or “UE” may include any type of wireless/wired device or any computing device including a wireless communications interface.


The term “network element” as used herein refers to physical or virtualized equipment and/or infrastructure used to provide wired or wireless communication network services. The term “network element” may be considered synonymous to and/or referred to as a networked computer, networking hardware, network equipment, network node, router, switch, hub, bridge, radio network controller, RAN device, RAN node, gateway, server, virtualized VNF, NFVI, and/or the like.


The term “computer system” as used herein refers to any type interconnected electronic devices, computer devices, or components thereof. Additionally, the term “computer system” and/or “system” may refer to various components of a computer that are communicatively coupled with one another. Furthermore, the term “computer system” and/or “system” may refer to multiple computer devices and/or multiple computing systems that are communicatively coupled with one another and configured to share computing and/or networking resources.


The term “appliance,” “computer appliance,” or the like, as used herein refers to a computer device or computer system with program code (e.g., software or firmware) that is specifically designed to provide a specific computing resource. A “virtual appliance” is a virtual machine image to be implemented by a hypervisor-equipped device that virtualizes or emulates a computer appliance or otherwise is dedicated to provide a specific computing resource.


The term “element” refers to a unit that is indivisible at a given level of abstraction and has a clearly defined boundary, wherein an element may be any type of entity including, for example, one or more devices, systems, controllers, network elements, modules, etc., or combinations thereof. The term “device” refers to a physical entity embedded inside, or attached to, another physical entity in its vicinity, with capabilities to convey digital information from or to that physical entity. The term “entity” refers to a distinct component of an architecture or device, or information transferred as a payload. The term “controller” refers to an element or entity that has the capability to affect a physical entity, such as by changing its state or causing the physical entity to move.


The term “cloud computing” or “cloud” refers to a paradigm for enabling network access to a scalable and elastic pool of shareable computing resources with self-service provisioning and administration on-demand and without active management by users. Cloud computing provides cloud computing services (or cloud services), which are one or more capabilities offered via cloud computing that are invoked using a defined interface (e.g., an API or the like). The term “computing resource” or simply “resource” refers to any physical or virtual component, or usage of such components, of limited availability within a computer system or network. Examples of computing resources include usage/access to, for a period of time, servers, processor(s), storage equipment, memory devices, memory areas, networks, electrical power, input/output (peripheral) devices, mechanical devices, network connections (e.g., channels/links, ports, network sockets, etc.), operating systems, virtual machines (VMs), software/applications, computer files, and/or the like. A “hardware resource” may refer to compute, storage, and/or network resources provided by physical hardware element(s). A “virtualized resource” may refer to compute, storage, and/or network resources provided by virtualization infrastructure to an application, device, system, etc. The term “network resource” or “communication resource” may refer to resources that are accessible by computer devices/systems via a communications network. The term “system resources” may refer to any kind of shared entities to provide services, and may include computing and/or network resources. System resources may be considered as a set of coherent functions, network data objects or services, accessible through a server where such system resources reside on a single host or multiple hosts and are clearly identifiable.


The term “channel” as used herein refers to any transmission medium, either tangible or intangible, which is used to communicate data or a data stream. The term “channel” may be synonymous with and/or equivalent to “communications channel,” “data communications channel,” “transmission channel,” “data transmission channel,” “access channel,” “data access channel,” “link,” “data link,” “carrier,” “radiofrequency carrier,” and/or any other like term denoting a pathway or medium through which data is communicated. Additionally, the term “link” as used herein refers to a connection between two devices through a RAT for the purpose of transmitting and receiving information. As used herein, the term “communication protocol” (either wired or wireless) refers to a set of standardized rules or instructions implemented by a communication device and/or system to communicate with other devices and/or systems, including instructions for packetizing/depacketizing data, modulating/demodulating signals, implementation of protocols stacks, and/or the like.


The terms “instantiate,” “instantiation,” and the like as used herein refers to the creation of an instance. An “instance” also refers to a concrete occurrence of an object, which may occur, for example, during execution of program code.


The terms “coupled,” “communicatively coupled,” along with derivatives thereof are used herein. The term “coupled” may mean two or more elements are in direct physical or electrical contact with one another, may mean that two or more elements indirectly contact each other but still cooperate or interact with each other, and/or may mean that one or more other elements are coupled or connected between the elements that are said to be coupled with each other. The term “directly coupled” may mean that two or more elements are in direct contact with one another. The term “communicatively coupled” may mean that two or more elements may be in contact with one another by a means of communication including through a wire or other interconnect connection, through a wireless communication channel or ink, and/or the like.


The term “information element” refers to a structural element containing one or more fields. The term “field” refers to individual contents of an information element, or a data element that contains content.


The term “admission control” refers to a validation process in communication systems where a check is performed before a connection is established to see if current resources are sufficient for the proposed connection.


The term “workload” refers to an amount of work performed by a computing system, device, entity, etc., during a period of time or at a particular instant of time. A workload may be represented as a benchmark, such as a response time, throughput (e.g., how much work is accomplished over a period of time), and/or the like. Additionally or alternatively, the workload may be represented as a memory workload (e.g., an amount of memory space needed for program execution to store temporary or permanent data and to perform intermediate computations), processor workload (e.g., a number of instructions being executed by the processor 102 during a given period of time or at a particular time instant), an I/O workload (e.g., a number of inputs and outputs or system accesses during a given period of time or at a particular time instant), database workloads (e.g., a number of database queries during a period of time), a network-related workload (e.g., a number of network attachments, a number of mobility updates, a number of radio link failures, a number of handovers, an amount of data to be transferred over an air interface, etc.), and/or the like. Various algorithms may be used to determine a workload and/or workload characteristics, which may be based on any of the aforementioned workload types.


The term “SMTC” refers to an SSB-based measurement timing configuration configured by SSB-MeasurementTimingConfiguration.


The term “SSB” refers to an SS/PBCH block.


The term “a “Primary Cell” refers to the MCG cell, operating on the primary frequency, in which the UE either performs the initial connection establishment procedure or initiates the connection re-establishment procedure.


The term “Primary SCG Cell” refers to the SCG cell in which the UE performs random access when performing the Reconfiguration with Sync procedure for DC operation.


The term “Secondary Cell” refers to a cell providing additional radio resources on top of a Special Cell for a UE configured with CA.


The term “Secondary Cell Group” refers to the subset of serving cells comprising the PSCell and zero or more secondary cells for a UE configured with DC.


The term “Serving Cell” refers to the primary cell for a UE in RRC_CONNECTED not configured with CA/DC there is only one serving cell comprising of the primary cell. When a UE in RRC_CONNECTED configured with CA/DC, the term “serving cell” refers to the set of cells comprising the Special Cell(s) and all secondary cells.


The term “Special Cell” refers to the PCell of the MCG or the PSCell of the SCG for DC operation; otherwise, the term “Special Cell” refers to the Pcell.

Claims
  • 1. One or more non-transitory computer-readable media (NTCRM) having instructions, stored thereon, that when executed by one or more processors cause a user equipment (UE) to: communicate with a master node (MN) and a secondary node (SN) using multi-radio access technology (MR)-dual connectivity (DC);detecting a radio link failure (RLF) on the MN; andencoding an indication of the detected RLF for transmission to a secondary node (SN) over signalling radio bearer type 3 (SRB3).
  • 2. The one or more NTCRM of claim 1, wherein, to encode the indication, the UE is to generate a failure indication message used for split signalling radio bearer type 1 (SRB1) to be the failure indication for the SRB3.
  • 3. The one or more NTCRM of claim 2, wherein the instructions, when executed, are further to cause the UE to receive a reconfiguration message or recovery message over SRB3, wherein the reconfiguration message or recovery message is a same reconfiguration message or recovery message sent over a split SRB1.
  • 4. The one or more NTCRM of claim 3, wherein the instructions, when executed, are further to cause the UE to process the reconfiguration message or recovery message in a same manner as the UE is configured to process the reconfiguration message or recovery message of the split SRB1.
  • 5. The one or more NTCRM of claim 3, wherein the failure indication message and the reconfiguration message or recovery message are tunneled transparently between the UE and a master node (MN) over SRB3.
  • 6. One or more non-transitory computer-readable media (NTCRM) having instructions, stored thereon, that when executed by one or more processors cause a secondary node (SN) to: communicate with a user equipment (UE) using multi-radio access technology (MR)-dual connectivity (DC) in association with a master node (MN);receive, from the UE via signalling radio bearer type 3 (SRB3), an indication of a radio link failure on the MN; andforward the indication to the MN.
  • 7. The one or more NTCRM of claim 6, wherein the MN is a different radio access technology (RAT) than the SN.
  • 8. The one or more NTCRM of claim 6, wherein the indication is received and forwarded in a failure indication message format that is used for split signalling radio bearer type 1 (SRB1).
  • 9. The one or more NTCRM of claim 6, wherein the instructions, when executed, are further to cause the SN to: receive a reconfiguration message or recovery message from the MN; andforward the reconfiguration message or recovery message to the UE over SRB3, wherein the reconfiguration message or recovery message is a same reconfiguration message or recovery message format that is sent over a split SRB1.
  • 10. The one or more NTCRM of claim 9, wherein the failure indication message and the reconfiguration message or recovery message are tunneled transparently by the SN between the MN and the UE over SRB3.
CROSS REFERENCE TO RELATED APPLICATION

The present application claims priority to U.S. Provisional Patent Application No. 62/909,493, which was filed Oct. 2, 2019; the disclosure of which is hereby incorporated by reference.

Provisional Applications (1)
Number Date Country
62909493 Oct 2019 US