This invention relates to a recovery mechanism for point-to-multipoint (P2MP) traffic paths in a connection-oriented network, such as a Generalised Multi-Protocol Label Switching (GMPLS), Multi-Protocol Label Switching (MPLS) or Multi-Protocol Label Switching Transport Profile (MPLS-TP) network.
Multi-Protocol Label Switching Transport Profile (MPLS-TP) is a joint International Telecommunications Union (ITU-T)/Internet Engineering Task Force (IETF) effort to include an MPLS Transport Profile within the IETF MPLS architecture to support the capabilities and functionalities of a packet transport network as defined by ITU-T.
Many carriers have Synchronous Digital Hierarchy (SDH) networks. One goal of MPLS-TP is to allow a smooth migration from existing SDH networks to packet networks, thereby minimising the cost to carriers. Existing SDH networks are often based on a ring topology and it is desirable that MPLS-TP solutions work with this kind of network topology. Existing carrier networks have recovery mechanisms to detect and recover from a failure in the network and it is desirable that MPLS-TP networks also have resilience to failures. However, the recovery mechanism used in existing SDH networks cannot be directly applied to networks which use label switched paths.
RFC4872 describes signalling to support end-to-end GMPLS recovery, but the scope of this document is limited to point-to-point (P2P) paths.
WO 2008/080418A1 describes a protection scheme for an MPLS network having a ring topology. A primary path connects an ingress node to a plurality of egress nodes. A pre-configured secondary path also connects the ingress node to the plurality of egress nodes. In the event of a failure, traffic is sent along both the primary path and the secondary path, thus ensuring that each egress node receives traffic via the primary path or the secondary path.
An IETF Internet-Draft “P2MP traffic protection in MPLS-TP ring topology”, draft-ceccarelli-mpls-tp-p2 mp-ring-00, D. Ceccarelli et al, January 2009, describes a data plane-driven solution for the distribution and recovery of P2MP traffic over ring topology networks.
The present invention seeks to provide an alternative method of traffic recovery.
An aspect of the present invention provides a method of operating a first node in a connection-oriented network to provide traffic recovery according to claim 1.
The first node can select a backup path which is matched to the position of the failure, thereby efficiently re-routing traffic when a failure occurs. This minimises, or avoids, the need to send traffic over communication links in forward and reverse directions, as can often occur in the MPLS Fast Rerouting (FRR) technique which is implemented at the data plane level of the network. The use of a backup path which is used instead of the working path, and which connects to destination nodes of the working path, avoids a situation where a node receives the same packet of data via a working path and a backup path.
Advantageously, only one of the plurality of backup paths is used at a time. This allows the set of backup paths to share a common set of reserved resources, particularly in the case of a ring topology. The point-to-multipoint backup path makes efficient use of network resources compared to using a set of point-to-point (P2P) paths.
The first node can be the source node, or head node, of the point-to-multipoint working path. This is the most efficient arrangement as it minimises the number of communication links that are traversed in forward and reverse directions when traffic is sent along a backup path. However, in an alternative arrangement the first node can be positioned downstream of the source node along the working path.
Another aspect of the invention provides a method of traffic recovery in a connection-oriented network according to claim 11.
The methods can be applied to a range of different network topologies, such as meshed networks, but are particularly advantageous when applied to ring topologies.
Advantageously, the recovery scheme is used within a network having a Generalised Multi-Protocol Label Switching (GMPLS) or a Multi-Protocol Label Switching (MPLS) control plane. Data plane connections can be packet based or can use any of a range of other data plane technologies such as: wavelength division multiplexed traffic (lambda); or time-division multiplexed (TDM) traffic such as Synchronous Digital Hierarchy (SDH). The data plane can be an MPLS or an MPLS-TP data plane. The recovery scheme can also be applied to other connection-oriented technologies such as connection-oriented Ethernet or Provider Backbone Bridging Traffic Engineering (PBB-TE), IEEE 802.1Qay.
Further aspects of the invention provide apparatus for performing the methods.
The functionality described here can be implemented in software, hardware or a combination of these. The functionality can be implemented by means of hardware comprising several distinct elements and by means of a suitably programmed processing apparatus. The processing apparatus can comprise a computer, a processor, a state machine, a logic array or any other suitable processing apparatus. The processing apparatus can be a general-purpose processor which executes software to cause the general-purpose processor to perform the required tasks, or the processing apparatus can be dedicated to the perform the required functions. Another aspect of the invention provides machine-readable instructions (software) which, when executed by a processor, perform any of the described methods. The machine-readable instructions may be stored on an electronic memory device, hard disk, optical disk or other machine-readable storage medium. The machine-readable instructions can be downloaded to a processing apparatus via a network connection.
Embodiments of the invention will be described, by way of example only, with reference to the (accompanying drawings in which:
Node A is called the head node of the ring and is the root node of the P2MP LSP 10. The communication links 11 of the path are monitored to detect a failure in a communications link or node. Failure detection can be performed using the Operations, Administration and Management (OAM) tools provided by MPLS-TP, or by any other suitable mechanism. One form of failure detection mechanism periodically exchanges a Continuity Check message between a pair of nodes. If a reply is not received within a predetermined time period, an alarm is raised.
Now consider that a failure affects the link between nodes C and D. This failure affects the P2MP LSP 10, as it prevents traffic from reaching nodes D, E, F.
Node A is provided with a set of pre-computed and pre-signalled backup P2MP LSPs, one for each possible point of failure in the network. The extent to which the backup paths are configured is described below, and varies depending on whether “restoration” or “protection” is required. The full set of possible backup LSPs for the working path LSP of
In the event of a node or link failure a signalling message is sent from a node detecting the failure (in
The backup LSP can protect the ring from a link failure (e.g. link C-D) and a node failure (e.g. node D). Node failure may be detected using the same mechanisms used for link detection (e.g. OAM, RSVP-TE hello). In the event of node failure it is not possible to route traffic to, or through, the failed node.
The signalling message sent from the node that detects a failure can be a ReSource ReserVation Protocol-Traffic Engineering (RSVP-TE) Notify message. This message is sent via the Control Plane of the network.
There are two possible ways of operating: (i) restoration and (ii) protection.
In the restoration scheme, resources required for the backup paths 21-25 are not cross-connected at the data plane level prior to a failure. This allows other LSPs to use the bandwidth of the backup paths until they are needed. This scheme requires some additional time, following failure detection, to signal to nodes along the backup path to cross-connect resources. The selected backup LSP is activated by cross-connecting resources at the data plane level at each node. Traffic is then switched from the working LSP 10 to the backup LSP 20 that has just been prepared for use. The backup LSP can be activated using a modified Path message with the S bit set to 0 in the PROTECTION object. At this point, the link and node resources must be allocated for this LSP that becomes a primary LSP (ready to carry normal traffic).
At the initial stage of setting up the backup paths (pre-failure), the backup LSP is signalled but no resources are committed at the data plane level. The resources are pre-reserved only at the control plane level only. Signalling is performed by indicating in the Path message (in the PROTECTION object) that the LSPs are of type “working” and “protecting”, respectively. To make the bandwidth pre-reserved for the backup (not activated) LSP available for extra-traffic, this bandwidth could be included in the advertised Unreserved Bandwidth at priority lower (means numerically higher) than the Holding Priority of the protecting LSP. In addition, the Max LSP Bandwidth field in the Interface Switching Capability Descriptor sub-TLV should reflect the fact that the bandwidth pre-reserved for the protecting LSP is available for extra traffic. LSPs for extra-traffic then can be established using the bandwidth pre-reserved for the protecting LSP by setting (in the Path message) the Setup Priority field of the SESSION_ATTRIBUTE object to X (where X is the Setup Priority of the protecting LSP), and the Holding Priority field to at least X+1. Also, if the resources pre-reserved for the protecting LSP are used by lower-priority LSPs, these LSPs should be pre-empted when the protecting LSP is activated.
In the protection scheme resources required for the backup paths are cross-connected at the data plane level prior to a failure. This allows a quick switch to a required one of the backup paths but it incurs a penalty in terms of bandwidth, as the resources of the backup paths are reserved. The reserved resources of a backup path can be used to carry other traffic, such as “best efforts” traffic, until a time at which the reserved resources are required to carry traffic along the backup path.
In the case where the backup LSP has the same bandwidth as the working path LSP 10, the set of backup paths shown in
When the working path has recovered from the failure which originally caused the protection switch traffic is returned to the working path LSP 10. Nodes detect that working path is up in the same way they detect the fails (e.g. OAM-CC, RSVP-TE hello). When a node detects the failure ends, it may notify the information to the ingress node using an RSVP-TE NOTIFY message.
The operation of a node in the network will now be described in more detail.
Controller 52 comprises a set of functional modules 53-57 which control operation of the LSR. A Control Plane module 53 exchanges signalling and routing messages with other network nodes and can incorporate functions for IP routing and Label Distribution Protocol. The Control Plane module 53 can support RSVP-TE signalling, allowing the LSR 40 to signal to other nodes to implement the traffic recovery operation by signalling the occurrence of a failure and activating a required backup LSP. A Management Plane module 54 (if present) performs signalling with a Network Management System, allowing LSPs to be set up. An OAM module 55 supports OAM signalling, such as Continuity Check signalling, to detect the occurrence of a link or node failure. A Data Plane forwarding module 56 performs label look up and switching to support forwarding of received transport units (packets). The Data Plane forwarding module 56 uses the forwarding data stored in the LFIB 51. A combination of the Data Plane forwarding module 56 and LFIB 51 perform the cross-connect function shown in
Although a single storage entity 50 is shown in
For a restoration scheme, the method proceeds to step 73 and signals to nodes. The signalling may include instructing nodes to reserve suitable resources, such as bandwidth, to support the backup paths. However, nodes are not instructed to cross-connect resources at the data plane level. This means that the back-up path is not fully established, and requires further signalling at the time of failure detection to fully establish the backup path.
For a protection scheme, the method proceeds to step 74 and signals to nodes. The signalling instructs nodes to fully establish the backup paths in readiness for use. This includes reserving suitable resources, such as bandwidth, to support the backup paths. The nodes are also instructed to cross-connect resources at the data plane level. This means that the back-up path is fully established, and may not require any further signalling at the time of failure detection to carry traffic.
The example P2MP working path LSP 10 shown in
The backup paths only need to connect to destination nodes of the working path, and nodes which must be transited to reach the destination nodes. In the example shown in
Modifications and other embodiments of the disclosed invention will come to mind to one skilled in the art having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. Therefore, it is to be understood that the invention is not to be limited to the specific embodiments disclosed and that modifications and other embodiments are intended to be included within the scope of this disclosure. Although specific terms may be employed herein, they are used in a generic and descriptive sense only and not for purposes of limitation.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/EP09/59150 | 7/16/2009 | WO | 00 | 4/27/2012 |