Non-stop forwarding in a multi-chassis router

Information

  • Patent Grant
  • 8483048
  • Patent Number
    8,483,048
  • Date Filed
    Thursday, September 23, 2010
    14 years ago
  • Date Issued
    Tuesday, July 9, 2013
    11 years ago
Abstract
State information is synchronized between a plurality of routing engines in a multi-chassis router according to a synchronization gradient. An example multi-chassis router is described that includes a primary routing engine and a standby routing engine in each chassis. According to the synchronization gradient, the primary routing engine of a control node updates state information on the standby routing engine of the control node prior to updating the primary routing engines of the other chassis. The primary routing engines of the other chassis update state information in respective standby routing engines prior to updating state information in consumers. If a primary routing engine fails, the corresponding standby routing engine assumes control of the primary routing engine's duties. Upon assuming control, a standby routing engine resumes updating state information without having to resend state information or interrupt packet forwarding.
Description
TECHNICAL FIELD

The invention relates to computer networks and, more particularly, to systems for routing packets within computer networks.


BACKGROUND

A computer network is a collection of interconnected computing devices that can exchange data and share resources. In a packet-based network, the computing devices communicate data by dividing the data into small blocks called packets, which are individually routed across the network from a source device to a destination device. The destination device extracts the data from the packets and assembles the data into its original form. Dividing the data into packets enables the source device to resend only those individual packets that may be lost during transmission.


Certain devices within the network, such as routers, maintain tables of information that describe routes through the network. A “route” can generally be defined as a path between two locations on the network. Upon receiving an incoming data packet, the router examines destination information within the packet to identify the destination for the packet. Based on the destination, the router forwards the packet in accordance with the routing table.


The physical connection between devices within the network is generally referred to as a link. A router uses interface cards (IFCs) for receiving and sending data packets via network links. These IFCs are installed in ports known as interfaces and are configured using interface configurations.


Generally, a router maintains state information. For example, a router may maintain state information representing the current state of the interfaces between the router and the network. Such state information may include information representing the state of one or more IFCs, such as the current configuration of the IFCs. As additional examples, a router may maintain state information representing the state of one or more packet forwarding engines (PFEs), one or more routing engines, or other resources within the router.


In particular, a process, e.g., a control node known as a “routing engine,” operating within a router may maintain the state information and communicate changes to the state information to various other processes or components within the router. These other processes or components are sometimes referred to as “consumers,” because they receive and utilize the state information maintained by the operating system. These consumers make use of the state information when performing their various functions.


As the complexity of conventional networks has increased in recent years, management of the state information within a router or other network device has likewise become a significant challenge. Some existing methods for managing state information involve caching the information within the operating system, and issuing state change notification messages to software modules executing within the router. In response, the software modules retrieve the state information from the operating system.


To increase reliability, some routers may include a primary routing engine and one or more standby routing engines. Both primary and standby routing engines may require state information. In the event that the primary routing engine fails, one of the standby routing engines assumes control of the routing resources to continue operation of the router. The process of switching control of routing functions between the primary and standby routing engines is often referred to as failover. In some instances, to assume proper control and ensure operation, the standby routing engine is forced to “relearn” the lost state information from each resource, e.g., by power cycling the router resources to a known state. This causes an interruption in packet forwarding while the router resources restart operations.


Routers have not only developed to be more reliable, but also to meet increasing bandwidth demands. One way to meet increasing bandwidth needs is to use multi-chassis routers, i.e., routers in which multiple routing devices are physically coupled and configured to operate as a single router. For example, a multi-chassis router may contain multiple line card chassis (LCCs), which include one or more IFCs, and a central switch card chassis (SCC), which forward packets between the LCCs and provides top-down management of the multi-chassis router. Because multi-chassis routers combine resources of multiple routing devices, multi-chassis routers typically have much higher bandwidth capabilities than standalone routers. The use of multi-chassis routers can simplify and improve routing on a service provider network by consolidating routing functions onto fewer routers. However, forwarding state information to each consumer in a multi-chassis router can be significantly more difficult than in a standalone router.


SUMMARY

Techniques for providing non-stop forwarding on a multi-chassis router are described. For example, a multi-chassis router may include one or more master routing engines on one device of the multi-chassis router, and one or more local routing engines on the other device of the multi-chassis router. For example, a multi-chassis router may include a central switch card chassis (SCC) having a master routing engine and one or more line card chassis (LCCs), each having one or more local routing engines. The local routing engines forward state information received from router resources within their device directly to the master routing engine without processing or recognizing the state information. The master routing engine, which may be included in a control node, manages state information for the entire multi-chassis router and propagates the state information back to each local routing engine for distribution to consumers. The master routing engine also provides state information for consumers within its own device, e.g., the SCC.


Each device may also have one or more standby routing engines to allow for failover in that event that that device's primary routing engine fails. The routing engines of the multi-chassis router may enforce synchronization of state information in accordance with a “synchronization gradient” in which the primary master routing engine distributes state information to one or more standby master routing engines prior to distributing the state information to primary local routing engines. Upon receipt, primary local routing engines distribute the state information to standby local routing engines prior to distributing it to consumers within each chassis. In this manner, updated state information is first synchronized from a primary master routing engine of a central control node within the multi-chassis system to the standby master routing engine of the control node and then the local routing engines before distributing the state information to the consumers. This synchronization gradient ensures that, upon failover, the standby routing engine contains state information that is at least as current as the state information provided to all other consumers of the state information. Consequently, the standby routing engines are able to readily assume responsibility in the event of a failover, and can continue updating the other consumers with the state information as necessary. In this manner, the standby routing engine of one chassis of a multi-chassis system may assume control of router resources within that chassis without needing to “relearn” state information, e.g., by power cycling the router resources to a known state.


In one embodiment, a method for distributing state information within a multi-chassis router comprises receiving updates to the state information with a first local routing engine from a third routing engine of the multi-chassis router, communicating the updates from the first local routing engine to a second local routing engine and communicating the updates from the first local routing engine to a consumer of the state information.


In another embodiment, a system comprises a first node having a first routing engine and a second routing engine and a second node having a first consumer. The first routing engine manages state information and communicates updates to the state information to the second routing engine before communicating the updates to the first consumer.


In an embodiment, a computer-readable medium comprises instructions for causing a first consumer to receive state information from a master routing engine and communicate updates to the state information in accordance with an order that requires the updates to be communicated to a second consumer before communicating the updates to a third consumer of the state information.


In an embodiment a method comprises forwarding state updates from a first control unit of a first chassis of a multi-chassis network device to an intermediary consumer within a second chassis of the multi-chassis network device, transferring state update operations from the first control unit to a second control unit of the first chassis and substantially maintaining network functions during the transfer.


In a different embodiment, a method comprises receiving state updates with a first control unit of a first chassis of a multi-chassis network system from a control unit of a control node of the multi-chassis network system, forwarding the state updates from the first control unit to a consumer, transferring state update operations from the first control to a second control unit of the first chassis and substantially maintaining network functions during the transfer.


The details of one or more embodiments of the invention are set forth in the accompanying drawings and the description below. Other features, objects, and advantages of the invention will be apparent from the description and drawings, and from the claims.





BRIEF DESCRIPTION OF DRAWINGS


FIG. 1 is block diagram of an example computing environment in which a service-provider network includes a multi-chassis router.



FIG. 2 is a block diagram illustrating an exemplary multi-chassis router consistent with the principles of the invention.



FIG. 3 is a block diagram illustrating an exemplary control plane of the multi-chassis router of FIG. 2 consistent with the principles of the invention.



FIGS. 4A and 4B are block diagrams illustrating an example data structure in for maintaining state information within a multi-chassis router.



FIG. 5 is a flowchart illustrating an exemplary state update operation in a multi-chassis router consistent with the principles of the invention.



FIG. 6 is a flowchart illustrating failover of a master routing engine consistent with the principles of the invention.



FIG. 7 is a flowchart illustrating failover of a local routing engine consistent with the principles of the invention.





DETAILED DESCRIPTION


FIG. 1 is a block diagram illustrating an example computing environment 2 in which service provider network 6 includes a multi-chassis router 4. In this example, multi-chassis router 4 communicates with edge routers 5A and 5B (“edge routers 5”) to provide customer networks 8A-8C (“customer networks 8”) with access to network 6. In one embodiment, multi-chassis router 4 includes an SCC that operates as a control node and one or more line card chassis LCCs that operate as packet routing devices. The LCCs may contain all the physical interfaces for coupling to other devices within network 6, while the SCC controls and routes traffic between the LCCs.


Although not illustrated, service provider network 6 may be coupled to one or more networks administered by other providers, and may thus form part of a large-scale public network infrastructure, e.g., the Internet. Consequently, customer networks 8 may be viewed as edge networks of the Internet. Service provider network 6 may provide computing devices within customer networks 8 with access to the Internet, and may allow the computing devices within customer networks 8 to communicate with each other. Service provider network 6 may include a variety of network devices other than multi-chassis router 4 and edge routers 5, such as additional routers, switches, servers, or other devices.


In the illustrated embodiment, edge router 5A is coupled to customer network 8A via access link 9A and edge router 5B is coupled to customer networks 8B and 8C via access links 9B and 9C, respectively. Customer networks 8 may be networks for geographically separated sites of an enterprise. Customer networks 8 may include one or more computing devices (not shown), such as personal computers, laptop computers, handheld computers, workstations, servers, switches, printers, or other devices. The configuration of network 2 illustrated in FIG. 1 is merely exemplary. For example, service provider network 6 may be coupled to any number of customer networks 8. Nonetheless, for ease of description, only customer networks 8A-8C are illustrated in FIG. 1.


Consistent with the principles of the inventions, multi-chassis router 4 provides for failover by including a primary routing engine as well as one or more standby routing engines. For example, an SCC may contain primary and standby master routing engines, and one or more LCCs may contain primary and standby local routing engines. State information is pushed downward from the primary master routing engine to the standby master engine prior to forwarding the state information to the primary local routing engines in the LCCs. Similarly, the primary local routing engines forward the state information to one or more standby local routing engines prior to forwarding the state information to “consumers” within their chassis. In this manner, multi-chassis router 4 enforces a synchronization gradient when communicating state information throughout the multi-chassis environment. In the event a primary routing engine fails, a standby routing engine in the same chassis assumes control over routing resources and routing functionality for that chassis. Moreover, because state information is delivered to a standby routing engine prior to forwarding the state information to a consumers, a standby routing engine can take up forwarding state information to consumers at the same place where the primary routing engine left off. U.S. patent application Ser. No. 10/678,280 titled “Syncronizing State Information Between Control Units”, filed Oct. 3, 2003, describes techniques for a synchronization gradient within a standalone router and is hereby incorporated by reference.


In this manner, the primary and standby routing engines synchronize their respective state information to allow the standby routing engine to assume control of the router resources without having to relearn state information. For exemplary purposes, the principles of the invention will be described in reference to multi-chassis router 4. However, the principles of the invention are applicable to any multi-chassis network device. For example, the principles of the invention could be applied to edge routers 5, enterprise appliances, session border controllers, intelligent switches or hubs, firewalls or any other network device having a multi-chassis operating environment.


As described in further detail below, operating systems executing within the primary and standby routing engines of multi-chassis router 4 manage data structures and inform consumers of any change to the state information. Consumers may comprise software processes executing within components of multi-chassis router 4, such as chassis management processes, configuration management processes, or other processes in multi-chassis router 4. Additionally, consumers of the state information may comprise hardware components, or combinations of software, hardware or firmware, such as one or more forwarding engines, IFCs or other hardware. It should be noted that, as to the master routing engine, the local routing engines are viewed as consumers with respect to state information. Similarly, as to local components within the LCCs, the local routing engines are providers of state information. Because of their dual role, the local routing engines may be viewed as intermediate consumers. Likewise, the standby routing engines in the SCC and the LCCs also operate as consumers as to the primary routing engines until taking over state information forwarding, e.g., in event of a failover.


Multi-chassis router 4 may manage state information within a hierarchically-ordered and temporally-linked data structures. One example of the use of hierarchically-ordered and temporally-linked data structures is described in U.S. patent application Ser. No. 10/457,814 titled “Managing State Information in a Computing Environment” by David M. Katz and Dennis C. Ferguson, filed Jun. 9, 2003, hereby incorporated by reference in its entirety.


During normal operation, in the event state information changes, the primary master routing engine of a control node of multi-chassis router 4 (i.e., the SCC in this example) synchronizes state information with operating systems executing on the one or more standby master routing engines. Specifically, the primary master routing engine replicates the state updates and transmits the state update in message form to the standby master routing engines. The standby master routing engines receive the message and update their corresponding data structures to record the state updates.


Upon recording the state updates, the standby master routing engines may transmit an acknowledgement to the primary master routing engine to indicate successful state information synchronization. Once the primary master routing engine has received an acknowledgement from the standby master routing engines, it forwards the state information changes to the consumers of the SCC. For example, this includes any consumers located within the SCC. In addition, the primary master routing engine forwards the state information to the primary local routing engines on the LCCs, which are viewed as consumers to the primary master routing engine. The primary local routing engines then forward the state update information to one or more standby local routing engines on the LCCs. However, each LCC is not required to have a standby local routing engine, and some or all of the LCCs may only have a primary local routing engine. After receiving an acknowledgement from each standby local routing engine on that LCC, a primary local routing engine may then transmit the state update messages to consumers on the LCC.


In this manner, multi-chassis router 4 enforces the requirement that the standby master routing engines are updated with state information changes before consumers, and that the standby routing engines of the individual chassis are updated with the state information before any local consumers within the individual chassis. Therefore, if any of the primary routing engines fail, a corresponding standby routing engine can readily assume routing and state management functionality if needed. Once one of the standby routing engines assumes control that standby routing engine is no longer a standby routing engine, but becomes a primary routing engine; however, that routing engine will maintain its functional status as either a local or a master routing engine. In this regard, multi-chassis router 4 can be viewed as maintaining a “synchronization gradient” such that the primary master routing engine of the control node (i.e., the SCC) receives state updates first, followed by the standby master routing engines, and followed by consumers of the SCC including the primary local routing engines of the other chassis, followed by standby local routing engines, followed by the consumers within the other chassis. This synchronization gradient ensures that upon failover, a standby routing engine within any of the chassis contains enough state information to assume control of the router resources without having to relearn a substantial portion of the state information.


In some embodiments, the primary master and primary local routing engines may use commit markers and commit proposals within their data structures to track the state updates given to consumers. As further described below, a commit marker may be inserted into the state update data structure after the most recent object within the temporally-ordered data structure which a consumer has acknowledged receiving. A commit proposal may be inserted into at a point in the state update date chain to indicate the next point in the state chain to send a acknowledgement request for all information up to that point in the chain. The primary master and primary local engines also replicate the position of commit markers and commit proposals within their data structure to the standby master and standby local routing engines. In this manner the standby routing engines also know which state updates have been received by which consumers.


In response to the state update messages from a local routing engine, consumers issue requests to retrieve the updated state information. When the primary local routing engine receives such a request, the primary local routing engine traverses the hierarchically-ordered and temporally-linked data structure and issues state update messages to the requesting consumers. The primary local routing engine then updates respective commit markers and commit proposals associated with the requesting shared consumers to reflect transmission of the state updates. The primary local routing engine again synchronizes state information with the one or more standby local routing engines so that the commit markers and commit proposals within the state information maintained by the local routing engines on a chassis are uniform.


In this manner, one of the standby local routing engines may assume control of a chassis, and can deterministically identify the state information of which each consumer on that chassis has already been informed, i.e., consumed. As a result, the standby local routing engines may need only update the consumers with limited amount of state information, and need not rely on relearning state information from the master routing engine of multi-chassis router 4. Once one of the standby local routing engines assumes control of a chassis that standby local routing engine is no longer a standby local routing engine, but becomes a primary local routing engine.


The combination of failover techniques for the local routing engines and the master routing engine extends a synchronization gradient for a standalone router for use in multi-chassis router 4. The failure of a single routing engine in multi-chassis router 4 will not interrupt packet-forwarding and state information updates can be resumed with limited interruption and redundancy. In this manner, multi-chassis router 4 has a non-stop forwarding capability equivalent to non-stop forwarding in a standalone router.



FIG. 2 is a block diagram illustrating an exemplary multi-chassis router 120 that operates consistent with the principles of the invention. Multi-chassis router 120 routes data packets between network devices across a network. In this example, multi-chassis router 120 comprises four substantially identical LCCs 128A-128D (“LCCs 128”) and SCC 122 that operates as a central control node. In other embodiments, a multi-chassis router may include more or less LCCs. SCC 122 provides centralized switching and control for multi-chassis router 120. LCCs 128 provide interfaces to a network using IFC sets 134A-134D (“IFCs 134”).


SCC 122 includes switch fabric 124 and two routing engines: primary master routing engine 126 and standby master routing engine 127. Switch fabric 124 provides a back-side connection, i.e. a connection separate from the network, between switch fabric 125 of LCCs 128. Functions of primary master routing engine 126 include maintaining routing information to describe a topology of a network, and using that information to derive forwarding information bases (FIBs). Routing engine 126 controls packet forwarding throughout multi-chassis router 120 by installing an FIB in LCCs 128 via communication with local routing engines 130 and/or 131 over cables 137. An FIB for one of LCCs 128 may be the same or different than an FIB for other LCCs 128 and SCC 122. Because cables 137 provide a dedicated connection, i.e., separate from a data packet forwarding connection provided by cables 136, between SCC 122 and LCCs 128, FIBs in LCC routing engines 130 can be updated without interrupting packet forwarding performance of multi-chassis router 120. LCCs 128 each contain one of primary local routing engines 130A-130D (“routing engines 130”), one of standby local routing engines 131A-131D (“routing engines 131”), one of switch fabrics 125A-D (“switch fabric 125”), at least one packet forwarding engine (PFE), shown as PFEs 132A-132D (“PFEs 132”), and one or more IFCs 134.


Multi-chassis router 120 performs routing functions in the following manner. An incoming data packet is first received from a network by one of IFCs 134, e.g., 134B, which directs it to one of PFEs 132, e.g., PFE 132B. The PFE then determines a proper route for the data packet using the FIB provided by the primary local routing engine, e.g., routing engine 130B. If the data packet is destined for an outbound link associated with the one of IFCs 134 that initially receive the packet, the PFE forwards the packet to the outbound link. In this manner, packets sent out by the same PFE on which they were received from the network bypass switch fabric 124 and switch fabric 125.


Otherwise, the PFE sends the data packet to switch fabric 125, where it is directed to switch fabric 124, where it follows a route to one of the other PFEs 132, e.g., PFE 132D. This PFE, e.g., PFE 132D, sends the data packet across the network via one of IFCs 134, e.g., IFC 134D. Thus an incoming data packet received by one of LCCs 128 may be sent by another one of LCCs 128 to its destination. Other multi-chassis routers that operate in a manner consistent with the principles of the invention may use different switching and routing mechanisms.


As described with respect to multi-chassis router 4 in FIG. 1, in some embodiments, primary master routing engine 126 forwards state information updates to consumers using a hierarchically-ordered and temporally-linked data structure according to a synchronization gradient. With respect to primary master routing engine 126, standby master routing engine 127 and primary local routing engines 130 are consumers.


Standby master routing engine 127 is substantially similar to primary master routing engine 126. For example, standby master routing engine 127 may include the same hardware and operating system and other software applications as primary master routing engine 126. If primary master routing engine 126 fails or is taken off-line, standby master routing engine 127 carries on the functions of primary master routing engine 126. In this example, if primary master routing engine 126 comes back on-line, it could either serve as a standby master routing engine subservient to routing engine 127 now acting as a primary, or resume operation as the primary master routing engine of SCC 122. In either situation, at any one time, only one of primary master routing engine 126 and standby master routing engine 127 provides control of multi-chassis router 120.


Primary local routing engines 130 control and manage LCCs 128, but are subservient to primary master routing engine 126 of SCC 122. For example, after receiving state information updates from primary master routing engine 126, primary local routing engines 130 forward the state information update to consumers on LCCs 128 using the hierarchically-ordered and temporally-linked data structure. For example, consumers that receive state information updates from primary local routing engines 130 include standby local routing engines 131, PFEs 132 and IFCs 134. Primary local routing engines 130 also distribute the FIB derived by primary master routing engine 126 to PFEs 132.


Standby local routing engines 131 may be substantially similar to primary local routing engines 130. For example, standby local routing engines 131 may include the same hardware and operating system and other software applications as primary master routing engine 126. If one of primary local routing engines 130 fails or is taken off-line, the corresponding standby local routing engines 131 would carry on the functions of the failed one of routing engines 130. For example, assume primary local routing engine 130B of LCC 128B fails. In this case, standby local routing engine 131B takes over control of LCC 128B. As a result, standby local routing engine 131B begins forwarding state information updates to consumers on LCC 128B. In this example, if routing engine 130B were to come back on-line, the routing engine could either serve as a standby local routing engine to standby local routing engine 131B or resume control of LCC 128B. In either case, at any one time, only one of primary local routing engine 130B and standby local routing engine 131B provide control of LCC 128B. While LCCs 128 are shown with exactly one of standby local routing engines 131, in other embodiments, some of LCCs 128 may operate without a standby local routing engine 128 or include more than one of standby local routing engines 128.


In some embodiments, operating systems executing on the primary and standby routing engines of each chassis encode commit markers and commit proposals within the hierarchically-ordered and temporally-linked data structures, which are used to track the distribution of the state information to the various consumers within multi-chassis router 120. In particular, in some embodiments, the operating systems executing on the primary and standby routing engines of each chassis maintain a commit marker and a commit proposal to track the consumption of state information for each consumer associated with a shared router resource. A shared router resource, as referred to herein, is any router resource, such as a PFE or other consumer, that is shared by both the primary routing engine and the one or more of the standby routing engines on a chassis. For example, primary master routing engine 126 encodes commit markers and commit proposals for standby master routing engine 127 as well as each of primary local routing engines 130. Primary local routing engines 130 contain commit markers only for consumers on the same chassis. For example, primary local routing engine 128A contains a separate commit marker for each of standby local routing engine 131A, PFE 132A and IFCs 134A, each of which reside on LCC 128A. In contrast, primary and standby routing engines may each contain their own software processes or daemons, which also operate as consumers of state information. In some embodiments, routing engines do not include commit markers or commit proposals for non-shared router resources.


A commit marker may be used to mark the last state update received and acknowledged by the shared consumer, while a commit proposal for a consumer indicates a position in a state update corresponding to a request for acknowledgement sent to the consumer. Consequently, the commit marker and commit proposal may be a pointer, status bit, or other data structure capable of marking a location within the hierarchically-ordered and temporally-linked data structure.


In order to be capable of assuming control of state information updates, the standby routing engines also encode the commit markers and commit proposals into their data structure. The standby local routing engines encode a copy of the commit markers encoded in the corresponding primary local routing engine, except for its own commit marker, into its own data structure. By encoding the commit markers and commit proposals within the data structure, standby routing engines 127 and 130 can ensure the extent to which consumers have been updated with state information in the event of a failover. As a result, standby routing engines 127 and 130 need only update each consumer with a limited amount of the state information that is bounded by its respective commit markers and commit protocols encoded within the state information data structure.


On a failover, when the sequence number matches either the commit marker or the commit proposal, the standby routing engine sends updates starting from the state information after the commit marker. If there are more updates after the commit proposal, it will continue to send those updates after sending updates between the commit marker and the commit proposal.


Primary routing engines 126 and 130 may also encode a separate status bit marker on each object in their state information update chains. The status bit markers may be used to indicate that the corresponding standby routing engine has indeed received that object. These status bit markers may be used in addition to the commit marker and commit proposal for the standby routing engine. As described, a primary routing engine first provides an object in the state information chain to a corresponding standby routing engine. Once the standby routing engine acknowledges receipt of an object in the data chain, that object is marked with the status bit marker and released to other consumers, such as lower level routing engines of other chassis or local consumers. The status bit marker may be necessary, for example, when a routing engine is first operating without a standby routing engine and then a standby routing engine comes online. The status bit marker forces the primary operating system to allow the standby routing engine to catch-up on state information updates within memory of the primary routing engine immediately, so that the standby routing engine becomes ready to take over state information updates as soon as possible once activated. In this manner, maintaining a status bit marker ensures that a standby routing engine receives state updates before any other consumer. This is required so to allow a standby routing engine to resume state update from wherever the primary routing engine left off in a failover.


When there is not an active corresponding standby routing engine, a primary routing engine does not use the status bit marker or wait for a standby routing engine to acknowledge receipt of an object before providing it to other consumers. While the status bit marker is described as such, it may also be a pointer or other data structure capable of marking state information objects. A status bit marker is merely an example embodiment.


Although described in reference to a two-level multi-chassis router 120, the principles of the invention may readily be applied to a multi-chassis router having three or more levels (tiers) of routing control. For example, each of local routing engines 131 of LCCs 128 of multi-chassis router 120 may operate as master routing engines over one or more additional chassis having primary local and standby routing engines. The techniques for ensuring the synchronization gradient for distributing state information may be applied to ensure that standby routing engines at any level are able to take over control for respective primary routing engines in the event of a failure.



FIG. 3 is a block diagram illustrating an exemplary control plane of multi-chassis router 120 in accordance with principles of the invention. As shown, SCC 122 includes primary master routing engine 126 and standby master routing engine 127. Each of LCCs 128 include one of primary local routing engines 130 and one of standby local routing engines 131, but for illustrative purposes, this detail is only shown on LCC 128A as primary local routing engine 130A and standby local routing engine 131A. LCCs 128B-D may be substantially similar to LCC 128A.


As described with respect to FIG. 2, routing engines 126, 127, 130 and 131 may maintain state information according to a hierarchically-ordered and temporally-linked data structure in a state chain. A portion of a state chain may represent, for example, the interface between multi-chassis router 120 and the network, which may include the current configuration of IFCs 134 and PFEs 132. State chains 150, 152, 154A and 156A may be stored in memory, such as RAM, located on respective routing engines 126, 127, 130A, 131A or external to respective routing engines 126, 127, 130A, 131A.


The state information on routing engines 126, 127, 130 and 131 is referred to as a “state chain” because the information is propagated to routing engines 126, 127, 130 and 131 according to the temporal links in the data, i.e., state objects are distributed to routing engines 127 and 130 in the order they are received by primary master routing engine 126. Likewise, primary local routing engines 126 forward the state information objects to standby local routing engines 127 using the same first-in-first-out methodology. In contrast, the hierarchical order of the state information is used when a consumer requires state information update to be provided in a certain order. For example, chassis daemon (“chassisd”) 129 may require any available state information updates regarding PFE 132A prior to receiving a state information update regarding IFCs 134A. However, generally, consumers receive state updates in a temporary order unless a state has a dependency on another state. Each consumer may have unique dependencies for receiving state information. For example dependencies could be the result of required state information for a consumer.


Consumers also maintain state information. For example, FIG. 3 shows chassisd 129, chassisd 173, PFE 132A, chassisd 135A and chassisd 175A maintaining state information 161, 165, 162A, 164A and 184A respectively. State information 161, 165, 162A,164A and 184A may, for example, represent the current state of field replaceable units, such as interface cards, encryption cards, accounting service cards, and the like. Again, state information 161, 165, 162A, 164A and 184A may be stored in memory, such as RAM, located within or external to the shared consumers. As examples, FIG. 3 shows shared consumers PFE 132A and shared intermediary consumers LCCs 128. FIG. 3 also shows non-shared consumers chassisd 129, chassisd 173, chassisd 135A and chassisd 175A. Multi-chassis router 120 also includes additional shared and non-shared consumers not shown in FIG. 3 that require state information, e.g., IFCs 134.


Because routing engines 126, 127, 130 and 131 push state data to consumers as the state information changes, rather than requiring consumers to poll for the updated state information, control plane data traffic and convergence time.


In order to prevent overloading of a consumer in the event of a large number of state updates, consumers are informed of a state update and then allowed to retrieve it once the consumer's resources allow it. In some embodiments, some consumers may be serviced indirectly by a proxy in the routing engine. For example, primary master routing engine 126 may include a proxy for each of LCCs 128. Other consumers may directly retrieve state information from state chain 150. For example, non-shared consumers, e.g., Chassisd 129, may directly retrieve state information from state chain 150.


To illustrate the process of state information updates in multi-chassis router as shown in FIG. 3, assume primary master and primary local routing engines 126 and 130 are in control and standby master and standby local routing engines 127 and 131 are serving as back-ups. Primary master routing engine 126 receives event messages indicative of a change in the state of the shared resource, e.g., PFE 132A. When primary master routing engine 126 receives an event message, primary master kernel 140 in primary master routing engine 126 updates state chain 150 by reconfiguring the data structure and updating the data stored within the data structure. In addition to updating the data based on the change of state, primary master kernel 140 may add, delete or move commit markers and/or commit proposals to various positions within the data structure in the event that the event message relates to the state of a shared routing resources, e.g., PFE 132A.


Primary master kernel 140 then replicates the state update by transmitting the state information in message form to ksync daemon 144, a state synchronization process executing on standby master routing engine 127. Primary master kernel 140 may also send an acknowledgement request to standby master routing engine 127 and may also move the commit proposal representing standby master routing engine 127 on state chain 150. Ksync daemon 144 extracts the state information and transmits the state information to standby master kernel 141. Standby master kernel 141 receives this state update and updates state chain 152 in accordance with the state update.


Similar to primary master kernel 140, standby master kernel 141 may reconfigure and update the data structures of state chain 152 based on the state information. Standby master kernel 141 may also add, delete or move commit markers and/or commit proposals to various positions within these data structures. After being updated, state chain 152 is substantially similar to state chain 150. In other words, state chain 152 and state chain 150 are synchronized. If an acknowledgement request was sent by primary master kernel 140, standby master kernel 141 transmits an acknowledgement to primary master kernel 140 via ksync daemon 144 to indicate this synchronized state. In this manner, ksync daemon 144 provides an interface between primary master kernel 140 and standby master kernel 141 that allows for the synchronization of state information. Upon receiving the acknowledgement, if any, indicating the synchronized state in standby master routing engine 127, primary master kernel 140 moves its commit marker representing standby master routing engine 127 to the current location of the commit proposal in master routing engine 127. Primary master kernel 140 then marks each of the objects represented by the acknowledgement from standby master routing engine 127 with a status bit marker to allow distribution of the state information to LCCs 128 and consumers on SCC 122. The status bit marker ensures that state information distributed from state chain 150 to other consumers has first been received by standby master routing engine 127.


Standby master routing engine 127 then updates state information in its internal daemons. For example, standby master routing engine 127 updates state information 165 in chassisd 173 in substantially the same manner as primary master routing engine 126 updates state information 161 in chassisd 129.


Once standby master routing engine 127 is synchronized with primary master routing engine 126, primary master routing engine 126 provides state information updates as necessary to various consumers within SCC 122. For example, primary master kernel 140 may issue alerts to chassisd 129 to indicate a change in state information 160. In response, primary master kernel 140 receives requests from the consumers for state information. Primary master kernel 140 services each request by traversing the hierarchically-ordered and temporally-linked data structure of state chain 150 and issuing update messages to the requesting consumer. Primary master kernel 140 generates the update messages to contain state information that the consumer has not already received based on the respective commit proposal and commit marker for the consumer. The consumers respond to acknowledge receipt of the state data up to an acknowledgement request from primary master kernel 140. Upon updating the consumers and receiving their acknowledgements, primary master kernel 140 moves commit markers within the data structure of state chain 150 to reflect the updates.


After standby master routing engine 127 is synchronized with primary master routing engine 126, primary master kernel 140 again replicates the state update by transmitting the state information in message form to ksync daemons 145, one in each of primary local routing engines 130, as represented by ksync daemon 145A. As state updates occur, primary master kernel 140 may also send an acknowledgement request once it reaches a commit proposal in the state chain. The position of commit proposal in the state chain may not be consistently spaced. For example, the position of commit proposals for a consumer may depend on time lapsed since a previous commit proposal or other variable.


Ksync daemons 145 extract the state information and transmit the state information to primary kernels 142, as represented by primary local kernel 142A. Primary kernels 142 receive this state update and update state chains 154, as represented by state chain 154A, in accordance with the state update. In addition to updating state chain 154A with the state update information, primary local kernel 142A adds an additional object referred to as a “ksync object” corresponding to an acknowledgement request received from primary kernel 140 to state chain 154A. The ksync object on state chain 154A is required to remind primary kernel 142A to respond to the acknowledgement request made by master routing engine 126 once the state update has replicated to all consumers that require it on LCC 128A. This is because primary kernel 142A must wait until primary kernel 142A receives acknowledgements from all consumers on LCC 128A that require the state update before providing an acknowledgement for the state update to master routing engine 126.


After being synchronized with state chain 150, state chains 154 differ from state chain 150 in that primary local kernels 142 encode different sets of commit markers and commit proposals within each of state chains 154. Each of state chains 154 contains its own set of commit markers and commit proposals, which represent consumers on their particular chassis. For example, primary local kernel 142A will encode a separate commit marker for each of standby local routing engine 128A, PFE 132A and IFCs 134A.


Once primary local routing engines 130 are synchronized with primary master routing engine 126, they replicate the state update to standby local routing engines 131 by transmitting the state information in message form to ksync daemons 146, represented by ksync daemon 146A. Ksync daemon 146A extracts the state information and transmits the state information to standby local kernel 143A. Standby local kernel 143A receives this state update and updates state chain 156A in accordance with the state update. Once standby local kernel 143A has updated state chain 156A, if requested, it acknowledges receipt of the state update to primary local kernel 142A.


Upon receipt of the acknowledgement, primary local kernel 142A moves the commit marker representing standby local routing engine 131A to the commit proposal and marks all state objects in state chain 154A between the commit marker and commit proposal with a status bit marker signifying that the objects are available to be replicated to consumers on LCC 128A.


Primary local kernel 142A issues alerts to consumers on LCC 128A that require the state updates represented by one or more objects in state chain 154A. In response, the consumers request state information and primary kernel 142 services those requests by traversing the hierarchically-ordered and temporally-linked data structure of state chain 154A and issuing update messages to the requesting consumer.


Once each commit marker on LCC 128A passes a ksync object in state chain 154A, primary local kernel 142A responds to the acknowledgement request form primary master kernel 140. Primary master kernel 140 then moves the commit marker for LCC 128A within state chain 150 to the commit proposal for LCC 128A.


The described processes for distribution of state update information on multi-chassis router 120 may be performed when new state information is created or at defined time intervals. State updates occur often in multi-chassis router 120, and synchronization may occur continuously or nearly continuously. Because synchronization does not occur instantaneously, multi-chassis router 120 may perform multiple iterations of state updates simultaneously. For example, primary master kernel 140 may simultaneously update primary kernels 142 in primary local routing engines 130 with a first state update while updating standby master kernel 141 in standby master routing engine 127 with a second state update because standby master kernel 141 has already acknowledged the first state update.


Synchronization of state information in state chains 150, 152, 154 and 156 continues in this manner until a failover occurs. Failover may occur for primary master routing engine 126 and/or for any of primary local routing engines 130. Once a failover occurs, the standby routing engine corresponding to the failed primary routing engine assumes the responsibilities of the primary routing engine. For example, if failover occurs while a primary master routing engine 126 is updating primary local routing engines 130, standby master routing engine 127 resumes the updates. In particular, standby master kernel 141 uses the commit markers and commit proposals stored and synchronized within state chain 152 to continue the updates from the same point that primary master kernel 140 left off.


In the event of a failover in one of LCCs 128, e.g. LCC 128A, standby local routing engine 131A resumes updating consumers on LCC 128A. Standby local routing engine 131A operates without a back-up routing engine until primary local routing engine 130A comes on-line. Because there is no back-up, upon receiving a state update, standby local routing engine 131A does not maintain a status bit marker to signify an object has been acknowledged by a standby routing engine. Instead, standby local routing engine 131A immediately updates consumers.


Standby local routing engine 131A begins sending state updates from the hierarchically-ordered and temporally-linked data structures of state chain 156A from the commit markers corresponding to each consumer. For example, standby local routing engine 131A reaches a commit marker corresponding to PFE 132A. Then, standby local routing engine 131A issues state update messages to PFE 132A. PFE 132A begins receiving state objects in state chain 162A that follow its commit marker. PFE 132A may receive redundant state update messages if primary local routing engine 130A sent the same state update message but failed prior to receiving an acknowledgement from PFE 132A. In this case, PFE 132A ignores the redundant state update message other than to respond to an acknowledgement request regarding the redundant state update.


Standby master routing engine 127 maintains state chain 152 to be synchronous with state chain 150 of primary master routing engine 126. Standby local routing engines 131 maintain state chains 156 to be synchronous with state chains 154 of primary local routing engines 130. Therefore, standby routing engines 127 and 131 may facilitate failover by assuming control without having to learn state information updates, e.g., by restarting the chassis, or even multi-chassis router 120 from a known state. Moreover, standby routing engines 127 and 131 can update the consumers with regard to the limited amount of the state information bounded by their respective commit markers and commit proposals encoded within the state information data structure.


As described, in the event of a failover, one or more standby routing engines 127 and 131 assumes the role of a primary routing engine. Later, the primary routing engine may return to an operational status, e.g., after being reset, and assume the role of a standby routing engine. In this case, the primary routing engine initiates a state synchronization process to synchronize its state information with state information of the standby routing engine, now operating as the primary routing engine.


Routing engines 126, 127, 130 and 131 may operate according to executable instructions fetched from one or more computer-readable media. Examples of such media include random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), electrically erasable programmable read-only memory (EEPROM), flash memory, and the like. The functions of multi-chassis router 120 may be implemented by executing the instructions of the computer-readable medium with one or more processors, discrete hardware circuitry, firmware, software executing on a programmable processor, or combinations thereof.



FIGS. 4A and 4B (“FIG. 4”) are block diagrams illustrating exemplary propagation of state updates to state chains 150, 152, 154A, 156A, 162A and 164A in primary master routing engine 126, standby master routing engine 127, primary local routing engine 130A, standby local routing engine 131A, PFE 132A and chassisd 135A respectively. As illustrated, state information may be stored within multiple objects 250. As state updates occur, new objects are created and distributed to the routing engines and shared consumers on multi-chassis router 120. Moreover, objects 250 may be propagated to routing engines 126, 127, 130 and 131 according to the temporal-links in the data structure. Objects 250 may store information corresponding to routes, firewalls, interface cards, and other components.


As shown in FIG. 4A, primary master kernel 140 in primary master routing engine 126 maintains state information in the form of state chain 150 including objects 250A-250H, as illustrated in FIG. 4A. Primary master kernel 140 associates commit marker 201 and commit proposal 211 with standby master routing engine 127. Commit proposal 212 and commit marker 202 correspond to primary local routing engine 130A. State chain 150 may contain additional commit markers and commit proposals (not shown) for primary local routing engines on each of LCCs 128B-128D. State chain 150 may additionally include commit markers and commit proposals (not shown) for shared consumers on SCC 122.


Standby master kernel 141 also maintains state chain 152 including objects 250A-250H. State chain 152 is synchronized with state chain 150. Objects on state chain 150 include a status bit marker to indicate whether that object has been received by standby master routing engine 127. As shown in FIG. 4A, objects 250A-250G in state chain 150 are marked as having been added to state chain 152. While state chain 152 currently includes object 250F, standby master routing engine 127 has not yet acknowledged receipt of object 250F and will do so in response to commit proposal 211. Until then, object 250F in state chain 150 is not marked as having been received by standby master routing engine 127.


With reference to FIG. 4A, state chain 152 is substantially similar to state chain 150, and is in the process of being synchronized with state chain 150 and includes all objects in state chain 150. State chain 152 includes commit marker 202 and commit proposal 212 corresponding to primary local routing engine 130A. State chain 152 includes additional commit markers and commit proposals (not shown) that are part of on state chain 150, except for commit marker 201 and commit proposal 211, which correspond to standby local routing engine 127. In contrast to state chain 150, objects on state chain 152 do not include a status bit marker to show if a standby master routing engine has received each object. In the event that standby master routing engine 127 becomes the primary master routing engine and is backed-up by another routing engine serving as a standby master routing engine, status bit markers would then be added to each object in state chain 152.


State chain 154A is maintained by primary local routing engine 130A in LCC 128A, and is representative of state chains in primary local routing engines 130 in each of LCCs 128. In this example, state chain 154A includes commit markers 203 and 204 for standby local routing engine 131A and PFE 132A, respectively. State chain 154A also includes commit proposal 214 for PFE 132A. As shown, state chain 154A does not include a commit proposal for standby local routing engine 131A. At other times during state update operation state chain 154A may also include a commit proposal for standby local routing engine 131A. Similar to state chain 150 on primary master routing engine 126, state chain 154A includes a status bit marker on each object to indicate when the object has been acknowledged by standby local routing engine 131A. In addition to objects 250, primary local kernel 142A inserts or moves ksync object 260 in state chain 154A upon receipt of an acknowledgement request from primary master routing engine 126. Ksync object 260 in state chain 154A reminds primary local routing engine to respond to the acknowledgement request once all commit markers in state chain 154A have passed the ksync object.


In the example of FIG. 4A, commit marker 202 marks object 250C as holding the last state update acknowledged by primary local routing engine 130A. Commit proposal 212 marks object 250E as holding the most recent acknowledgement request sent to primary local routing engine 130A. Accordingly, primary local kernel 142A has already acknowledged receipt of object 250C, after which primary master kernel 140 moved commit proposal 212 just after object 250C in state chain 150. Primary master kernel 140 forwarded the acknowledgement to standby master routing engine 127, where standby master kernel 141 also moved commit marker 202 beyond object 250C in state chain 152. The additional commit markers (not shown) for primary local routing engines on each of LCCs 128B-128D are manipulated in the same manner. Commit marker 201 and commit proposal 211 operate in a similar manner with respect to standby master routing engine 127, but only exist on state chain 150.


Again with reference to FIG. 4A, state chain 154A of primary local routing engine 128A includes objects 250A-250G and ksync object 260. Commit marker 203 indicates object 250G as holding the most recent, in temporal relation, state information acknowledged by standby local routing engine 131A. Similarly, commit marker 204 indicates object 250E as holding the most recent state information acknowledged by PFE 132A. Commit proposal 214 indicates object 250G as the most recent object in an acknowledgement request sent to PFE 132A.


Primary local kernel 142A inserted ksync object 260 into state chain 154A after object 250E to mark the receipt of an acknowledgement request for objects up to and including object 250E from primary master routing engine 126. Ksync object 260 functions to remind primary local routing engine 130A to respond to the acknowledgement request once all consumers on LCC 128A acknowledged receipt of all objects they require up to and including object 250E in the temporally-linked state chain 154A. Primary local kernel 142A moves ksync object 260 within state chain 154A to the end of state chain 154A every time primary master routing engine 126 sends an acknowledgement request to LCC 128A. Primary master routing engine 126 wait to send a new acknowledgement request until receiving an acknowledgement to the previous acknowledgement request. Each acknowledgement request received by LCC 128A from primary master routing engine 126 results in a new location for ksync object 260 and a corresponding new location for commit proposal 212.


In multi-chassis router 120, before acknowledging a state update from primary master routing engine 126, primary local routing engine 130A forwards the update to standby local routing engine 131A and every consumer on LCC 128A that requires the update. Furthermore, primary local routing engine 130A must wait to receive acknowledgements from standby local routing engine 131A and each shared consumer on LCC 128A before sending the acknowledgement to the master routing engine. For example, in some embodiments, acknowledgements may not be required for non-shared consumers. For example, commit marker 202 in state chains 150 and 152 will remain between objects 250C and 250D until every commit marker in state chain 154A reaches or passes ksync object 260. In the event of a failover, there may be a significant amount of state update information received, but not acknowledged, by primary local routing engine 130A.


Primary master kernel 140 encodes a random number “N” (not shown), referred to as a sequence number, into commit proposal 212 in state chain 150. For example, N may be a random three-digit number. The acknowledgement request to primary local routing engine corresponding to object 250C also includes the random number N. Primary local kernel 142A encodes the number N into ksync object 260. Once synchronized, ksync object 160A in state chain 156A of standby local routing engine 131A also includes the sequence number N.


After primary local kernel 142A responds to the acknowledgement request corresponding to ksync object 260, primary master kernel 140 will later send a new acknowledgement request to corresponding to an object further along state chain 150 and move commit proposal 212 to that object. At this location, primary master kernel 140 encodes a number N+1 into commit proposal 212. In this manner, primary master kernel 140 counts the number of times that an acknowledgement request is sent to LCC 128A. Primary master kernel 140 includes an updated sequence number in every subsequent acknowledgement request.


In the event of a failover, either master or local, upon reconnect with SCC 122, the active local routing engine in LCC 128A sends its most recent sequence number to the active master routing engine in SCC 122. This has the effect of informing the master routing engine that LCC 128A received the state corresponding to that sequence number. Assuming the sequence number sent by LCC 128A matches the sequence number of either the corresponding commit marker or commit proposal in the state chain of the active master kernel, the active master routing engine can resume state updates from the commit marker. If the sequence numbers do not match, then the active master routing engine sends an error signal. In the event of an error, state chain 154 is re-synchronized with the active master routing engine, for example, by way of clearing all state information from LCC 128A and relearning it from SCC 122.


A sequence number mismatch, while rare, may occur, for example, when a standby routing engine, either master or local, assumes the duties of the respective primary routing engine after the primary routing engine produces a new sequence number, but before the primary routing engine was able to update the sequence number encoded in the state chain of the standby routing engine. For example, primary master routing engine 126 updates the sequence number in standby master routing engine 127 after the position of commit marker 212 moves in state chain 150 when it updates the position of the commit marker in state chain 152. Primary local routing engine 130A sends an updated sequence number to standby local routing engine 131A with each new acknowledgement request, which is then included within the corresponding ksync object in state chain 156A. By comparing sequence numbers before resuming state information updates after a failover, multi-chassis router 120 ensures that state chains 154 and 156 in local routing engines 130 and 131 contain the same set of objects 250 as state chains 150 and 152 in master routing engines 126 and 127.


Even if sequence numbers match, the local routing engine may receive objects already available in its state chain. For example, state chain 154A includes two objects after ksync object 260. In the event of a master routing engine failover, standby master routing engine 127 receives the sequence number N+1 from primary local routing engine 130A. This sequence number corresponds to the sequence number on commit proposal 212 on state chain 152. Therefore, standby master routing engine 127 knows that primary local routing engine 130A received all objects up to and including object 250E. However, standby master routing engine 127 does not know that primary local routing engine 130A has also received objects 250F and 250G. Therefore, standby master routing engine 127 resumes state updates for LCC 128A with object 250F. Primary local kernel 142A simply ignores objects 250F and 250G before adding a new object to state chain 154A, e.g., object 250H. Depending on the circumstances, the first new object to state chain 154A may be either ksync object 260 or a state update object 250.


During the course of forwarding state updates to standby master routing engine 127 and primary local routing engines 130, primary master kernel 140 also receives new state update objects. State update objects may be created by processes in any chassis of multi-chassis router 120. However, every state update is forwarded to SCC 122 and distributed according to the techniques described prior to being acted upon in any chassis of multi-chassis router 120. For example, a state update produced on LCC 128A is forwarded to SCC 122, where it is inserted in state chain 150 and copied to state chain 152 before being forwarded back to primary local routing engine 130A in LCC 128A. In this manner, primary master routing engine 126 maintains centralized control of multi-chassis router 120.



FIG. 4B illustrates the propagation of state updates from state chain 154A of primary local routing engine 130A to state chain 156A in standby local routing engine 131A, and state information 162A and 164A for PFE 132A and chassisd 135A, respectively. The state information updates shown in FIG. 4B correspond with the state information updates shown in FIG. 4A.


As shown in FIG. 4B, PFE 132A and chassisd 135A may receive objects according to the hierarchical-order of the data structure. Furthermore, PFE 132A and chassisd 135A do not each need to receive every object, but only those updates relevant to their operation. In FIG. 4B, state information 162A and 164A for PFE 132A and chassisd 135A is exemplary of state information updates provided to consumers in any of LCCs 128. For example, LCC 128A includes additional consumers that require state information updates. Updates to other consumers in LCCs 128 and SCC 122 occur in a substantially similar manner to exemplary consumers PFE 132A and chassisd 135A.


State chain 156A is synchronized with state chain 154A. Once synchronized, state chain 156A includes all the objects in state chain 154A. However, because state update information is regularly created in multi-chassis router 120, state chain 156A may often require synchronization with state chain 154A. State chain 156A includes commit marker 204 and commit proposals 214 corresponding to PFE 132A. State chain 156A also includes additional commit markers and commit proposals (not shown) that are part of state chain 150, except for commit marker 203 and commit proposal 213, which correspond to standby local routing engine 131A. In contrast to state chain 154A, objects on state chain 156A do not include a status bit marker to show if a standby local routing engine has received each object. In the event that standby local routing engine 131A becomes the primary local routing engine on LCC 128A and is backed-up by another routing engine serving as a standby local routing engine, status bit markers would then be added to each object in state chain 156A.


State chain 156A includes objects 250A-250G, ksync object 260, commit marker 204 and commit proposal 214. Each of objects 250A-250G are marked as having been acknowledges by standby local routing engine 131A. For example, standby local routing engine 131A acknowledged receipt of ksync object 260 in response to an acknowledgement request. After receiving acknowledgement of ksync object 260 from standby local routing engine 131A, primary local routing engine 130A marked objects 250A-250E and ksync object 260 as having been acknowledged by standby local routing engine 131A. Primary local routing engine 130A then forwarded objects 250A-250E, as necessary, to PFE 132A and chassisd 135A. As shown in FIG. 4B, PFE 132A does not require object 250B.


LCC 128A will not respond to the acknowledgment request corresponding to commit proposal 212 on state chain 150 until each of the commit markers on state chain 154A pass or reach ksync object 260. At that point, primary local routing engine 130A will send an acknowledgment to primary master routing engine 126 and primary master kernel 140 will move commit marker 202 to replace commit proposal 212. Primary master kernel 140 may then reinsert commit proposal 212 at a later position along state chain 150.


The order which consumers receive state information updates may be dependent on the hierarchical-order of the state update data structure. As shown in FIG. 4B, PFE 132A and chassisd 135A received state updates according to the hierarchical-order of the state update data structure. For example, chassisd 135A received state information updates in the following order: object 250A, object 250C, object 250B, object 250F and object 250D. Consumers may require state update objects in a particular order as defined by the hierarchical-order of the state update data structure. For example, a consumer may require state update information regarding PFEs 132 prior to state update information regarding IFCs 134.


The propagation of state updates continues indefinitely in multi-chassis router 120. For example, state chain 154A in primary local routing engine 130A includes object 250G, which has not yet been forwarded to consumers on LCC 128A. Also, primary master routing engine 126 includes object 250H, which has not yet been forwarded to LCC 128A. Primary master routing engine 126 must wait for an acknowledgement from standby master routing engine 127 in order to forward objects 250H to LCC 128A.


As described, processes in every chassis of multi-chassis router 120 may produce state update information, and state update information is produced regularly during operation of multi-chassis router 120. However, in the event of a failure of a consumer, e.g., chassisd 135A in LCC 128A, in order to restart, the consumer may require receiving substantially all state update information required to operate. Chassisd 135A in LCC 128A, for example, may require information on routing engines 130A and 131A, PFE 132A and IFCs 134A. Chassisd 135A requires state information to be resent from primary local kernel 142A. Upon restarting after a failure, chassisd 135A informs primary local kernel 142A that it requires state information updates. Primary local kernel 142A then resends chassisd 135A all the state information it requires that is available in state chain 154A. As before, this state information is sent according to the particular hierarchical data structure necessary for chassisd 134A.



FIG. 5 is a flowchart illustrating exemplary operation of multi-chassis router 120 (FIGS. 2 and 3) when synchronizing and processing state updates in accordance with the principles of the invention. Initially, multi-chassis router 120 and, more particularly, primary master kernel 140 receive information regarding changes to the state of the network or resources with multi-chassis router 120 (354). As one example, primary master kernel 140 receives state information from chassisd 129. In another example, state information may define changes to the network topology, such as the addition of network routes.


In response, primary master kernel 140 updates state chain 150 to reflect the changes in state of multi-chassis router 120 (356). Primary master kernel 140 updates state chain 150 by adding and/or removing objects and adding, deleting or moving commit markers and commit proposals.


Next, primary master kernel 140 synchronizes state chain 152 with state chain 150. In particular, primary master kernel 140 replicates the state information of state chain 150 and transmits the state information to standby master kernel 141 via ksync daemon 144 (358). Upon receiving the updates, standby master kernel 141 processes the state information and executes the necessary changes to synchronize state chain 152 (360). Standby master kernel 141 waits for an acknowledgement request from primary master kernel 140 and then transmits an acknowledgement to primary master kernel 140 via ksync daemon 144 (362).


Once primary master kernel 140 receives the acknowledgement, the primary kernel marks the acknowledged objects with a status bit marker and replicates the state update information to primary local kernels 142 and any local consumers within SCC 122. State update information is replicated via ksync daemons 145 to primary local kernels 142 on and directly to any local consumers within SCC 122 (364). Upon receiving the updates, primary local kernels 142 process the state information and execute the necessary changes to synchronize state chains 154 (366).


State chains 156 are separately synchronized with state chains 154 in each of LCCs 128. Primary local kernels 142 replicate the state information of state chains 154 and transmit the state information to standby kernels 143 via ksync daemons 146 (368). Upon receiving the updates, standby master kernel 141 processes the state information and executes the necessary changes to synchronize state chain 152 (370). Standby kernels 143 then transmit acknowledgements to primary kernels 142 via ksync daemons 146 (372).


Primary local kernels 142 each receive a separate acknowledgment and proceed to transmit the updated state information to consumers, such as PFE 132A (374). In particular, primary local routing engines 130 issue alerts to indicate to the consumers that state updates exist. Upon receiving requests from the consumers, primary kernels 142 transmit the updated state information based upon the locations of the respective commit proposals and commit markers associated with the consumers. The consumers receive the updated state information and make the necessary changes to their respective state information. For example, PFE 132A may receive updated state information from primary local kernel 142A and update state information 162A.


Once a primary local kernel 142 receives acknowledgements from each shared consumer that requires the state update, it sends an acknowledgement to master routing engine 126 (376). Once master routing engine 126 receives an acknowledgement for a state update, it may dequeue the update. For example, in some embodiments, master routing engine 126 may delete the least recent objects in state chain 150 after those objects have been acknowledged by all consumers of primary master routing engine 126, e.g., LCCs 128. The deletions may also be propagated to standby master routing engine 127 and local routing engines LCCs 128. In other embodiments, routing engines may maintain a state information object until it becomes obsolete due to a state update.


As illustrated, multi-chassis router 120 synchronizes and updates state information in accordance with a defined “synchronization gradient,” whereby primary master kernel 140 receives state updates, followed by standby master kernel 141, followed by primary local kernels 142A, followed by standby local kernels 143A, followed by consumers in each chassis of multi-chassis router 120. This synchronization gradient may ensure that upon failover, standby routing engines 127 and 131 contain state information that is at least as current as the state information provided to all consumers of state information. Consequently, standby routing engines 127 and 131 are able to readily assume responsibility in the event of a failover, and can continue updating the consumers with the state information as necessary.



FIG. 6 is a flowchart illustrating exemplary failover of a master routing engine in a multi-chassis router that allows for non-stop forwarding. The failover process is described with reference to multi-chassis router 120 of FIGS. 2 and 3. For exemplary purposes, first assume that primary master routing engine 126 fails (602). For example, failure of primary master routing engine 126 may result from a software failure, hardware failure or an administrator taking primary master routing engine 126 offline. Because of the failure, the connection between primary master routing engine 126 and primary local routing engines 130 also fails (604). Primary local routing engines 130 sever connections with PFEs 132 (606). This prevents primary local routing engines 130 from having to store state updates produced in PFEs 132 until re-connecting with a master routing engine.


Next, primary local routing engines 130 open connections with standby master routing engine 127 (608). Once the connections have been established, primary local routing engines 130 each send their unique sequence number to standby master routing engine 127 (610). Standby master routing engine 127 compares each sequence number received from primary local routing engines 130 with the sequence numbers in commit proposals and commit markers for LCCs 128 in state chain 152 (612). For example, as shown in FIG. 4, commit proposal 212 corresponds to LCC 128A. If the sequence number from one of LCCs 128 matches either the corresponding commit proposal or commit marker each case, standby master kernel 141 begins forwarding state updates to the one of LCCs 128 from the commit marker (616).


In the rare occurrence that sequence numbers do not match, master routing engine 127 sends an error signal to the LCC. In the event of an error, state chain 154 in one or more of LCCs 128 must be re-synchronized with state chain 152, for example, by clearing all state information from the LCC and resending the state information in state chain 152 (614). Clearing all state information from one or more of LCCs 128 may interrupt packet forwarding in multi-chassis router 120. After resetting one or more of LCCs 128 to a known state, master routing engine 127 resumes state updates (616).


In this manner, multi-chassis router 120 maintains packet forwarding during a failover of master routing engine 126. State updates are temporarily interrupted during the failover process while local routing engines 130 establish connections with master routing engine 127, but multi-chassis router 120 continues to send and receive packets during this time according to already known state information. Only in the rare event of a sequence number mismatch might packet forwarding be interrupted in multi-chassis router 120.


While the described failover process describes master routing engine 126 as the primary master routing engine and master routing engine 127 as the standby master routing engine, master routing engines 126 and 127 are substantially similar. That is, both master routing engines 126 and 127 may act as either the primary master routing engine or the standby master routing engine. Therefore, failover may also occur by transferring central control of multi-chassis router 120 from master routing engine 127 to master routing engine 126.



FIG. 7 is a flowchart illustrating exemplary failover of a local routing engine in a multi-chassis router that allows for non-stop forwarding. The failover process is described with reference to multi-chassis router 120 of FIGS. 2 and 3 and, in particular, with reference to LCC 128A. For exemplary purposes, first assume that primary local routing engine 130A fails (702). For example, failure of primary local routing engine 130A may result from a software failure, hardware failure or an administrator taking primary local routing engine 130A offline. Because of the failure, the connection between primary local routing engine 130A and primary master routing engine 126 fails (704). The connection between primary local routing engine 130A and PFEs 132A also fails (706).


Next, standby local routing engine 131A opens a connection with primary master routing engine 126 (708). Once the connection has been established, standby local routing engine 131A sends the sequence number contained in the most recent ksync object in state chain 156A to primary master routing engine 126 (710). Primary master routing engine 126 compares this sequence number with the sequence numbers in commit marker 202 and commit proposal 212. If the sequence number matches either the commit proposal or commit marker, primary master kernel 140 begins forwarding state updates to LCC 128A, beginning with the first object in state chain 150 following the commit marker (716).


In the rare occurrence that sequence numbers do not match, master routing engine 127 sends an error signal to standby local routing engine 131A. In the event of an error, state chain 156A must be re-synchronized with state chain 150, for example, by clearing all state information from LCC 128A (714). Optionally, only LCC 128A may need to be reset to a known state. LCCs 128B-D may continue to receive state updates according to their commit markers and commit proposals in state chain 150. Resetting LCC 128A to a known state may interrupt packet forwarding in multi-chassis router 120. After resetting LCC 128A to a known state, master routing engine 126 resumes state updates (616). This may require requesting state information from processes throughout multi-chassis router 120.


In this manner, multi-chassis router 120 maintains packet forwarding during a failover of local routing engine 130A. State updates for LCC 128A are temporarily interrupted during the failover process while standby local routing engine 130 established a connection with master routing engine 126, but LCC 128A continues to send and receive packets during this time according to already known state information held by consumers of LCC 128A. Only in the rare event of a sequence number mismatch might packet forwarding be interrupted in multi-chassis router 120.


While the described failover process describes local routing engine 130A as the primary local routing engine and local routing engine 131A as the standby local routing engine, local routing engines 130A and 131A are substantially similar. That is, both local routing engines 130A and 131A may act as either the primary local routing engine or the standby local routing engine. Furthermore, the described local routing engine failover techniques are further applicable to primary and standby local routing engines contained within each of LCCs 128.


One or more of the techniques described herein may be partially or wholly executed in software. For example, a computer-readable medium may store or otherwise comprise computer-readable instruction, i.e., program code that can be executed by a processor to carry out one or more of the techniques described above. For example, the computer-readable medium may comprise random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), electrically erasable programmable read-only memory (EEPROM), flash memory, magnetic or optical media, or the like.


Various embodiments of the invention have been described. Although described in reference to a multi-chassis router, which each chassis including a plurality of routing engines, the techniques may be applied to any multi-chassis device having a plurality of control nodes in at least one chassis. Examples of other devices include switches, gateways, intelligent hubs, firewalls, workstations, file servers, database servers, and computing devices generally. Furthermore, the described embodiments refer to hierarchically-ordered and temporally-linked data structures, but other embodiments may use different data structures. These and other embodiments are within the scope of the following claims.

Claims
  • 1. A method for distributing state information within a multi-chassis router comprising a first master routing engine and a second master routing engine, wherein the second master routing engine acts as a standby routing engine for the first master routing engine, the multi-chassis router further comprising a local routing engine, wherein the first master routing engine, the second master routing engine, and the local routing engine are physically coupled within the multi-chassis router and configured to operate as a single router within a network, the method comprising: communicating updates to the state information with the first master routing engine to the second master routing engine of the multi-chassis router, wherein the updates are communicated according to a hierarchically-ordered and temporally-linked data structure, wherein communicating the updates to the second master routing engine comprises replicating the hierarchically-ordered and temporally-linked data structure within the second master routing engine; andafter communicating the updates from the first master routing engine to the second master routing engine, communicating the updates from the first master routing engine to the local routing engine.
  • 2. The method of claim 1, wherein the local routing engine comprises a first local routing engine, wherein the multi-chassis router further comprises a second local routing engine, wherein the second local routing engine acts as a standby routing engine for the first local routing engine, the method further comprising: receiving the updates to the state information with the first local routing engine of the multi-chassis router from the first master routing engine, wherein the updates are received according to the hierarchically-ordered and temporally-linked data structure;communicating the updates from the first local routing engine to the second local routing engine, wherein communicating the updates to the second local routing engine comprises replicating the hierarchically-ordered and temporally-linked data structure within the second local routing engine; andafter communicating the updates from the first local routing engine to the second local routing engine, communicating the updates from the first local routing engine to one or more consumers of the state information, wherein the one or more consumers comprise a processor or module within the multi-chassis router that receives and utilizes state information from the first local routing engine.
  • 3. The method of claim 2, further comprising receiving an acknowledgement for the updates from the second local routing engine prior to communicating the updates to the one or more consumers, wherein communicating the updates from the first local routing engine to the one or more consumers comprises communicating the updates to the one or more consumers in response to the acknowledgement from the second local routing engine.
  • 4. The method of claim 1, wherein communicating the updates to the second master routing engine comprises communicating the updates in accordance with an order that requires the updates to be communicated to the second master routing engine prior to communicating the updates to the local routing engine.
  • 5. The method of claim 1, wherein communicating the updates to the local routing engine comprises communicating the updates to the local routing engine in accordance with the hierarchically-ordered and temporally-linked data structure.
  • 6. The method of claim 5, further comprising: setting a commit proposal to identify a most recent object of the hierarchically-ordered and temporally-linked data structure that is the subject of an acknowledgement request sent to one or more consumers that comprise a processor or module within the multi-chassis router; andafter receiving an acknowledgement from each of the one or more consumers in response to the acknowledgement request, setting a commit marker to identify the object of the hierarchically-ordered and temporally-linked data structure identified by the commit proposal.
  • 7. The method of claim 6, further comprising: replicating the commit proposal and the commit marker to the second master routing engine; andcommunicating a portion of the hierarchically-ordered and temporally-linked data structure following the replicated commit marker to the local routing engine from the second master routing engine in the event the first master routing engine fails.
  • 8. The method of claim 7, further comprising issuing a communication from the first master routing engine to cause the second master routing engine to set the replicated commit proposal to identify a most recent portion of the hierarchically-ordered and temporally-linked data structure that is the subject of the acknowledgement request.
  • 9. The method of claim 7, further comprising issuing a communication from the first master routing engine to cause the second master routing engine to set the replicated commit marker to identify a most recent portion of the hierarchically-ordered and temporally-linked data structure that has been communicated to the local routing engine and for which an acknowledgement has been received from the local routing engine.
  • 10. The method of claim 6, further comprising sending an acknowledgement of the received updates to the master routing engine for a most recent portion of the hierarchically-ordered and temporally-linked data structure prior to the commit marker, after receiving the acknowledgement from each of the one or more consumers.
  • 11. The method of claim 5, further comprising: receiving, with the local routing engine, an acknowledgement request from the first master routing engine; andinserting, with the local routing engine, a marker in the hierarchically-ordered and temporally-linked data structure to store the acknowledgement request.
  • 12. The method of claim 11, further comprising: receiving, with the local routing engine, a first acknowledgement for all objects less recent than the marker; andsending, with the local routing engine, a second acknowledgement to the master routing engine in response to the acknowledgement request, after receiving the first acknowledgement.
  • 13. The method of claim 5, further comprising receiving, with the local routing engine from the first master routing engine, instructions to delete a portion of the hierarchically-ordered and temporally-linked data structure.
  • 14. The method of claim 1, further comprising: receiving an acknowledgement for the updates from the second local routing engine; andmarking objects in the hierarchically-ordered and temporally linked data structure to indicate the updates have been acknowledged by the second local routing engine.
  • 15. The method of claim 1, further comprising receiving, with the first local routing engine, a first acknowledgement request from the first master routing engine to acknowledge receipt of the updates to the state information, the first acknowledgement including a sequence number, wherein the sequence number represents the first acknowledgement request in a count of acknowledgement requests received from the first master routing engine.
  • 16. The method of claim 15, further comprising: counting, with the first master routing engine, the number of acknowledgement requests sent to the first local routing engine from the first master routing engine; andstoring, with the first master routing engine, a value representing the count within objects that store the acknowledgement requests.
  • 17. The method of claim 1, wherein the multi-chassis router comprises a chassis comprising the local routing engine,wherein the local routing engine comprises a primary local routing engine for the chassis of the multi-chassis router, andwherein the first master routing engine comprises a primary master routing engine for the control node within the multi-chassis router.
  • 18. The method of claim 2, wherein at least one of the consumers comprises an interface card for the local routing engine or a packet forwarding component for the first local routing engine.
  • 19. A multi-chassis network device comprising: a first master routing engine;a second master routing engine configured as a standby master routing engine for the first master routing engine;a first local routing engine;a second local routing engine configured as a standby routing engine for the first local routing engine; anda plurality of consumers, wherein at least one of the plurality of consumers comprises a packet forwarding component for the first local routing engine,wherein the first master routing engine, the second master routing engine, the first local routing engine, and the second local routing engine are physically coupled within the multi-chassis network device and configured to operate as a single router within a network,wherein the first master routing engine manages state information and communicates updates to the state information to the second master routing engine before communicating the updates to the first local routing engine,wherein the first local routing engine communicates the updates to the second local routing engine before communicating the updates to the plurality of consumers,wherein the first local routing engine issues a communication that causes the second local routing engine to set a commit proposal to identify a most recent portion of the state information that is the subject of an acknowledgement request sent to at least one of the plurality of consumers, andwherein the first local routing engine issues a communication that causes the second local routing engine to set a commit marker to identify a the portion of the state information identified by the commit proposal after an acknowledgement has been received from the at least one of plurality of consumers in response to the acknowledgement request.
  • 20. The multi-chassis network device of claim 19, wherein the first master routing engine manages the state information within a hierarchically-ordered and temporally-linked data structure and communicates the updates by replicating the hierarchically-ordered and temporally-linked data structure within the first local routing engine, andwherein the first local routing engine replicates the commit proposal and the commit marker to the second local routing engine, and the second local routing engine communicates a portion of the replicated hierarchically-ordered and temporally-linked data structure after the commit marker to the at least one of the plurality of consumers in the event the first local routing engine fails.
  • 21. The multi-chassis network device of claim 20, wherein the first local routing engine communicates the updates to the second local routing engine in accordance with the hierarchically-ordered and temporally-linked data structure.
  • 22. The multi-chassis network device of claim 21, wherein the first local routing engine utilizes a commit proposal and a commit marker to identify the portion of the state information within the hierarchically-ordered and temporally-linked data structure that has been communicated to the second local routing engine.
  • 23. The multi-chassis network device of claim 22, wherein the first local routing engine sets the commit proposal to identify a most recent potion of the updates that is the subject of an acknowledgement request sent to the second local routing engine, andsets the commit marker to identify the portion of the updates that has been communicated to the second local routing engine and for which an acknowledgement has been received from the second local routing engine in response to the acknowledgement request.
Parent Case Info

This application is a continuation of U.S. Pat. No. 7,804,769, filed Dec. 1, 2005, the entire content of which is incorporated herein by reference.

US Referenced Citations (112)
Number Name Date Kind
5454080 Fasig et al. Sep 1995 A
5684984 Jones et al. Nov 1997 A
5848227 Sheu Dec 1998 A
6154847 Schofield et al. Nov 2000 A
6161151 Sudhakaran et al. Dec 2000 A
6205502 Endo et al. Mar 2001 B1
6219706 Fan et al. Apr 2001 B1
6253266 Ohanian Jun 2001 B1
6285679 Dally et al. Sep 2001 B1
6308176 Bagshaw Oct 2001 B1
6370145 Dally et al. Apr 2002 B1
6594229 Gregorat Jul 2003 B1
6594721 Sakarda et al. Jul 2003 B1
6597689 Chiu et al. Jul 2003 B1
6618805 Kampe Sep 2003 B1
6633544 Rexford et al. Oct 2003 B1
6654381 Dally et al. Nov 2003 B2
6671694 Baskins et al. Dec 2003 B2
6751191 Kanekar et al. Jun 2004 B1
6799319 Van Loo et al. Sep 2004 B2
6816936 Wu et al. Nov 2004 B1
6854069 Kampe et al. Feb 2005 B2
6865737 Lucas et al. Mar 2005 B1
6910148 Ho et al. Jun 2005 B1
6941487 Balakrishnan et al. Sep 2005 B1
6948088 Sharan Sep 2005 B1
6965903 Agarwal et al. Nov 2005 B1
6966058 Earl et al. Nov 2005 B2
6980543 Kastenholz et al. Dec 2005 B1
6981173 Ferguson et al. Dec 2005 B2
6983317 Bishop et al. Jan 2006 B1
7000153 Inagawa et al. Feb 2006 B2
7010617 Kampe et al. Mar 2006 B2
7046683 Zhao May 2006 B1
7054924 Harvey et al. May 2006 B1
7054926 Abidi et al. May 2006 B1
7061858 Di Benedetto et al. Jun 2006 B1
7069354 Pooni et al. Jun 2006 B2
7076555 Orman et al. Jul 2006 B1
7076696 Stringer Jul 2006 B1
7096383 Talaugon et al. Aug 2006 B2
7111035 McClellan et al. Sep 2006 B2
7117311 Rankin et al. Oct 2006 B1
7146448 Davies et al. Dec 2006 B2
7151741 Elliot et al. Dec 2006 B1
7154902 Sikdar Dec 2006 B1
7155534 Meseck et al. Dec 2006 B1
7159017 Oi et al. Jan 2007 B2
7162737 Syvanne et al. Jan 2007 B2
7164676 Chakraborty Jan 2007 B1
7167479 Christensen et al. Jan 2007 B2
7178052 Hebbar et al. Feb 2007 B2
7233975 Gerraty et al. Jun 2007 B1
7263091 Woo et al. Aug 2007 B1
7274702 Toutant et al. Sep 2007 B2
7275081 Katz et al. Sep 2007 B1
7277018 Reyes et al. Oct 2007 B2
7406038 Oelke et al. Jul 2008 B1
7411925 Nain et al. Aug 2008 B2
7430735 Balakrishnan et al. Sep 2008 B1
7461374 Balint et al. Dec 2008 B1
7518986 Chadalavada et al. Apr 2009 B1
7552262 Turner et al. Jun 2009 B1
7606241 Raghunathan et al. Oct 2009 B1
7747999 Eng et al. Jun 2010 B1
7804769 Tuplur et al. Sep 2010 B1
20010017863 McMillian et al. Aug 2001 A1
20020060986 Fukushima et al. May 2002 A1
20020103921 Nair et al. Aug 2002 A1
20020133400 Terry et al. Sep 2002 A1
20020147974 Wookey Oct 2002 A1
20020191619 Shafer Dec 2002 A1
20020198974 Shafer Dec 2002 A1
20030007493 Oi et al. Jan 2003 A1
20030084371 Mongazon-Cazavet et al. May 2003 A1
20030099247 Toutant et al. May 2003 A1
20030147376 Coutinho et al. Aug 2003 A1
20030179703 Levy et al. Sep 2003 A1
20030185226 Tang et al. Oct 2003 A1
20030200330 Oelke et al. Oct 2003 A1
20030217096 McKelvie et al. Nov 2003 A1
20030235195 Shenoy et al. Dec 2003 A1
20040030766 Witkowski Feb 2004 A1
20040031030 Kidder et al. Feb 2004 A1
20040034871 Lu et al. Feb 2004 A1
20040042395 Lu et al. Mar 2004 A1
20040062196 Gunner et al. Apr 2004 A1
20040073646 Cho et al. Apr 2004 A1
20040078619 Vasavada Apr 2004 A1
20040078621 Talaugon et al. Apr 2004 A1
20040111578 Goodman et al. Jun 2004 A1
20040199808 Freimuth et al. Oct 2004 A1
20040230759 Braun et al. Nov 2004 A1
20050013308 Wybenga et al. Jan 2005 A1
20050041658 Mayhew et al. Feb 2005 A1
20050060413 Oyadomari et al. Mar 2005 A1
20050063308 Wise et al. Mar 2005 A1
20050175017 Christensen et al. Aug 2005 A1
20050216910 Marchand Sep 2005 A1
20050226144 Okita Oct 2005 A1
20050257213 Chu et al. Nov 2005 A1
20060007944 Movassaghi et al. Jan 2006 A1
20060039384 Dontu et al. Feb 2006 A1
20060056285 Krajewski, III et al. Mar 2006 A1
20060089985 Poletto Apr 2006 A1
20060101375 Loughlin et al. May 2006 A1
20060120342 Christensen et al. Jun 2006 A1
20060223466 Behzad Oct 2006 A1
20070006056 Lehner et al. Jan 2007 A1
20070019642 Lu et al. Jan 2007 A1
20080044181 Sindhu Feb 2008 A1
20080275975 Pandey et al. Nov 2008 A1
Foreign Referenced Citations (4)
Number Date Country
1663196 Aug 2005 CN
1202598 May 2002 EP
1231538 Aug 2002 EP
1396968 Mar 2004 EP
Non-Patent Literature Citations (24)
Entry
“Juniper's TX Waits Its Turn,” Light Reading, Sep. 27, 2004, 2 pgs.
“Boston See Party,” Light Reading, Jul. 8, 2004, 2 pgs.
“Cisco CRS-1 24-Slot Fabric-Card Chassis,” Data Sheet, Cisco Systems, retrieved May 9, 2007, 6 pgs.
Xiaozhe Zhang, “ARS: An Sychronization Algorithm Maintaining Single Image Among Nodes' Forwarding Tables of Clustered Router,” Sep. 6, 2005, Springer Berlin/Heidelberg, vol. 3619/2005, pp. 752-761.
“T640 Routing Node and TX Matrix™ Platform: Architecture,” White Paper, Juniper Networks, Inc., (Printed Mar. 3, 2005), 21 pgs.
“TX Matrix™ Platform: Hardware Components and Cable System,” White Paper, Juniper Networks, Inc., (Printed Mar. 15, 2005), 12 pgs.
Gary Matthews and Richard Hendricks, “Chapter Four—Routing Matrix,” JUNOS 7.0 Feature Guide, Juniper Networks, Inc., Oct. 6, 2004, pp. 197-241.
“TX Matrix Platform Migration Guide,” Juniper Networks, Inc., Revision 3, Dec. 9, 2004, 65 pgs.
“Cisco Carrier Routing System,” Cisco Systems, May 19, 2004, 12 pgs.
Avici Systems Press Release, “Avici Demonstrates Carrier Class System Scalability and Reliability of its Routing Platform,” 1 page, Jun. 22, 2004.
Avici Systems Press Release, “Avici and Ciena Demonstrate Dynamic Bandwidth Provisioning Across Intelligent Optical Network,” 1 page, Jun. 5, 2001.
Network World, “New ways of routing the Internet,” 2 pages, Jan. 19, 1998.
The Yankee Group Report, “Core Competence: New Requirements for Core Routing,” 8 pages, Apr. 2002.
Yue (Edward) Yao, “Scalability of a class of Switch Networks,” Department of Engineering Mathematics, Dalhousie University, 8 pages, May 16-18, 2005.
Translation of Office Action from Chinese application No. 200610160913.5, dated Oct. 12, 2010, 7 pp.
Wagner, “Multichassis-Core-Router fuer Next-Generation-IP-Netze”, Nachrichtentechnische Zeitschrift, Verlag GmbH, DE, vol. 55, No. 7/08, Jan. 1, 2002, pp. 26.
Search Report from European application No. EP 06256057, dated Jun. 9, 2010, 9 pp.
U.S. Appl. No. 12/581,664, by Sriram Raghunathan, filed Oct. 19, 2009.
U.S. Appl. No. 12/028,455, by Steven W. Turner, filed Oct. 19, 2009.
U.S. Appl. No. 12/411,113, by Bharani Chadalavada, filed Oct. 19, 2009.
U.S. Appl. No. 11/235,468, by Sriram Raghunathan, filed Oct. 19, 2009.
U.S. Appl. No. 12/825,675, by Eng Wing, filed Oct. 19, 2009.
Communication pursuant to Article 94(3) EPC dated Mar. 14, 2012, received in corresponding EP Application No. 06256057.8, 4 pgs.
Response filed Jun. 21, 2012 to the Communication pursuant to Article 94(3) EPC dated Mar. 14, 2012, in corresponding EP Application No. 06256057.8, 30 pgs.
Related Publications (1)
Number Date Country
20110013508 A1 Jan 2011 US
Continuations (1)
Number Date Country
Parent 11292204 Dec 2005 US
Child 12889209 US