Fault tolerant routing in a non-hot-standby configuration of a network routing system

Information

  • Patent Grant
  • 8412982
  • Patent Number
    8,412,982
  • Date Filed
    Friday, April 8, 2011
    13 years ago
  • Date Issued
    Tuesday, April 2, 2013
    11 years ago
Abstract
Methods and systems for facilitating fault tolerance in a non-hot-standby configuration of a network routing system are provided. According to one embodiment, a failover method is provided. One or more processing engines of a network routing system are configured to function as active processing engines, each of which having one or more software contexts. A control blade is configured to monitor the active processing engines. One or more of the processing engines are identified to function as non-hot-standby processing engines, each of which having no pre-created software contexts corresponding to the software contexts of the active processing engines. The control blade monitors the active processing engines. Responsive to detecting a fault associated with an active processing engine the active processing engine is dynamically replaced with a non-hot-standby processing engine by creating one or more replacement software contexts within the non-hot-standby processing engine corresponding to those of the active processing engine.
Description
COPYRIGHT NOTICE

Contained herein is material that is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction of the patent disclosure by any person as it appears in the Patent and Trademark Office patent files or records, but otherwise reserves all rights to the copyright whatsoever. Copyright©2002-2010, Fortinet, Inc.


BACKGROUND

1. Field


Embodiments of the present invention generally relate to data communications. In particular, embodiments of the present invention relate to network routing and routing systems, and more particularly to fault-tolerant routing.


2. Description of the Related Art


It is desirable to provide reliable packet delivery between nodes in the network connected by a functional physical path. Interconnected networks vary in the number of redundant paths they provide between nodes. Many conventional routing systems use an active replication technique to provide for failures. With active replication, recovery from failures may be quick, but there is a large overhead in ordinary execution. Active replication uses a redundant structure consisting of two processor resources (e.g., two processors and memory). One problem with active replication is that because all the replicas must be pre-created when the system is running, the processor resources are used wastefully. Another problem with active replication is that because it complicates object management, flexible management and flexible construction are difficult.


Thus, there is a need for systems and methods that provide reliable packet delivery between all pairs of nodes. There is also a need for systems and methods that detect when a fault has occurred and alert the operating system. There is also a need for systems and methods that provide a mechanism to reconfigure a network around faulty areas to ensure quick and reliable packet delivery to all non-faulty areas of the network. There is also a need for systems and methods that are able to reconfigure a network within a short period of time after a failure. There is also a need for systems and methods that provide quick recovery from failure, do not require extra or dedicated hardware as in a hot-standby configuration, and provide for fault tolerant routing without the need to reboot.


SUMMARY

Methods and systems are described for facilitating fault tolerance in a non-hot-standby configuration of a network routing system. According to one embodiment, a failover method is provided. One or more processing engines associated with multiple server blades of a network routing system are configured to function as active processing engines, each of which having one or more software contexts. A control blade of the server blades is configured to monitor the active processing engines. One or more of the processing engines are identified to function as one or more non-hot-standby processing engines, each of which having no pre-created software contexts corresponding to the software contexts of the active processing engines. The control blade monitors the active processing engines. Responsive to detecting a fault associated with an active processing engine the active processing engine is dynamically replaced with a non-hot-standby processing engine by creating one or more replacement software contexts within the non-hot-standby processing engine corresponding to those of the active processing engine.


Other features of embodiments of the present invention will be apparent from the accompanying drawings and from the detailed description that follows.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of the present invention are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:



FIG. 1 is a simplified functional block diagram of a network routing system in accordance with an embodiment of the present invention;



FIG. 2 is a simplified functional block diagram of control blade in accordance with an embodiment of the present invention;



FIG. 3 is a flow chart of a virtual router failover procedure in accordance with an embodiment of the present invention;



FIG. 4 illustrates an example distribution of protocol modules within objects in accordance with an embodiment of the present invention;



FIG. 5 illustrates a distinction between an Object Class and an Object Group in accordance with an embodiment of the present invention;



FIG. 6 illustrates VPN and VR replication using remote object references in accordance with an embodiment of the present invention;



FIG. 7 illustrates a user interface layer showing the saving of VPN and VR information in a configuration file in accordance with an embodiment of the present invention;



FIG. 8 illustrates the layout of the ASCII-text configuration in accordance with an embodiment of the present invention;



FIG. 9 is a diagram illustrating the generation of a list of VPN & VR IDs present in a primary processor engine in accordance with an embodiment of the present invention;



FIG. 10 illustrates an example of a command line interface (CLI) engine using a VPN/VR ID and slot/PE ID of a failed processing engine to find a corresponding command line set in accordance with an embodiment of the present invention;



FIG. 11 illustrates a CLI engine replaying a command line set for VPN and VR reconstruction in accordance with an embodiment of the present invention;



FIG. 12 illustrates VPN and VR object reconstruction during PE failover in accordance with an embodiment of the present invention; and



FIG. 13 illustrates a fault management system in accordance with an embodiment of the present invention.





DETAILED DESCRIPTION

Methods and systems are described for passive replication to facilitate fault tolerance in a network routing system. In one embodiment, a virtual routing system and method provides reliable packet delivery between all pairs of nodes. In another embodiment, a virtual routing system and method detects when a fault has occurred, and alert the operating system. In another embodiment, a virtual routing system and method reconfigures the network around faulty areas of network to ensure quick and reliable packet delivery to all non-faulty areas of the network. In yet another embodiment, a virtual routing system and method are able to reconfigure the network within a short period of time after a failure. In several embodiments, a virtual routing system and method provide for quick recovery from failure, do not require additional dedicated hardware as in non-hot standby, and provide for fault tolerant routing without a need to reboot.


In the following description, numerous specific details are set forth in order to provide a thorough understanding of embodiments of the present invention. It will be apparent, however, to one skilled in the art that embodiments of the present invention may be practiced without some of these specific details. In other instances, well-known structures and devices are shown in block diagram form.


Embodiments of the present invention include various steps, which will be described below. The steps may be performed by hardware components or may be embodied in machine-executable instructions, which may be used to cause a general-purpose or special-purpose processor programmed with the instructions to perform the steps. Alternatively, the steps may be performed by a combination of hardware, software, firmware and/or by human operators.


Embodiments of the present invention may be provided as a computer program product, which may include a machine-readable medium having stored thereon instructions, which may be used to program a computer (or other electronic devices) to perform a process. The machine-readable medium may include, but is not limited to, floppy diskettes, optical disks, compact disc read-only memories (CD-ROMs), and magneto-optical disks, ROMs, random access memories (RAMs), erasable programmable read-only memories (EPROMs), electrically erasable programmable read-only memories (EEPROMs), magnetic or optical cards, flash memory, or other type of media/machine-readable medium suitable for storing electronic instructions. Moreover, embodiments of the present invention may also be downloaded as a computer program product, wherein the program may be transferred from a remote computer to a requesting computer by way of data signals embodied in a carrier wave or other propagation medium via a communication link (e.g., a modem or network connection).


TERMINOLOGY

Brief definitions of terms used throughout this application are given below.


The terms “connected” or “coupled” and related terms are used in an operational sense and are not necessarily limited to a direct connection or coupling.


The phrases “in one embodiment,” “according to one embodiment,” and the like generally mean the particular feature, structure, or characteristic following the phrase is included in at least one embodiment of the present invention, and may be included in more than one embodiment of the present invention. Importantly, such phases do not necessarily refer to the same embodiment.


If the specification states a component or feature “may”, “can”, “could”, or “might” be included or have a characteristic, that particular component or feature is not required to be included or have the characteristic.


The term “responsive” includes completely or partially responsive.



FIG. 1 is a simplified functional block diagram of a network routing system in accordance with an embodiment of the present invention. Network routing system 100, among other things, may provide hardware-based network processor capabilities and high-end computing techniques, such as parallel processing and pipelining. In embodiment of the present invention, network routing system 100 may implement one or more virtual private networks (VPNs) and one or more associated virtual routers (VRs), and in some embodiments, system 100 may implement hundreds and even thousands of VPNs and VRs. Network routing system 100 may include one or more line interfaces 102, one or more virtual routing engines (VREs) 104, one or more virtual service engines (VSEs) 106, and one or more advanced security engines (ASEs) 108 coupled by switching fabric 110. Network routing system 100 may also include interface 112 which may interface with other routing systems. Network routing system 100 may also include one or more control blades 114 to create VPNs and/or VRs to operate on VREs 104.


In one embodiment, several VPNs and/or VRs may, for example, run on one of processing engines (PEs) 116 of VRE 104. A VPN or VR may be a software context comprised of a set of objects that are resident in the processing engine's memory system. The software context may include the state and processes found in a conventional router, however hundreds or more of these virtual router contexts may be overlaid onto a single processing engine and associated memory system. Accordingly, one of processing engines 116 may provide the context of many VRs to be shared allowing one piece of hardware, such as network routing system 100, to function as up to a hundred or even a thousand or more routers.


Line interface 102 may receive packets of different packet flows from an external network over a communication channel. VREs 104 may perform packet classification, deep packet inspection, and service customization. In one embodiment, VRE 104 may support up to one million or more access control list (ACL) level packet flows. VREs 104 may include a virtual routing processor (not illustrated) to provide hardware assisted IP packet forwarding, multi-protocol label switching (MPLS), network address translation (NAT), differentiated services (DiffServ), statistics gathering, metering and marking VREs 104 and VSEs 106 may include a virtual service controller (not illustrated) to support parallel processing and pipelining for deep packet inspection and third-party application computing. VSEs 106 may perform parallel processing and/or pipelining, and other high-end computing techniques, which may be used for third party applications such as firewall services and anti-virus services. ASEs 108 may provide for hardware and hardware assisted acceleration of security processing, including encryption/decryption acceleration for IP security protocol type (IPSec) packet flows and virtual private networks (VPNs). Switching fabric 110 may be a high-capability non-blocking switching fabric supporting rates of up to 51.2 Gbps and greater.


Line interface 102 may include a flow manager (not illustrated) to load-balance service requests to VSEs 106 and VREs 104, and may support robust priority and/or weighted round robin queuing. In one embodiment, the flow manager may provide for service load balancing and may dynamically determine one of VREs 104, which may best handle a certain packet flow. Accordingly, all packets of a particular flow may be sent to the same VRE 104. Line interface 102 may identify one of the VREs to process packets of a packet flow based on a physical interface and virtual channel from which the packets of the packet flow were received. The identified VRE may perform ingress metering, header transformation and egress metering for packets of the packet flow. In one embodiment, hardware based metering and marking using a dual token bucket scheme assists in rate-control capabilities of system 100. This may allow for granular application level support and the ability to provide strong performance based service level agreements (SLAs).


Different packets may take different paths through network routing system 100 and may not necessarily require the resources of all the various functional elements of network routing system 100. In one embodiment, a packet, such as a virtual local area network (VLAN) Ethernet packet, may arrive at an input port of line interface 102. The input port may be a gigabit Ethernet input port, which may be one of several input ports. The flow manager may program a steering table look-up to determine which VLAN is associated with a particular one of VREs 104. The flow manager may tag the packet with an internal control header and may transfer the packet from line interface 102 across switching fabric 110 to the selected VRE 104. A service controller of VRE 104 may perform deep packet classification and extract various fields on the packet header. A flow cache may be looked up to determine whether the packet should be processed in hardware or software. If the packet is to be processed in hardware, an index to the packet processing action cache may be obtained.


The packet may be deposited via a high-speed direct access memory (DMA) into the VRE's main memory. A routing processor may retrieve the packet, identify the packet processing actions and may perform actions, such as time-to-live decrementation, IP header and checksum updating, and IP forwarding patch matching. Egress statistics counters may also be updated. The packet may be forwarded to one of ASEs 108 for security operations. The packet may also be forwarded to another one of VREs 104.


In accordance with embodiments of the present invention, control blade 114 provides for redundancy and failover of the virtual routers instantiated by objects running on processing engines 116 of VREs 104. In one embodiment, control blade 114 may detect a failure of one processing engines 116, and may identify the VPNs and/or VRs operating on a failed processing engine. Control blade 114 may also identify a set of command lines corresponding with the identified VPNs and VRs, and replay the set of command lines with an identity of a new processing engine to recreate the identified VPNs and VRs on the new processing engine. This is described in more detail below.


Although system 100 is illustrated as having several separate functional elements, one or more of the functional elements may be combined and may be implemented by combinations of software configured elements, such as processors including digital signal processors (DSPs), and/or other hardware elements.



FIG. 2 is a simplified functional block diagram of control blade in accordance with an embodiment of the present invention. Control blade 200 may be suitable for use as control blade 114 (FIG. 2) although other control blades and systems may also be suitable. Control blade 200 may be comprised of one or more processing engines 202, which may be configured to operate fault manager 204, object manager 206, configuration manager 208 and command line interface (CLI) engine 210.


Configuration manager 208 may request the creation of one or more VPNs and associated VRs, which may be defined by objects, and object groups as further described below. Object manager 206 manages the objects that define particular VPNs and VRs. In one embodiment, fault manager 204 may monitor keep-alive messages from one or more processing engines of VREs 104 (FIG. 1) to detect failures of one or more processing engines. Fault manager 204 may query object manager 206 for a list of VPNs and VRs operating on the failed processing engine. Fault manager 104 may identify the failed processing engine to object manager 206 by a slot ID and processing engine ID. The slot ID may identify a particular one of a plurality virtual routing engines (VREs) located at a particular chassis slot of system 100 (FIG. 1). The processing engine ID may identify a particular one of the processing engines of the identified VRE. Object manager 206 may query object manager database 212 to generate a list of VPNs and VRs affected by the failed processing engine. After fault manager 204 receives the list of VPNs and VRs affected by the failed processing engine, fault manager 204 may store the list in memory 214, and may identify a new or backup processing engine to replace the failed processing engine. Fault manager 204 may provide the list to CLI engine 210, along with information to identify the new and failed processing engine. This identity information may include, for example, the slot ID and processing engine ID of the new and failed processing engines. CLI engine 210 may find a set of command lines in configuration file 216 that correspond with the affected VPNs and VRs and the slot ID and processing engine ID of the failed processing engine. CLI engine 110 may replay the matching command lines with the slot ID and processing engine ID of the new processing engine substituted for the failed processing engine, which may activate the identified VPNs and VRs on the new processing engine. In this way, objects that instantiate the VPNs and VRs on the failed processing engine are reconstructed on a new processing engine.


Although control blade 200 is illustrated as having several separate functional elements, one or more of the functional elements may be combined and may be implemented by combinations of software configured elements, such as processors including digital signal processors (DSPs), and/or other hardware elements.



FIG. 3 is a flow chart of a virtual router failover procedure in accordance with an embodiment of the present invention. Procedure 300 may be performed by a control blade, such as control blade 200 (FIG. 2), although other systems and control blades may also be suitable for performing procedure 300. Procedure 300 may be used to create VPNs and VRs on a network routing system, such as network routing system 100 (FIG. 1) and provides for automated recovery of VPNs and VRs that are operating on a failed processing engine. Although the individual operations of procedure 300 are illustrated and described as separate operations, one or more of the individual operations may be performed concurrently and nothing requires that the operations be performed in the order illustrated.


In operation 302, VPNs and associated VRs may be created. Each VPN and VR may be defined by one or more objects and object groups, which may be identified for a particular VPN and VR in an object manager database. A configuration manager, such as configuration manager 208 (FIG. 2) may request the creation of one or more VPNs and VRs by object manager 206 (FIG. 2). A CLI engine, such as CLI engine 210 (FIG. 2) may generate a configuration file containing portions for each VPN and VR to be instantiated on a processing engine, such as one of processing engines 116 (FIG. 1). Operation 302 may also include executing the configuration file to instantiate the VPNs and VRs on the processing engines.


In operation 304, a fault manager, such as fault manager 204 (FIG. 2) may monitor messages from the processing engines operating the VPNs and VRs to detect a failure. In one embodiment, the fault manager monitors keep-alive messages from the processing engines. The keep-alive messages may be viewed as heart beats generated by each processing engine on a regular basis and may be communicated between the various processing engines and the control blade in a peer-to-peer networking arrangement through a switching fabric, such as switching fabric 110 (FIG. 1).


In operation 306, when a processing engine failure is detected, operation 308 is performed. When operation 306 does not detect a processing engine failure, operation 304 continues to monitor the keep-alive messages until a processing engine failure is detected. Although not illustrated as separate operations of procedure 300, the network routing system may concurrently operate VPNs and VRs during the performance of the various operations of procedure 300.


In operation 308, the fault manager may query an object manager, such as object manager 206 (FIG. 2) for a list of VPNs and VRs operating on the failed processing engine. The object manager may generate the list using object manager database 309, which may correspond with object manager database 212 (FIG. 2). In operation 310, a new processing engine may be identified to replace the failed processing engine. In one embodiment, a replacement processing engine may be pre-determined during the creation of VPNs and VRs in operation 302, and may be identified in the object manager database as a backup for the failed processing engine.


In operation 312, the fault manager may provide a list of VPNs and VRs identified by the object manager to a command line interface engine, such as CLI engine 210 (FIG. 2). As part of operation 312, the fault manager may also provide information identifying the failed processing engine and new processing engine to the CLI engine. The identity information of a processing engine may include a slot ID and PE ID which identifies the blade location, such as a particular VRE 104 (FIG. 1) and particular processing engine on the blade.


In operation 314, the command line interface engine may identify a command line set from a configuration file corresponding with the particular VPNs and VRs for the failed processing engine. In operation 316, the command line interface engine may substitute the slot ID and PE ID of the new processing engine for that of the failed processing engine in the command line set identified in operation 314. In operation 318, the command line interface engine may replay (e.g., execute) the identified command line set with the new slot ID and PE ID to recreate the VPNs and VRs on the new processing engine.


In operation 320, the object manager database may be updated to correlate the particular VPNs and VRs as operating on the new processing engine. After the completion of operation 320, the routing system may continue to operate the VPNs and VRs accordingly and monitor keep-alive messages in accordance with operation 304.


Embodiments of the present invention provide for fault-tolerant routing. Fault tolerant routing may be viewed as providing reliable packet delivery between nodes in the network connected by a functional physical path. Interconnected networks vary in the number of redundant paths they provide between nodes. Three basic classes of faults that may be considered include link failure, processor node failure, and virtual router failure. Link failures refer to faults in the physical network link between routers. This class of fault includes physical failures caused by the unplugging of a routing cable or the unseating of a router board from the chassis. Link failure resulting in either the corruption of data or the loss of communication between routers may be easily detected. A link that is found to be faulty may be removed from service, and a redundant link if any may be activated and put into service. Processor node failures refer to any failure, which may cause a processing node that runs the router software to refuse to accept packets. The node may appear dead to the network when this fault occurs. These faults may be easily detected, although it may not be possible to isolate the fault further than identifying which processing node is faulty. Such failures may reside either in the node itself or the communication coprocessor. A processor node that is found to be faulty may be removed from service, and a redundant node if any, may be activated and put into service.


A virtual router may fail in different ways, each possibly resulting in varying loss of service. A failure in which packets are routed successfully with proper data, but uses an incorrect channel or experiences unnecessary delay may be difficult to detect because such behavior is within the realm of normal operation. Fortunately, such failures may not be catastrophic because they do not result in corruption or loss of data. Some failures may result in certain channels of the router becoming inoperable, and may appear to be the same as link failures and be treated as such. Furthermore, a router may fail to operate at all. This may be the most severe form of fault in the network because this fault may render all network links to the router inoperable and may isolate the attached processing node from the rest of the network.



FIG. 4 illustrates a possible distribution of protocol modules within objects in accordance with an embodiment of the present invention. Embodiment of the present invention may implement objects within the various elements of system 100 (FIG. 1). Objects may encompass protocol modules and networking services. Objects may enable the implementation of a VPN and VR. Each VR may comprise a set of objects that provide routing and network services. This may allow the operation of multiple protocol stacks within a single address space with isolation between VRs. Objects may be viewed as containers for modules. Distribution 400 includes objects 402 which represent a basic unit of management for purposes of fault tolerance, computational load balancing etc. One or more adjacent protocol modules 404 may reside in a single object. It is also possible that a module is split across two objects. Collections of objects (e.g. Object-1, Object-2 and Object-3) may be related by the module relationships that they encompass. Such collections of objects may be referred to as object groups.



FIG. 5 illustrates a distinction between an object class and an object group in accordance with an embodiment of the present invention. Both are a collection of objects. An object class may be a set of objects that have the same type signature and behavior. Distribution 500 is illustrated with applications object class 502, TCP/IP object class 504 and interface object class 506. However, for an object group, the constituent objects do not necessarily have the same type signature and behavior (e.g. object groups 508, 510 and 512). There may be multiple objects of the same class in an object group. For example, object group 510 has two objects of interface object class 506. On the other hand, an object group need not have an object of each class. For example, object group 512 does not have an object of interface object class 506. All objects may be managed by object manager 204 (FIG. 2), which may reside control blade 114 (FIG. 1), which is a management blade of system 100. Control blade 114 (FIG. 1) may also be responsible for the creation and deletion of objects driven by the text-based configuration.


An object of system 100 may not relate directly to a VR component. Multiple objects may correspond to a single component, just as a single object may correspond to multiple components. A VR, for example, may be composed of more than ten objects. Creating a VR may result in the creation of all objects and its components. VR component creation is dynamic, thus a VR may include only of the minimum set of classes of objects that may be required to realize all the components that are provisioned for the VR.


Network communication services may require processing platforms that afford high performance while reducing the cost of equipment and application development. In particular, networking systems should be able to continue operating in the presence of equipment failure. Embodiments of the present invention provide distributed object-oriented network-processing platform including software, which concurrently runs several objects. Each concurrently running object may be viewed as an object instance that encapsulates data and their functions with an independent execution unit. These concurrent objects provide services by communicating with each other without knowing each other's physical locations or internal structures. Each node may be multiple objects, network operating system, and a real-time kernel. The distributed processing platform provides a fault-tolerant mechanism in which fault-tolerant objects may be handled flexibly, with its internal structures hidden and indicated by logical IDs and the execution overhead for fault tolerance may be minimized.


In accordance with an embodiment, the execution units (e.g., the objects) may be replicated and managed in the platform by means of a passive replication approach. Object manager 206 (FIG. 2), for example, may use pseudo objects to manage these passive replicas. The pseudo object may not be the actual object, but instead may be an object image created by the actions of the platform.


Fault tolerance may utilize replication. Replication may allow local access to a resource and may provide a network routing system with an aggregate computing power, that is, the overall computation from all processors taken as a whole. Furthermore, replication may also offer better availability since even though one or more replicas are unavailable, services may be available from the up replicas. Since failures of processors can cause the partial or total loss of system functionality, a level of availability needs to be supported. The fault tolerant system of embodiments of the present invention allows the failure of processors and enhances availability by masking the failure of the master processor to a new processor. Failed processors may be excluded from the set of processors by a failure detection protocol.


With fault tolerance based on replication, there may be a trade-off problem between the overhead of ordinary execution and the recovery time. Two methods for managing the replicas include active replication and passive replication. With active replication, recovery from failures may be quick, but requires large overhead in ordinary execution. With passive replication, recovery may be slower, but resources may be used more efficiently. Active replication conventionally uses a redundant structure consisting of two processor resources and associated memory. One problem with active replication is that because all the replicas must be pre-created when the system is running, the processor resources are used wastefully. Another problem with active replication is that because it complicates object management, flexible management and flexible construction are difficult. With recent gains in processor performance, the fault recovery time with passive replication may be significantly be shortened, and may even exceed that of active replication. In addition, passive replication may solve the problems of active replication while effectively managing processor resources. Accordingly, passive replication supports the development of a large-scale system.


In accordance with embodiments of the present invention, in the case of a blade-level failover (i.e., one of VREs 104FIG. 1) in which all PEs 116 of a particular blade fails, the blade may reboot and all affected VRs may be recreated on backup PEs. In the case of a PE-level failover, (i.e., if one or more of PEs 116 fails), affected VPNs and VRs may be recreated on a backup PEs. This is described in more detail below.


In accordance with embodiments of the present invention, two substantially identical processor engines may be used to create a primary and a secondary PE system that may run multiple instances of virtual routers and services. Both primary and secondary PEs may consume the same input information, and both connect to the same outputs, though only the primary may be active and controlling those outputs. Both processor engines may be are linked to the same ring network. This allows them to maintain synchronous communication with each other including the management processor engine. When the primary is no longer capable of control due to an internal fault or communication loss, the standby takes over and runs with configuration identical to that on the primary. There are cases where service provider dedicates a single access or trunk blade to their golden customer. To provide resilient network services to this type of customer, a backup blade may be provided that may help ensure continuous operations, even if some devices completely fail. In the blade backup system, two substantially identical blades may be used to create a primary and a secondary blade system that may run multiple instances of virtual routers and services. A blade may comprise multiple processor engines and network modules. If the primary blade becomes non-operational due to internal fault or communication loss in one of the processor engines, the standby blade takes over and runs with configuration identical to that on the primary blade. Embodiments of the present invention may allow a user to specify any available processor when configuring a primary processor engine with a secondary processor engine.


In accordance with embodiments of the present invention, to perform PE failover, a primary-backup pair of processors may be determined prior to creating VPN and VR. The primary-backup pair may be created either through command line interface or GUI-based management interface. Any type of processor engine may be suitable as a standby as long as it is compatible with the primary processor engine. The system may disallow the use of incompatible processor engine as a standby. Once both primary and standby PEs are configured, the fault manager may set the primary PE redundancy state to a primary-active state and the standby PE may be set to a standby state. PEs that are active (e.g., created with VPNs/VRs) but configured without a hot-standby PE may have a redundancy state set to active. PEs that are not active may have a state set to idle.



FIG. 6 illustrates VPN and VR replication using remote object references in accordance with an embodiment of the present invention. VPN/VR object may be replicated as follows: Configuration manager 608 may be an agent that resides in control blade 600 drives the creation of a VPN 604 and/or VR 602, and services within VR 602. A VR, such as VR 602, may be an object group 614 and may an aggregation point for all objects that comprises the VR. A model of replicated object management provides a way of maintaining information about replicated objects in a consistent manner. Remote object references based on IDs (e.g., vpn-id, vr-id, obj-grp-id and object id) may be used. These IDs may allow for distributed object identification. The remote object references are stored in OM database 612 and may be used by configuration manager 608 to manage VPN and VR information query and dynamic VR creation. These remote object references may be used to recreate VPNs and VRs, and their components during PE failover. Object manager 606 may be a module that manages the VPN and VR objects and object groups. Object Manager 606 may create a VPN descriptor every time the configuration manager request VPN creation. Every VPN may be identified by a unique VPN ID. A VR may be identified by a VR ID, which may be the IP Address of the VR. The VR ID may be unique in the VPN context. In one embodiment, objects may be identified using an obj-group-id and an object id.


In accordance with embodiments of the present invention, control blade 600 may correspond with control blade 200 (FIG. 2), configuration manager 608 may correspond with configuration manager 208 (FIG. 2), object manager 606 may correspond with object manager 206 (FIG. 2) and object manager database 612 may correspond with object manager database 212 (FIG. 2). In addition, processing engine 616 may correspond with one of processing engines 116 (FIG. 1).



FIG. 7 illustrates a user interface layer illustrating the saving of VPN and VR information in configuration file in accordance with an embodiment of the present invention. User interface layer 700 may include command line interface (CLI) 702 which may sit on top of SNMP access functions 704, which may make it in the same level with SNMP engine 706. Accordingly, any configuration that is possible with SNMP may also be done with CLI and vice versa. CLI 702 may use a transaction based commit model. This feature may enables a user to configure and commit configuration on a per object basis instead of per command line used by most CLI on other networking equipments. This may eliminate the possibility of leaving an object in an unstable state due to incomplete configuration. CLI 702 may also communicate with object manager 708 since it deals with objects 710 during VPN and VR creation. Object manager 708, which corresponds with object manager 206 (FIG. 2). In accordance with embodiments of the present invention, CLI 702 may correspond with CLI engine 210 (FIG. 2), and object manager 708 may correspond with object manager 206 (FIG. 2).



FIG. 8 illustrates the layout of the ASCII-text configuration in accordance with an embodiment of the present invention. Object manager 206 (FIG. 2), may store the remote VPN and VR object references in OM database 212 (FIG. 2). When a user invokes a save of the configuration, CLI queries the object manager for VPN and VR object references, may transform the result into ASCII-text information and may save the text output in the configuration file. The use of VPN and VR IDs and the ASCII-text information in the PE failover process is discussed in more detail below.


In one embodiment of the present invention, processor engine failure detection may be implemented using a heartbeat mechanism. A management processor engine, such as fault manager 204 (FIG. 2) that resides in a control blade monitors the health of all processors. The primary PE may periodically send a keep-alive packet to the management PE over a ring network that is marked for internal communication. The management PE raises a failure suspicion event after it detects a number of consecutive missing heartbeats from the primary PE. Once a failure suspicion is raised, a standard voting protocol can be used to confirm that the primary process is out of service, because the primary PE is down, the application process has crashed, or the communication link from the primary is broken. To implement a fast failover system, the failure detection mechanism should detect failures quickly and precisely.


In one embodiment, system 100 may use a distributed messaging layer (DML) for sending keep-alive messages and monitoring processor membership. A tradeoff exists where a tighter deadline for keep-alive message responses can result in false failures. How tight the timeout tolerances can be tuned may be dependent on a number of factors including provisioning and peak load. In one embodiment, a processor engine may send a heartbeat every two seconds and the detection period for a failure suspicion may be 24 seconds (e.g., a total of 12 keep-alive retries). In one embodiment, the keep-alive timeout value may be fixed, and in other embodiments, it may be made customer tunable. Once a primary failure is confirmed, it may generate an event in the management PE for the backup process to take over from the primary.



FIG. 9 is a diagram illustrating the generation of a list of VPN & VR IDs present in the primary processor engine that failed in accordance with an embodiment of the present invention. Once the control blade detects a primary PE failure, fault management 902 may query object manager 904 for list of VPNs and VRs that were created in the failed PE identified by a slot/PE ID. Object manager 904 may query object manager database 910 and return list 906 that includes VPN and VR IDs. Fault manager 902 may save this list in memory 908. FIG. 9 illustrates how the fault manager generates a list of VPN & VR IDs—for example VPN 1 and VR 1.1.1.1 were present in the primary processor engine (slot 3 PE 1) that failed. After the list of affected VPNs and VRs has been generated, fault manager 902 may pass this list to a CLI engine for processing along with the slot/PE ID of the failed PE and slot/PE ID of the new active PE. The list contains VPN & VR IDs. CLI engine uses these IDs and the slot/PE ID of the failed PE to find the set of command lines from the configuration file that correspond to the VPN/VR IDs in the list and the slot/PE ID of the failed PE.



FIG. 10 illustrates an example of a CLI engine using a VPN/VR ID and slot/PE ID of a failed PE to find a corresponding command line set in accordance with an embodiment of the present invention. In FIG. 10, fault manger 1010 may retrieve the list from memory 1020 and may provide information 1012 to CLI engine 1014. CLI engine 1014 uses the VPN/VR ID and slot/PE ID of a failed PE to fine a corresponding command line set 1018 from configuration file 1016.



FIG. 11 illustrates a CLI engine replaying a command line set for VPN and VR reconstruction in accordance with an embodiment of the present invention. CLI engine 1102 may replay each matching command line, but before it fetches the command string 1104 to the command line processing module 1106 it may substitute the destination slot/PE ID with the ID of the new active PE. Therefore, all VPNs, VRs and VR components may be recreated in the new active PE 1112 after CLI engine 1102 finishes replaying the command line set. Object manager 1114 may manage the new object associations.



FIG. 12 illustrates VPN and VR object reconstruction during PE failover in accordance with an embodiment of the present invention. Objects and object groups of VPN 1202 and VR 1204 from failed processing engine 1206 are recreated on new processing engine 1208 by a processing engine of control bland 1210. When a fault manager detects a failure of processing engine 1206, object manager 1212 may identify VRs operating on a failed processing engine using object manager database 1214, the command line interface engine may identify a set of command lines from configuration file 1210 corresponding with the identified VRs, and to replay the set of command lines with an identity of a new processing engine to recreate the identified VRs on new processing engine 1208. Configuration manager may manage the new configuration creating the VPNs and VRs.



FIG. 13 illustrates a fault management system in accordance with an embodiment of the present invention. The fault management system 1300 may be implemented by fault manager 204 (FIG. 2). Fault manager 1302 may include a fault tolerance (FT) configurator 1304, fault detector 1306, blade/PE state machine 1308, and failure recovery agent 1309. Fault manager 1302 may be capable of scaling to large number of processor engines. Fault manager 1302 may identify faults on processing engines 1310 and may restore affected virtual routers and applications accurately. Fault manager 1302 may identify problems in a timely fashion, so that responses and corrective actions may be taken as soon as possible to meet failover requirements. Fault manager 1302 may utilize low overhead and monitoring so as not to have a significant impact on the performance of virtual routers and application processes. Fault manager 1302 may also support a range of application-specific fault detection policies and usage models. For example, applications may wish to control which entities are monitored, how often they are monitored, the criteria used to report failure, and where failures are reported.


Fault tolerance configurator 1304 may be a module that interfaces with CLI or SMS user interface 1312 to configure the fault tolerance properties. Fault detector 1306 may be a module responsible for monitoring the health of the processor engines 1312 (e.g., through keep-alive messages 1307) and reporting faults. Blade/PE state machine 1308 may be a module is responsible for maintaining the state of blades and processor engines. The fault recovery agent 1309 may be a module that is responsible for processing and interfacing with CLI engine 1314.


While embodiments of the invention have been illustrated and described, it will be clear that the invention is not limited to these embodiments only. Numerous modifications, changes, variations, substitutions, and equivalents will be apparent to those skilled in the art, without departing from the spirit and scope of the invention, as described in the claims.

Claims
  • 1. A computer-implemented failover method comprising: configuring one or more of a plurality of processing engines associated with a plurality of server blades of a network routing system to function as one or more active processing engines, each of the one or more active processing engines having one or more software contexts;configuring a control blade of the plurality of server blades, to monitor the one or more active processing engines;identifying one or more of the plurality of processing engines to function as one or more non-hot-standby processing engines, each of the one or more non-hot-standby processing engines having no pre-created software contexts corresponding to the software contexts of the one or more active processing engines;monitoring, by the control blade, the one or more active processing engines; andresponsive to detecting a fault associated with an active processing engine of the one or more active processing engines, dynamically replacing the active processing engine with a non-hot-standby processing engine of the one or more non-hot-standby processing engines by creating one or more replacement software contexts within the non-hot-standby processing engine corresponding to the one or more software contexts of the active processing engine.
  • 2. The method of claim 1, wherein the fault comprises a link failure to or from the active processing engine.
  • 3. The method of claim 1, wherein the fault comprises a hardware or software failure associated with the active processing engine.
  • 4. The method of claim 1, wherein one of the one or more software contexts of the active processing engine includes a set of objects implementing a virtual router (VR).
  • 5. The method of claim 1, wherein said dynamically replacing the active processing engine with the non-hot-standby processing engines involves use of a network-management protocol.
  • 6. The method of claim 5, wherein the network-management protocol comprises Simple Network Management Protocol (SNMP).
  • 7. The method of claim 1, further comprising monitoring, by the control blade, a health of the one or more active processing engines by tracking keep-alive messages received from the one or more active processing engines.
  • 8. The method of claim 1, wherein said dynamically replacing the active processing engine with the non-hot-standby processing engine involves use of a transaction-based commit model.
  • 9. The method of claim 1, wherein the active processing engine and the non-hot-standby processing engine comprise substantially identical processing engines.
  • 10. The method of claim 1, wherein said dynamically replacing the active processing engine with a non-hot-standby processing engine comprises: identifying virtual private networks (VPNs) and virtual routers (VRs), defined by one or more objects and objects groups, that were operating on the active processing engine prior to detection of the fault, andrecreating the VPNs and VRs within the non-hot-standby processing engine.
  • 11. The method of claim 10, wherein the fault comprises failure of one or more of the VPNs or VRs.
  • 12. A non-transitory computer-readable storage medium tangibly embodying a set of instructions, which when executed by one or more processors associated with a control blade of a plurality of server blades of a network routing system or a plurality of processing engines of the plurality of server blades cause the one or more processors to perform a failover method comprising: configuring one or more of the plurality of processing engines to function as one or more active processing engines, each of the one or more active processing engines having one or more software contexts;identifying one or more of the plurality of processing engines to function as one or more non-hot-standby processing engines, each of the one or more non-hot-standby processing engines having no pre-created software contexts corresponding to the software contexts of the one or more active processing engines;monitoring, by the control blade, the one or more active processing engines; andresponsive to detecting a fault associated with an active processing engine of the one or more active processing engines, dynamically replacing the active processing engine with a non-hot-standby processing engine of the one or more non-hot-standby processing engines by creating one or more replacement software contexts within the non-hot-standby processing engine corresponding to the one or more software contexts of the active processing engine.
  • 13. The computer-readable storage medium of claim 12, wherein the fault comprises a link failure to or from the active processing engine.
  • 14. The computer-readable storage medium of claim 12, wherein the fault comprises a hardware or software failure associated with the active processing engine.
  • 15. The computer-readable storage medium of claim 12, wherein one of the one or more software contexts of the active processing engine includes a set of objects implementing a virtual router (VR).
  • 16. The computer-readable storage medium of claim 12, wherein said dynamically replacing the active processing engine with the non-hot-standby processing engines involves use of a network-management protocol.
  • 17. The computer-readable storage medium of claim 16, wherein the network-management protocol comprises Simple Network Management Protocol (SNMP).
  • 18. The computer-readable storage medium of claim 12, wherein the method further comprises monitoring a health of the one or more active processing engines by tracking keep-alive messages received from the one or more active processing engines.
  • 19. The computer-readable storage medium of claim 12, wherein said dynamically replacing the active processing engine with the non-hot-standby processing engine involves use of a transaction-based commit model.
  • 20. The computer-readable storage medium of claim 12, wherein the active processing engine and the non-hot-standby processing engine comprise substantially identical processing engines.
  • 21. The computer-readable storage medium of claim 12, wherein said dynamically replacing the active processing engine with a non-hot-standby processing engine comprises: identifying virtual private networks (VPNs) and virtual routers (VRs), defined by one or more objects and objects groups, that were operating on the active processing engine prior to detection of the fault, andrecreating the VPNs and VRs within the non-hot-standby processing engine.
  • 22. The computer-readable storage medium of claim 21, wherein the fault comprises failure of one or more of the VPNs or VRs.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. application Ser. No. 12/838,487, filed on Jul. 18, 2010, which is a continuation of U.S. application Ser. No. 12/554,977, filed on Sep. 7, 2009, now U.S. Pat. No. 7,761,743, which is a continuation of U.S. application Ser. No. 11/848,298, filed on Sep. 3, 2007, now U.S. Pat. No. 7,587,633, which is a continuation of U.S. application Ser. No. 11/466,098, filed on Aug. 21, 2006, now U.S. Pat. No. 7,278,055, which is a continuation of U.S. application Ser. No. 10/232,979, filed on Aug. 29, 2002, now U.S. Pat. No. 7,096,383, all of which are hereby incorporated by reference in their entirety for all purposes.

US Referenced Citations (279)
Number Name Date Kind
4667287 Allen et al. May 1987 A
4667323 Engdahl et al. May 1987 A
4726018 Bux et al. Feb 1988 A
5473599 Li et al. Dec 1995 A
5490252 Macera et al. Feb 1996 A
5491691 Shtayer et al. Feb 1996 A
5550816 Hardwick et al. Aug 1996 A
5581705 Passint et al. Dec 1996 A
5592610 Chittor et al. Jan 1997 A
5598414 Walser et al. Jan 1997 A
5633866 Callon May 1997 A
5745778 Alfieri Apr 1998 A
5748882 Huang et al. May 1998 A
5825772 Dobbins et al. Oct 1998 A
5841973 Kessler et al. Nov 1998 A
5841990 Picazo et al. Nov 1998 A
5875290 Bartfai et al. Feb 1999 A
5892924 Lyon et al. Apr 1999 A
5920705 Lyon et al. Jul 1999 A
5963555 Takase et al. Oct 1999 A
5964847 Booth et al. Oct 1999 A
5978933 Wyld et al. Nov 1999 A
5983371 Lord et al. Nov 1999 A
5987521 Arrowood et al. Nov 1999 A
6014382 Takihiro et al. Jan 2000 A
6014669 Slaughter et al. Jan 2000 A
6032193 Sullivan Feb 2000 A
6047330 Stracke et al. Apr 2000 A
6069895 Ayandeh et al. May 2000 A
6085238 Yuasa et al. Jul 2000 A
6098110 Witkowski et al. Aug 2000 A
6101327 Holte et al. Aug 2000 A
6108699 Moiin Aug 2000 A
6108701 Davies et al. Aug 2000 A
6118791 Fichou et al. Sep 2000 A
6137777 Vaid et al. Oct 2000 A
6169739 Isoyama Jan 2001 B1
6169793 Godwin et al. Jan 2001 B1
6173399 Gilbrech Jan 2001 B1
6175867 Taghadoss Jan 2001 B1
6192051 Lipman et al. Feb 2001 B1
6212556 Arunachalam Apr 2001 B1
6220768 Barroux Apr 2001 B1
6226296 Lindsey et al. May 2001 B1
6226788 Schoening et al. May 2001 B1
6243580 Garner Jun 2001 B1
6249519 Rangachar Jun 2001 B1
6256295 Callon Jul 2001 B1
6260072 Rodriguez Jul 2001 B1
6260073 Walker et al. Jul 2001 B1
6266695 Huang et al. Jul 2001 B1
6278708 Von Hammerstein et al. Aug 2001 B1
6286038 Reichmeyer et al. Sep 2001 B1
6295297 Lee Sep 2001 B1
6298130 Galvin Oct 2001 B1
6317748 Menzies et al. Nov 2001 B1
6324583 Stevens Nov 2001 B1
6330602 Law et al. Dec 2001 B1
6338092 Chao et al. Jan 2002 B1
6339782 Gerard et al. Jan 2002 B1
6370656 Olarig et al. Apr 2002 B1
6405262 Vogel et al. Jun 2002 B1
6414595 Scrandis et al. Jul 2002 B1
6424657 Voit et al. Jul 2002 B1
6434619 Lim et al. Aug 2002 B1
6438612 Ylonen et al. Aug 2002 B1
6449650 Westfall et al. Sep 2002 B1
6453406 Sarnikowski et al. Sep 2002 B1
6459682 Ellesson et al. Oct 2002 B1
6463061 Rekhter et al. Oct 2002 B1
6466976 Alles et al. Oct 2002 B1
6487666 Shanklin Nov 2002 B1
6493349 Casey Dec 2002 B1
6496935 Fink et al. Dec 2002 B1
6526056 Rekhter et al. Feb 2003 B1
6532088 Dantu Mar 2003 B1
6542466 Pashtan et al. Apr 2003 B1
6542515 Kumar et al. Apr 2003 B1
6553423 Chen Apr 2003 B1
6556544 Lee Apr 2003 B1
6597956 Aziz et al. Jul 2003 B1
6608816 Nichols Aug 2003 B1
6611498 Baker et al. Aug 2003 B1
6611522 Zheng et al. Aug 2003 B1
6625169 Tofano Sep 2003 B1
6625650 Stelliga Sep 2003 B2
6631519 Nicholson et al. Oct 2003 B1
6635014 Starkweather et al. Oct 2003 B2
6636516 Yamano Oct 2003 B1
6638516 Becker et al. Oct 2003 B1
6639897 Shiomoto et al. Oct 2003 B1
6658013 de Boer et al. Dec 2003 B1
6668282 Booth et al. Dec 2003 B1
6697359 George Feb 2004 B1
6697360 Gai et al. Feb 2004 B1
6701449 Davis et al. Mar 2004 B1
6738371 Ayres May 2004 B1
6738821 Wilson et al. May 2004 B1
6763236 Sirén Jul 2004 B2
6769124 Schoening et al. Jul 2004 B1
6775267 Kung Aug 2004 B1
6775284 Calvignac et al. Aug 2004 B1
6778502 Ricciulli Aug 2004 B2
6785691 Hewett et al. Aug 2004 B1
6807181 Weschler Oct 2004 B1
6816462 Booth et al. Nov 2004 B1
6839866 Lerman Jan 2005 B2
6856676 Pirot et al. Feb 2005 B1
6865157 Scott et al. Mar 2005 B1
6868082 Allen et al. Mar 2005 B1
6883170 Garcia Apr 2005 B1
6894994 Grob et al. May 2005 B1
6907039 Shen Jun 2005 B2
6915445 Navar et al. Jul 2005 B2
6920580 Cramer et al. Jul 2005 B1
6922774 Meushaw et al. Jul 2005 B2
6934880 Hofner Aug 2005 B2
6938097 Vincent Aug 2005 B1
6944128 Nichols Sep 2005 B2
6944168 Paatela et al. Sep 2005 B2
6954429 Horton et al. Oct 2005 B2
6959194 Brouwer et al. Oct 2005 B2
6971044 Geng et al. Nov 2005 B2
6980526 Jang et al. Dec 2005 B2
6982987 Cain Jan 2006 B2
6985438 Tschudin Jan 2006 B1
6985956 Luke et al. Jan 2006 B2
6990103 Gollamudi Jan 2006 B1
7010715 Barbas et al. Mar 2006 B2
7020143 Zdan Mar 2006 B2
7028333 Tuomenoksa et al. Apr 2006 B2
7042843 Ni May 2006 B2
7042848 Santiago et al. May 2006 B2
7054291 Balazinski et al. May 2006 B2
7054311 Norman et al. May 2006 B2
7058716 Sundaresan et al. Jun 2006 B1
7062642 Langrind et al. Jun 2006 B1
7068656 Sainomoto et al. Jun 2006 B2
7079481 Kramer Jul 2006 B2
7082477 Sadhasivam et al. Jul 2006 B1
7089293 Grosner et al. Aug 2006 B2
7096383 Talaugon Aug 2006 B2
7096495 Warrier et al. Aug 2006 B1
7111072 Matthews et al. Sep 2006 B1
7116665 Balay et al. Oct 2006 B2
7116679 Ghahremani Oct 2006 B1
7117241 Bloch et al. Oct 2006 B2
7127049 Godse et al. Oct 2006 B2
7159031 Larkin et al. Jan 2007 B1
7159035 Garcia-Luna-Aceves et al. Jan 2007 B2
7161904 Hussain et al. Jan 2007 B2
7174372 Sarkar Feb 2007 B1
7177311 Hussain et al. Feb 2007 B1
7181547 Millet Feb 2007 B1
7181766 Bendinelli et al. Feb 2007 B2
7187676 DiMambro Mar 2007 B2
7197553 Roberts et al. Mar 2007 B2
7203192 Desai et al. Apr 2007 B2
7221945 Milford May 2007 B2
7225259 Ho et al. May 2007 B2
7243371 Kasper et al. Jul 2007 B1
7263106 Matthews et al. Aug 2007 B2
7266120 Cheng et al. Sep 2007 B2
7272643 Sarkar et al. Sep 2007 B1
7278055 Talaugon Oct 2007 B2
7293255 Kumar Nov 2007 B2
7313614 Considine et al. Dec 2007 B2
7316029 Parker et al. Jan 2008 B1
7324489 Iyer Jan 2008 B1
7337221 Radi et al. Feb 2008 B2
7340535 Alam Mar 2008 B1
7376125 Hussain et al. May 2008 B1
7378827 Stoichita May 2008 B2
7386010 Solomon et al. Jun 2008 B2
7389358 Matthews et al. Jun 2008 B1
7409577 Wing et al. Aug 2008 B2
7463633 Endo et al. Dec 2008 B2
7587633 Talaugon Sep 2009 B2
7639632 Sarkar et al. Dec 2009 B2
7720053 Hussain May 2010 B2
7761743 Talaugon Jul 2010 B2
7843813 Balay Nov 2010 B2
7869361 Balay Jan 2011 B2
7876683 Balay Jan 2011 B2
7881244 Balay Feb 2011 B2
7925920 Talaugon Apr 2011 B2
7933269 Cheng Apr 2011 B2
7957407 Desai Jun 2011 B2
7961615 Balay Jun 2011 B2
20010024425 Tsunoda et al. Sep 2001 A1
20010028636 Skog et al. Oct 2001 A1
20010033580 Dorsey et al. Oct 2001 A1
20010034733 Prompt et al. Oct 2001 A1
20010043571 Jang et al. Nov 2001 A1
20010048661 Clear et al. Dec 2001 A1
20010052013 Munguia et al. Dec 2001 A1
20020023171 Garrett et al. Feb 2002 A1
20020062344 Ylonen et al. May 2002 A1
20020066034 Schlossberg et al. May 2002 A1
20020071389 Seo Jun 2002 A1
20020075901 Perlmutter et al. Jun 2002 A1
20020097872 Maliszewski Jul 2002 A1
20020099849 Alfieri et al. Jul 2002 A1
20020150093 Ott et al. Oct 2002 A1
20020150114 Sainomoto et al. Oct 2002 A1
20020152373 Sun et al. Oct 2002 A1
20020186661 Santiago et al. Dec 2002 A1
20020191604 Mitchell et al. Dec 2002 A1
20030012209 Abdelilah et al. Jan 2003 A1
20030026262 Jarl Feb 2003 A1
20030033401 Poisson et al. Feb 2003 A1
20030063590 Mohan et al. Apr 2003 A1
20030108041 Aysan Jun 2003 A1
20030115308 Best et al. Jun 2003 A1
20030117954 De Neve et al. Jun 2003 A1
20030128663 Kramer Jul 2003 A1
20030131228 Twomey Jul 2003 A1
20030140273 Kamalvanshi et al. Jul 2003 A1
20030169747 Wang Sep 2003 A1
20030185221 Deikman et al. Oct 2003 A1
20030200295 Roberts et al. Oct 2003 A1
20030212735 Hicok et al. Nov 2003 A1
20030223406 Balay Dec 2003 A1
20040006601 Bernstein et al. Jan 2004 A1
20040037279 Zelig et al. Feb 2004 A1
20040042416 Ngo et al. Mar 2004 A1
20040078772 Balay et al. Apr 2004 A1
20040095932 Astarabadi et al. May 2004 A1
20040095934 Cheng et al. May 2004 A1
20040141521 George Jul 2004 A1
20040160900 Lund et al. Aug 2004 A1
20040199567 Lund Oct 2004 A1
20040199568 Lund Oct 2004 A1
20040199569 Kalkunte et al. Oct 2004 A1
20050002417 Kelly et al. Jan 2005 A1
20050047407 Desai et al. Mar 2005 A1
20050055306 Miller et al. Mar 2005 A1
20050081059 Bandini et al. Apr 2005 A1
20050083955 Guichard et al. Apr 2005 A1
20050108340 Gleeson et al. May 2005 A1
20050113114 Asthana May 2005 A1
20050213589 Shih Sep 2005 A1
20050243798 Dunn et al. Nov 2005 A1
20060087969 Santiago et al. Apr 2006 A1
20060206713 Hickman et al. Sep 2006 A1
20060265519 Millet Nov 2006 A1
20070058648 Millet Mar 2007 A1
20070064704 Balay et al. Mar 2007 A1
20070073733 Matthews et al. Mar 2007 A1
20070083528 Matthews et al. Apr 2007 A1
20070104119 Sarkar et al. May 2007 A1
20070109968 Hussain et al. May 2007 A1
20070110062 Balay et al. May 2007 A1
20070115979 Balay et al. May 2007 A1
20070121579 Matthews et al. May 2007 A1
20070127382 Hussain et al. Jun 2007 A1
20070147368 Desai et al. Jun 2007 A1
20070291755 Cheng et al. Dec 2007 A1
20080013470 Kopplin Jan 2008 A1
20080028456 O'Rourke et al. Jan 2008 A1
20080117917 Balay et al. May 2008 A1
20080259936 Hussain et al. Oct 2008 A1
20080317040 Balay et al. Dec 2008 A1
20080317231 Balay et al. Dec 2008 A1
20080320553 Balay et al. Dec 2008 A1
20090007228 Balay et al. Jan 2009 A1
20090073977 Hussain et al. Mar 2009 A1
20090131020 van de Groenendaal May 2009 A1
20090225759 Hussain et al. Sep 2009 A1
20090238181 Desai et al. Sep 2009 A1
20090279567 Ta et al. Nov 2009 A1
20100094980 Sarkar et al. Apr 2010 A1
20100142527 Balay et al. Jun 2010 A1
20100146098 Ishizakl et al. Jun 2010 A1
20100146627 Lin Jun 2010 A1
20100189016 Millet Jul 2010 A1
20100220732 Hussain et al. Sep 2010 A1
20100220741 Desai et al. Sep 2010 A1
20110122872 Balay May 2011 A1
Foreign Referenced Citations (5)
Number Date Country
0051290 Aug 2000 WO
0076152 Dec 2000 WO
0163809 Aug 2001 WO
0223855 Mar 2002 WO
03103237 Dec 2003 WO
Non-Patent Literature Citations (141)
Entry
Notice of Allowance dated Dec. 1, 2004 for U.S. Appl. No. 09/661,636.
Amendment and Response filed on Sep. 2, 2004 for U.S. Appl. No. 09/661,636.
Office Action dated May 28, 2004 for U.S. Appl. No. 09/661,636.
Amendment and Response filed on Mar. 22, 2004 for U.S. Appl. No. 09/661,636.
Office Action dated Nov. 18, 2003 U.S. Appl. No. 09/661,636.
Amendment and Response filed on Apr. 29, 2007 for U.S. Appl. No. 09/661,130.
Office Action dated Dec. 28, 2006 for U.S. Appl. No. 09/661,130.
Amendment and Response filed on Mar. 6, 2006 for U.S. Appl. No. 09/661,130.
Office Action dated Oct. 18, 2004 for U.S. Appl. No. 09/661,130.
Amendment and Response filed on Apr. 9, 2004 for U.S. Appl. No. 09/661,130.
Office Action dated Nov. 5, 2003 for U.S. Appl. No. 09/661,130.
Notice of Allowance dated Jun. 14, 2007 for U.S. Appl. No. 10/067,106.
Amendment and Response filed on Mar. 10, 2007 for U.S. Appl. No. 10/067,106.
Office Action dated Nov. 16, 2006 for U.S. Appl. No. 10/067,106.
Amendment and Response filed on Aug. 28, 2006 for U.S. Appl. No. 10/067,106.
Office Action dated Mar. 27, 2006 for U.S. Appl. No. 10/067,106.
Amendment and Response filed on Nov. 6, 2006 for U.S. Appl. No. 09/663,483.
Office Action dated Jul. 6, 2006 for U.S. Appl. No. 09/663,483.
Amendment and Response filed on Mar. 13, 2006 for U.S. Appl. No. 09/663,483.
Advisory Action dated Nov. 12, 2004 for U.S. Appl. No. 09/663,483.
Amendment and Response filed on Oct. 8, 2004 for U.S. Appl. No. 09/663,483.
Office Action dated Jun. 3, 2004 for U.S. Appl. No. 09/663,483.
Amendment and Response filed on Feb. 26, 2004 for U.S. Appl. No. 09/663,483.
Office Action dated Aug. 21, 2003 for U.S. Appl. No. 09/663,483.
Amendment and Response filed on Mar. 13, 2006 for U.S. Appl. No. 09/952,520.
Office Action dated Mar. 14, 2005 for U.S. Appl. No. 09/952,520.
Notice of Allowance dated Jul. 30, 2007 for U.S. Appl. No. 09/663,485.
Amendment and Response filed on Jun. 11, 2007 for U.S. Appl. No. 09/663,485.
Office Action dated Jan. 11, 2007 for U.S. Appl. No. 09/663,485.
Amendment and Response filed on Aug. 28, 2006 for U.S. Appl. No. 09/663,485.
Office Action dated Jul. 26, 2007 for U.S. Appl. No. 09/663,485.
Amendment and Response filed on Feb. 2, 2006 for U.S. Appl. No. 09/663,485.
Office Action dated Dec. 21, 2004 for U.S. Appl. No. 09/663,485.
Amendment and Response filed on Nov. 16, 2004 for U.S. Appl. No. 09/663,485.
Office Action dated May 14, 2004 for U.S. Appl. No. 09/663,485.
Amendment and Response filed on Mar. 15, 2004 for U.S. Appl. No. 09/663,485.
Office Action dated Sep. 8, 2003 for U.S. Appl. No. 09/663,485.
Office Action dated Aug. 8, 2007 for U.S. Appl. No. 09/663,457.
Amendment and Response filed on Jul. 11, 2007 for U.S. Appl. No. 09/663,457.
Office Action dated May 17, 2007 for U.S. Appl. No. 09/663,457.
Amendment and Response filed on Oct. 2, 2006 for U.S. Appl. No. 09/663,457.
Office Action dated Apr. 22, 2005 for U.S. Appl. No. 09/663,457.
Office Action dated Aug. 27, 2004 for U.S. Appl. No. 09/663,457.
Amendment and Response filed on Jun. 21, 2004 for U.S. Appl. No. 09/663,457.
Office Action dated Dec. 11, 2003 for U.S. Appl. No. 09/663,457.
Notice of Allowance dated Nov. 21, 2006 for U.S. Appl. No. 09/663,484.
Amendment and Response filed on Aug. 24, 2006 for U.S. Appl. No. 09/663,484.
Office Action dated Feb. 24, 2006 for U.S. Appl. No. 09/663,484.
Amendment and Response filed on Feb. 7, 2006 for U.S. Appl. No. 09/663,484.
Office Action dated Apr. 6, 2005 for U.S. Appl. No. 09/663,484.
Lawrence, J. Lang et al.“Connecting Remote FDDI Installations with Single-Mode Fiber, Dedicated Lines, or SMDS.” Jul. 1990; ACM SIGCOMM Computer Communication Review. vol. 20, Issue 3; pp. 72-82.
IEEE Potentials Publication; “Local Area Networks” Dec. 1995/Jan. 1996; pp. 6. http://www.ece.uc.edu/-paw/potentials/sample.
Office Action dated Oct. 18, 2007 for U.S. Appl. No. 09/663,483.
Office Action dated Oct. 16, 2007 for U.S. Appl. No. 09/661,130.
Office Action dated Nov. 28, 2007 for U.S. Appl. No. 09/952,520.
A lightweight Protocol for Interconnection Heterogenous Devices in Dynamic Environments, (c) 1999, obtained from the Internet at : http//ieeexplore.ieee.org/iel5/6322/16898/00778477.pdf.
The Guide to Computing Literature, Jairo A.: A Framework and Lightweight Protocol for Multimedia Network Management, vol. 8, Issue 1, published 2000, ISSN: 1064-7570.
Bookfinder4u.com: High Performance Networks by Ahmed N. Tantawy, ISBN-10: 0792393716, Published 1993, Lightweight Protocols.
Ipinfusion white paper: Virtual Routing for Provide Edge Application, obtained from the Internet at: http://www.ipinfusion.com/pdf/VirtualRouting—app-note—3rev0302.pdf, pp. 1-8.
Non-Final Office Action for U.S. Appl. No. 10/991,969, dated Feb. 20, 2008.
Non-Final Office Action for U.S. Appl. No. 10/273,669, dated Feb. 20, 2008.
Non-Final Office Action for U.S. Appl. No. 10/949,943 dated Feb. 14, 2008.
Restriction Requirement for U.S. Appl. No. 11/556,697, dated Mar. 13, 2008.
Office Action dated Aug. 1, 2007 for U.S. Appl. No. 10/163,260.
Amendment and Response filed on May 23, 2007 for U.S. Appl. No. 10/163,260.
Office Action dated Apr. 13, 2007 for U.S. Appl. No. 10/163,260.
Amendment and Response filed on Mar. 13, 2007 for U.S. Appl. No. 10/163,260.
Office Action dated Dec. 21, 2006 for U.S. Appl. No. 10/163,260.
Amendment and Response filed on Sep. 18, 2006 U.S. Appl. No. 10/163,260.
Office Action dated May 18, 2006 for U.S. Appl. No. 10/163,260.
Office Action dated Aug. 22, 2007 for U.S. Appl. No. 10/650,298.
Response to Restriction Requirement Apr. 26, 2004 for U.S. Appl. No. 09/663,483.
Restriction Requirement dated Mar. 22, 2004 for U.S. Appl. No. 09/663,483.
Office Action dated Sep. 11, 2007 for U.S. Appl. No. 09/661,637.
Amendment and Response filed on Jun. 20, 2007 for U.S. Appl. No. 09/661,637.
Office Action dated Feb. 8, 2007 for U.S. Appl. No. 09/661,637.
Amendment and Response filed on Mar. 6, 2006 for U.S. Appl. No. 09/661,637.
Office Action dated Dec. 23, 2004 for U.S. Appl. No. 09/661,637.
Amendment and Response filed on Aug. 5, 2004 for U.S. Appl. No. 09/661,637.
Office Action dated May 5, 2004 for U.S. Appl. No. 09/661,637.
Supplemental Amendment and Response filed on Sep. 17, 2007, for U.S. Appl. No. 09/663,457.
Amendment and Response filed on Nov. 12, 2004 for U.S. Appl. No. 09/663,484.
Office Action dated May 6, 2004 for U.S. Appl. No. 09/663,484.
Amendment and Response filed on Feb. 18, 2004 for U.S. Appl. No. 09/663,484.
Office Action dated Aug. 12, 2003 for U.S. Appl. No. 09/663,484.
Notice of Allowance dated Jan. 4, 2007 for U.S. Appl. No. 09/894,471.
Amendment and Response filed on Nov. 2, 2006 for U.S. Appl. No. 09/894,471.
Office Action dated Oct. 26, 2006 for U.S. Appl. No. 09/894,471.
Amendment and Response filed on Mar. 10, 2006 for U.S. Appl. No. 09/894,471.
Office Action dated Dec. 14, 2004 for U.S. Appl. No. 09/894,471.
Notice of Allowance dated Nov. 7, 2006 for U.S. Appl. No. 09/771,346.
Amendment and Response filed on Oct. 18, 2006 for U.S. Appl. No. 09/771,346.
Office Action dated Jul. 18, 2006 for U.S. Appl. No. 09/771,346.
Amendment and Response filed on Mar. 13, 2006 for U.S. Appl. No. 09/771,346.
Office Action dated Jan. 25, 2005 for U.S. Appl. No. 09/771,346.
Amendment and Response filed on Oct. 14, 2004 for U.S. Appl. No. 09/771,346.
Office Action dated Mar. 26, 2004 for U.S. Appl. No. 09/771,346.
Notice of Allowance dated Nov. 19, 2006 for U.S. Appl. No. 10/163,162.
Amendment and Response filed on Aug. 5, 2006 for U.S. Appl. No. 10/163,162.
Office Action dated May 5, 2006 for U.S. Appl. No. 10/163,162.
Notice of Allowance dated Jan. 4, 2007 for U.S. Appl. No. 10/163,261.
Amendment and Response filed on Nov. 9, 2006 for U.S. Appl. No. 10/163,261.
Office Action dated Nov. 3, 2006 for U.S. Appl. No. 10/163,261.
Amendment and Response filed on Aug. 22, 2006 for U.S. Appl. No. 10/163,261.
Office Action dated May 22, 2006 for U.S. Appl. No. 10/163,261.
Notice of Allowance dated Jul. 27, 2006 for U.S. Appl. No. 10/163,073.
Office Action dated May 30, 2007 for U.S. Appl. No. 10/273,669.
Amendment and Response filed on Mar. 9, 2007 for U.S. Appl. No. 10/273,669.
Office Action dated Sep. 21, 2006 for U.S. Appl. No. 10/273,669.
Amendment and Response filed on Jun. 21, 2006 for U.S. Appl. No. 10/273,669.
Office Action dated Feb. 21, 2006 for U.S. Appl. No. 10/273,669.
Notice of Allowance dated Aug. 14, 2007 for U.S. Appl. No. 10/163,071.
Amendment and Response filed on Jul. 17, 2007 for U.S. Appl. No. 10/163,071.
Office Action dated Jul. 3, 2007 for U.S. Appl. No. 10/163,071.
Amendment and Response filed on May 6, 2007 for U.S. Appl. No. 10/163,071.
Office Action dated Nov. 7, 2006 for U.S. Appl. No. 10/163,071.
Amendment and Response filed on Sep. 1, 2006 for U.S. Appl. No. 10/163,071.
Office Action dated Jun. 1, 2006 for U.S. Appl. No. 10/163,071.
Amendment and Response filed on Mar. 6, 2006 for U.S. Appl. No. 10/163,071.
Office Action dated Dec. 2, 2005 for U.S. Appl. No. 10/163,071.
Notice of Allowance dated Nov. 29, 2006 for U.S. Appl. No. 10/163,079.
Amendment and Response filed on Nov. 1, 2006 for U.S. Appl. No. 10/163,079.
Office Action dated Oct. 27, 2006 for U.S. Appl. No. 10/163,079.
Amendment and Response filed on Aug. 17, 2006 for U.S. Appl. No. 10/163,079.
Office Action dated May 17, 2006 for U.S. Appl. No. 10/163,079.
Notice of Allowance dated Jul. 17, 2007 for U.S. Appl. No. 10/298,815.
Amendment and Response filed on Mar. 9, 2007 for U.S. Appl. No. 10/298,815.
Office Action dated Feb. 23, 2007 for U.S. Appl. No. 10/298,815.
Notice of Allowance dated Jun. 27, 2005 for U.S. Appl. No. 10/232,979.
Notice of Allowance dated Jul. 5, 2007 for U.S. Appl. No. 11/466,098.
Amendment and Response filed on Aug. 10, 2007 for U.S. Appl. No. 10/163,260.
Non-Final Office Action for U.S. Appl. No. 11/849,352 mailed Jul. 17, 2009.
Non-Final Office Action for U.S. Appl. No. 10/991,970 mailed May 18, 2009.
Final Office Action for U.S. Appl. No. 12,467,609 mailed Apr. 19, 2011.
Non-Final Office Action for U.S. Appl. No. 12/328,858 mailed Apr. 15, 2011.
Non-Final Rejection for U.S. Appl. No. 12/477,124 mailed May 23, 2011. (1920).
Non-Final Rejection for U.S. Appl. No. 12/637,140, mailed Sep. 17, 2010.
Non-Final Rejection for U.S. Appl. No. 12/537,898, mailed Sep. 9, 2010.
Final Rejection for U.S. Appl. No. 12/202,223, mailed Sep. 16, 2010.
Non-Final Rejection for U.S. Appl. No. 12/202,233 mailed Jun. 21, 2010.
Non-Final Rejection for U.S. Appl. No. 11/460,977, mailed Jul. 2, 2010.
Related Publications (1)
Number Date Country
20110185221 A1 Jul 2011 US
Continuations (5)
Number Date Country
Parent 12838487 Jul 2010 US
Child 13083291 US
Parent 12554977 Sep 2009 US
Child 12838487 US
Parent 11849298 Sep 2007 US
Child 12554977 US
Parent 11466098 Aug 2006 US
Child 11849298 US
Parent 10232979 Aug 2002 US
Child 11466098 US