Multi-fabric manager

Information

  • Patent Grant
  • 9807005
  • Patent Number
    9,807,005
  • Date Filed
    Tuesday, March 17, 2015
    9 years ago
  • Date Issued
    Tuesday, October 31, 2017
    7 years ago
Abstract
One embodiment of the present invention provides a computing system. The computing system includes processing circuitry, one or more ports, a persistent storage module, and a management module. The persistent storage module stores a plurality of persistent storage instances associated with a plurality of switch groups. A respective persistent storage instance stores configuration information associated with a switch group in a data structure. The management module identifies at least two switch groups, which are associated with a logical network, from the plurality of switch groups and configures a respective virtual network representing the logical network in a respective identified switch group. The management module also maintains a mapping between a respective virtual network and the logical network.
Description
RELATED APPLICATION

The present disclosure is related to U.S. Pat. No. 8,867,552, titled “Virtual Cluster Switching,” by inventors Suresh Vobbilisetty and Dilip Chatwani, issued 21 Oct. 2014; and U.S. patent application Ser. No. 14/512,268, titled “Distributed Configuration Management in a Switch Group,” by inventors Vidyasagara R. Guntaka, Suresh Vobbilisetty, Manjunath A. G. Gowda, and Himanshu Varshney, filed 10 Oct. 2014, the disclosures of which are incorporated by reference herein.


BACKGROUND

Field


The present disclosure relates to communication networks. More specifically, the present disclosure relates to a method for managing multiple scalable systems.


Related Art


The exponential growth of the Internet has made it a popular delivery medium for a variety of applications running on physical and virtual devices. Such applications have brought with them an increasing demand for bandwidth. As a result, equipment vendors race to build larger and faster switches with versatile capabilities. However, the size of a switch cannot grow infinitely. It is limited by physical space, power consumption, and design complexity, to name a few factors. Furthermore, switches with higher capability are usually more complex and expensive. More importantly, because an overly large and complex system often does not provide economy of scale, simply increasing the size and capability of a switch may prove economically unviable due to the increased per-port cost.


A flexible way to improve the scalability of a switch system is to build a fabric switch. A fabric switch is a collection of individual member switches. These member switches form a single, logical switch that can have an arbitrary number of ports and an arbitrary topology. As demands grow, customers can adopt a “pay as you grow” approach to scale up the capacity of the fabric switch. However, a respective fabric switch is typically managed individually. As a result, if a tenant (e.g., a customer of a network provider) has a network that spans multiple fabric switches across multiple locations, managing such a network requires configuration at multiple switch groups individually. Additionally, a respective configuration change leads to individual configurations of a respective fabric switch, which can be tedious and error-prone.


Meanwhile, a switch, an individual or a member switch of a fabric switch, continues to store more configuration information as the switch participates in network virtualizations, partitions, and switch groups, and operates on a plurality of network protocols of different network layers. This configuration needs to be applied to the switch when the switch powers up, and thus, should be persistent. A switch typically stores such configuration information in a local storage in an unstructured format. The switch reads the information during booting up (i.e., powering up), and loads the information into memory. Managing persistent storage in unstructured format is inefficient and requires runtime structuring.


While persistent storage brings many desirable features to a fabric switch, some issues remain unsolved in efficiently managing multiple fabric switches.


SUMMARY

One embodiment of the present invention provides a computing system. The computing system includes processing circuitry, one or more ports, a persistent storage module, and a management module. The persistent storage module stores a plurality of persistent storage instances associated with a plurality of switch groups. A respective persistent storage instance stores configuration information associated with a switch group in a data structure. The management module identifies at least two switch groups, which are associated with a logical network, from the plurality of switch groups and configures a respective virtual network representing the logical network in a respective identified switch group. The management module also maintains a mapping between a respective virtual network and the logical network.


In a variation on this embodiment, the computing system also includes a tunnel module which establishes a tunnel between the identified switch groups.


In a further variation, the tunnel is terminated at a logical interface, which is associated with one or more switches of an identified switch group.


In a variation on this embodiment, the management module obtains a media access control (MAC) address learned at a first switch group of the identified switch groups and stores the MAC address in a local data structure in association with a switch group identifier which identifies the first switch group.


In a variation on this embodiment, the management module programs reachability of a MAC address learned at one of the identified switch groups in another identified switch group.


In a variation on this embodiment, a header of a packet belonging to the logical network includes an identifier of a first virtual network representing the logical network in a first switch group of the identified switch group.


In a further variation, the header of the packet belonging to the logical network includes an identifier of a second virtual network representing the logical network in a second switch group of the identified switch group.


In a variation on this embodiment, the computing system also includes a switch group module which maintains a membership in a switch group. The switch group includes a plurality of switches and operates as a single switch.


In a variation on this embodiment, the management module generates a first control message comprising configuration information associated with a switch group for a principal switch of the switch group and a second control message indicating whether the configuration information is prepared for storage in a persistent storage instance of the switch group. The management module then commits the configuration information to the persistent storage instance based on an instruction message from the principal switch.


In a variation on this embodiment, the management module verifies whether configuration information is prepared for storage in a persistent storage instance of a replica of the computing system. If prepared, the management module generates an instruction message instructing to commit the configuration information to the persistent storage instance of the replica of the computing system.


In a variation on this embodiment, the persistent storage is an object relational database, and a data structure in the persistent storage is a table in the database.





BRIEF DESCRIPTION OF THE FIGURES


FIG. 1A illustrates an exemplary network with persistent storage framework support, in accordance with an embodiment of the present invention.



FIG. 1B illustrates an exemplary persistent storage framework support in a switch, in accordance with an embodiment of the present invention.



FIG. 2 illustrates an exemplary object identifier generated by a persistent storage framework in a switch, in accordance with an embodiment of the present invention.



FIG. 3 illustrates exemplary base classes for supporting a persistent storage framework in a switch, in accordance with an embodiment of the present invention.



FIG. 4A illustrates an exemplary Unified Modeling Language (UML) model of classes of a switch with a persistent storage framework, in accordance with an embodiment of the present invention.



FIG. 4B illustrates an exemplary Extensible Markup Language (XML) representation of a class corresponding to a switch with a persistent storage framework, in accordance with an embodiment of the present invention.



FIG. 4C illustrates exemplary tables generated by a persistent storage framework in a switch, in accordance with an embodiment of the present invention.



FIG. 4D illustrates an exemplary table representing a one-to-many association, which is generated by in a persistent storage framework in a switch, in accordance with an embodiment of the present invention.



FIG. 5A presents a flowchart illustrating the process of a persistent storage framework of a switch generating auxiliary tables for an inheritance chain in a persistent storage, in accordance with an embodiment of the present invention.



FIG. 5B presents a flowchart illustrating the process of a persistent storage framework of a switch generating class tables in a persistent storage, in accordance with an embodiment of the present invention.



FIG. 5C presents a flowchart illustrating the process of a persistent storage framework of a switch generating an auxiliary table representing an one-to-many relationship in a persistent storage, in accordance with an embodiment of the present invention.



FIG. 5D presents a flowchart illustrating the process of a persistent storage framework of a switch updating tables in a persistent storage, in accordance with an embodiment of the present invention.



FIG. 6A illustrates an exemplary switch group with persistent storage framework and distributed configuration management support, in accordance with an embodiment of the present invention.



FIG. 6B illustrates an exemplary switch group with multi-input and multi-principal switch support, in accordance with an embodiment of the present invention.



FIG. 7 illustrates an exemplary distributed configuration management, in accordance with an embodiment of the present invention.



FIG. 8A illustrates an exemplary multi-fabric manager (MFM) managing a plurality of switch groups, in accordance with an embodiment of the present invention.



FIG. 8B illustrates an exemplary console of a multi-fabric manager, in accordance with an embodiment of the present invention.



FIG. 8C illustrates an exemplary persistent storage framework of a multi-fabric manager managing a switch group, in accordance with an embodiment of the present invention.



FIG. 9A illustrates an exemplary logical network, which spans a plurality of switch groups and is created by a multi-fabric manager, in accordance with an embodiment of the present invention.



FIG. 9B illustrates exemplary tunnel establishment and global virtual local area network (GVLAN) translation for a logical network spanning a plurality of switch groups, in accordance with an embodiment of the present invention.



FIG. 9C illustrates exemplary overlapping logical networks, which span a plurality of switch groups and are created by a multi-fabric manager, in accordance with an embodiment of the present invention.



FIG. 10A presents a flowchart illustrating the process of a multi-fabric manager associating with a switch group, in accordance with an embodiment of the present invention.



FIG. 10B presents a flowchart illustrating the process of a multi-fabric manager creating a logical network, in accordance with an embodiment of the present invention.



FIG. 11A presents a flowchart illustrating the distributed commit process of a principal persistent storage framework applying configuration information, in accordance with an embodiment of the present invention.



FIG. 11B presents a flowchart illustrating the distributed commit process of a managed persistent storage framework applying configuration information received from a principal persistent storage framework, in accordance with an embodiment of the present invention.



FIG. 12 illustrates an exemplary multi-fabric manager, in accordance with an embodiment of the present invention.





In the figures, like reference numerals refer to the same figure elements.


DETAILED DESCRIPTION

The following description is presented to enable any person skilled in the art to make and use the invention, and is provided in the context of a particular application and its requirements. Various modifications to the disclosed embodiments will be readily apparent to those skilled in the art, and the general principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of the present invention. Thus, the present invention is not limited to the embodiments shown, but is to be accorded the widest scope consistent with the claims.


Overview

In embodiments of the present invention, the problem of managing multiple switch groups (e.g., fabric switches) is solved by associating a multi-fabric manager (MFM) with the switch groups. The multi-fabric manager maintains a copy of the configuration information of a respective switch group it manages in a local persistent storage instance. Based on the configuration information, these switch groups can be managed from the multi-fabric manager.


A switch can be configured by logging in to the switch (e.g., via telnet) or via a console port (e.g., an RS-232 port). Such configuration can be related to network virtualizations, partitions, and switch groups, and a plurality of network protocols of different network layers. The attribute values (e.g., parameters) of the configuration information are stored in a local persistent storage and applied to the switch (e.g., loaded to the switch modules). In some embodiments, a switch can be in a switch group (e.g., a fabric switch). Configuration and state information of a switch group can be stored in a persistent storage of a respective member switch of the switch group. One of the switches in the switch group is elected as a principal switch, which manages and distributes configuration information in the switch group. The rest of the switches of the switch group can be referred to as managed switches. When a user accesses this principal switch, the principal switch allows the user to configure the global configuration information of the switch group. The principal switch, in turn, synchronizes that global configuration information with a respective managed switch. Furthermore, the principal switch also allows the user to configure a respective managed switch in the switch group. The principal switch, in turn, synchronizes that local configuration information to that managed switch.


With existing technologies, a respective switch group is managed individually. For example, a user can manage a switch group based on the configuration information in the persistent storage in a switch group (e.g., in a respective member switch). However, that persistent storage only includes configuration information of that switch group. As a result, if a tenant (e.g., a customer of a network provider) has a tenant network that spans multiple switch groups across multiple locations, managing such a tenant network requires individual configuration of multiple switch groups. Additionally, a respective configuration change to that tenant network leads to individual configurations of a respective switch group, which can be tedious and error-prone.


To solve this problem, a multi-fabric manager facilitates a single point of management for a plurality of switch groups. In some embodiments, the multi-fabric manager becomes a member switch of these switch groups. As a result, the principal switch of a respective switch group synchronizes its local persistent storage with the multi-fabric manager. In this way, the multi-fabric manager obtains and maintains an instance of the persistent storage of a respective switch group it manages. This allows the multi-fabric manager to have a copy of the configuration information of a respective switch group it manages and facilitate configuration changes to the switch groups. Furthermore, a respective entry of a persistent storage can be associated with an object identifier comprising a group identifier. The multi-fabric manager can use this group identifier portion of the object identifier to distinguish among the configuration information associated with different switch groups.


Since the multi-fabric manager appears as a member switch of a respective switch group, a tenant network spanning multiple switch groups can be created from the multi-fabric manager. During operation, a user may configure the tenant network in the multiple switch groups from the multi-fabric manager. The user can provide the configuration information locally (e.g., via an input device and a display device) or remotely (e.g., via telnet or a remote shell). Upon receiving the configuration information, the multi-fabric manager provides (e.g., sends via a control message) the configuration information to the principal switch of a respective switch group associated with the tenant network.


The principal switch, in turn, performs a distributed commit in the local switch group. This distributed commit includes updating the attribute values of the configuration information in the persistent storage and applying the attribute values to the switch modules of a respective member switch in the switch group. Since the multi-fabric manager operates as a member switch of a respective switch group, the copy of the persistent storage of that switch group is also updated at the multi-fabric manager. However, the multi-fabric manager may not apply the attribute values to its local device modules since the multi-fabric manager may not participate in frame forwarding for the switch group. In this way, configuration information from the multi-fabric manager is propagated to multiple switch groups, thereby facilitating a single point of management for those switch groups.


In some embodiments, a switch in the switch group can be referred to as a member switch. A respective member switch of the switch group can be equipped with a persistent storage framework which stores the configuration information in a local persistent storage. Such a persistent storage can be an object relational database. The configuration information is loaded from this persistent storage to the switch (or device) modules (e.g., the application-specific integrated circuit (ASIC) chips of the switch). In this disclosure, the description in conjunction with FIGS. 1-5 is associated with persistent storage in the switch; the description in conjunction with FIGS. 6-7 is associated with distributed configuration management in a switch group; and the description in conjunction with FIG. 8 and onward provides more details on a multi-fabric manager.


In some embodiments, the framework supports Model Driven Architecture (MDA), Object Oriented Programming (OOP), and/or Model/View/Controller (MVC) design patterns to facilitate modular development and operation of the units. The framework can also support class frameworks based on Unified Modeling Language (UML). Upon receiving class models (e.g., class name, attributes, and methods) and their relations based on UML, the framework automatically generates the corresponding code, thereby ensuring structure in the operational units of a switch. In some embodiments, the class models are expressed in YANG, which is a data modeling language used to model configuration and state data manipulated by the Network Configuration Protocol (NETCONF).


Since the units operate on the framework in a modular way, their associated attribute values can be stored in a persistent storage in a structured way. In some embodiments, the framework uses Object-Relational Mapping to store the attribute values of the units in a structured way in an object relational database. The framework allows different classes to be defined for a unit based on MDA, OOP, and/or MVC design patterns. The framework then seamlessly maps a respective class to a database table and vice-versa. Furthermore, the framework also seamlessly represents the relationships among the classes (e.g., an association or a composition) in the database. As a result, when a unit becomes operational on the switch, attribute values associated with a respective class in that unit is automatically loaded from the database. Moreover, if a class changes (e.g., a new attribute or a new relationship), the framework seamlessly incorporates that change into the database.


In some embodiments, the switch group can be a fabric switch. The switch can include one or more units which allow the switch to join and operate as a member switch of the fabric switch. These units can also run on the framework. In a fabric switch, any number of switches coupled in an arbitrary topology may logically operate as a single switch. The fabric switch can be an Ethernet fabric switch or a virtual cluster switch (VCS), which can operate as a single Ethernet switch. Any member switch may join or leave the fabric switch in “plug-and-play” mode without any manual configuration. In some embodiments, a respective switch in the fabric switch is a Transparent Interconnection of Lots of Links (TRILL) routing bridge (RBridge). In some further embodiments, a respective switch in the fabric switch is an Internet Protocol (IP) routing-capable switch (e.g., an IP router).


It should be noted that a fabric switch is not the same as conventional switch stacking. In switch stacking, multiple switches are interconnected at a common location (often within the same rack), based on a particular topology, and manually configured in a particular way. These stacked switches typically share a common address, e.g., an IP address, so they can be addressed as a single switch externally. Furthermore, switch stacking requires a significant amount of manual configuration of the ports and inter-switch links. The need for manual configuration prohibits switch stacking from being a viable option in building a large-scale switching system. The topology restriction imposed by switch stacking also limits the number of switches that can be stacked. This is because it is very difficult, if not impossible, to design a stack topology that allows the overall switch bandwidth to scale adequately with the number of switch units.


In contrast, a fabric switch can include an arbitrary number of switches with individual addresses, can be based on an arbitrary topology, and does not require extensive manual configuration. The switches can reside in the same location, or be distributed over different locations. These features overcome the inherent limitations of switch stacking and make it possible to build a large “switch farm,” which can be treated as a single, logical switch. Due to the automatic configuration capabilities of the fabric switch, an individual physical switch can dynamically join or leave the fabric switch without disrupting services to the rest of the network.


Furthermore, the automatic and dynamic configurability of the fabric switch allows a network operator to build its switching system in a distributed and “pay-as-you-grow” fashion without sacrificing scalability. The fabric switch's ability to respond to changing network conditions makes it an ideal solution in a virtual computing environment, where network loads often change with time.


In switch stacking, multiple switches are interconnected at a common location (often within the same rack), based on a particular topology, and manually configured in a particular way. These stacked switches typically share a common address, e.g., an IP address, so they can be addressed as a single switch externally. Furthermore, switch stacking requires a significant amount of manual configuration of the ports and inter-switch links. The need for manual configuration prohibits switch stacking from being a viable option in building a large-scale switching system. The topology restriction imposed by switch stacking also limits the number of switches that can be stacked. This is because it is very difficult, if not impossible, to design a stack topology that allows the overall switch bandwidth to scale adequately with the number of switch units.


It should also be noted that a fabric switch is distinct from a virtual local area network (VLAN). A fabric switch can accommodate a plurality of VLANs. A VLAN is typically identified by a VLAN tag. In contrast, the fabric switch is identified a fabric identifier (e.g., a VCS identifier), which is assigned to the fabric switch. A respective member switch of the fabric switch is associated with the fabric identifier. Furthermore, when a member switch of a fabric switch learns a media access control (MAC) address of an end device (e.g., via layer-2 MAC address learning), the member switch generates a notification message, includes the learned MAC address in the payload of the notification message, and sends the notification message to all other member switches of the fabric switch. In this way, a learned MAC address is shared among a respective member switch of the fabric switch.


The term “fabric switch” refers to a number of interconnected physical switches which form a single, scalable logical switch. These physical switches are referred to as member switches of the fabric switch. In a fabric switch, any number of switches can be connected in an arbitrary topology, and the entire group of switches functions together as one single, logical switch. This feature makes it possible to use many smaller, inexpensive switches to construct a large fabric switch, which can be viewed as a single logical switch externally. Although the present disclosure is presented using examples based on a fabric switch, embodiments of the present invention are not limited to a fabric switch. Embodiments of the present invention are relevant to any computing device that includes a plurality of devices operating as a single device.


The term “end device” can refer to any device external to a fabric switch. Examples of an end device include, but are not limited to, a host machine, a conventional layer-2 switch, a layer-3 router, or any other type of network device. Additionally, an end device can be coupled to other switches or hosts further away from a layer-2 or layer-3 network. An end device can also be an aggregation point for a number of network devices to enter the fabric switch. An end device hosting one or more virtual machines can be referred to as a host machine. In this disclosure, the terms “end device” and “host machine” are used interchangeably.


The term “switch” is used in a generic sense, and it can refer to any standalone or fabric switch operating in any network layer. “Switch” should not be interpreted as limiting embodiments of the present invention to layer-2 networks. Any device that can forward traffic to an external device or another switch can be referred to as a “switch.” Any physical or virtual device (e.g., a virtual machine/switch operating on a computing device) that can forward traffic to an end device can be referred to as a “switch.” Examples of a “switch” include, but are not limited to, a layer-2 switch, a layer-3 router, a TRILL RBridge, or a fabric switch comprising a plurality of similar or heterogeneous smaller physical and/or virtual switches.


The term “edge port” refers to a port on a fabric switch which exchanges data frames with a network device outside of the fabric switch (i.e., an edge port is not used for exchanging data frames with another member switch of a fabric switch). The term “inter-switch port” refers to a port which sends/receives data frames among member switches of a fabric switch. The terms “interface” and “port” are used interchangeably.


The term “switch identifier” refers to a group of bits that can be used to identify a switch. Examples of a switch identifier include, but are not limited to, a media access control (MAC) address, an Internet Protocol (IP) address, and an RBridge identifier. Note that the TRILL standard uses “RBridge ID” (RBridge identifier) to denote a 48-bit intermediate-system-to-intermediate-system (IS-IS) System ID assigned to an RBridge, and “RBridge nickname” to denote a 16-bit value that serves as an abbreviation for the “RBridge ID.” In this disclosure, “switch identifier” is used as a generic term, is not limited to any bit format, and can refer to any format that can identify a switch. The term “RBridge identifier” is also used in a generic sense, is not limited to any bit format, and can refer to “RBridge ID,” “RBridge nickname,” or any other format that can identify an RBridge.


The term “packet” refers to a group of bits that can be transported together across a network. “Packet” should not be interpreted as limiting embodiments of the present invention to layer-3 networks. “Packet” can be replaced by other terminologies referring to a group of bits, such as “message,” “frame,” “cell,” or “datagram.”


Network Architecture



FIG. 1A illustrates an exemplary network with persistent storage framework support, in accordance with an embodiment of the present invention. As illustrated in FIG. 1A, a network 100 includes switches 101, 102, 103, 104, and 105. An end device 112 is coupled to switch 102. In some embodiments, end device 112 is a host machine, hosting one or more virtual machines. End device 112 can include a hypervisor, which runs one or more virtual machines. End device 112 can be equipped with a Network Interface Card (NIC) with one or more ports. End device 112 couples to switch 102 via the ports of the NIC.


In some embodiments, network 100 is a TRILL network and a respective switch of network 100, such as switch 102, is a TRILL RBridge. Under such a scenario, communication among the switches in network 100 is based on the TRILL protocol. For example, upon receiving an Ethernet frame from end device 112, switch 102 encapsulates the received Ethernet frame in a TRILL header and forwards the TRILL packet. In some further embodiments, network 100 is an IP network and a respective switch of network 100, such as switch 102, is an IP-capable switch, which calculates and maintains a local IP routing table (e.g., a routing information base or RIB), and is capable of forwarding packets based on its IP addresses. Under such a scenario, communication among the switches in network 100 is based on IP. For example, upon receiving an Ethernet frame from end device 112, switch 102 encapsulates the received Ethernet frame in an IP header and forwards the IP packet.


In some embodiments, network 100 is a fabric switch (under such a scenario, network 100 can also be referred to as fabric switch 100). Fabric switch 100 is assigned with a fabric switch identifier. A respective member switch of fabric switch 100 is associated with that fabric switch identifier. This allows the member switch to indicate that it is a member of fabric switch 100. In some embodiments, whenever a new member switch joins fabric switch 100, the fabric switch identifier is automatically associated with that new member switch. Furthermore, a respective member switch of fabric switch 100 is assigned a switch identifier (e.g., an RBridge identifier, a Fibre Channel (FC) domain ID (identifier), or an IP address). This switch identifier identifies the member switch in fabric switch 100.


Switches in fabric switch 100 use edge ports to communicate with end devices (e.g., non-member switches) and inter-switch ports to communicate with other member switches. For example, switch 102 is coupled to end device 112 via an edge port and to switches 101, 103, 104, and 105 via inter-switch ports and one or more links. Data communication via an edge port can be based on Ethernet and via an inter-switch port can be based on the IP and/or TRILL protocol. It should be noted that control message exchange via inter-switch ports can be based on a different protocol (e.g., the IP or FC protocol).


A switch, such as switch 102, stores configuration information needed to operate switch 102 as an individual switch or as a member switch of fabric switch 100. Furthermore, switch 102 can participate in various services and operations, such as network virtualization (e.g., a virtual local area networks (VLAN)), switch partitioning, and link aggregations (e.g., a multi-chassis trunk). Furthermore, switch 102 operates on a plurality of network protocols of different network layers (e.g., Ethernet, TRILL, FC, and IP). As a result, switch 102 runs protocol daemons for each of these protocols. However, to incorporate the services and operations, the protocol daemons need to be directly modified, which can lead to conflicts and errors.


Furthermore, each of the operations, services, and the protocols is associated with one or more attributes. These attribute values (e.g., parameters) is typically applied to switch 102 when switch 102 powers up. As a result, these attribute values are stored in a persistent storage so that these values can be retrieved even when switch 102 is powered off or restarts. With existing technologies, switch 102 may store such attribute values in a local storage in an unstructured format (e.g., a string comprising the attribute values). During the boot up process, switch 102 reads and parses the attribute values in the unstructured format, and loads the attribute values into switch memory. Managing persistent storage in unstructured format is inefficient and requires runtime structuring.


To solve this problem, switch 102 is equipped with a persistent storage framework 120 which facilitates structured persistent storage to the attribute values associated with different operational units of switch 102 (e.g., modules and services of switch 102). It should be noted that other switches of network 100 can be equipped with a persistent storage framework and support persistent storage. In some embodiments, some switch of network 100 may not be equipped with a persistent storage framework. Different units of switch 102, each of which facilitates an aspect of switch 102's operations, operate on framework 120 in a structured and modular way. This allows a respective unit to be independently introduced to framework 120 in such a way that the unit can interoperate with other units (e.g., modules and services) of switch 102. Framework 120 supports MDA, OOP, and/or MVC design patterns to facilitate structured development and operation of the units in switch 102.


Since the units operate on framework 120 in a structured way, their associated attribute values can be stored in a persistent storage in a structured way. In some embodiments, framework 120 uses Object-Relational Mapping to store the attribute values of the units in a structured way in an object relational database. Framework 120 allows different classes to be defined for a unit during development based on MDA, OOP, and/or MVC design patterns. Framework 120 supports class models based on UML. In some embodiments, class models are expressed in YANG, which is a data modeling language used to model configuration and state data manipulated by NETCONF. Upon receiving class models (e.g., class name, attributes, and methods) and their relationships based on UML, framework 120 automatically generates the corresponding code, thereby ensuring structure in the operational units of switch 102.


Framework 120 seamlessly maps a respective class to a database table and vice-versa. Furthermore, framework 120 also seamlessly represents the relations among the classes (e.g., an association or a composition) in the database. As a result, when a unit becomes operational on switch 102 (e.g., when switch 102 powers up), attribute values associated with a respective class in that unit is automatically loaded from the database. Moreover, if a class changes (e.g., a new attribute or a new relationship), framework 120 seamlessly incorporates that change into the database.


Persistent Storage Framework



FIG. 1B illustrates an exemplary persistent storage framework in a switch, in accordance with an embodiment of the present invention. In this example, persistent storage framework 120 of switch 102 provides structured persistent storage to the operational units of switch 102. In some embodiments, switch 102 is coupled to an end device 114, which can operate as an administrative terminal for switch 102. Switch 102 runs one or more protocol daemons 140. For example, switch 102 can run respective protocol daemons for Ethernet, TRILL, FC, and IP. A protocol daemon facilitates the services and operations of a corresponding protocol for switch 102.


Switch 102 further includes an input interface 122 to switch 102 (e.g., a graphical user interface (GUI) and/or a command line interface (CLI). A user can access input interface 122 via end device 114. The user can obtain information from and provide instruction to switch 102 via input interface 122. Switch 102 also includes a configuration daemon 124, which can receive configuration (e.g., an IP address) for switch 102 from end device 114 (e.g., from a user) via input interface 122. Configuration daemon 124 provides this configuration information to framework 120. Framework 120 can include a configuration daemon gateway module 132, which communicates with configuration daemon 124. Upon receiving the configuration information, framework 120 can identify different attribute values (e.g., a VLAN identifier) and assigns those attribute values to the corresponding attribute of an operational unit of switch 102.


On the other hand, switch 102 can receive an instruction via input interface 122 to provide its configuration associated with one or more units. For example, a user can issue a command to show the IP addresses assigned to switch 102 from end device 114. Input interface 122 provides this instruction to configuration daemon 124, which in turn, sends an internal command to configuration daemon gateway module 132 for the requested configuration information. In response, framework 120 identifies the attributes (e.g., IP addresses) associated with the requested configuration information and obtains the corresponding attribute values (e.g., assigned IP addresses to switch 120) from a persistent storage. Configuration daemon gateway module 132 provides the obtained attribute values to configuration daemon 124. Upon receiving the attribute values, configuration daemon 124 provides the attribute values as the requested configuration information to input interface 122, which in turn, provides the configuration information to end device 114.


Framework 120 includes a core management module 130, which facilitates structured persistent storage to the attribute values associated with different operational units of switch 102 (e.g., modules and services of switch 102). Different units of switch 102 operate on core management module 130 in a structured way. This allows a respective unit to be independently introduced to framework 120 such a way that the unit can interoperate with other units (e.g., modules and services) of switch 102. Framework 120 supports MDA, OOP, and/or MVC design pattern to facilitate structured development and operation of the units in switch 102.


For example, instead of modifying protocol daemons 140, switch 102 can have plug-ins 134 for protocol daemons 140. Core management module 130 facilitates inter-operations between plug-in 134 and protocol daemons 140. Suppose that a modification to standard Ethernet protocol is needed. Instead of modifying the native protocol daemon of Ethernet, a plug-in for the protocol daemon of Ethernet can be introduced to core management module 130. Similarly, to facilitate membership to a fabric switch, fabric switch module 136 can be introduced to core management module 130. Fabric switch module 136 allows switch 102 to run a control plane with automatic configuration capability and join a fabric switch based on the control plane. Plug-ins 134 and fabric switch module 136 can be developed using MDA, OOP, and/or MVC design patterns, supported by framework 120.


Since the units of switch 102 operate core management module 130 in a structured way, their associated attribute values can be stored in a persistent storage in a structured way. In some embodiments, core management module 130 uses Object-Relational Mapping to store the attribute values of the units in a structured way in an object relational database 150. Core management module 130 allows different classes to be defined for a unit during development based on MDA, OOP, and/or MVC design patterns and expressed as a UML model, and seamlessly maps a respective class to a database table in database 150 and vice-versa.


Furthermore, core management module 130 also seamlessly represents the relations among the classes (e.g., an association or a composition) in database 150. As a result, when a unit becomes operational on switch 102 (e.g., when switch 102 powers up), attribute values associated with a respective class in that unit is automatically loaded from database 150. Moreover, if a class changes (e.g., a new attribute or a new relationship), core management module 130 seamlessly incorporates that change into database 150. It should be noted that a class defined by a user may not include explicit instructions (e.g., a Structured Query Language (SQL) query) for inserting and retrieving attribute values from database 150. The class simply includes an instruction indicating that persistent storage is required for some operations and core management module 130 facilitates the object relational mapping, and the corresponding database operations (e.g., SQL insert and select).


Attribute Data Types


To facilitate seamless object relational mapping, a persistent storage framework defines a set of data types for different categories of attributes. These attributes can be used to define class attributes of different operational units of a switch. In some embodiments, the framework can identify these class attributes expressed in a UML model. It should be noted that such expression can be represented in various forms, such as graphical, textual, XML, etc. The framework ensures these attributes are compatible with an object relational database. As a result, during operation, the framework can seamlessly map the class attributes to an object relational database and provide persistent storage to the attributes.


A data type of an attribute is basic entity provided by the framework that can be persisted or transported in the object relational database. A data type is associated with an identifier (e.g., a name). A data type can be, persisted or ephemeral, configuration or operational and read-only or read-write. The framework can serialize or de-serialize a data type to or from: XML, remote procedure call (RPC), SQL, JavaScript Object Notation (JSON), and Open vSwitch Database (OVSDB) management protocol.


The framework supports different categories of attributes. Such categories include, but are not limited to, integers and numbers, string, date and time, messaging, UML relations, network, and others. In addition, the framework supports user defined data types and corresponding attributes. Table 1 includes different categories of attributes and their corresponding data types supported by the framework. It should be noted that the categories and data types listed in Table 1 is not exhaustive. The framework can support more categories and data types.









TABLE 1







Data types supported by Persistent Storage Framework








Category
Data Types





Integers and
8-bit Unsigned Integer (UI8), 8-bit Signed Integer (SI8),


Numbers
UI16, SI16, UI32, SI32, UI64, SI64, 64-bit decimal



(Decimal64)



Vector variants of all of the above



User-configured variants of all of the above



UI32Range


String
String, StringVector, StringVectorVector, StringUC


Date and Time
Date, Time, DateTime



Vector variants of all of the above and



User-configured variants of all of the above


Messaging
ServiceId, ResourceId, ResourceEnum



MessageType, MessagePriority, LocationId,



SerializableObjectType


UML Relations
Association, Aggregation, Composition


Network
Universally Unique Identifier (UUID), World Wide



Name (WWN), IPv4Address, IPv4AddressNetworkMask,



IPv6Address, IPv6AddressNetworkMask, IPvXAddress,



IPvXAddressNetworkMask, Uniform Resource



Identifier (URI), MACAddress, MACAddress2, Host,



SNMPObjectId (Simple Network Management



Protocol (SNMP))



Vector variants of all of the above and



User-configured variants of all of the above


SQL
SQLIn, SQLBetween, SQLComparator, SQLExists


Other
Union, Bool, BoolUC, BoolVector,



SerializableObejct, SerializableObjectVector



ManagedObject, ManagedObjectVector, Enumeration



ObjectId, ObjectIdVector



LargeObject, Map, XML









The framework provides extensive list of built-in data types, as described in conjunction with Table 1. These data types capture the attribute values (e.g., data fields) of objects. In some embodiments, the framework includes one or more attributes that provide run time introspection that allows runtime identification of classes. Since attributes can be serialized to and de-serialized from a variety of formats, the framework provides extensive support for custom behavior overriding in serialization and de-serialization. Furthermore, the framework supports user defined data types.


Object Identifier


In the example in FIG. 1B, framework 120 stores attribute values of different classes in database 150. During operation, a class is instantiated in switch 102 (e.g., in the memory of switch 102), and one or more attributes of that instance are assigned corresponding values. For example, if the class represents a line card switch 102, an attribute can be a MAC address of a port in that line card (e.g., MACAddress data type). When the line card becomes active, an instance of the class, which can be referred to as an object, is created in the memory of switch 102, and framework 120 stores the attribute values of that object in a table associated with the class in database 150.


However, a switch can have a plurality of line cards. For another line card, another object (i.e., another instance) of the class is created in the memory of switch 102, and framework 120 stores the attribute values of that other object in the table associated with the class in database 150. In this way, the same table can store attribute values of different objects of the same class. To identify different objects of a class in the table, framework 120 generates and assigns an object identifier (object ID or OID) to a respective object of a respective class. This object identifier operates as the primary identifier of that object. In the database table, this primary identifier is the primary key of that table. It should be noted that an object identifier is referred to be associated with a class in a generic sense, which indicates an object identifier of an object of the class.



FIG. 2 illustrates an exemplary object identifier generated by a persistent storage framework in a switch, in accordance with an embodiment of the present invention. During operation, an object 200 of a class is created in the memory of a switch. The persistent storage framework of the switch creates an object identifier 210 for object 200. This object identifier 210 can be the primary identifier for object 210 in the persistent storage. If the persistent storage is an object relational database, the database can include a table corresponding to the class. The attribute values of object 200 and object identifier 210 are inserted into the table. Object identifier 210 can be the primary key for that table.


In some embodiments, object identifier includes a class identifier (a class ID or CID) 220 and an instance identifier (an instance ID or IID) 230. Class identifier 220 represents the class from which the object is instantiated. In some embodiments, class identifier 220 is generated based on a hash function (e. g., Rabin Polynomial hash function) applied to the name of the class. Instance identifier 230 represents that particular instance of the object. Hence, if two objects of the same class are created, class identifier 220 of object identifier 210 remains the same for both the objects. However, the two objects differ in their respective instance identifier 230. Typically, class identifier 220 and instance identifier 230 are 32 and 64 bits long, respectively. However, these lengths can vary.


In some embodiments, instance identifier 230 includes a group identifier 232, a location identifier 234, a management module identifier 236, and a serial identifier 238. Group identifier 232 identifies a group in which the switch is a member. For example, if the switch is a member switch of a fabric switch, group identifier 232 can be a fabric switch identifier, which identifies a fabric switch. Location identifier 234 identifies the switch in the group. For example, if the switch is a member switch of a fabric switch, location identifier 234 can be a switch identifier, which identifies the switch within that fabric switch. Typically, group identifier 232 and location identifier 234 are 10 and 20 bits long, respectively.


Management module identifier 236 identifies the type of management module is operating the switch. For example, if the switch is participating in an active-standby high availability protocol (e.g., Virtual Router Redundancy Protocol (VRRP) and Virtual Switch Redundancy Protocol (VSRP)), management module identifier 236 can indicate whether the switch is an active or a standby switch. Typically, management module identifier 236 is 1 bit long. However, length of management module identifier 236 can be increased by incorporating adjacent bits from location identifier 234.


Serial identifier 238 provides identification of a specific instance of an object and can be a wrapped-around monotonically increasing number (e.g., an unsigned integer). Typically, serial identifier 238 is 32 bits long. In this way, object identifier 210 uniquely identifies an object of a class created by a management module in a switch, which can be in a fabric switch. In other words, object identifier 210 can be unique among objects, classes, management modules, fabric switches, and switch locations within a corresponding fabric switch.


Base Classes


In the example in FIG. 1B, persistent storage framework 120 maps classes to object relational tables in database 150, and inserts attribute values of an object of the class into the table. Framework 120 provides a set of base classes from which a class created for an operational unit of switch 102 can be inherited from. These base classes provide a development framework for the operational units and ensure that the operational units of switch 102 remain structured during operation. In this way, framework 120 can facilitate structured persistent storage to the attribute values of the operational units.


The framework supports a set of base classes and multiple inheritance from these base classes. FIG. 3 illustrates exemplary base classes for supporting a persistent storage framework in a switch, in accordance with an embodiment of the present invention. In some embodiments, the most base class 302 is the PersistableObject class. This class outlines the most fundamental operations supported by the persistent storage framework of a switch. Another base class 304, denoted as the ManagedObject class, is virtually derived from the PersistableObject class. Any object instantiated from an inheritance chain of the ManagedObject class can be referred to as a managed object. The framework provides seamless persistent storage support to these managed objects.


Class 304 outlines the most common attributes and operations of the objects managed by the framework. In other words, all class hierarchies derive virtually from the PersistableObject class. Since a class can inherit from multiple classes and each of these classes can inherit from the PersistableObject class, there can potentially be a conflict during execution of a managed object. This problem is generally referred to as the diamond problem. To solve this problem, the framework supports virtual derivation from the PersistableObject class. Another base class 306, denoted as the LocalManagedObjectBase class, outlines the attributes and operations locally managed within a switch. For example, a port is locally managed in a switch.


Base class 308, denoted as the LocalManagedObject class, is virtually derived from the ManagedObject class and the ManagedObjectBase class. Hence, the LocalManagedObject class outlines the attributes and operations of a switch which are locally and globally managed. For example, a port is locally managed within a switch and a VLAN configured for the port is managed globally. In some embodiments, an application (e.g., a protocol plug-in) running on a switch can specify more base classes for that application. Typically, base classes are not directly mapped to the tables of the object relational database. These base classes provide object relational mapping support. The attributes (i.e., the data fields) of these classes become part of a respective managed object derived from these base classes. As a result, the managed objects can share states and behavior.


In some embodiments, the attributes of a managed object can be any of the attribute data types supported by the framework, as described in conjunction with Table 1. The framework also supports vector variants (e.g., arrays and lists) for a number of the data types. Furthermore, the framework provides support to check whether a particular attribute is user configured. As described in conjunction with FIG. 3, the framework supports hierarchical managed objects based on inheritance. The framework also supports weak and strong references to objects. A weak reference does not protect the referenced object from being destroyed (e.g., by a garbage collector), unlike a strong reference, which protects the object from being destroyed.


Object Relational Mapping


In some embodiments, a persistent storage framework of a switch supports, both one-to-one and one-to-many, association, aggregation, and composition UML relationships. Association and aggregation are supported via ObjectID and ObjectIDVector data types, and ObjectIDAssociation and ObjectIDAssociationVector attributes, respectively. On the other hand, composition is supported via ManagedObjectPointer and ManagedObjectPointerVector data types and corresponding attributes. In some embodiments, the framework supports smart pointers and vector to facilitate seamless development.



FIG. 4A illustrates an exemplary UML model of classes of a switch with a persistent storage framework, in accordance with an embodiment of the present invention. In this example, a class 404, denoted as the Node class, represents network nodes, such as a switch or a router. Attributes for the Node class includes a NodeID, which represents an identifier for a node. Since a switch can be a member of a switch group (e.g., a fabric switch), the Node class has a relationship with class 402, denoted as the SwitchGroup class, which represents a group of switches. A switch can be in one such switch group and a switch group aggregates a plurality of switches. Hence, the relationship between the Node class and the SwitchGroup class is a one-to-many aggregation, which is denoted as “isMemberOf.” In this relationship, the SwitchGroup class can be referred to as the container class since a switch group “contains” a switch. On the other hand, the Node class can be referred to as a related class.


Similarly, a switch can include one or more line cards. Hence, the Node class has a relationship with class 406, denoted as the LineCard class, which represents a line card. A line card can be in one switch and a switch includes (i.e., is composed of) a plurality of line cards. Hence, the relationship between the Node class and the LineCard class is a one-to-many composition, which is denoted as “includes.” On the other hand, a switch typically has a power source, which may not be inside of the switch. So, the Node class has a relationship with class 408, denoted as the PowerSource class, which represents a power source of a node. Suppose that, at a time, a power source can power one switch and a switch can receive power from one source. Hence, the relationship between the Node class and the PowerSource class is a one-to-one association, which is denoted as “getsPower.”


A power source can be based on alternating current (AC) or direct current (DC). So, class 408-A, denoted as the ACPowerSource class, and class 408-B, denoted as the DCPowerSource class, are derived from the PowerSource class. The ACPowerSource class and the DCPowerSource class represent AC and DC power sources, respectively. Hence, based on the getsPower association, a Node can get power from a generic power source, an AC power source, or a DC power source. In this UML diagram, since the relationship between the Node class and class 408 is one-to-one, an object of the Node class can have only one of the power sources. In this example, the PowerSource class, the ACPowerSource class, and the DCPowerSource class can be referred to as the inheritance chain of the PowerSource class (class 408).


The framework can receive the UML diagram of FIG. 4A and generate the corresponding classes in a supported programming language (e.g., C++, Java, C#, etc). Furthermore, the framework generates an object relational table for the classes in the model. Furthermore, the framework can generate corresponding auxiliary tables to represent one-to-many relationships, as well as tables for classes in an inheritance chain (e.g., class derivations) and for their corresponding instances (i.e., objects), as described in conjunction with FIGS. 4C and 4D. In some embodiments, the framework receives XML representation of classes and their relationship (e.g., from a user), and generates the corresponding classes and tables. FIG. 4B illustrates an exemplary XML representation of a class corresponding to a switch with a persistent storage framework, in accordance with an embodiment of the present invention. In this example, XML definition 400 represents the Node class (class 404 of the UML model in FIG. 4A). XML definition 400 represents class Node as a ManagedObject with name “Node.”


XML definition 400 includes a respective attribute, such as NodeID, and its type (i.e., data type, as described in conjunction with Table 1). XML definition 400 also includes one-to-one and one-to-many relationships for which the Node class is a container class. For example, a node contains line cards. Hence, XML definition 400 specifies aggregation “includes” as an attribute, its type, and the class to which Node is related. It should be noted that the isMemberOf relationship is not represented in XML definition 400 even though the isMemberOf relationship to the Node class. This is because the container class for the isMemberOf relationship is the SwitchGroup class. Hence, the isMemberOf relationship is represented in an XML definition corresponding to the SwitchGroup class (not shown in FIG. 4B).


Persistent Storage in a Switch


Upon receiving XML definitions associated with the classes of a UML model, the framework creates a respective table for a respective class, their derivations, their instances (i.e., objects), and their one-to-many relationships in an object relational database. FIG. 4C illustrates exemplary tables generated by a persistent storage framework in a switch, in accordance with an embodiment of the present invention. During operation, the persistent storage framework of the switch generates a table 420 for the Node class in an object relational database. Table 420 includes a column 421 for an object identifier associated with the Node class. Column 421 includes two columns 422 and 423 for class identifier and instance identifier, respectively, of the object identifier associated with the Node class.


Table 420 also includes a column for a respective attribute of the Node class. For example, table 420 includes a column 424 for a NodeID of the Node class. Furthermore, since the Node class has a one-to-one association with the PowerSource class, for which the Node class is the container class, the framework includes a column 425 for an object identifier of an object of the PowerSource class (i.e., an object associated with the PowerSource class). Column 425 includes two columns 426 and 427 for the class identifier and instance identifier, respectively, of the object identifier associated with the PowerSource class. The framework also creates a table 410 for the PowerSource class, comprising column 411 for the object identifier associated with the PowerSource class. Column 411 includes two columns 412 and 413 for the class identifier and instance identifier, respectively, of the object identifier of the PowerSource class.


Similarly, the framework also creates a table 440 for the ACPowerSource class, comprising column 441 for an object identifier of an object of the ACPowerSource class (i.e., an object associated with the ACPowerSource class). Column 441 includes two columns 442 and 443 for the class identifier and instance identifier, respectively, of the object identifier associated with the ACPowerSource class. In the same way, the framework also creates a table 450 for the DCPowerSource class, comprising column 451 for an object identifier of an object of the PowerSource class. Column 451 includes two columns 452 and 453 for the class identifier and instance identifier, respectively, of the object identifier associated with the DCPowerSource class.


In some embodiments, the framework creates auxiliary tables to enforce consistency on columns 426 and 427. For example, the framework creates an auxiliary table 430 for the derivations of the PowerSource class (e.g., based on the UML model in FIG. 4A). In this example, table 430 corresponds to the PowerSource, ACPowerSource, and DCPowerSource classes. Table 430 includes a column 431 for the class identifier associated with the derivations of the PowerSource class. Similarly, the framework creates an auxiliary table 460 for the objects instantiated from the derivations of the PowerSource class. In this example, table 460 corresponds to the PowerSource, ACPowerSource, and DCPowerSource classes. Table 460 includes a column 461 for the instance identifiers of the objects instantiated from the derivations of the PowerSource class.


When a class identifier is generated for any class of the inheritance chain of the PowerSource class, that class identifier is inserted into table 430. The framework identifies the PowerSource, the ACPowerSource, and the DCPowerSource classes of the inheritance chain of the PowerSource class from the UML model in FIG. 4A and generates class identifiers 432, 433, and 434, respectively. The framework then inserts class identifiers 432, 433, and 434 into table 430. In this example, an entry in a table is denoted with dotted lines. Column 431 of table 430 provides consistency enforcement to column 426 of table 420 (denoted with a dashed arrow). In some embodiments, consistency enforcement of column 426 is based on a foreign key constraint on column 431 of table 430. On the other hand, when the framework identifies an object of the PowerSource, ACPowerSource, or the DCPowerSource class, the framework generates a corresponding object identifier and inserts the object identifier into table 410, 440, or 450, respectively.


When an object identifier is inserted into table 410, 440, or 450, the instance identifier of the object identifier is concurrently inserted into table 460 (denoted with dotted arrow). Suppose that, upon detecting an object in the memory of the switch, the framework inserts an object identifier comprising a class identifier 432 and instance identifier 435 into table 410. Similarly, an object identifier comprising a class identifier 433 and instance identifier 444, and an object identifier comprising a class identifier 433 and instance identifier 445 are inserted into table 440. An object identifier comprising a class identifier 434 and instance identifier 454 is inserted into table 450. The framework ensures that instance identifiers 435, 444, 445, and 454 are also inserted into table 460. Column 461 of table 460 provides consistency enforcement to column 426 of table 420 (denoted with a dashed arrow). In some embodiments, consistency enforcement to column 427 is based on a foreign key constraint on column 461 of table 460.


During operation, an object of the Node class is instantiated in the memory of the switch. The framework identifies the object in the memory and generates an object identifier for the object comprising a class identifier 464 and an instance identifier 465. The framework identifies the attribute values of the object, which includes NodeID 466 and an object identifier of a power source object. Suppose that the power source for the switch is an AC power source, and the object identifier comprises a class identifier 433 and an instance identifier 444, as stored in table 440 corresponding to the ACPowerSource class. The framework creates an entry in table 420 by inserting class identifier 464, instance identifier 465, NodeID 466, class identifier 433, and instance identifier 444 into table 420. Since consistency is enforced on columns 426 and 427, the relational database ensures that class identifier 433 and instance identifier 444 appear in columns 431 and 461, respectively.


However, even though the Node class is related to the LineCard class, since it is a one-to-many relationship, table 420 does not include an object identifier associated with the LineCard class. The framework creates an auxiliary table to represent the relationship the Node class and the LineCard class. FIG. 4D illustrates an exemplary table representing a one-to-many association, which is generated by a persistent storage framework in a switch, in accordance with an embodiment of the present invention. Upon detecting the LineCard class in the UML model in FIG. 4A, the persistent storage framework of the switch generates a table 470 for the LineCard class in an object relational database. Table 470 includes a column 471 for an object identifier associated with the LineCard class. Column 471 includes two columns 472 and 473 for corresponding class identifier and instance identifier, respectively, of the object identifier associated with the LineCard class.


During operation, an object of the LineCard class is instantiated in the memory of the switch. The framework identifies the object in memory and generates an object identifier comprising a class identifier 474 and an instance identifier 475 for the object. The framework then creates an entry in table 470 by inserting the object identifier into table 470. Similarly, the framework generates an object identifier comprising a class identifier 474 and an instance identifier 476 for another object of the LineCard class, and a third object identifier comprising a class identifier 474 and an instance identifier 477 for an object of the LineCard class. The framework creates respective entries in table 470 by inserting these object identifiers into table 470.


In some embodiments, the framework creates an auxiliary table 480 to represent the one-to-many “includes” relationship between the Node class and the LineCard class. In the relationship, the Node class is the container class and the LineCard class is the related class. Table 480 includes a column 481 for an object identifier associated with the Node class, and a column 484 for an object identifier associated with the LineCard class. Column 481 includes two columns 482 and 483 for the class identifier and instance identifier, respectively, of the object identifier associated with the Node class. Similarly, column 484 includes two columns 485 and 486 for the class identifier and instance identifier, respectively, of the object identifier associated with the LineCard class.


Suppose that the object of the Node class, which is associated with class identifier 464 and instance identifier 465, includes two line cards. Hence, the object of the Node class include two objects (e.g., an ManagedObjectVector) of the LineCard class. Suppose that instance identifiers 475 and 476 belong to these two objects. As a result, the framework inserts class identifier 464, instance identifier 465, class identifier 474, and instance identifier 475 into table 480. The framework also inserts class identifier 464, instance identifier 465, class identifier 474, and instance identifier 476 into table 480. In this way, the relationship between the object of the Node class (associated with instance identifier 465) and two objects of the LineCard class (associated with instance identifier 475 and 476) is stored in table 480.


In some embodiments, similar to tables 430 and 460, the framework creates auxiliary table 490 for the derivations of the Node class (e.g., based on the UML model in FIG. 4A). In this example, table 490 corresponds to the Node class (and its derivations, if any). Table 490 includes a column 491 for the class identifier associated with the derivations of the Node class. Similarly, the framework creates an auxiliary table 492 for the objects instantiated from the derivations of the Node class. In this example, table 492 corresponds to the Node class (and its derivations, if any). Table 492 includes a column 493 for the instance identifiers of the objects instantiated from the derivations of the Node class.


In the same way, the framework creates auxiliary table 495 for the derivations of the LineCard class (and its derivations, if any). Table 495 includes a column 496 for the class identifier associated with the derivations of the LineCard class. Similarly, the framework creates an auxiliary table 497 for the objects instantiated from the derivations of the LineCard class. In this example, table 497 corresponds to the LineCard class (and its derivations, if any). Table 497 includes a column 498 for the instance identifiers of the objects instantiated from the derivations of the LineCard class.


When a class identifier is generated for the Node class or the LineCard class, that class identifier is inserted into table 490 or 495, respectively. The framework inserts class identifiers 464 and 474 associated with the Node and the LineCard classes, respectively, into tables 490 and 495, respectively. In this example, an entry in a table is denoted with dotted lines. Column 491 of table 490 provides consistency enforcement to column 482 of table 480 (denoted with a dashed arrow). In some embodiments, consistency enforcement of column 482 is based on a foreign key constraint on column 491 of table 490. In the same way, column 496 of table 495 provides consistency enforcement to column 485 of table 480 (denoted with a dashed arrow). In some embodiments, consistency enforcement of column 485 is based on a foreign key constraint on column 496 of table 495.


On the other hand, when the framework identifies objects of the Node or the LineCard class, the framework generates a corresponding object identifier and inserts the object identifier, comprising a class identifier and an instance identifier, into table 420 or 470, respectively. When an object identifier is inserted into table 420 or 470, the instance identifier of the object identifier is concurrently inserted into table 492 or 497, respectively (denoted with dotted arrow). For example, when the framework inserts an object identifier comprising a class identifier 464 and instance identifier 465 into table 420, instance identifier 465 is inserted into table 492. In the same way, when the framework inserts an object identifier comprising a class identifier 474 and instance identifier 475 into table 470, instance identifier 475 is inserted into table 497.


Similar to table 480, the framework also creates an auxiliary table to represent the one-to-many “isMemberOf” relationship between the Node class and the SwitchGroup class, as described in conjunction with FIG. 4A. That table includes a column for an object identifier associated with the container class, which is the SwitchGroup class, and a column for an object identifier associated with the related class, which is the Node class. The column for the object identifier associated with the SwitchGroup class includes two columns corresponding to class identifier and instance identifier, respectively, of the object identifier. Similarly, the column for the object identifier associated with the Node class includes two columns corresponding to class identifier and instance identifier, respectively, of the object identifier.


It should be noted that the framework distinguishes between a composition relationship (e.g., “includes” in FIG. 4A) and an aggregation relation (e.g., “isMemberOf” in FIG. 4A). In some embodiments, for a composition relationship, the class definition of the container class includes an attribute of data type ManagedObject (and/or ManagedObjectPointer), as described in conjunction with Table 1. In this disclosure, the phrases “class” and “class definition” are used interchangeably. If the relationship is one-to-many, the date type can be ManagedObjectVector (and/or ManagedObjectPointerVector). In this way, when an object of the container class is instantiated, the related objects are created and included in that instantiated object of the container class. On the other hand, for an aggregation relationship, the class definition of the container class includes an attribute of data type ObjectId. If the relationship is one-to-many, the date type can be ObjectIdVector. In this way, the objects are created separately, and when an object of the container class is instantiated, only references to those related objects are included in that instantiated object of the container class.


Operations of a Persistent Storage Framework



FIG. 5A presents a flowchart illustrating the process of a persistent storage framework of a switch generating auxiliary tables for an inheritance chain in a structured persistent storage, in accordance with an embodiment of the present invention. During operation, the framework identifies a respective class of a non-base class inheritance chain (operation 502). The framework generates a respective class identifier for a respective identified class (operation 504). The framework generates an auxiliary table for the classes of the inheritance chain comprising a column for the class identifiers of the inheritance chain (operation 506) and updates the table for the classes of the inheritance chain by inserting the generated class identifiers (operation 508). The framework also generates an auxiliary table for the objects (i.e., the instantiated objects) of the classes of the inheritance chain, each comprising a column corresponding to the instance identifiers associated with the classes of the inheritance chain (operation 510).



FIG. 5B presents a flowchart illustrating the process of a persistent storage framework of a switch generating class tables in a structured persistent storage, in accordance with an embodiment of the present invention. During operation, the framework identifies a non-base class and generates a class table for the identified class (operation 532). In some embodiments, the framework identifies the class, and the attributes and operations (e.g., data members and methods) of the class from a class model (e.g., a UML model). The framework can receive the UML model from a graphical or textual input (e.g., a GUI, CLI, or XML file). In some embodiments, the table is named based on a hash function (e. g., a Rabin Polynomial hash function) calculated on the name of the class. The table can also have a prefix “T.” For example, if the name of the class is Node and hash(“Node”)=xxx, the table name can be Txxx. The framework adds a column comprising columns for a class identifier and an instance identifier to the class table for an object identifier (operation 534), as described in conjunction with FIG. 4C.


The framework identifies an attribute of the identified class (operation 536). It should be noted that the relationships for which the class is a container class are can also be attributes, as described in conjunction with FIG. 4A. The framework then checks whether the attribute is a simple attribute (e.g., not a relationship) (operation 538). If the attribute is a simple attribute, the framework adds a column for the identified attribute to the class table (operation 540). If the attribute is not a simple attribute (e.g., an attribute representing a relationship), the framework checks whether the attribute corresponds to a one-to-one relationship (operation 544). If the attribute corresponds to a one-to-one relationship, the framework adds a column, which is for an object identifier, comprising columns for class identifier and instance identifier of the object identifier (operation 546), as described in conjunction with FIG. 4C.


The framework enforces consistency on the class identifier and the instance identifier based on the corresponding auxiliary tables of the related classes (operation 548), as described in conjunction with FIG. 4C. In some embodiments, the consistency is enforced based on a foreign key constraint. If the attribute does not correspond to a one-to-one relationship (i.e., corresponds to a one-to-many relationship), the framework generates an auxiliary table for the one-to-many relationship (operation 550) and enforce consistency on object identifiers in the auxiliary table for the one-to-many relationship (operation 552). Upon adding a column for the identified attribute (operation 540), enforcing consistency on the class identifier and the instance identifier (operation 548), or enforcing consistency on the object identifier (operation 552), the framework checks whether all attributes have been checked (operation 542). If not, the framework continues to identify an attribute of the identified class (operation 536).



FIG. 5C presents a flowchart illustrating the process of a persistent storage framework of a switch generating an auxiliary table representing an one-to-many relationship in a structured persistent storage, in accordance with an embodiment of the present invention. Operations described in FIG. 5C elaborates operation 550 of FIG. 5B. During operation, the framework generates an auxiliary table for the one-to-many relationship (operation 562). In some embodiments, the name of the auxiliary table is based on the container table name, related table name, and the relationship name. For example, if the container table name is Txxx, related table name is Tyyy, and the relationship name is ABC, the name of the auxiliary table can be TxxxABCTyyy.


The framework adds a column for an object identifier comprising columns for class identifier and instance identifier of the container class (operation 564), as described in conjunction with FIG. 4D. The framework enforces consistency on the object identifier (i.e., both the class identifier and the instance identifier) of the container class based on the corresponding columns of the container class table (operation 566). Similarly, the framework adds a column for an object identifier comprising columns for class identifier and instance identifier of the related class (operation 568), as described in conjunction with FIG. 4D. The framework enforces consistency on the object identifier (i.e., both the class identifier and the instance identifier) of the related class based on the corresponding columns of the related class table (operation 570).



FIG. 5D presents a flowchart illustrating the process of a persistent storage framework of a switch updating tables in a persistent storage, in accordance with an embodiment of the present invention. During operation, the framework monitors the memory of the switch for object generation of the inheritance chain (operation 582) and checks whether a new object has been detected (operation 584). If a new object has not been detected, the framework continues to monitor the memory of the switch (operation 582). If a new object has been detected, the framework generates an object identifier comprising a class identifier and an instance identifier for the new object (operation 516). The frame creates an entry comprising the object identifier in the table of a class associated with the object (i.e., the class from which the object has been instantiated) (operation 588). The framework creates an entry comprising the class identifier, instance identifier, or both in corresponding auxiliary tables associated with the object (operation 590) and continues to monitor the memory of the switch (operation 582).


Distributed Configuration Management


In the example in FIG. 1A, network 100 can be a switch group (e.g., an Ethernet fabric switch). Under such a scenario, network 100 can also be referred to as switch group 100. In some embodiments, a switch group is a fabric switch. FIG. 6A illustrates an exemplary switch group with persistent storage framework and distributed configuration management support, in accordance with an embodiment of the present invention. In this example, switch 102, which is in switch group 100, can be configured by logging in to switch 102 using a management IP address (MIPA) 632 via a management port 602. Management port 602 can be an edge port, which is used for communicating configuration and control traffic. Switch 102 may not use management port 602 for data traffic. A user (e.g., an administrator) can use an end device 612 to log in to switch 102 based on management IP address 632 (e.g., using telnet via network 620). Switch 102 can also be configured via a console port (e.g., an RS-232 port) 604, which couples an end device 614 (e.g., a workstation). The user can access an input interface (e.g., a textual or graphical interface) of switch 102 from end device 612 or 614.


With existing technologies, a respective switch in switch group 100 is configured individually and the configuration process is repeated for a respective switch in switch group 100. Such configuration can be related to network virtualizations, partitions, and switch groups, and a plurality of network protocols of different network layers. The attribute values (e.g., parameters) of the configuration are applied to a respective switch, such as switch 102 (e.g., loaded to the switch modules, such as the processing hardware of switch 102). This process is repeated for switches 101, 103, 104, and 105. This process can be repetitive, tedious, and error-prone (e.g., human errors, such spelling mistake).


Furthermore, some of the configuration information is applicable for entire switch group 100 (e.g., a VLAN for switch group 100). Such configuration can be referred to as global configuration. This global configuration is repeated for a respective switch in switch group 100. This can lead to inconsistency among the global configuration in different switches in switch group 100. For example, the user repeats global configuration for switch 101, 102, 103, 104, and 105 in switch group 100. The user may, mistakenly, configure VLAN 110 in switch 102 and VLAN 111 in switch 101. As a result, traffic belonging to VLAN 110 may be erroneously dropped at switch 101.


To solve this problem, one of the switches in switch group 100 is elected as a principal switch, which manages and distributes configuration information in switch group 100. Suppose that switch 102 is elected as the principal switch. In some embodiments, a respective switch in switch group 100 broadcasts its information in switch group 100, and the switch with the lowest (or highest) identifier value is elected as the principal switch. Other switches of switch group 100, such as switches 101, 103, 104, and 105, can be referred to as managed switches. When the user accesses switch 102 (e.g., via end device 112 or 114), switch 102 allows the user to configure the global configuration of switch group 100. Switch 102, in turn, provides that global configuration to switches 101, 103, 104, and 105. Furthermore, switch 102 also allows the user to configure switches 101, 103, 104, and 105 in the switch group. Switch 102, in turn, provides the corresponding local configuration information to switches 101, 103, 104, and 105, respectively.


In some embodiments, a respective switch of switch group 100 is equipped with a persistent storage framework, as described in conjunction with FIG. 1B. The framework stores the configuration information, comprising both local and global configuration information, in a local persistent storage, which can be an object relational database. For example, in switch 102, framework 120 stores the configuration information in object relational database 150. The configuration information is loaded from database 150 to the switch modules (e.g., the ASIC chips of switch 102). Framework 120 maintains the global configuration of switch group 100 in database 150. Framework 120 also maintains respective local configuration of a respective switch, including switch 102, of switch group 100 in database 150. Switch 102 then provides a copy of database 150 to switches 101, 103, 104, and 105, each of which in turn, stores that copy in its object relational database. In this way, a respective switch in switch group 100 has the same local and global configuration. This allows a user to access the configuration of switch group 100 from any member switch.


Since switch 102 can also be accessed via console port 604 via end device 614, the user can configure switches of switch group 100 without the presence of a management network. It should be noted that, even though a user can configure the switches of switch group 100 from principal switch 102, the user can obtain the local configuration information of a respective switch of switch group 100 via any of the member switches of the switch group. For example, switch 101 can be accessed by logging in to switch 101 using a management IP address 631 via a management port 606. The user can use an end device 616 to log in to switch 101 based on management IP address 631. Since switch 101 includes the configuration information in object relational database 640 of switch 101, the user can obtain the global configuration information as well the local configuration information of a respective switch of switch group 100 from switch 101.


In some embodiments, switch group 100 is associated with a virtual IP address 610. During operation, virtual IP address 610 is associated with the principal switch of switch group 100. As a result, the user can access principal switch of group 100 via virtual IP address 610 without knowing the management IP address 632. Furthermore, if switch 102 becomes unavailable (e.g., due to a failure), another switch in switch group 100, such as switch 101, is elected as the principal switch and virtual IP address 610 becomes associated with switch 101. Since switch 101 has the configuration information in database 640 is the same as the configuration information in database 150, switch 101 can continue to facilitate distributed configuration for the switches of switch group 100. As a result, the user can continue to access the principal switch of switch group 100, which is switch 101, via virtual IP address 610 and configure the switches of switch group 100.


In some embodiments, a respective switch in switch group 100 is associated with an internal control IP address (ICIPA), which is used for communication among member switches of switch group 100. For example, an internal control IP address can be used for communication via an inter-switch port. When the principal switch of a switch group provides the configuration information to a managed switch, the principal switch communicates with that managed switch based on respective internal control IP addresses. Suppose that internal control IP addresses 621 and 622 are associated with switches 101 and 102, respectively. When switch 102 provides the configuration information to switch 101, switch 102 uses IP addresses 621 and 622 as the destination and source addresses for the packets comprising the configuration information.


In some embodiments, a respective switch in a switch group can receive configuration information from a user. FIG. 6B illustrates an exemplary switch group with multi-input and multi-principal switch support, in accordance with an embodiment of the present invention. In this example, switches 101, 102, 103, 104, and 105 are associated with management IP addresses 631, 632, 633, 634, and 635, respectively. A user can provide configuration information to a respective switch (e.g., via a console of the switch) in switch group 100 by logging in to that switch using the corresponding management IP address via a management port. For example, a user can log in to switch 103 using management IP address 633 and provides configuration information 664. Configuration information 664 can be global configuration information associated with switch group 100 or local configuration information associated with any of the member switches. Similarly, the user can log in to switch 101 using management IP address 633 and provides configuration information 662.


Suppose that configuration information 662 and 664 can cause a conflict in the switch. For example, configuration information 662 and 664 can be associated with conflicting quality of service settings for switch 105. If both switches 101 and 103 are allowed to operate as a principal switch, switches 101 and 103 can individually apply these conflicting configurations to switch 105. This can lead to race condition in switch group 100. To solve this problem, a switch group internally has one principal switch, and other switches forward their received configuration information to that principal switch. That principal switch receives the configuration information, serializes them, resolves any conflict, and applies them based on distributed commit. In this way, it would appear to the user that the switch group is being configured from any of the member switches, though internally, the switch group is configured from the principal switch.


For example, switches 101 and 103 forward configuration information 662 and 664, respectively, to principal switch 102 of switch group 100. Switch 102 receives configuration information 662 and 664, and serializes them (e.g., based on the order of arrival or an associated priority). In some embodiments, if switch 102 detects any conflict, switch 102 resolves the conflict based on its serialization. For example, configuration information 664 is serialized before configuration information 662, switch 102 resolves the conflict by applying the configuration of configuration information 664 to switch 105 based on distributed commit, and discarding configuration information 662. In this way, it would appear to the user that switch 105 is being configured switches 101 and/or 103, though internally, switch 102 configures switch 105.


In some embodiments, in switch group 100, different member switch can be the principal switch for different services. A service can be a modular component which manages a related set of configuration. In the example in FIG. 1B, a plug-in can correspond to a service. A principal switch manages all global and local configuration information of the service the switch is associated with. Suppose that switches 101 and 102 are principal switches for port management and network virtualization, respectively. As a result, in switch group 100, switch 101 processes all global and local configuration information associated with port management, and switch 102 processes all global and local configuration information associated with network virtualizations. In some embodiments, switches in switch group 100 elect a principal switch for a respective service. The election process can be based on load balancing such that different services are associated with different principal switches.


During operation, a user logs in to switch 102 using management IP address 632 and provides configuration information 666, which is associated with port management (e.g., creating a port channel). Even though switch 102 is a principal switch, switch 102 forwards configuration information 666 to switch 101 since switch 101 is the principal switch for port management. On the other hand, suppose that configuration information 662 is associated with network virtualization (e.g., creating a new VLAN). Even though switch 101 is a principal switch, switch 101 forwards configuration information 662 to switch 102 since switch 102 is the principal switch for network virtualization.



FIG. 7 illustrates an exemplary distributed configuration management, in accordance with an embodiment of the present invention. As described in conjunction with FIG. 6A, suppose that switch 102 has been elected the principal switch of switch group 100. Switch 102 manages configuration information of a respective managed switch in switch group 100. In the example in FIG. 7, switch 102 manages configuration information of a managed switch 101. Database 150 of switch 102 includes global configuration information 701 and local configuration information 702. Database 640 of switch 101 includes global configuration information 703 and local configuration information 704.


Suppose that switch 101 is joining switch group 100 as a member switch (e.g., during the formation of switch group 100). During operation, switch 102 pauses the configuration updates to database 150 (operation 711) and stores the new updates in buffer 710. This provides consistency during the joining process. Switch 101 sends a message comprising global configuration information 703 to switch 102 (operation 712). Communication between switches 101 and 102 can be based on internal control IP addresses 621 and 622. Switch 102 receives the request and compares the local and received global configurations (operation 713). If the configuration matches, switch 102 determines that switch 101 can join switch group 100. It should be noted that a user configuring switch 102 via a console (e.g., console 650) may not be aware of this pause and continue to provide configuration information, which in turn is stored in buffer 710.


Switch 102 instructs switch 101 to clear its persistent storage (i.e., database 640) (operation 714). Switch 101 creates a local backup 709 of local configuration information 704 and removes configuration information 703 and 704 from database 640 (denoted with a “cross” sign) (operation 715). Switch 102 then sends an image of database 150 to switch 101 (operation 716). Upon receiving the image, switch 101 updates database 640 with the received image (operation 717). At this point, databases 150 and 640 have the same information, which includes configuration information 701 and 702. Switch 101 acknowledges the update of database 640 (e.g., by sending an update confirmation message) to switch 102 (operation 718). From this point on, configuration information in database 150 should match the configuration information in database 640. Hence, switch 102 initiates a distributed commit (operation 719) (i.e., uses the distributed commit to update configuration information, as described in conjunction with FIGS. 11A-11B).


Switch 102 resumes updates to database 150 (operation 720) by incorporating the update(s) stored in buffer 710. Switch 102 performs distributed commit(s) on the update(s) in buffer 710 (operation 721). To incorporate local configuration information 704 in the switch group, switch 101 sends the backed-up local configuration information 704 to switch 102 (operation 722). Upon receiving local configuration information 704, switch 102 performs distributed commit on local configuration information 704 (operation 723). During this distributed commit process, switch 102 includes local configuration information 704 in its local database 150 as well as in database 640 of switch 101.


At this point, databases 150 and 640 have the same configuration information. In this way, a principal switch forms a switch group and ensures a respective member switch of the switch group has the same configuration information in its local persistent storage. In some embodiments, global configuration information 701 is only active in database 150 of principal switch 102, and in a standby mode in database 640 of managed switch 101. Under such a scenario, attribute values of global configuration information 701 is only extracted from database 150 and applied in switch group 100. Global configuration information 701 becomes active in database 640 if switch 102 becomes unavailable and switch 101 is elected the principal switch of switch group 100.


A new switch can join switch group 100 after it has converged. As described in conjunction with FIG. 7, the new switch backs up its local configuration information, clears its local persistent storage, and receives the database image of database 150 from principal switch 102. The new switch then updates its local persistent storage with the database image of database 150. When switch group 100 converges, the database image of database 150 comprises the global configuration of switch group 100 and the local configuration of a respective member switch. As a result, the configuration information of switch group 100 becomes readily available to the new switch. Similar to operations 720 and 721, principal switch 102 receives the local configurations information of the new switch and performs a distributed commit to the local configurations information, which then becomes stored in the persistent storage of a respective member switch. The new switch can self-configure using the received information.


A switch can rejoin switch group 100 after it has converged. For example, switch 101 can leave (e.g., gets disconnected) and rejoin switch group 100. If its configuration information has not changed, switch 101 simply receives the database image of database 150, which comprises the global configuration of switch group 100 and the local configuration of a respective member switch, from switch 102 and self-configures using the received information. However, if the global configuration information of rejoined switch 101 has changed, switch 102 may not allow switch 101 to rejoin, as described in operation 713. On the other hand, if the local configuration information of rejoined switch 101 has changed, switch 102 receives the changed local configuration information and performs a distributed commit on the changed local configuration information. In this way, the local configuration information of rejoined switch 101 is updated in the persistent storage of a respective member switch, including databases 150 and 640. Hence, the local configuration information of a rejoined switch has higher priority over existing local configuration information of that switch.


A switch can be removed from switch group 100 even after it has converged. Under such a scenario, principal switch 102 performs a distributed commit to remove the local configuration information of the removed switch. For example, if switch 102 leaves switch group 100, principal switch 102 performs a distributed commit to remove local configuration information 704 from the persistent storage of a respective member switch.


Multi-Fabric Manager



FIG. 8A illustrates an exemplary multi-fabric manager managing a plurality of switch groups, in accordance with an embodiment of the present invention. In this example, switch groups 100, 810, and 820 are coupled to each other via a network 802 via one or more links. Examples of network 802 include, but are not limited to, a LAN, a wide area network (WAN), and the Internet. Switch group 810 includes member switches 812, 814, and 816, and switch group 820 includes member switches 822, 824, 826, and 828. Suppose that switches 812 and 822 are the principal switches of switch groups 810 and 820, respectively. Persistent storage 832 in switch 812 stores the global and local configuration information associated with switch group 810, and persistent storage 834 in switch 822 stores the global and local configuration information associated with switch group 820. Switch groups 810, 820, and 100 can be at different locations. In some embodiments, switch groups 810 and 820 are fabric switches. Switch group 810 or 820 can be a TRILL network and a respective switch of switch group 810 or 820 can be an RBridge. Switch group 810 or 820 can also be an IP network and a respective switch of switch group 810 or 820 can be a layer-3 device (e.g., an IP-capable switch or a router).


With existing technologies, switch groups 100, 810, and 820 are managed individually. For example, a user can manage switch group 810 based on the configuration information in persistent storage 832 in switch 812. Similarly, a user can manage switch group 100 and 820 based on the configuration information in persistent storages 150 and 834 in switches 102 and 822, respectively. However, persistent storage 832 only includes configuration information of switch group 810. Similarly, persistent storage 834 only includes configuration information of switch group 820 and persistent storage 150 only includes configuration information of switch group 100. As a result, if a tenant needs a tenant network that spans switch groups 810 and 100, managing such a tenant network requires individual configuration of switch groups 810 and 100. Additionally, a respective configuration change to that tenant network leads to individual configurations of switch groups 810 and 100, which can be tedious and error-prone.


To solve this problem, a multi-fabric manager 804 is associated with switch groups 100, 810, and 820. Multi-fabric manager 804 facilitates a single point of management for switch groups 100, 810, and 820. Multi-fabric manager 804 can be any physical or virtual device (e.g., a virtual machine). Multi-fabric manager 804 can be a standalone switch or member of a switch group. Multi-fabric manager 804 can be coupled to a switch group via a WAN (e.g., network 802) or be coupled to one or more switches of a switch group (denoted with dotted lines). In some embodiments, multi-fabric manager 804 includes a persistent storage 836, which includes configuration information of multi-fabric manager 804. Such configuration information includes identifiers of the switch groups multi-fabric manager 804 is managing (e.g., switch groups 100, 810, and 820), tenant networks spanning multiple switch groups, and the global VLAN (GVLAN) mappings.


Multi-fabric manager 804 includes a persistent storage framework 840, which provides corresponding configuration information to switch groups 100, 810, and 820. Furthermore, a respective entry of persistent storage 150, 832, and 834 can be associated with an object identifier comprising a group identifier, as described in conjunction with FIG. 2. Multi-fabric manager 804 can use this group identifier portion (e.g., group identifier 232 in FIG. 2) of the object identifier (e.g., object identifier 210 in FIG. 2) to distinguish among the configuration information associated with switch groups 100, 810, and 820. For example, upon retrieving an entry from persistent storage 150, multi-fabric manager 804 can determine that the attribute values stored in the entry is for switch group 100 based on the group identifier in the object identifier of the entry. In other words, multi-fabric manager 804 can distinguish among attribute values, which are associated with different objects of different switch groups, in persistent storage 150, 832, and 834 based on the corresponding object identifiers.


During operation, multi-fabric manager 804 becomes a member switch of switch groups 100, 810, and 820. For example, multi-fabric manager 804 sends a join request to switch group 100 to join as a member switch. Principal switch 102 of switch group 100 receives the request and processes the request to allow multi-fabric manager 804 to join switch group 100 as a member switch. Joining a switch group, such as a fabric switch, is described in U.S. patent application Ser. No. 13/092,724, titled “Fabric formation for virtual cluster switching,” by inventors Shiv Haris and Phanidhar Koganti, filed 22 Apr. 2011, the disclosure of which is incorporated by reference herein.


When multi-fabric manager 804 becomes a member switch of switch group 100, persistent storage framework 120 in principal switch 102 synchronizes the configuration information in persistent storage 150 with persistent storage framework 840 in multi-fabric manager 804, as described in conjunction with FIG. 7. In this way, multi-fabric manager 804 obtains and maintains an instance of persistent storage 150. Furthermore, since a respective switch in a switch group shares the learned MAC addresses with all other member switches, multi-fabric manager 804 receives the MAC addresses learned by the switches of switch group 100. Multi-fabric manager 804 stores these MAC addresses in a local table in association with a switch group identifier (e.g., a fabric identifier), which identifies switch group 100.


In the same way, multi-fabric manager 804 joins switch groups 810 and 820, and receives respective instances of persistent storages 832 and 834. Multi-fabric manager 804 receives the learned MAC addresses of switch groups 810 and 820, and stores them in association with switch group identifiers of switch groups 810 and 820, respectively. Since multi-fabric manager 804 operates as a member switch of switch groups 100, 810, and 820, a tenant network spanning multiple switch groups (e.g., switch groups 100 and 810) can be created from multi-fabric manager 804, as described in conjunction with FIG. 6B. During operation, a user may configure the tenant network in switch groups 100 and 810 from multi-fabric manager 804. For example, the user can create respective GVLANs in switch groups 100 and 810.


Upon receiving the configuration information, multi-fabric manager 804 provides (e.g., sends via a control message) the configuration information associated with switch group 100 to principal switch 102 and the configuration information associated with switch group 810 to principal switch 812. In switch group 100, principal switch 102 receives the configuration information and performs a distributed commit in switch group 100. The distributed commit includes updating the persistent storage with the attribute values of the configuration information and applying the attribute values to the switch modules of a respective member switch in switch group 100.


Since multi-fabric manager 804 operates as a member switch of switch group 100, the instance of persistent storage 150 of switch group 100 is also updated in multi-fabric manager 804. Similarly, principal switch 812 of switch group 810 receives the configuration information and performs a distributed commit in switch group 810. This also updates the instance of persistent storage 832 of switch group 810 in multi-fabric manager 804. In this way, configuration information from multi-fabric manager 804 is propagated to switch groups 100 and 810, and creates the tenant network from a single point of management. However, if multi-fabric manager 804 does not participate in frame forwarding for a switch group, multi-fabric manager 804 may not apply the configuration information of that switch group in the local device modules (e.g., local ASIC chips).


In some embodiments, multi-fabric manager 804 can be replicated a plurality of times. For example, multi-fabric manager 804 can be replicated at multiple physical devices (e.g., on different servers or different disks), which can be at multiple physical locations. Multi-fabric manager 804 can also be replicated at multiple virtual locations (e.g., in different virtual machines or disk partitions). One of the replicas can operate as the principal multi-fabric manager, and the other replicas operate as managed multi-fabric manager. Any of the replica can receive configuration information from a user and the principal multi-fabric manager applies the received configuration information to the corresponding switch group, as described in conjunction with FIG. 6B. The principal multi-fabric manager performs a distributed commit to ensure a respective replica of the multi-fabric manager is synchronized and has the same copy of the persistent storage. In some embodiments, a respective replica of multi-fabric manager 804 can join the switch groups it is managing (e.g., switch groups 100, 810, and 820), and its instance of the persistent storage of a switch group is synchronized based on a distributed commit from the principal switch of the switch group.


If the principal multi-fabric manager becomes unavailable, another replica of multi-fabric manager 804 becomes the principal multi-fabric manager. If multi-fabric manager 804 is unavailable for a distributed commit from the principal switch of a switch group, such as switch 102, persistent storage framework 120 maintains the set of transactions for which multi-fabric manager 804 has not been updated. If that set of transactions is smaller than a threshold value, which can be configured by a user or automatically generated, switch 102 updates the instance of persistent storage 150 in multi-fabric manager 804 with that set of transactions. Otherwise, switch 102 considers multi-fabric manager 804 to be a rejoining member switch and updates the instance of persistent storage 150 in multi-fabric manager 804 accordingly, as described in conjunction with FIG. 7.



FIG. 8B illustrates an exemplary console of a multi-fabric manager, in accordance with an embodiment of the present invention. In this example, multi-fabric manager 804 has a console 850. Upon accessing, multi-fabric manager 804 presents console 850 to the user. In some embodiments, when the user accesses multi-fabric manager 804, console 850 provides a command line interface shell 852 to the user. The user can type commands to shell 852. Shell 852 can be the initial screen which appears when the user accesses multi-fabric manager 804. The user can provide a command to shell 852 to gain access to switch group 100 (e.g., “switchgroup_FID100,” wherein “FID100” can be a switch group identifier, such as a fabric identifier, which identifies switch group 100). The user then provides another command (e.g., “config terminal”) to shell 852 to gain access to a configuration terminal for switch group 100 in shell 852.


The user can use this configuration terminal to provide global configuration associated with switch group 100 and local configuration associated with any switch in switch group 100. For example, the user can issue a command to the terminal to create a VLAN 110. This VLAN 110 is created across switch group 100, and hence, is part of the global configuration. On the other hand, user can also issue a command to configure a specific port of a specific switch in switch group 100. This port configuration is a local configuration for that switch. In some embodiments, a port is identified by a port identifier, which can be in a “switch identifier/chassis number (e.g., line card number)/port number” format. For example, the user can configure a 10 Gigabit Ethernet port identified by port identifier “1/2/16.” If switch 101 is associated with identifier 1, the port identifier indicates that the port is port number 16 in line card number 2 of switch 101. The user can add VLAN 110 to that specific port. This VLAN configuration of the port is a local configuration of switch 101.


Similarly, the user can provide a command to shell 852 to gain access to switch group 810 (e.g., “switchgroup_FID810,” wherein “FID810” can be a switch group identifier, such as a fabric identifier, which identifies switch group 810). The user then provides another command (e.g., “config terminal”) to shell 852 to gain access to a configuration terminal for switch group 810 in shell 852. The user can issue a command to the terminal to create a VLAN 110. This VLAN 110 is created across switch group 810, and hence, is part of the global configuration. The user can also configure a 10 Gigabit Ethernet port identified by port identifier “2/0/7.” A chassis number 0 can indicate that the switch has a “pizza box” form factor, often implemented in a single rack. If switch 814 is associated with identifier 2, the port identifier indicates that the port is port number 7 of switch 814. The user can add VLAN 110 to that specific port. This VLAN configuration of the port is a local configuration of switch 814.



FIG. 8C illustrates an exemplary persistent storage framework of a multi-fabric manager managing a switch group, in accordance with an embodiment of the present invention. In this example, persistent storage framework 840 of multi-fabric manager 804 provides structured persistent storage to a respective switch group it manages (e.g., switch group 100). Similar to a member switch of a switch group, multi-fabric manager 804 runs one or more protocol daemons. For example, multi-fabric manager 804 can run respective protocol daemons for Ethernet, TRILL, FC, and IP. A protocol daemon facilitates the services and operations of a corresponding protocol for multi-fabric manager 804. Framework 840 includes a core management module 860, which facilitates structured persistent storage to the attribute values associated with different operational units of a switch in a switch group (e.g., modules and services of switch 102). Core management module 860 facilitates inter-operations between plug-ins 864 and the protocol daemons of multi-fabric manager 804.


Multi-fabric manager 804 further includes an input interface 872 to multi-fabric manager 804 (e.g., a graphical user interface (GUI) and/or a command line interface (CLI)). A user can access input interface 872 via a display device (e.g., a monitor or projector) and using an input device (e.g., a keyboard and/or a pointing device). The user can obtain information from and provide instruction to multi-fabric manager 804 via input interface 872. Multi-fabric manager 804 also includes a configuration daemon 874, which can receive configuration information (e.g., of the switch groups it is associated with) for multi-fabric manager 804 via input interface 872. Configuration daemon 874 provides this configuration information to framework 840. Framework 840 can include a configuration daemon gateway module 862, which communicates with configuration daemon 874. Upon receiving the configuration information, framework 840 can identify different attribute values (e.g., a VLAN identifier) and assigns those attribute values to the corresponding attribute of an operational unit of a switch in a switch group.


Persistent storage framework 120 of switch 102 includes configuration daemon gateway module 132, which provides configuration information to plug-ins 134 for the protocol daemons of switch 102, as described in conjunction with FIG. 1B. On the other hand, configuration daemon gateway module 862 of persistent storage framework 840 can update both plug-ins 864 and the plug-ins of the persistent storage framework at a respective principal switch (e.g., plug-ins 134 of switch 102). For example, as described in conjunction with FIG. 8B, multi-fabric manager 804 can receive configuration information for the protocol daemons of a switch, such as switch 102, in switch group 100. Configuration daemon gateway module 862 determines that the configuration information is not for plug-ins 864 but for plug-ins 134 in switch 102.


Configuration daemon gateway module 862 constructs a control message, includes the configuration information in the control message (e.g., in the payload), and sends the control message to switch 102. Sending the message includes identifying an output port for the message and transmitting via the output port. Upon receiving the message, persistent storage framework 120 extracts the configuration information and provides the configuration information to plug-ins 134. It should be noted that, since plug-ins 134 can receive instructions from a CLI of switch 102, or remotely via NETCONF or SNMP, plug-ins 134 may not distinguish between the modes via which it has received configuration information. Hence, configuration daemon gateway module 862 can directly send the configuration information to plug-ins 134 using NETCONF or SNMP.


Multi-Fabric Networks


As described in conjunction with FIG. 8A, a user can create a tenant network that spans multiple switch groups from multi-fabric manager 804. FIG. 9A illustrates an exemplary logical network, which spans a plurality of switch groups and is created by a multi-fabric manager, in accordance with an embodiment of the present invention. During operation, a user configures a tenant network spanning switch groups 100 and 810 at multi-fabric manager 804, which, in turn, creates a logical network 902 spanning switch groups 100 and 810. This logical network 902 represents the tenant network. In some embodiments, the user provides the configuration information of logical network 902 via configuration console 850, as described in conjunction with FIG. 8B.


Multi-fabric manager 804 provides configuration information of logical network 902 to principal switches 102 and 812 of switch groups 100 and 810, respectively. Switches 102 and 812 apply the configuration information in switch groups 100 and 810 using a distributed commit. In this way, switches in switch groups 100 and 810 are configured with logical network 902. Furthermore, the configuration information is also updated in the local instances of persistent storages 150 and 832 in multi-fabric manager 804. Multi-fabric manager 804 then establishes a tunnel 912 between switch groups 100 and 810. Examples of a tunnel include, but are not limited to, Virtual Extensible LAN (VXLAN), Generic Routing Encapsulation (GRE), Network Virtualization using GRE (NVGRE), and openvSwitch GRE.


Suppose that end device 952, which is coupled to switch 102, and end device 954, which is coupled to switch 812, are in logical network 902. When switch 102 learns the MAC address of end device 952, switch 102 shares that MAC address with a respective switch of switch group 100. Similarly, when switch 812 learns the MAC address of end device 954, switch 812 shares that MAC address with a respective switch of switch group 810. When multi-fabric manager 804 starts managing switch group 100 and 810, multi-fabric manager 804 obtains the learned MAC addresses and their associated information (e.g., respective switch identifiers, port identifiers, and VLAN tags) of end devices 952 and 954. Multi-fabric manager 804 programs switch group 100 to indicate that end device 954 is reachable via tunnel 912. This programming can include providing the reachability information to principal switch 102, which, in turn, synchronizes the information in switch group 100. In the same way, multi-fabric manager 804 programs switch group 810 to indicate that end device 952 is reachable via tunnel 912.


This allows packets belonging to logical network 902 to travel between switch groups 100 and 810. Suppose that end device 952 sends a layer-2 frame (e.g., an Ethernet frame) to end device 954 in logical network 902. When the frame reaches the tunnel end point in switch group 100 (e.g., switch 101), the frame is encapsulated in a tunnel header and sent via tunnel 912. For example, if tunnel 912 is a VXLAN tunnel, the tunnel header can include an outer Ethernet header, an IP header, and a VXLAN header. The tunnel end point of switch group 810 (e.g., switch 816) receives the tunnel-encapsulated frame, removes the tunnel header, and forwards the frame in logical network 902 in switch group 810 to end device 954. In this way, tunnel 912 facilitates frame forwarding across switch groups 100 and 810 in logical network 902. It should be noted that the layer-2 frame is forwarded in a switch group based on a fabric encapsulation associated with that switch group. For example, the frame can be encapsulated in a TRILL or IP packet when forwarded in switch group 100 or 810.



FIG. 9B illustrates exemplary tunnel establishment and global virtual local area network translation for a logical network spanning a plurality of switch groups, in accordance with an embodiment of the present invention. In this example, tunnel 912 is established between two logical interfaces 922 and 924 of switch groups 100 and 810, respectively. A logical interface is a virtualized network interface which can be shared by one or more switches. It should be noted that a logical interface can be distinct from port partitioning where a physical port is divided into virtual port. Multi-fabric manager 804 creates logical interfaces 922 and 924 for switch groups 100 and 810, respectively, during the establishment of tunnel 912. In some embodiments, switch groups 100 and 810 are associated with virtual IP addresses 942 and 944, respectively. One or more switches in a respective switch group can operate at tunnel end points (i.e., are capable of terminating a tunnel), and the virtual IP address of that switch group is associated with those switches. For example, if switches 101 and 102 are tunnel end points for switch group 100, virtual IP address 942 of switch group 100 is associated with switches 101 and 102. In other words, tunnel 912 is established between the tunnel end points associated with virtual IP addresses 942 and 944.


In some embodiments, logical network 902 is represented as a global VLAN in a respective switch group. Suppose that logical network 902 is represented as global VLANs 932 and 934 in switch groups 100 and 810, respectively. A global VLAN is a virtualized network in a switch group. Typically, one or more VLANs are mapped to a global VLAN based on one or more conditions, such as quality of service, tenant separation, security, etc. When a frame is forwarded in a switch group, such as a fabric switch, the frame is encapsulated in a fabric encapsulation. An identifier of a corresponding global VLAN can be included in the fabric encapsulation header. This allows different tenants to use the same VLAN tags in a switch group as long as their global VLAN tags are different. Typically, the scope of a global VLAN is within a switch group. The same VLAN can be mapped to two different global VLANs in two different switch groups.


Upon establishing tunnel 912, multi-fabric manager 804 programs a respective switch in switch group 100 to indicate that end device 954 is reachable via logical interface 922. In some embodiments, a respective switch has an entry in its local forwarding table for interface 922. The forwarding table entry can include a port identifier of an output port and an associated switch identifier. The forwarding table entry can further include a global VLAN tag as well. Similarly, multi-fabric manager 804 programs a respective switch in switch group 810 to indicate that end device 952 is reachable via logical interface 924.


Suppose that switches 101 and 816 are the tunnel end points for tunnel 912 in switch groups 100 and 810, respectively. In some embodiments, a respective switch in switch group 100 is programmed to indicate that logical interface is reachable via switch 101. Switches 101 and 816 can maintain a mapping between global VLAN 932 and logical network 902, and global VLAN 934 and logical network 902. Multi-fabric manager 804 can also maintain this mapping. Virtual IP addresses 942 and 944 are associated with switches 101 and 816, respectively. During operation, end device 952 sends a layer-2 frame 962 (e.g., an Ethernet frame) to end device 954 in logical network 902. Frame 962 has the MAC address of end device 954 as the destination MAC address in its layer-2 header. Switch 102 receives frame 962 and determines that the MAC address of end device 954 is reachable via logical interface 922. Switch 102 encapsulates frame 962 in a fabric encapsulation header 964, which can be based on TRILL or IP protocol, and sets a switch identifier of logical interface 922 as the egress switch identifier of encapsulation header 964. Since frame 962 is in logical network 902, switch 102 includes an identifier (e.g., a VLAN tag) of global VLAN 932 in encapsulation header 964 and forwards fabric-encapsulated frame 962 to switch 101.


Upon receiving fabric-encapsulated frame 962, switch 101 determines that fabric-encapsulated frame 962 is destined to logical interface 922. Switch 101 then further encapsulates the frame in a tunnel encapsulation header 966 and sets virtual IP addresses 942 and 944 as source and destination IP addresses of tunnel encapsulation header 966. In some embodiments, switch 101 uses a tunnel identifier associated with tunnel 912 to indicate that the frame is for logical network 902. Switch 101 forwards the tunnel- and fabric-encapsulated frame 962 via tunnel 912. Switch 816 receives the frame and determines that virtual IP address 944 is associated with the local switch. Hence, switch 816 removes tunnel encapsulation header 966 and extracts fabric-encapsulated frame 962.


Switch 816 determines that fabric encapsulation header 964 includes an identifier of global VLAN 932, which is mapped to logical network 902, and its egress switch identifier is associated with logical interface 922. Based on the determination, switch 816 further determines that the frame is associated with logical network 902, which is mapped to global VLAN 934 in switch group 810. As a result, switch 816 removes fabric encapsulation header 864 and re-encapsulates frame 962 with fabric encapsulation header 968. Since frame 962 has the MAC address of end device 954 as the destination MAC address in its header, switch 816 determines that the MAC address of end device 954 is reachable via switch 812. Switch 816 then sets the switch identifier of switch 816 as the egress switch identifier of encapsulation header 968. Since frame 962 is in logical network 902, switch 816 includes an identifier of global VLAN 934 in encapsulation header 968. In this way, switch 816 translates between identifiers of global VLANs 932 and 934. Switch 816 then forwards fabric-encapsulated frame 962 to switch 812.


Switch 812 receives fabric-encapsulated frame 962 and determines that the egress switch identifier of fabric encapsulation header 968 corresponds to the local switch. Switch 812 hence removes fabric encapsulation header 968, extracts frame 962, and determines that the destination MAC address of frame 962 is associated with a local port (i.e., learned from a local edge port). Switch 812 forwards frame 962 via the determined local port to end device 954. In this way, layer-2 frame 962 can be forwarded in logical network 902 across switch groups 100 and 810 even when logical network 902 is represented as global VLANs 932 and 934 in switch groups 100 and 810, respectively.



FIG. 9C illustrates exemplary overlapping logical networks, which span a plurality of switch groups and are created by a multi-fabric manager, in accordance with an embodiment of the present invention. During operation, a user configures another tenant network spanning switch groups 100, 810, and 820 at multi-fabric manager 804, which, in turn, creates a logical network 904 spanning switch groups 100, 810, and 820. This logical network 904 represents the tenant network. In some embodiments, the user provides the configuration information of logical network 904 via configuration console 850, as described in conjunction with FIG. 8B.


Multi-fabric manager 804 provides configuration information of logical network 904 to principal switches 102, 812, and 822 of switch groups 100, 810, and 820, respectively. Switches 102, 812 and 822 apply the configuration information in switch groups 100, 810, and 820, respectively, using a distributed commit. In this way, switches in switch groups 100, 810, and 820 are configured with logical network 904. Furthermore, the configuration information is also updated in the local instances of persistent storages 150, 832, and 834 in multi-fabric manager 804. Logical network 904 is represented by local global VLANs in switch groups 100, 810, and 820. Switches in a switch group include an identifier of the global VLAN of that switch group in the fabric encapsulation header of the switch group. Because logical network 904 also includes switch group 820, multi-fabric manager 804 then creates a tunnel 914 between switch groups 100 and 820, and a tunnel 916 between switch groups 810 and 820.


In some embodiments, multi-fabric manager 804 determines that a tunnel 912 has already been established between switch groups 100 and 810. As a result, multi-fabric manager 804 reuses tunnel 912 for logical network 904. Since logical network 904 is represented by different global VLANs in switch groups 100 and 810, tunnel end point switches 101 and 816 can distinguish between frames belonging to logical networks 902 and 904 even when those frames use the same tunnel 912. For example, upon receiving a frame via tunnel 912, switch 816 can determine whether a frame is in logical network 902 or 904 based on the global VLAN identifier in the fabric encapsulation header of the frame. In some embodiments, multi-fabric manager 804 creates a separate tunnel for logical network 904 between switch groups 100 and 810.


Operations



FIG. 10A presents a flowchart illustrating the process of a multi-fabric manager associating with a switch group, in accordance with an embodiment of the present invention. During operation, the multi-fabric manager identifies the one or more switch groups to be managed (operation 1002). In some embodiments, a switch group can be a fabric switch. The multi-fabric manager then joins a respective switch group as a member switch (operation 1004), as described in conjunction with FIG. 8A. As a result, the multi-fabric manager obtains an instance of the persistent storage of a respective identified switch group (operation 1006), as described in conjunction with FIG. 7. The multi-fabric manager then imports the learned MAC addresses (and associated information) from a respective identified switch group (operation 1008) and stores the imported learned MAC addresses (and associated information) in association with a switch group identifier of the corresponding switch group (operation 1010). The learned MAC addresses can be imported from the forwarding table, which can also be referred to as a MAC database, of the principal switch of the switch group. The multi-fabric manager can also store a global VLAN identifier with the learned MAC addresses.



FIG. 10B presents a flowchart illustrating the process of a multi-fabric manager creating a logical network, in accordance with an embodiment of the present invention. During operation, the multi-fabric manager receives an instruction for a logical network (operation 1032) and identifies the switch groups associated with the logical network (operation 1034). In some embodiments, a user can specify the sites that should be included in the logical network and the multi-fabric manager can identify the corresponding switch groups automatically. The multi-fabric manager then configures the logical network in a respective switch group based on a distributed commit (operation 1036). Configuring the logical network in a switch group can include creating a global VLAN corresponding to the logical network in the switch group. During the distributed commit process, the multi-fabric manager provides the configuration information to the principal switch of a switch group. The principal switch, in turn, performs the distributed commit for that switch group. Since the multi-fabric manager joins as a member switch to that switch group, the distributed commit also updates the local instance of the persistent storage of that switch group.


The multi-fabric manager then identifies a switch group pair in the logical network (i.e., from the identified switch groups) (operation 1038) and checks whether individual tunnels are supported between that switch group pair (operation 1040). If individual tunnels are supported, the multi-fabric manager creates a separate tunnel for the logical network for that switch group pair. If individual tunnels are not supported, the multi-fabric manager checks whether a tunnel has already been established between the switch group pair (operation 1042). If individual tunnels are supported or a tunnel has not been established, the multi-fabric manager establishes a tunnel between the respective logical interfaces of the switch group pair (operation 1044).


If a tunnel has already been established (operation 1042 or 1044), the multi-fabric manager programs the MAC addresses of one switch group in the other switch group as reachable via the logical interface of the local switch group (operation 1046). In some embodiments, the logical interface is mapped to one or more ports of one or more switches. A frame destined to the logical interface is forwarded via one of the mapped ports. The multi-fabric manager configures the identifier of the global VLAN in a respective switch group to map to the logical network (e.g., map to an identifier of the logical network) (operation 1048) and checks whether all switch group pairs have been configured (operation 1050). If not, the multi-fabric manager continues to identify another switch group pair in the logical network (operation 1038). If all switch group pairs have been configured, the multi-fabric manager configures the tunnels and the global VLAN translations in the respective identified switch group based in a distributed commit (operation 1052). In some embodiments, upon being a principal multi-fabric manager, the multi-fabric manager updates the persistent storages of a respective replica of the multi-fabric manager via a distributed commit (operation 1054).


Distributed Commit



FIG. 11A presents a flowchart illustrating the distributed commit process of a principal persistent storage framework applying configuration information, in accordance with an embodiment of the present invention. The persistent storage framework of a principal switch or a principal multi-fabric manager can be called a principal persistent storage framework. During operation, the framework receives configuration information of logical network (operation 1102). In some embodiments, this configuration information includes the attribute values of the persistent storage of a switch group. The framework then generates and sends an instruction message to a respective managed framework instructing to prepare their local persistent storage for the configuration information (operation 1104). The persistent storage framework of a managed switch or a managed multi-fabric manager can be called a managed persistent storage framework. The framework receives a response from a respective managed framework (operation 1106) and checks whether all persistent storages have been prepared (operation 1108). If all persistent storages have not been prepared, the framework rolls back the configuration information from the local persistent storage (operation 1112).


The framework also generates a control message instructing to roll back the configuration information from the persistent storage and sends the message to a respective managed framework with successful storage preparation (operation 1114) (e.g., the framework may not send the control message to the managed frameworks which have not succeeded in preparing their corresponding local persistent storage). If all persistent storages have been prepared, the framework generates a control message instructing to commit the configuration information to the persistent storage (and the switch modules), and sends the message to a respective framework (operation 1110). It should be noted that the instruction for the managed switches is for committing the configuration information to the persistent storage and the switch modules, whereas the instruction for the managed multi-fabric managers can be for committing to the persistent storage.



FIG. 11B presents a flowchart illustrating the distributed commit process of a managed persistent storage framework applying configuration information received from a principal persistent storage framework, in accordance with an embodiment of the present invention. During operation, the framework receives an instruction message from a principal framework instructing to prepare the local persistent storage for the configuration information (operation 1152) and prepares the local persistent storage for the configuration information (operation 1154). The framework then checks whether the local persistent storage has been prepared (operation 1156). If prepared, the framework generates a response message indicating that the local persistent storage has been prepared for the configuration information and sends the message to the principal framework (operation 1160).


Depending on whether other managed frameworks have succeeded or not, the framework may or may not commit the configuration information. If all other managed frameworks have succeeded (denoted with a dotted arrow), the framework receives an instruction message from the principal framework instructing to commit the configuration information to the local persistent storage (and the switch modules) (operation 1162) and commits the configuration information to the local persistent storage (and the switch modules) (operation 1164). It should be noted that the instruction for the managed switches is for committing the configuration information to the persistent storage and the switch modules, whereas the instruction for the managed multi-fabric managers can be for committing to the persistent storage. For example, if a managed multi-fabric manager is a switch in a switch group, configuration information of the switch group is committed to the switch modules as well.


If at least one other managed framework has not succeeded (denoted with a dashed arrow), the framework receives an instruction message from the principal framework instructing to roll back the configuration information from the local persistent storage (operation 1166) and rolls back the configuration information from the local persistent storage (operation 1168). If the local persistent storage has not been prepared, the framework generates a response message indicating a failure and sends the message to the principal framework (operation 1158).


Exemplary Multi-Fabric Manager



FIG. 12 illustrates an exemplary multi-fabric manager, in accordance with an embodiment of the present invention. In this example, a computing system 1200 operates as a multi-fabric manager. Computing system 1200 includes a number of communication ports 1202, a packet processor 1210, a persistent storage module 1230, a management module 1232, and a storage device 1250. Computing system 1200 can also include device modules 1260 (e.g., processing hardware of computing system 1200, such as its ASIC chips), which includes information based on which computing system 1200 processes packets (e.g., determines output ports for packets). Packet processor 1210 extracts and processes header information from the received frames. Packet processor 1210 can identify a switch identifier associated with the switch in header of a packet.


In some embodiments, computing system 1200 maintains a membership in a switch group, which can be a fabric switch, as described in conjunction with FIG. 1. Computing system 1200 then includes a switch group module 1220. Switch group module 1220 maintains a configuration database in storage device 1250 that maintains the configuration state of every switch within the fabric switch. Switch group module 1220 maintains the state of the fabric switch, which is used to join other switches. In some embodiments, management module 1232 generates a first control message comprising configuration information associated with the switch group for the principal switch of the switch group and a second control message indicating whether the configuration information is prepared for storage in a persistent storage instance of the switch group. Management module 1232 then commits the configuration information to the persistent storage instance based on an instruction message from the principal switch.


Communication ports 1202 can include inter-switch communication channels for communication within the fabric switch. This inter-switch communication channel can be implemented via a regular communication port and based on any open or proprietary format. Communication ports 1202 can also include one or more extension communication ports for communication between neighbor fabric switches. Communication ports 1202 can include one or more TRILL ports capable of receiving frames encapsulated in a TRILL header. Communication ports 1202 can also include one or more IP ports capable of receiving IP packets. An IP port is capable of receiving an IP packet and can be configured with an IP address. Packet processor 1210 can process TRILL-encapsulated frames and/or IP packets.


During operation, persistent storage module 1230 stores a plurality of persistent storage instances associated with a plurality of switch groups in object relational database 1240 in storage device 1250. Management module 1232 identifies at least two switch groups, which are associated with a logical network, from the plurality of switch groups and configures a respective virtual network representing the logical network in a respective identified switch group. Management module 1232 also maintains a mapping between a respective virtual network and the logical network. In some embodiments, computing system 1200 also includes a tunnel module 1270, which establishes a tunnel between the identified switch groups. The tunnel can be terminated at a logical interface, which is associated with one or more switches of an identified switch group.


In some embodiments, management module 1232 obtains a MAC address learned at a first switch group of the identified switch groups and stores the MAC address in a local data structure in association with a switch group identifier (e.g., a fabric identifier) which identifies the first switch group. Management module 1232 then programs reachability of the MAC address in a second switch group of the identified switch groups. In some embodiments, computing system 1200 can be the principal multi-fabric manager among a plurality of replicas of the multi-fabric manager. Management module 1232 then verifies whether configuration information in database 1240 is prepared for storage in a persistent storage instance of a replica of the multi-fabric manager. If prepared, management module 1232 generates an instruction message instructing to commit the configuration information to the persistent storage instance of the replica of the multi-fabric manager.


Note that the above-mentioned modules can be implemented in hardware as well as in software. In one embodiment, these modules can be embodied in computer-executable instructions stored in a memory which is coupled to one or more processors in computing system 1200. When executed, these instructions cause the processor(s) to perform the aforementioned functions.


In summary, embodiments of the present invention provide a computing system and a method for facilitating online restoration of a switch snapshot. In one embodiment, the computing system includes processing circuitry, one or more ports, a persistent storage module, and a management module. The persistent storage module stores a plurality of persistent storage instances associated with a plurality of switch groups. A respective persistent storage instance stores configuration information associated with a switch group in a data structure. The management module identifies at least two switch groups, which are associated with a logical network, from the plurality of switch groups and configures a respective virtual network representing the logical network in a respective identified switch group. The management module also maintains a mapping between a respective virtual network and the logical network.


The methods and processes described herein can be embodied as code and/or data, which can be stored in a computer-readable non-transitory storage medium. When a computer system reads and executes the code and/or data stored on the computer-readable non-transitory storage medium, the computer system performs the methods and processes embodied as data structures and code and stored within the medium.


The methods and processes described herein can be executed by and/or included in hardware modules or apparatus. These modules or apparatus may include, but are not limited to, an application-specific integrated circuit (ASIC) chip, a field-programmable gate array (FPGA), a dedicated or shared processor that executes a particular software module or a piece of code at a particular time, and/or other programmable-logic devices now known or later developed. When the hardware modules or apparatus are activated, they perform the methods and processes included within them.


The foregoing descriptions of embodiments of the present invention have been presented only for purposes of illustration and description. They are not intended to be exhaustive or to limit this disclosure. Accordingly, many modifications and variations will be apparent to practitioners skilled in the art. The scope of the present invention is defined by the appended claims.

Claims
  • 1. A computing system, comprising: processing circuitry;one or more ports;persistent storage circuitry configured to store a first and a second persistent storage instances storing configuration information associated with first and a second networks of interconnected switches, respectively, wherein the first and the second networks of interconnected switches are identified by a first and a second fabric identifiers, respectively; anda management circuitry configured to: identify a logical network representing a tenant network associated with the first and second storage instances;configure a first virtual network representing the logical network in the first network of interconnected switches and a second virtual network representing the logical network in the second network of interconnected switches; andmaintain a first mapping between the first virtual network and the logical network, and a second mapping between the second virtual network and the logical network.
  • 2. The computing system of claim 1, further comprising tunnel circuitry configured to establish a tunnel between the first and second networks of interconnected switches.
  • 3. The computing system of claim 2, wherein the tunnel is terminated at a logical interface, wherein the logical interface is associated with one or more switches of one of the first and second networks of interconnected switches.
  • 4. The computing system of claim 1, wherein the management circuitry is further configured to: obtain a media access control (MAC) address learned at the first network of interconnected switches; andstore the MAC address in a local data structure in association with a switch group identifier which identifies the first network of interconnected switches.
  • 5. The computing system of claim 1, wherein the management circuitry is further configured to program reachability of a MAC address learned at the first network of interconnected switches in the second network of interconnected switches.
  • 6. The computing system of claim 1, wherein a header of a packet belonging to the logical network includes an identifier of the first virtual network.
  • 7. The computing system of claim 6, wherein the header of the packet belonging to the logical network includes an identifier of the second virtual network.
  • 8. The computing system of claim 1, further comprising switch group circuitry configured to maintain a membership in the first and second networks of interconnected switches.
  • 9. The computing system of claim 1, wherein the management circuitry is further configured to: generate a first control message comprising configuration information associated with the first network of interconnected switches for a principal switch of the first network of interconnected switches;generate a second control message indicating whether the configuration information is prepared for storage in the first persistent storage instance; andcommit the configuration information to the first persistent storage instance based on an instruction message from the principal switch.
  • 10. The computing system of claim 1, wherein the management circuitry is further configured to: verify whether configuration information is prepared for storage in a persistent storage instance of a replica of the computing system; andin response to verifying that the configuration information is prepared for storage in the persistent storage instance of the replica of the computing system, generate an instruction message instructing to commit the configuration information to the persistent storage instance of the replica of the computing system.
  • 11. The computing system of claim 1, wherein persistent storage instance is an object relational database instance, and wherein a data structure in a persistent storage instance is a table in the database instance.
  • 12. A method, comprising: storing a first and a second persistent storage instances storing configuration information associated with a first and a second networks of interconnected switches, respectively, wherein the first and the second networks of interconnected switches are identified by a first and a second fabric identifiers, respectively;identifying a logical network representing a tenant network associated with the first and second storage instances;configuring a first virtual network representing the logical network in the first network of interconnected switches and a second virtual network representing the logical network in the second network of interconnected switches; andmaintaining a first mapping between the first virtual network and the logical network, and a second mapping between the second virtual network and the logical network.
  • 13. The method of claim 12, further comprising establishing a tunnel between first and second networks of interconnected switches.
  • 14. The method of claim 13, wherein the tunnel is terminated at a logical interface, wherein the logical interface is associated with one or more switches of one of the first and second networks of interconnected switches.
  • 15. The method of claim 12, further comprising: obtaining a media access control (MAC) address learned at the first network of interconnected switches; andstoring the MAC address in a local data structure in association with a switch group identifier which identifies the first network of interconnected switches.
  • 16. The method of claim 12, further comprising programming reachability of a MAC address learned at the first network of interconnected switches in the second network of interconnected switches.
  • 17. The method of claim 12, wherein a header of a packet belonging to the logical network includes an identifier of the first virtual network.
  • 18. The method of claim 17, wherein the header of the packet belonging to the logical network includes an identifier of the second virtual network.
  • 19. The method of claim 12, further comprising maintaining a membership in the first and second networks of interconnected switches.
  • 20. The method of claim 12, further comprising: generating a first control message comprising configuration information associated with the first network of interconnected switches for a principal switch of the first network of interconnected switches;generating a second control message indicating whether the configuration information is prepared for storage in the first persistent storage instance; andcommitting the configuration information to the first persistent storage instance based on an instruction message from the principal switch.
  • 21. The method of claim 12, further comprising: verifying whether configuration information is prepared for storage in a replica persistent storage instance of the first persistent storage instance; andin response to verifying that the configuration information is prepared for storage in the replica persistent storage instance, generating an instruction message instructing to commit the configuration information to the replica persistent storage instance.
  • 22. The method of claim 12, wherein a persistent storage instance is an object relational database instance, and wherein a data structure in a persistent storage instance is a table in the database instance.
  • 23. A non-transitory computer-readable storage medium storing instructions that when executed by a computing system cause the computing system to perform a method for managing a group of switches, the method comprising: storing a first and a second persistent storage instances storing configuration information associated with a first and a second networks of interconnected switches, respectively, wherein the first and the second networks of interconnected switches are identified by a first and a second fabric identifiers, respectively;identifying a logical network representing a tenant network associated with the first and second storage instances;configuring a first virtual network representing the logical network in the first network of interconnected switches and a second virtual network representing the logical network in the second network of interconnected switches; andmaintaining a first mapping between the first virtual network and the logical network, and a second mapping between the second virtual network and the logical network.
US Referenced Citations (573)
Number Name Date Kind
829529 Keathley Aug 1906 A
5390173 Spinney Feb 1995 A
5802278 Isfeld Sep 1998 A
5878232 Marimuthu Mar 1999 A
5879173 Poplawski Mar 1999 A
5959968 Chin Sep 1999 A
5973278 Wehrill, III Oct 1999 A
5983278 Chong Nov 1999 A
5995262 Hirota Nov 1999 A
6041042 Bussiere Mar 2000 A
6085238 Yuasa Jul 2000 A
6092062 Lohman Jul 2000 A
6104696 Kadambi Aug 2000 A
6185214 Schwartz Feb 2001 B1
6185241 Sun Feb 2001 B1
6295527 McCormack Sep 2001 B1
6331983 Haggerty Dec 2001 B1
6438106 Pillar Aug 2002 B1
6498781 Bass Dec 2002 B1
6542266 Phillips Apr 2003 B1
6553029 Alexander Apr 2003 B1
6571355 Linnell May 2003 B1
6583902 Yuen Jun 2003 B1
6633761 Singhal Oct 2003 B1
6636963 Stein Oct 2003 B1
6771610 Seaman Aug 2004 B1
6870840 Hill Mar 2005 B1
6873602 Ambe Mar 2005 B1
6937576 DiBenedetto Aug 2005 B1
6956824 Mark Oct 2005 B2
6957269 Williams Oct 2005 B2
6975581 Medina Dec 2005 B1
6975864 Singhal Dec 2005 B2
7016352 Chow Mar 2006 B1
7061877 Gummalla Jun 2006 B1
7062177 Grivna Jun 2006 B1
7173934 Lapuh Feb 2007 B2
7197308 Singhal Mar 2007 B2
7206288 Cometto Apr 2007 B2
7310664 Merchant Dec 2007 B1
7313637 Tanaka Dec 2007 B2
7315545 Chowdhury et al. Jan 2008 B1
7316031 Griffith Jan 2008 B2
7330897 Baldwin Feb 2008 B2
7380025 Riggins May 2008 B1
7397794 Lacroute Jul 2008 B1
7430164 Bare Sep 2008 B2
7453888 Zabihi Nov 2008 B2
7477894 Sinha Jan 2009 B1
7480258 Shuen Jan 2009 B1
7508757 Ge Mar 2009 B2
7558195 Kuo Jul 2009 B1
7558273 Grosser Jul 2009 B1
7571447 Ally Aug 2009 B2
7599901 Mital Oct 2009 B2
7688736 Walsh Mar 2010 B1
7688960 Aubuchon Mar 2010 B1
7690040 Frattura Mar 2010 B2
7706255 Kondrat et al. Apr 2010 B1
7716370 Devarapalli May 2010 B1
7720076 Dobbins May 2010 B2
7729296 Choudhary Jun 2010 B1
7787480 Mehta Aug 2010 B1
7792920 Istvan Sep 2010 B2
7796593 Ghosh Sep 2010 B1
7801021 Triantafillis Sep 2010 B1
7808992 Homchaudhuri Oct 2010 B2
7836332 Hara Nov 2010 B2
7843906 Chidambaram et al. Nov 2010 B1
7843907 Abou-Emara Nov 2010 B1
7860097 Lovett Dec 2010 B1
7898959 Arad Mar 2011 B1
7912091 Krishnan Mar 2011 B1
7924837 Shabtay Apr 2011 B1
7937438 Miller May 2011 B1
7937756 Kay May 2011 B2
7945941 Sinha May 2011 B2
7949638 Goodson May 2011 B1
7957386 Aggarwal Jun 2011 B1
8018938 Fromm Sep 2011 B1
8027354 Portolani Sep 2011 B1
8054832 Shukla Nov 2011 B1
8068442 Kompella Nov 2011 B1
8078704 Lee Dec 2011 B2
8090805 Chawla Jan 2012 B1
8102781 Smith Jan 2012 B2
8102791 Tang Jan 2012 B2
8116307 Thesayi Feb 2012 B1
8125928 Mehta Feb 2012 B2
8134922 Elangovan Mar 2012 B2
8155150 Chung Apr 2012 B1
8160063 Maltz Apr 2012 B2
8160080 Arad Apr 2012 B1
8170038 Belanger May 2012 B2
8175107 Yalagandula May 2012 B1
8194674 Pagel Jun 2012 B1
8195774 Lambeth Jun 2012 B2
8204061 Sane Jun 2012 B1
8213313 Doiron Jul 2012 B1
8213336 Smith Jul 2012 B2
8230069 Korupolu Jul 2012 B2
8239960 Frattura Aug 2012 B2
8249069 Raman Aug 2012 B2
8270401 Barnes Sep 2012 B1
8295291 Ramanathan Oct 2012 B1
8295921 Wang Oct 2012 B2
8301686 Appajodu Oct 2012 B1
8339994 Gnanasekaran Dec 2012 B2
8351352 Eastlake Jan 2013 B1
8369335 Jha Feb 2013 B2
8369347 Xiong Feb 2013 B2
8392496 Linden Mar 2013 B2
8451717 Venkataraman May 2013 B2
8462774 Page Jun 2013 B2
8467375 Blair Jun 2013 B2
8520595 Yadav Aug 2013 B2
8553710 White Oct 2013 B1
8599850 Jha Dec 2013 B2
8599864 Chung Dec 2013 B2
8615008 Natarajan Dec 2013 B2
8619788 Sankaran Dec 2013 B1
8705526 Hasan Apr 2014 B1
8706905 McGlaughlin Apr 2014 B1
8717895 Koponen May 2014 B2
8724456 Hong May 2014 B1
8792501 Rustagi Jul 2014 B1
8798045 Aybay Aug 2014 B1
8804736 Drake Aug 2014 B1
8806031 Kondur Aug 2014 B1
8826385 Congdon Sep 2014 B2
8918631 Kumar Dec 2014 B1
8937865 Kumar Jan 2015 B1
8948181 Kapadia Feb 2015 B2
8995272 Agarwal Mar 2015 B2
9178793 Marlow Nov 2015 B1
9350680 Thayalan May 2016 B2
9401818 Venkatesh Jul 2016 B2
9438447 Basso Sep 2016 B2
20010005527 Vaeth Jun 2001 A1
20010055274 Hegge Dec 2001 A1
20020019904 Katz Feb 2002 A1
20020021701 Lavian Feb 2002 A1
20020039350 Wang Apr 2002 A1
20020054593 Morohashi May 2002 A1
20020087723 Williams Jul 2002 A1
20020091795 Yip Jul 2002 A1
20030026290 Umayabashi Feb 2003 A1
20030041085 Sato Feb 2003 A1
20030093567 Lolayekar May 2003 A1
20030097464 Martinez May 2003 A1
20030097470 Lapuh May 2003 A1
20030123393 Feuerstraeter Jul 2003 A1
20030147385 Montalvo Aug 2003 A1
20030152075 Hawthorne, III Aug 2003 A1
20030174706 Shankar Sep 2003 A1
20030189905 Lee Oct 2003 A1
20030189930 Terrell Oct 2003 A1
20030208616 Laing Nov 2003 A1
20030216143 Roese Nov 2003 A1
20030223428 Blanquer Gonzalez Dec 2003 A1
20030233534 Bernhard Dec 2003 A1
20040001433 Gram Jan 2004 A1
20040003094 See Jan 2004 A1
20040008868 Bornowski Jan 2004 A1
20040010600 Baldwin Jan 2004 A1
20040049699 Griffith Mar 2004 A1
20040057430 Paavolainen Mar 2004 A1
20040081171 Finn Apr 2004 A1
20040095900 Siegel May 2004 A1
20040117508 Shimizu Jun 2004 A1
20040120326 Yoon Jun 2004 A1
20040156313 Hofmeister et al. Aug 2004 A1
20040165595 Holmgren Aug 2004 A1
20040165596 Garcia Aug 2004 A1
20040205234 Barrack Oct 2004 A1
20040213232 Regan Oct 2004 A1
20040225725 Enomoto Nov 2004 A1
20040243673 Goyal Dec 2004 A1
20050007951 Lapuh Jan 2005 A1
20050044199 Shiga Feb 2005 A1
20050074001 Mattes Apr 2005 A1
20050094568 Judd May 2005 A1
20050094630 Valdevit May 2005 A1
20050108375 Hallak-Stamler May 2005 A1
20050122979 Gross Jun 2005 A1
20050152335 Lodha Jul 2005 A1
20050157645 Rabie et al. Jul 2005 A1
20050157751 Rabie Jul 2005 A1
20050169188 Cometto Aug 2005 A1
20050195813 Ambe Sep 2005 A1
20050207423 Herbst Sep 2005 A1
20050213561 Yao Sep 2005 A1
20050220096 Friskney Oct 2005 A1
20050259586 Hafid Nov 2005 A1
20050265330 Suzuki Dec 2005 A1
20050265356 Kawarai Dec 2005 A1
20050278565 Frattura Dec 2005 A1
20060007869 Hirota Jan 2006 A1
20060018302 Ivaldi Jan 2006 A1
20060023707 Makishima et al. Feb 2006 A1
20060029055 Perera Feb 2006 A1
20060034292 Wakayama Feb 2006 A1
20060036648 Frey Feb 2006 A1
20060036765 Weyman Feb 2006 A1
20060039366 Ghosh Feb 2006 A1
20060059163 Frattura Mar 2006 A1
20060062187 Rune Mar 2006 A1
20060072550 Davis Apr 2006 A1
20060083172 Jordan Apr 2006 A1
20060083254 Ge Apr 2006 A1
20060098589 Kreeger May 2006 A1
20060126511 Youn Jun 2006 A1
20060140130 Kalkunte Jun 2006 A1
20060168109 Warmenhoven Jul 2006 A1
20060184937 Abels Aug 2006 A1
20060209886 Silberman Sep 2006 A1
20060221960 Borgione Oct 2006 A1
20060227776 Chandrasekaran Oct 2006 A1
20060235995 Bhatia Oct 2006 A1
20060242311 Mai Oct 2006 A1
20060242398 Fontijn Oct 2006 A1
20060245439 Sajassi Nov 2006 A1
20060251067 DeSanti Nov 2006 A1
20060256767 Suzuki Nov 2006 A1
20060265515 Shiga Nov 2006 A1
20060285499 Tzeng Dec 2006 A1
20060291388 Amdahl Dec 2006 A1
20060291480 Cho Dec 2006 A1
20060294413 Filz Dec 2006 A1
20070036178 Hares Feb 2007 A1
20070053294 Ho Mar 2007 A1
20070061817 Atkinson Mar 2007 A1
20070074052 Hemmah Mar 2007 A1
20070081530 Nomura Apr 2007 A1
20070083625 Chamdani Apr 2007 A1
20070086362 Kato Apr 2007 A1
20070094464 Sharma Apr 2007 A1
20070097968 Du May 2007 A1
20070098006 Parry May 2007 A1
20070116224 Burke May 2007 A1
20070116422 Reynolds May 2007 A1
20070130295 Rastogi Jun 2007 A1
20070156659 Lim Jul 2007 A1
20070177525 Wijnands Aug 2007 A1
20070177597 Ju Aug 2007 A1
20070183313 Narayanan Aug 2007 A1
20070206762 Chandra Sep 2007 A1
20070211712 Fitch Sep 2007 A1
20070220059 Lu Sep 2007 A1
20070226214 Smits Sep 2007 A1
20070238343 Velleca Oct 2007 A1
20070258449 Bennett Nov 2007 A1
20070274234 Kubota Nov 2007 A1
20070280223 Pan Dec 2007 A1
20070289017 Copeland, III Dec 2007 A1
20080052487 Akahane Feb 2008 A1
20080056135 Lee Mar 2008 A1
20080056300 Williams Mar 2008 A1
20080057918 Abrant Mar 2008 A1
20080065760 Damm Mar 2008 A1
20080080517 Roy Apr 2008 A1
20080095160 Yadav Apr 2008 A1
20080101386 Gray May 2008 A1
20080112133 Torudbakken May 2008 A1
20080112400 Dunbar et al. May 2008 A1
20080133760 Berkvens Jun 2008 A1
20080159260 Vobbilisetty Jul 2008 A1
20080159277 Vobbilisetty Jul 2008 A1
20080172492 Raghunath Jul 2008 A1
20080181196 Regan Jul 2008 A1
20080181243 Vobbilisetty Jul 2008 A1
20080186981 Seto Aug 2008 A1
20080205377 Chao Aug 2008 A1
20080219172 Mohan Sep 2008 A1
20080225852 Raszuk et al. Sep 2008 A1
20080225853 Melman Sep 2008 A1
20080228897 Ko Sep 2008 A1
20080240129 Elmeleegy Oct 2008 A1
20080267179 LaVigne Oct 2008 A1
20080285458 Lysne Nov 2008 A1
20080285555 Ogasahara Nov 2008 A1
20080288020 Einav Nov 2008 A1
20080298248 Roeck Dec 2008 A1
20080304498 Jorgensen Dec 2008 A1
20080310342 Kruys Dec 2008 A1
20090022069 Khan Jan 2009 A1
20090024734 Merbach Jan 2009 A1
20090037607 Farinacci Feb 2009 A1
20090037977 Gai Feb 2009 A1
20090041046 Hirata Feb 2009 A1
20090042270 Dolly Feb 2009 A1
20090044270 Shelly Feb 2009 A1
20090067422 Poppe Mar 2009 A1
20090067442 Killian Mar 2009 A1
20090079560 Fries Mar 2009 A1
20090080345 Gray Mar 2009 A1
20090083445 Ganga Mar 2009 A1
20090092042 Yuhara Apr 2009 A1
20090092043 Lapuh Apr 2009 A1
20090094354 Rastogi Apr 2009 A1
20090106298 Furusho Apr 2009 A1
20090106405 Mazarick Apr 2009 A1
20090113408 Toeroe Apr 2009 A1
20090116381 Kanda May 2009 A1
20090122700 Aboba May 2009 A1
20090129384 Regan May 2009 A1
20090138577 Casado May 2009 A1
20090138752 Graham May 2009 A1
20090144720 Roush Jun 2009 A1
20090161584 Guan Jun 2009 A1
20090161670 Shepherd Jun 2009 A1
20090168647 Holness Jul 2009 A1
20090199177 Edwards Aug 2009 A1
20090204965 Tanaka Aug 2009 A1
20090213783 Moreton Aug 2009 A1
20090222879 Kostal Sep 2009 A1
20090225752 Mitsumori Sep 2009 A1
20090232031 Vasseur Sep 2009 A1
20090245112 Okazaki Oct 2009 A1
20090245137 Hares Oct 2009 A1
20090245242 Carlson Oct 2009 A1
20090246137 Hadida Ruah et al. Oct 2009 A1
20090249444 Macauley Oct 2009 A1
20090252049 Ludwig Oct 2009 A1
20090252061 Small Oct 2009 A1
20090252503 Ishigami Oct 2009 A1
20090260083 Szeto Oct 2009 A1
20090279558 Davis Nov 2009 A1
20090292858 Lambeth Nov 2009 A1
20090316721 Kanda Dec 2009 A1
20090323698 LeFaucheur Dec 2009 A1
20090323708 Ihle Dec 2009 A1
20090327392 Tripathi et al. Dec 2009 A1
20090327462 Adams Dec 2009 A1
20100002382 Aybay Jan 2010 A1
20100027420 Smith Feb 2010 A1
20100042869 Szabo Feb 2010 A1
20100046471 Hattori Feb 2010 A1
20100054260 Pandey Mar 2010 A1
20100061269 Banerjee Mar 2010 A1
20100074175 Banks Mar 2010 A1
20100085981 Gupta Apr 2010 A1
20100097941 Carlson Apr 2010 A1
20100103813 Allan Apr 2010 A1
20100103939 Carlson Apr 2010 A1
20100114818 Lier May 2010 A1
20100131636 Suri May 2010 A1
20100157844 Casey Jun 2010 A1
20100158024 Sajassi Jun 2010 A1
20100165877 Shukla Jul 2010 A1
20100165995 Mehta Jul 2010 A1
20100168467 Johnston Jul 2010 A1
20100169467 Shukla Jul 2010 A1
20100169948 Budko Jul 2010 A1
20100182920 Matsuoka Jul 2010 A1
20100189119 Sawada Jul 2010 A1
20100195489 Zhou Aug 2010 A1
20100215042 Sato Aug 2010 A1
20100215049 Raza Aug 2010 A1
20100220724 Rabie Sep 2010 A1
20100226368 Mack-Crane Sep 2010 A1
20100226381 Mehta Sep 2010 A1
20100246388 Gupta Sep 2010 A1
20100254703 Kirkpatrick Oct 2010 A1
20100257263 Casado Oct 2010 A1
20100258263 Douxchamps Oct 2010 A1
20100265849 Harel Oct 2010 A1
20100271960 Krygowski Oct 2010 A1
20100272107 Papp Oct 2010 A1
20100281106 Ashwood-Smith Nov 2010 A1
20100284414 Agarwal Nov 2010 A1
20100284418 Gray Nov 2010 A1
20100284698 McColloch Nov 2010 A1
20100287262 Elzur Nov 2010 A1
20100287548 Zhou Nov 2010 A1
20100290464 Assarpour Nov 2010 A1
20100290472 Raman Nov 2010 A1
20100290473 Enduri Nov 2010 A1
20100299527 Arunan Nov 2010 A1
20100303071 Kotalwar Dec 2010 A1
20100303075 Tripathi Dec 2010 A1
20100303083 Belanger Dec 2010 A1
20100309820 Rajagopalan Dec 2010 A1
20100309912 Mehta Dec 2010 A1
20100329110 Rose Dec 2010 A1
20110007738 Berman Jan 2011 A1
20110019678 Mehta Jan 2011 A1
20110032945 Mullooly Feb 2011 A1
20110035489 McDaniel Feb 2011 A1
20110035498 Shah Feb 2011 A1
20110044339 Kotalwar Feb 2011 A1
20110044352 Chaitou Feb 2011 A1
20110051723 Rabie Mar 2011 A1
20110055274 Scales et al. Mar 2011 A1
20110058547 Waldrop Mar 2011 A1
20110064086 Xiong Mar 2011 A1
20110064089 Hidaka Mar 2011 A1
20110072208 Gulati Mar 2011 A1
20110085560 Chawla Apr 2011 A1
20110085562 Bao Apr 2011 A1
20110085563 Kotha Apr 2011 A1
20110088011 Ouali Apr 2011 A1
20110110266 Li May 2011 A1
20110134802 Rajagopalan Jun 2011 A1
20110134803 Dalvi Jun 2011 A1
20110134925 Safrai Jun 2011 A1
20110142053 Van Der Merwe et al. Jun 2011 A1
20110142062 Wang Jun 2011 A1
20110149526 Turner Jun 2011 A1
20110158113 Nanda Jun 2011 A1
20110161494 McDysan Jun 2011 A1
20110161695 Okita Jun 2011 A1
20110176412 Stine Jul 2011 A1
20110188373 Saito Aug 2011 A1
20110194403 Sajassi Aug 2011 A1
20110194563 Shen Aug 2011 A1
20110228767 Singla Sep 2011 A1
20110228780 Ashwood-Smith Sep 2011 A1
20110231570 Altekar Sep 2011 A1
20110231574 Saunderson Sep 2011 A1
20110235523 Jha Sep 2011 A1
20110243133 Villait Oct 2011 A9
20110243136 Raman Oct 2011 A1
20110246669 Kanada Oct 2011 A1
20110255538 Srinivasan Oct 2011 A1
20110255540 Mizrahi Oct 2011 A1
20110261828 Smith Oct 2011 A1
20110268118 Schlansker Nov 2011 A1
20110268120 Vobbilisetty Nov 2011 A1
20110268125 Vobbilisetty Nov 2011 A1
20110273988 Tourrilhes Nov 2011 A1
20110273990 Rajagopalan Nov 2011 A1
20110274114 Dhar Nov 2011 A1
20110280572 Vobbilisetty Nov 2011 A1
20110286357 Haris Nov 2011 A1
20110286457 Ee Nov 2011 A1
20110286462 Kompella Nov 2011 A1
20110292947 Vobbilisetty Dec 2011 A1
20110296052 Guo Dec 2011 A1
20110299391 Vobbilisetty Dec 2011 A1
20110299413 Chatwani Dec 2011 A1
20110299414 Yu Dec 2011 A1
20110299527 Yu Dec 2011 A1
20110299528 Yu Dec 2011 A1
20110299531 Yu Dec 2011 A1
20110299532 Yu Dec 2011 A1
20110299533 Yu Dec 2011 A1
20110299534 Koganti Dec 2011 A1
20110299535 Vobbilisetty Dec 2011 A1
20110299536 Cheng Dec 2011 A1
20110317559 Kern Dec 2011 A1
20110317703 Dunbar et al. Dec 2011 A1
20120011240 Hara Jan 2012 A1
20120014261 Salam Jan 2012 A1
20120014387 Dunbar Jan 2012 A1
20120020220 Sugita Jan 2012 A1
20120027017 Rai Feb 2012 A1
20120033663 Guichard Feb 2012 A1
20120033665 Jacob Da Silva et al. Feb 2012 A1
20120033668 Humphries Feb 2012 A1
20120033669 Mohandas Feb 2012 A1
20120033672 Page Feb 2012 A1
20120039163 Nakajima Feb 2012 A1
20120042095 Kotha Feb 2012 A1
20120063363 Li Mar 2012 A1
20120075991 Sugita Mar 2012 A1
20120099567 Hart Apr 2012 A1
20120099602 Nagapudi Apr 2012 A1
20120099863 Xu Apr 2012 A1
20120102160 Breh Apr 2012 A1
20120106339 Mishra May 2012 A1
20120117438 Shaffer May 2012 A1
20120131097 Baykal May 2012 A1
20120131289 Taguchi May 2012 A1
20120134266 Roitshtein May 2012 A1
20120136999 Roitshtein May 2012 A1
20120147740 Nakash Jun 2012 A1
20120158997 Hsu Jun 2012 A1
20120163164 Terry Jun 2012 A1
20120170491 Kern Jul 2012 A1
20120177039 Berman Jul 2012 A1
20120210416 Mihelich Aug 2012 A1
20120221636 Surtani Aug 2012 A1
20120230225 Matthews Sep 2012 A1
20120239918 Huang Sep 2012 A1
20120243539 Keesara Sep 2012 A1
20120250502 Brolin Oct 2012 A1
20120260079 Mruthyunjaya Oct 2012 A1
20120275297 Subramanian Nov 2012 A1
20120275347 Banerjee Nov 2012 A1
20120278804 Narayanasamy Nov 2012 A1
20120287785 Kamble Nov 2012 A1
20120294192 Masood Nov 2012 A1
20120294194 Balasubramanian Nov 2012 A1
20120320800 Kamble Dec 2012 A1
20120320926 Kamath et al. Dec 2012 A1
20120327766 Tsai et al. Dec 2012 A1
20120327937 Melman et al. Dec 2012 A1
20130003535 Sarwar Jan 2013 A1
20130003549 Matthews Jan 2013 A1
20130003737 Sinicrope Jan 2013 A1
20130003738 Koganti Jan 2013 A1
20130028072 Addanki Jan 2013 A1
20130034015 Jaiswal Feb 2013 A1
20130034021 Jaiswal Feb 2013 A1
20130066947 Ahmad Mar 2013 A1
20130067466 Combs Mar 2013 A1
20130070762 Adams Mar 2013 A1
20130083693 Himura Apr 2013 A1
20130097345 Munoz Apr 2013 A1
20130114595 Mack-Crane et al. May 2013 A1
20130124707 Ananthapadmanabha May 2013 A1
20130127848 Joshi May 2013 A1
20130132296 Koppenhagen May 2013 A1
20130135811 Dunwoody May 2013 A1
20130136123 Ge May 2013 A1
20130148546 Eisenhauer Jun 2013 A1
20130156425 Kirkpatrick Jun 2013 A1
20130194914 Agarwal Aug 2013 A1
20130219473 Schaefer Aug 2013 A1
20130223221 Xu Aug 2013 A1
20130223449 Koganti Aug 2013 A1
20130250951 Koganti Sep 2013 A1
20130250958 Watanabe Sep 2013 A1
20130259037 Natarajan Oct 2013 A1
20130266015 Qu Oct 2013 A1
20130268590 Mahadevan Oct 2013 A1
20130272135 Leong Oct 2013 A1
20130294451 Li Nov 2013 A1
20130297757 Han Nov 2013 A1
20130301425 Udutha Nov 2013 A1
20130301642 Radhakrishnan Nov 2013 A1
20130308492 Baphna Nov 2013 A1
20130308647 Rosset Nov 2013 A1
20130315125 Ravishankar Nov 2013 A1
20130315586 Kipp Nov 2013 A1
20130322427 Stiekes Dec 2013 A1
20130332660 Talagala Dec 2013 A1
20130336104 Talla Dec 2013 A1
20130346583 Low Dec 2013 A1
20140013324 Zhang Jan 2014 A1
20140019608 Kawakami Jan 2014 A1
20140025736 Wang Jan 2014 A1
20140044126 Sabhanatarajan Feb 2014 A1
20140050223 Foo Feb 2014 A1
20140056298 Vobbilisetty Feb 2014 A1
20140059225 Gasparakis Feb 2014 A1
20140064056 Sakata Mar 2014 A1
20140086253 Yong Mar 2014 A1
20140105034 Sun Apr 2014 A1
20140157251 Hocker Jun 2014 A1
20140241147 Rajagopalan Aug 2014 A1
20140258446 Bursell Sep 2014 A1
20140269720 Srinivasan Sep 2014 A1
20140269733 Venkatesh Sep 2014 A1
20140298091 Carlen Oct 2014 A1
20140355477 Velayudhan Dec 2014 A1
20150010007 Matsuhira Jan 2015 A1
20150030031 Zhou Jan 2015 A1
20150110487 Fenkes Apr 2015 A1
20150127618 Alberti May 2015 A1
20150143369 Zheng May 2015 A1
20150172098 Agarwal Jun 2015 A1
20150188753 Anumala Jul 2015 A1
20150195093 Mahadevan et al. Jul 2015 A1
20150222506 Kizhakkiniyil Aug 2015 A1
20150248298 Gavrilov Sep 2015 A1
20150263991 MacChiano Sep 2015 A1
20150281066 Koley Oct 2015 A1
20150301901 Rath Oct 2015 A1
20150347468 Bester Dec 2015 A1
20160072899 Tung Mar 2016 A1
20170026197 Venkatesh Jan 2017 A1
Foreign Referenced Citations (35)
Number Date Country
1735062 Feb 2006 CN
101064682 Oct 2007 CN
101459618 Jun 2009 CN
101471899 Jul 2009 CN
101548511 Sep 2009 CN
101645880 Feb 2010 CN
102098237 Jun 2011 CN
102148749 Aug 2011 CN
102301663 Dec 2011 CN
102349268 Feb 2012 CN
102378176 Mar 2012 CN
102415065 Apr 2012 CN
102415065 Apr 2012 CN
102801599 Nov 2012 CN
102801599 Nov 2012 CN
102088388 Apr 2014 CN
0579567 May 1993 EP
0579567 Jan 1994 EP
0993156 Apr 2000 EP
0993156 Dec 2000 EP
1398920 Mar 2004 EP
1398920 Mar 2004 EP
1916807 Apr 2008 EP
2001167 Oct 2008 EP
2874359 May 2015 EP
2008056838 May 2008 WO
2009042919 Apr 2009 WO
2010111142 Sep 2010 WO
2010111142 Sep 2010 WO
2011132568 Oct 2011 WO
2011140028 Nov 2011 WO
2011140028 Nov 2011 WO
2012033663 Mar 2012 WO
2012093429 Jul 2012 WO
2014031781 Feb 2014 WO
Non-Patent Literature Citations (229)
Entry
Office Action dated Jun. 18, 215, U.S. Appl. No. 13/098,490, filed May 2, 2011.
Office Action dated Jun. 16, 2015, U.S. Appl. No. 13/048,817, filed Mar. 15, 2011.
Office Action dated Jul. 31, 2015, U.S. Appl. No. 13/598,204, filed Aug. 29, 2014.
Office Action dated Jul. 31, 2015, U.S. Appl. No. 14/473,941, filed Aug. 29, 2014.
Office Action dated Jul. 31, 2015, U.S. Appl. No. 14/488,173, filed Sep. 16, 2014.
Office Action dated Aug. 21, 2015, U.S. Appl. No. 13/776,217, filed Feb. 25, 2013.
Office Action dated Aug. 19, 2015, U.S. Appl. No. 14/156,374, filed Jan. 15, 2014.
Office Action dated Sep. 2, 2015, U.S. Appl. No. 14/151,693, filed Jan. 9, 2014.
Office Action dated Sep. 17, 2015, U.S. Appl. No. 14/577,785, filed Dec. 19, 2014.
Office Action dated Sep. 22, 2015 U.S. Appl. No. 13/656,438, filed Oct. 19, 2012.
Office Action dated Nov. 5, 2015, U.S. Appl. No. 14/178,042, filed Feb. 11, 2014.
Office Action dated Oct. 19, 2015, U.S. Appl. No. 14/215,996, filed Mar. 17, 2014.
Office Action dated Sep. 18, 2015, U.S. Appl. No. 13/345,566, filed Jan. 6, 2012.
Open Flow Switch Specification Version 1.1.0, Feb. 28, 2011.
Open Flow Switch Specification Version 1.0.0, Dec. 31, 2009.
Open Flow Configuration and Management Protocol 1.0 (OF-Config 1.0) Dec. 23, 2011.
Open Flow Switch Specification Version 1.2 Dec. 5, 2011.
Office action dated Feb. 2, 2016, U.S. Appl. No. 13/092,460, filed Apr. 22, 2011.
Office Action dated Feb. 2, 2016. U.S. Appl. No. 14/154,106, filed Jan. 13, 2014.
Office Action dated Feb. 3, 2016, U.S. Appl. No. 13/098,490, filed May 2, 2011.
Office Action dated Feb. 4, 2016, U.S. Appl. No. 13/557,105, filed Jul. 24, 2012.
Office Action dated Feb. 11, 2016, U.S. Appl. No. 14/488,173, filed Sep. 16, 2014.
Office Action dated Feb. 24, 2016, U.S. Appl. No. 13/971,397, filed Aug. 20, 2013.
Office Action dated Feb. 24, 2016, U.S. Appl. No. 12/705,508, filed Feb. 12, 2010.
Eastlake, D. et al., ‘RBridges: TRILL Header Options’, Dec. 24, 2009, pp. 1-17, TRILL Working Group.
Perlman, Radia et al., ‘RBridge VLAN Mapping’, TRILL Working Group, Dec. 4, 2009, pp. 1-12.
Touch, J. et al., ‘Transparent Interconnection of Lots of Links (TRILL): Problem and Applicability Statement’, May 2009, Network Working Group, pp. 1-17.
‘An Introduction to Brocade VCS Fabric Technology’, Brocade white paper, http://community.brocade.com/docs/DOC-2954, Dec. 3, 2012.
‘Switched Virtual Networks. Internetworking Moves Beyond Bridges and Routers’ Data Communications, McGraw Hill. New York, US, vol. 23, No. 12, Sep. 1, 1994 (Sep. 1, 1994), pp. 66-70,72,74, XP000462385 ISSN: 0363-6399.
Brocade, ‘Brocade Fabrics OS (FOS) 6.2 Virtual Fabrics Feature Frequently Asked Questions’, pp. 1-6, 2009 Brocade Communications Systems, Inc.
Brocade, ‘FastIron and TurboIron 24x Configuration Guide’, Feb. 16, 2010.
Brocade, ‘The Effortless Network: Hyperedge Technology for the Campus LAN’ 2012.
Brocade ‘Brocade Unveils ‘The Effortless Network’’, http://newsroom.brocade.com/press-releases/brocade-unveils-the-effortless-network-nasdaq-brcd-0859535, 2012.
Christensen, M. et al., ‘Considerations for Internet Group Management Protocol (IGMP) and Multicast Listener Discovery (MLD) Snooping Switches’, May 2006.
FastIron Configuration Guide Supporting Ironware Software Release 07.0.00, Dec. 18, 2009.
Foundary FastIron Configuration Guide, Software Release FSX 04.2.00b, Software Release FWS 04.3.00, Software Release FGS 05.0.00a, Sep. 2008.
Knight, ‘Network Based IP VPN Architecture using Virtual Routers’, May 2003.
Knight P et al: ‘Layer 2 and 3 Virtual Private Networks: Taxonomy, Technology, and Standardization Efforts’, IEEE Communications Magazine, IEEE Service Center, Piscataway, US, vol. 42, No. 6, Jun. 1, 2004 (Jun. 1, 2004), pp. 124-131, XP001198207, ISSN: 0163-6804, DOI: 10.1109/MCOM.2004.1304248.
Knight S et al: ‘Virtual Router Redundancy Protocol’ Internet Citation Apr. 1, 1998 (Apr. 1, 1998), XP002135272 Retrieved from the Internet: URL:ftp://ftp.isi.edu/in-notes/rfc2338.txt [retrieved on Apr. 10, 2000].
Kreeger, L. et al., ‘Network Virtualization Overlay Control Protocol Requirements draft-kreeger-nvo3-overlay-cp-00’, Jan. 30, 2012.
Lapuh, Roger et al., ‘Split Multi-link Trunking (SMLT)’, draft-lapuh-network-smlt-08, Jul. 2008.
Lapuh, Roger et al., ‘Split Multi-Link Trunking (SMLT)’, Network Working Group, Oct. 2012.
Louati, Wajdi et al., ‘Network-based virtual personal overlay networks using programmable virtual routers’, IEEE Communications Magazine, Jul. 2005.
Narten, T. et al., ‘Problem Statement: Overlays for Network Virtualization d raft-na rten-n vo3-over l ay-problem-statement-01’, Oct. 31, 2011.
Office Action for U.S. Appl. No. 13/092,752, filed Apr. 22, 2011, from Park, Jung H., dated Jul. 18, 2013.
Office Action for U.S. Appl. No. 13/351,513, filed Jan. 17, 2012, dated Jul. 24, 2014.
Office Action for U.S. Appl. No. 13/365,993, filed Feb. 3, 2012, from Cho, Hong Sol., dated Jul. 23, 2013.
Office Action for U.S. Appl. No. 13/742,207 dated Jul. 24, 2014, filed Jan. 15, 2013.
Office Action for U.S. Appl. No. 12/725,249, filed Mar. 16, 2010, dated Apr. 26, 2013.
Office Action for U.S. Appl. No. 12/950,968, filed Nov. 19, 2010, dated Jan. 4, 2013.
Office Action for U.S. Appl. No. 12/950,968, filed Nov. 19, 2010, dated Jun. 7, 2012.
Office Action for U.S. Appl. No. 12/950,974, filed Nov. 19, 2010, dated Dec. 20, 2012.
Office Action for U.S. Appl. No. 12/950,974, filed Nov. 19, 2010, dated May 24, 2012.
Office Action for U.S. Appl. No. 13/030,688, filed Feb. 18, 2011, dated Apr. 25, 2013.
Office Action for U.S. Appl. No. 13/030,806, filed Feb. 18, 2011, dated Dec. 3, 2012.
Office Action for U.S. Appl. No. 13/030,806, filed Feb. 18, 2011, dated Jun. 11, 2013.
Office Action for U.S. Appl. No. 13/042,259, filed Mar. 7, 2011, dated Mar. 18, 2013.
Office Action for U.S. Appl. No. 13/042,259, filed Mar. 7, 2011, dated Jul. 31, 2013.
Office Action for U.S. Appl. No. 13/044,301, filed Mar. 9, 2011, dated Feb. 22, 2013.
Office Action for U.S. Appl. No. 13/044,301, filed Mar. 9, 2011, dated Jun. 11, 2013.
Office Action for U.S. Appl. No. 13/044,326, filed Mar. 9, 2011, dated Oct. 2, 2013.
Office Action for U.S. Appl. No. 13/050,102, filed Mar. 17, 2011, dated Oct. 26, 2012.
Office Action for U.S. Appl. No. 13/050,102, filed Mar. 17, 2011, dated May 16, 2013.
Office Action for U.S. Appl. No. 13/087,239, filed Apr. 14, 2011, dated May 22, 2013.
Office Action for U.S. Appl. No. 13/092,460, filed Apr. 22, 2011, dated Jun. 21, 2013.
Office Action for U.S. Appl. No. 13/092,580, filed Apr. 22, 2011, dated Jun. 10, 2013.
Office Action for U.S. Appl. No. 13/092,701, filed Apr. 22, 2011, dated Jan. 28, 2013.
Office Action for U.S. Appl. No. 13/092,701, filed Apr. 22, 2011, dated Jul. 3, 2013.
Office Action for U.S. Appl. No. 13/092,724, filed Apr. 22, 2011, dated Feb. 5, 2013.
Office Action for U.S. Appl. No. 13/092,724, filed Apr. 22, 2011, dated Jul. 16, 2013.
Office Action for U.S. Appl. No. 13/092,752, filed Apr. 22, 2011, dated Feb. 5, 2013.
Office Action for U.S. Appl. No. 13/092,864, filed Apr. 22, 2011, dated Sep. 19, 2012.
Office Action for U.S. Appl. No. 13/092,873, filed Apr. 22, 2011, dated Jun. 19, 2013.
Office Action for U.S. Appl. No. 13/092,877, filed Apr. 22, 2011, dated Mar. 4, 2013.
Office Action for U.S. Appl. No. 13/092,877, filed Apr. 22, 2011, dated Sep. 5, 2013.
Office Action for U.S. Appl. No. 13/098,360, filed Apr. 29, 2011, dated May 31, 2013.
Office Action for U.S. Appl. No. 13/098,490, filed May 2, 2011, dated Jul. 9, 2013.
Office Action for U.S. Appl. No. 13/184,526, filed Jul. 16, 2011, dated Jan. 28, 2013.
Office Action for U.S. Appl. No. 13/184,526, filed Jul. 16, 2011, dated May 22, 2013.
Office Action for U.S. Appl. No. 13/365,808, filed Jul. 18, 2013, dated Jul. 18, 2013.
Perlman, Radia et al., ‘Challenges and Opportunities in the Design of TRILL: a Routed layer 2 Technology’, 2009.
Perlman, Radia et al., ‘RBridges: Base Protocol Specification; Draft-ietf-trill-rbridge-protocol-16.txt’, Mar. 3, 2010, pp. 1-117.
S. Nadas et al., ‘Virtual Router Redundancy Protocol (VRRP) Version 3 for IPv4 IPv6’, Internet Engineering Task Force, Mar. 2010.
TRILL Working Group Internet-Draft Intended status: Proposed Standard RBridges: Base Protocol Specificaiton Mar. 3, 2010.
‘RBridges: Base Protocol Specification’, IETF Draft, Perlman et al., Jun. 26, 2009.
U.S. Appl. No. 13/030,806 Office Action dated Dec. 3, 2012.
Office action dated Apr. 26, 2012, U.S. Appl. No. 12/725,249, filed Mar. 16, 2010.
Office action dated Sep. 12, 2012, U.S. Appl. No. 12/725,249, filed Mar. 16, 2010.
Office action dated Dec. 21, 2012, U.S. Appl. No. 13/098,490, filed May 2, 2011.
Office action dated Mar. 27, 2014, U.S. Appl. No. 13/098,490, filed May 2, 2011.
Office action dated Jul. 9, 2013, U.S. Appl. No. 13/098,490, filed May 2, 2011.
Office action dated May 22, 2013, U.S. Appl. No. 13/087,239, filed Apr. 14, 2011.
Office action dated Dec. 5, 2012, U.S. Appl. No. 13/087,239, filed Apr. 14, 2011.
Office action dated Apr. 9, 2014, U.S. Appl. No. 13/092,724, filed Apr. 22, 2011.
Office action dated Feb. 5, 2013, U.S. Appl. No. 13/092,724, filed Apr. 22, 2011.
Office action dated Jan. 10, 2014, U.S. Appl. No. 13/092,580, filed Apr. 22, 2011.
Office action dated Jun. 10, 2013, U.S. Appl. No. 13/092,580, filed Apr. 22, 2011.
Office action dated Jan. 16, 2014, U.S. Appl. No. 13/042,259, filed Mar. 7, 2011.
Office action dated Mar. 18, 2013, U.S. Appl. No. 13/042,259, filed Mar. 7, 2011.
Office action dated Jul. 31, 2013, U.S. Appl. No. 13/042,259, filed Mar. 7, 2011.
Office action dated Aug. 29, 2014, U.S. Appl. No. 13/042,259, filed Mar. 7, 2011.
Office action dated Mar. 14, 2014, U.S. Appl. No. 13/092,460, filed Apr. 22, 2011.
Office action dated Jun. 21, 2013, U.S. Appl. No. 13/092,460, filed Apr. 22, 2011.
Office action dated Jan. 28, 2013, U.S. Appl. No. 13/092,701, Apr. 22, 2011.
Office Action dated Mar. 26, 2014, U.S. Appl. No. 13/092,701, filed Apr. 22, 2011.
Office action dated Jul. 3, 2013, U.S. Appl. No. 13/092,701, filed Apr. 22, 2011.
Office Action dated Apr. 9, 2014, U.S. Appl. No. 13/092,752, filed Apr. 22, 2011.
Office action dated Jul. 18, 2013, U.S. Appl. No. 13/092,752, filed Apr. 22, 2011.
Office action dated Dec. 20, 2012, U.S. Appl. No. 12/950,974, filed Nov. 19, 2010.
Office action dated May 24, 2012, U.S. Appl. No. 12/950,974, filed Nov. 19, 2010.
Office action dated Jan. 6, 2014, U.S. Appl. No. 13/092,877, filed Apr. 22, 2011.
Office action dated Sep. 5, 2013, U.S. Appl. No. 13/092,877, filed Apr. 22, 2011.
Office action dated Mar. 4, 2013, U.S. Appl. No. 13/092,877, filed Apr. 22, 2011.
Office action dated Jan. 4, 2013, U.S. Appl. No. 12/950,968, filed Nov. 19, 2010.
Office action dated Jun. 7, 2012, U.S. Appl. No. 12/950,968, filed Nov. 19, 2010.
Office action dated Sep. 19, 2012, U.S. Appl. No. 13/092,864, filed Apr. 22, 2011.
Office action dated May 31, 2013, U.S. Appl. No. 13/098,360, filed Apr. 29, 2011.
Office action dated Oct. 2, 2013, U.S. Appl. No. 13/044,326, filed Mar. 9, 2011.
Office action dated Dec. 3, 2012, U.S. Appl. No. 13/030,806, Feb. 18, 2011.
Office action dated Apr. 22, 2014, U.S. Appl. No. 13/030,806, Feb. 18, 2011.
Office action dated Jun. 11, 2013, U.S. Appl. No. 13/030,806, filed Feb. 18, 2011.
Office action dated Apr. 25, 2013, U.S. Appl. No. 13/030,688, filed Feb. 18, 2011.
Office action dated Feb. 22, 2013, U.S. Appl. No. 13/044,301, filed Mar. 9, 2011.
Office action dated Jun. 11, 2013, U.S. Appl. No. 13/044,301, filed Mar. 9, 2011.
Office action dated Oct. 26, 2012, U.S. Appl. No. 13/050,102, filed Mar. 17, 2011.
Office action dated May 16, 2013, U.S. Appl. No. 13/050,102, filed Mar. 17, 2011.
Office action dated Aug. 4, 2014, U.S. Appl. No. 13/050,102, filed Mar. 17, 2011.
Office action dated Jan. 28, 2013, U.S. Appl. No. 13/148,526, filed Jul. 16, 2011.
Office action dated Dec. 2, 2013, U.S. Appl. No. 13/184,526, filed Jul. 16, 2011.
Office action dated May 22, 2013, U.S. Appl. No. 13/148,526, filed Jul. 16, 2011.
Office action dated Aug. 21, 2014, U.S. Appl. No. 13/184,526, filed Jul. 16, 2011.
Office action dated Nov. 29, 2013, U.S. Appl. No. 13/092,873, filed Apr. 22, 2011.
Office action dated Jun. 19, 2013, U.S. Appl. No. 13/092,873, filed Apr. 22, 2011.
Office action dated Jul. 18, 2013, U.S. Appl. No. 13/365,808, filed Feb. 3, 2012.
Office Action dated Mar. 6, 2014, U.S. Appl. No. 13/425,238, filed Mar. 20, 2012.
Office action dated Nov. 12, 2013, U.S. Appl. No. 13/312,903, filed Dec. 6, 2011.
Office action dated Jun. 13, 2013, U.S. Appl. No. 13/312,903, filed Dec. 6, 2011.
Office Action dated Jun. 18, 2014, U.S. Appl. No. 13/440,861, filed Apr. 5, 2012.
Office Action dated Feb. 28, 2014, U.S. Appl. No. 13/351,513, filed Jan. 17, 2012.
Office Action dated May 9, 2014, U.S. Appl. No. 13/484,072, filed May 30, 2012.
Office Action dated May 14, 2014, U.S. Appl. No. 13/533,843, filed Jun. 26, 2012.
Office Action dated Feb. 20, 2014, U.S. Appl. No. 13/598,204, filed Aug. 29, 2012.
Office Action dated Jun. 6, 2014, U.S. Appl. No. 13/669,357, filed Nov. 5, 2012.
Brocade ‘An Introduction to Brocade VCS Fabric Technology’, Dec. 3, 2012.
Huang, Nen-Fu et al., ‘An Effective Spanning Tree Algorithm for a Bridged LAN’, Mar. 16, 1992.
Lapuh, Roger et al., ‘Split Multi-link Trunking (SMLT) draft-lapuh-network-smlt-08’, Jan. 2009.
Office Action for U.S. Appl. No. 13/030,688, filed Feb. 18, 2011, dated Jul. 17, 2014.
Office Action for U.S. Appl. No. 13/042,259, filed Mar. 7, 2011, from Jaroenchonwanit, Bunjob, dated Jan. 16, 2014.
Office Action for U.S. Appl. No. 13/044,326, filed Mar. 9, 2011, dated Jul. 7, 2014.
Office Action for U.S. Appl. No. 13/092,752, filed Apr. 22, 2011, dated Apr. 9, 2014.
Office Action for U.S. Appl. No. 13/092,873, filed Apr. 22, 2011, dated Jul. 25, 2014.
Office Action for U.S. Appl. No. 13/092,877, filed Apr. 22, 2011, dated Jun. 20, 2014.
Office Action for U.S. Appl. No. 13/312,903, filed Dec. 6, 2011, dated Aug. 7, 2014.
Office Action for U.S. Appl. No. 13/425,238, filed Mar. 20, 2012, dated Mar. 6, 2014.
Office Action for U.S. Appl. No. 13/556,061, filed Jul. 23, 2012, dated Jun. 6, 2014.
Office Action for U.S. Appl. No. 13/950,974, filed Nov. 19, 2010, from Haile, Awet A., dated Dec. 2, 2012.
Office Action for U.S. Appl. No. 12/725,249, filed Mar. 16, 2010, dated Sep. 12, 2012.
Office Action for U.S. Appl. No. 13/098,490, filed May 2, 2011, dated Dec. 21, 2012.
Office Action for U.S. Appl. No. 13/098,490, filed May 2, 2011, dated Mar. 27, 2014.
Office Action for U.S. Appl. No. 13/312,903, filed Dec. 6, 2011, dated Jun. 13, 2013.
Office Action for U.S. Appl. No. 13/087,239, filed Apr. 14, 2011, dated Dec. 5, 2012.
Office Action for U.S. Appl. No. 13/092,873, filed Apr. 22, 2011, dated Nov. 29, 2013.
Office Action for U.S. Appl. No. 13/184,526, filed Jul. 16, 2011, dated Dec. 2, 2013.
Office Action for U.S. Appl. No. 13/351,513, filed Jan. 17, 2012, dated Feb. 28, 2014.
Office Action for U.S. Appl. No. 13/533,843, filed Jun. 26, 2012, dated Oct. 21, 2013.
Office Action for U.S. Appl. No. 13/598,204, filed Aug. 29, 2012, dated Feb. 20, 2014.
Office Action for U.S. Appl. No. 13/092,887, dated Jan. 6, 2014.
Perlman R: ‘Challenges and opportunities in the design of TRILL: a routed layer 2 technology’, 2009 IEEE GLOBECOM Workshops, Honolulu, HI, USA, Piscataway, NJ, USA, Nov. 30, 2009 (Nov. 30, 2009), pp. 1-6, XP002649647, DOI: 10.1109/GLOBECOM.2009.5360776 ISBN: 1-4244-5626-0 [retrieved on Jul. 19, 2011].
Zhai F. Hu et al. ‘RBridge: Pseudo-Nickname; draft-hu-trill-pseudonode-nickname-02.txt’, May 15, 2012.
Office action dated Aug. 14, 2014, U.S. Appl. No. 13/092,460, filed Apr. 22, 2011.
Office action dated Jul. 7, 2014, for U.S. Appl. No. 13/044,326, filed Mar. 9, 2011.
Office Action dated Dec. 19, 2014, for U.S. Appl. No. 13/044,326, filed Mar. 9, 2011.
Office Action for U.S. Appl. No. 13/092,873, filed Apr. 22, 2011, dated Nov. 7, 2014.
Office Action for U.S. Appl. No. 13/092,877, filed Apr. 22, 2011, dated Nov. 10, 2014.
Office Action for U.S. Appl. No. 13/157,942, filed Jun. 10, 2011.
Mckeown, Nick et al. “OpenFlow: Enabling Innovation in Campus Networks”, Mar. 14, 2008, www.openflow.org/documents/openflow-wp-latest.pdf.
Office Action for U.S. Appl. No. 13/044,301, dated Mar. 9, 2011.
Office Action for U.S. Appl. No. 13/184,526, filed Jul. 16, 2011, dated Jan. 5, 2015.
Office Action for U.S. Appl. No. 13/598,204, filed Aug. 29, 2012, dated Jan. 5, 2015.
Office Action for U.S. Appl. No. 13/669,357, filed Nov. 5, 2012, dated Jan. 30, 2015.
Office Action for U.S. Appl. No. 13/851,026, filed Mar. 26, 2013, dated Jan. 30, 2015.
Office Action for U.S. Appl. No. 13/786,328, filed Mar. 5, 2013, dated Mar. 13, 2015.
Office Action for U.S. Appl. No. 13/092,460, filed Apr. 22, 2011, dated Mar. 13, 2015.
Office Action for U.S. Appl. No. 13/425,238, dated Mar. 12, 2015.
Office Action for U.S. Appl. No. 13/092,752, filed Apr. 22, 2011, dated Feb. 27, 2015.
Office Action for U.S. Appl. No. 13/042,259, filed Mar. 7, 2011, dated Feb. 23, 2015.
Office Action for U.S. Appl. No. 13/044,301, filed Mar. 9, 2011, dated Jan. 29, 2015.
Office Action for U.S. Appl. No. 13/050,102, filed Mar. 17, 2011, dated Jan. 26, 2015.
Office action dated Oct. 2, 2014, for U.S. Appl. No. 13/092,752, filed Apr. 22, 2011.
Kompella, Ed K. et al., ‘Virtual Private LAN Service (VPLS) Using BGP for Auto-Discovery and Signaling’ Jan. 2007.
Rosen, E. et al., “BGP/MPLS VPNs”, Mar. 1999.
Office Action for U.S. Appl. No. 14/577,785, filed Dec. 19, 2014, dated Apr. 13, 2015.
Office Action for U.S. Appl. No. 13/425,238, filed Mar. 20, 2012, dated Mar. 12, 2015.
Abawajy J. “An Approach to Support a Single Service Provider Address Image for Wide Area Networks Environment” Centre for Parallel and Distributed Computing, School of Computer Science Carleton University, Ottawa, Ontario, K1S 5B6, Canada.
Mahalingam “VXLAN: A Framework for Overlaying Virtualized Layer 2 Networks over Layer 3 Networks” Oct. 17, 2013 pp. 1-22, Sections 1, 4 and 4.1.
Office action dated Apr. 30, 2015, U.S. Appl. No. 13/351,513, filed Jan. 17, 2012.
Office Action dated Apr. 1, 2015, U.S. Appl. No. 13/656,438, filed Oct. 19, 2012.
Office Action dated May 21, 2015, U.S. Appl. No. 13/288,822, filed Nov. 3, 2011.
Siamak Azodolmolky et al. “Cloud computing networking: Challenges and opportunities for innovations”, IEEE Communications Magazine, vol. 51, No. 7, Jul. 1, 2013.
Office Action dated Apr. 1, 2015 U.S. Appl. No. 13/656,438, filed Oct. 19, 2012.
Office action dated Jun. 8, 2015, U.S. Appl. No. 14/178,042, filed Feb. 11, 2014.
Office Action Dated Jun. 10, 2015, U.S. Appl. No. 13/890,150, filed May 8, 2013.
Office Action for U.S. Appl. No. 14/662,095, dated Mar. 24, 2017.
Office Action for U.S. Appl. No. 15/005,967, dated Mar. 31, 2017.
Office Action for U.S. Appl. No. 15/215,377, dated Apr. 7, 2017.
Office Action for U.S. Appl. No. 13/098,490, dated Apr. 6, 2017.
Office Action for U.S. Appl. No. 14/662,092, dated Mar. 29, 2017.
“Network based IP VPN Architecture using Virtual Routers” Paul Knight et al.
Yang Yu et al “A Framework of using OpenFlow to handle transient link failure”, TMEE, 2011 International Conference on, IEEE, Dec. 16, 2011.
Office Action for U.S. Appl. No. 15/227,789, dated Feb. 27, 2017.
Office Action for U.S. Appl. No. 14/822,380, dated Feb. 8, 2017.
Office Action for U.S. Appl. No. 14/704,660, dated Feb. 27, 2017.
Office Action for U.S. Appl. No. 14/510,913, dated Mar. 3, 2017.
Office Action for U.S. Appl. No. 14/473,941, dated Feb. 8, 2017.
Office Action for U.S. Appl. No. 14/329,447, dated Feb. 10, 2017.
Office Action dated Jan. 31, 2017, U.S. Appl. No. 13/184,526, filed Jul. 16, 2011.
Office Action dated Jan. 27, 2017, U.S. Appl. No. 14/216,292, filed Mar. 17, 2014.
Office Action dated Jan. 26, 2017, U.S. Appl. No. 13/786,328, filed Mar. 5, 2013.
Office Action dated Dec. 2, 2016, U.S. Appl. No. 14/512,268, filed Oct. 10, 2014.
Office Action dated Dec. 1, 2016, U.S. Appl. No. 13/899,849, filed May 22, 2013.
Office Action dated Dec. 1, 2016, U.S. Appl. No. 13/656,438, filed Oct. 19, 2012.
Office Action dated Nov. 30, 2016, U.S. Appl. No. 13/598,204, filed Aug. 29, 2012.
Office Action dated Nov. 21, 2016, U.S. Appl. No. 13/669,357, filed Nov. 5, 2012.
Office Action dated Feb. 8, 2017, U.S. Appl. No. 14/473,941, filed Aug. 29, 2014.
Office Action dated Feb. 8, 2017, U.S. Appl. No. 14/822,380, filed Aug. 10, 2015.
Office Action dated Jul. 6, 2016, U.S. Appl. No. 14/618,941, filed Feb. 10, 2015.
Office Action dated Jul. 20, 2016, U.S. Appl. No. 14/510,913, filed Oct. 9, 2014.
Office Action dated Jul. 29, 2016, U.S. Appl. No. 14/473,941, filed Aug. 29, 2014.
Office Action dated Jul. 28, 2016, U.S. Appl. No. 14/284,212, filed May 21, 2016.
Related Publications (1)
Number Date Country
20160277214 A1 Sep 2016 US