Integrating overlay logical networks into cloud provider networks introduces new challenges not found in either system alone. For example, virtual private clouds (VPCs) provided by some cloud providers do not allow for transitive routing using the cloud-provided network making it difficult to route traffic requiring a service provided at a service VPC to the service VPC as part of a route from a source VPC to a destination VPC. Additionally, some services provided by a cloud provider rely on the use of the cloud-provider assigned addresses and don't integrate well with compute nodes assigned addresses in an overlay logical network.
Some embodiments provide a centralized overlay-network cloud gateway and a set of centralized services in a transit virtual cloud network (VCN) connected to multiple other compute VCNs hosting compute nodes (VMs, containers, etc.) that are part of (belong to) the overlay network. The centralized overlay-network cloud gateway executing in the transit VCN, in some embodiments, is used in place of multiple overlay-network cloud gateways executing in multiple non-transit VCNs to reduce the cost and management overhead. The centralized overlay-network cloud gateway provides connectivity between compute nodes of the overlay network (e.g., a logical network spanning multiple VCNs) and compute nodes in external networks. In some embodiments, the centralized overlay-network cloud gateway also provides services for data messages it processes. The services, in some embodiments, are provided by some combination of the cloud gateway and third party services accessed through or by the gateway.
Some embodiments use the centralized overlay-network cloud gateway to provide transitive routing (e.g., routing through a transit VPC) in the absence of direct peering between source and destination VPCs. The cloud gateway, in some embodiments, is accessed from each VPC through a peering relationship established between the transit VPC and the VPC. In some embodiments, a managed forwarding element (MFE) implementing a logical routing element (LRE) that connects the logical switching elements of the logical network to a centralized logical routing element implemented by the cloud gateway performs a set of forwarding operations to direct a data message to the transit VPC. The overlay network, of some embodiments, uses the same subnetting and default gateway address for each compute node as the cloud provider network provided by the virtual private cloud provider. Using the same address space as the cloud provider network allows seamless integration of services provided by the cloud provider network and with the novel method of forwarding data messages disclosed herein also allows seamless integration with services provided by an overlay network.
In embodiments using the same address space in both the cloud provider and overlay networks, in order to direct traffic properly (i.e., either using the cloud provider network infrastructure or through tunnels used to implement the overlay network), some embodiments provide a method and arbiter module (e.g., as part of a managed forwarding element implementing a hybrid switching element that provides logical forwarding for both the overlay and cloud provider networks) for determining the proper method for forwarding data messages sent by the compute nodes. In some embodiments, the arbiter module determines whether a destination compute node is in a same or different VPC from the source compute node. When source and destination compute nodes are on a same VPC, the default behavior is to route using the cloud provider infrastructure. However, a user (e.g., an administrator) can override this default behavior and use the overlay network (i.e., tunneling to the cloud gateway in the transit VPC) if a centralized service is to be provided by a service engine in a transit VPC. When source and destination compute nodes are on different VPCs, the default behavior is to route the data message using the overlay network (i.e., the cloud gateway in the transit VPC). The method allows an administrator to override the default behavior for data messages (1) that do not need a service provided by a service engine in the transit VPC and (2) for which a peering (e.g., a tunnel provided by the cloud provider infrastructure) exists between a source VC and a destination VPC.
In some embodiments, the default behavior for data messages between compute nodes in VPCs connected by a cloud provider tunnel (e.g., peered VPCs) is to route the data message using the overlay logical network and the centralized overlay-network cloud gateway. In order to override the default and use the peering relationship to forward data messages between VPCs, in some embodiments, an administrator configures a policy-based routing (PBR) rule in an arbiter module of an MFE that directs the MFE to forward the data message using the peering between the VPCs. If no PBR rule or other method is used to override the default behavior, data messages for which a service provided in the transit VPC is required are routed using the overlay network based on a routing table of a distributed router implementation. In some embodiments, the routing decisions are configured by a PBR implementation that determines a next hop based on a policy that can be defined by characteristics of a data message or data message flow in addition to the destination address (e.g., a five-tuple specifying a source IP address/port number and destination IP address/port number as well as a protocol).
In some embodiments, the default behavior for data messages between compute nodes in VPCs connected by a cloud provider tunnel (e.g., peered VPCs) is to route the data message using the cloud provider tunnel. In such embodiments, data messages for which a service provided in the transit VPC is required are routed using the overlay network by having an administrator remove routes to the destination compute node (e.g., the route associated with the IP address or subnet of the destination compute node) from a routing table of a distributed router implementation. In some embodiments, the routing decisions are configured by a policy-based routing implementation that determines a next hop based on a policy that can be defined by characteristics of a data message or data message flow in addition to the destination address (e.g., a five-tuple specifying a source and destination IP and MAC address as well as a port or protocol).
The preceding Summary is intended to serve as a brief introduction to some embodiments of the invention. It is not meant to be an introduction or overview of all inventive subject matter disclosed in this document. The Detailed Description that follows and the Drawings that are referred to in the Detailed Description will further describe the embodiments described in the Summary as well as other embodiments. Accordingly, to understand all the embodiments described by this document, a full review of the Summary, Detailed Description, the Drawings and the Claims is needed. Moreover, the claimed subject matters are not to be limited by the illustrative details in the Summary, Detailed Description and the Drawing.
The novel features of the invention are set forth in the appended claims. However, for purposes of explanation, several embodiments of the invention are set forth in the following figures.
In the following detailed description of the invention, numerous details, examples, and embodiments of the invention are set forth and described. However, it will be clear and apparent to one skilled in the art that the invention is not limited to the embodiments set forth and that the invention may be practiced without some of the specific details and examples discussed.
Some embodiments provide a centralized overlay-network cloud gateway and a set of centralized services in a transit virtual cloud network (VCN) connected to multiple other compute VCNs hosting compute nodes (VMs, containers, etc.) that are part of (belong to) the overlay network. The centralized overlay-network cloud gateway executing in the transit VCN, in some embodiments, is used in place of multiple overlay-network cloud gateways executing in multiple non-transit VCNs to reduce the cost and management overhead. The centralized overlay-network cloud gateway provides connectivity between compute nodes of the overlay network (e.g., a logical network spanning multiple VCNs) and compute nodes in external networks. In some embodiments, the centralized overlay-network cloud gateway also provides services for data messages it processes. The services, in some embodiments, are provided by some combination of the cloud gateway and third party services accessed through or by the gateway.
A logical network as used in this application, refers to a particular logical abstraction of a network. In some embodiments, the logical abstraction includes logical counterparts to network elements of a physical network such as forwarding elements (e.g., switches, hubs, routers, bridges, etc.), load balancers, and firewalls. The logical forwarding elements (e.g., a logical switch or logical router) in some embodiments are implemented by a set of MFEs (e.g., physical or virtual/software switches, or routers) executing on host machines. A particular host machine may host data compute nodes (DCNs) (e.g., containers or virtual machines (VMs)) connected to multiple different logical networks and the set of MFEs implements all the logical networks to which the DCNs logically connect. Further details about logical networks and about logical networks implemented as overlay logical networks in a public cloud environment specifically can respectively be found in U.S. Pat. No. 9,787,605 and U.S. Patent publication number 2018/0062923 which are incorporated herein by reference.
The DCNs, in some embodiments, are hosted by a cloud provider (e.g., Amazon Web Services (AWS) or Microsoft Azure (Azure)) that provides VCNs (e.g., virtual private clouds (VPCs) for AWS or virtual networks (VNets) for Azure) for at least one tenant. Each VCN is a virtual network that is logically isolated from other physical networks and from other virtual networks (e.g., VCNs) provided by the cloud provider. In some embodiments, a cloud provider provides VCNs for multiple tenants with a set of tenants having multiple VCNs. For each VCN, a private range of IP addresses is specified (e.g. a range of IPv4 and/or IPv6 addresses) for communication within the VCN. Communication between compute nodes within a VCN uses the public cloud provider infrastructure and stay within the VCN at the level of the logical specification of the VCN. If a peering relationship is established between VCNs, the private IP addresses of a destination compute node in a different VCN can be used in sending a data message and the data message will not need to be processed by an internet gateway provided by the cloud provider. Instead, in some embodiments, the data message is forwarded using the same methodology as used for intra-VCN traffic.
In some embodiments, peering is implemented by the cloud provider by modifying the configuration of an underlay routing element (e.g., by adding routing entries, modifying security rules, etc.) such that compute nodes in peered VCNs communicate as if they are in a same VCN. In some embodiments, when two VCNs are not peered, the underlay routing elements enforce this by preventing direct communication between the compute nodes in the two unpeered VCNs (e.g., using security rules for preventing the private IP addresses of one VCN from communicating with the private IP addresses of the other VCN). It should also be noted that, in some embodiments, peering is not transitive and compute nodes in two VCNs peered with a common VCN (but not with each other) cannot communicate using the private IP addresses of the unpeered VCNs.
In some other embodiments, inter-VCN traffic traverses specific interfaces of the cloud provider underlay (e.g., gateway forwarding elements). That is, in such embodiments, these specific interfaces in the source and/or destination VCNs enforce the peering relationships through a set of routing tables and/or security rules. In such embodiments, the routing tables and security rules of each interface are modified to allow traffic that arrives at the cloud-provider gateway devices destined for a peered VCN to be forwarded to the cloud-provider gateway device of the peered VCN. Without peering, in such embodiments, the cloud-provider gateway devices would not forward traffic to the cloud-provider gateway device of an unpeered VCN.
A range of unique public IP addresses can additionally be specified in order to allow communications with an external network such as the internet through an internet gateway provided by the cloud provider. Communication between compute nodes in unpeered VCNs, in some embodiments, uses the internet gateway provided by the cloud provider and a public IP address associated with the public IP address of the unpeered VCN. In some embodiments, each VCN for a particular tenant has a different range of private IP addresses specified. A VCN, in some embodiments, spans multiple physical datacenters and is further broken up into availability zones that each span separate sets of physical datacenters such that no datacenter supports more than one availability zone.
By expanding a logical network into one or more public datacenters, a logical topology may be stretched across these datacenters. Some embodiments confine the VMs attached to a given logical switch to one physical datacenter, while other embodiments allow for even a single logical switch to be stretched across multiple datacenters (e.g., a set of datacenters supporting an availability zone).
A logical router, in some embodiments, comprises multiple logical routing components including a distributed logical router, a transit logical switch and a centralized (service) router. In some embodiments, a distributed logical router is implemented by a set of local MFEs (e.g., a local MFE executing in each VCN, or by each MFE implementing a logical switching element of the logical network). The distributed logical router performs routing for data message and can determine a logical switch to which the data message should be forwarded. In some instances, a transit logical switch is a logical switching element that is used only by the logical routing element to connect the distributed logical routing component to a centralized (service) logical routing component of the logical routing element. Both local and centralized MFEs, in some embodiments, implement distributed logical router components and transit logical switch components, while centralized MFEs implement the centralized (service) logical router. In some embodiments, the centralized logical router serves as a gateway to external networks for compute nodes in a tenant logical network.
The logical switches 110-125 attached to the logical router 105 are each assigned a subnet, in some embodiments, and thus the workload VMs created to attach to a particular logical switch should be assigned IP addresses in the appropriate subnet. In some embodiments, the logical topology is defined for both the overlay network and the VCN implementation such that the IP addresses (subnets and individual DCN IP addresses) assigned by the VCN can be, and are, used by the overlay logical switches and routers to allow the applications to continue operating with the IP addresses from the cloud provider, thereby facilitating easy seamless integration with other services provided by the cloud provider, such as storage or load balancing services. In some embodiments, this is accomplished by determining the cloud-provider assigned addresses using APIs provided by the cloud provider (i.e., the provider of the VCN) and using them for the overlay network as well (e.g., assigning the IP address assigned to a VM by the cloud provider to the tunnel endpoint created for the MFE operating on the VM). However, in some embodiments the IP address of the VM with regard to the cloud provider is different than the IP address of the logical port mapped to that VM, as the IP address facing the cloud provider network is that of the tunnel endpoint created for the MFE operating on the VM. For example, the logical switch 110 in some embodiments is assigned the subnet 192.168.1.0/24 in the overlay while being assigned 10.1.0.0/24 by the cloud provider. In addition, two VMs are shown attached to the logical switch 110.
As shown, each of the VMs is assigned a different workload IP (e.g., 10.1.0.2, 10.1.0.3, etc.) in the IP subnet assigned to the logical switch (e.g., 10.1.0.0/24) to which it connects, the IP subnet of the logical switch being an IP subnet of the subnet assigned to the VPC in which it is implemented (e.g., 10.1.0.0/16). These IP addresses, in some embodiments, are identical to the IP addresses assigned to the VM (or the virtual extensible local area network tunnel endpoint (VTEP) of the MFE operating on the VM) by the cloud provider. Forwarding operations performed by an MFE implementing a hybrid logical switch for such embodiments are discussed in more detail below.
In other embodiments, each of the VMs is assigned a different workload IP in a subnet of the logical network that is not the same as the logical switch subnet assigned by the cloud provider (e.g., the 192.168.1.0/24 subnet instead of the 10.1.1.0/24 subnet). When the workload application in these embodiments sends a packet, this different workload IP address will be the source IP used in the header of that packet. However, the MFEs operating on these VMs have VTEPs with different IP addresses on the VPC subnets (10.1.0.2, 10.1.0.3, 10.10.0.2, 10.10.0.3, etc.). The packets that exit the VM will thus be encapsulated using this VTEP IP address as the source IP address (after logical processing is performed by the MFE in the source VM), in order to be sent to other destinations in the VPC.
The figure also illustrates tunnels between these four MFEs and a cloud gateway 201 in the transit VPC 205. The tunnels in some embodiments use a peering relationship between compute VPC 210 and transit VPC 205 illustrated as network connection 250. The cloud gateway 201 provides connectivity for the overlay logical network to external network 260 and on-premises datacenter 270. As shown, the cloud gateway 201 connects to external network 260 through internet gateway 203. Internet gateway 203 is provided, in some embodiments, by the cloud provider as part of the deployment of transit VPC 205. The tunnels connecting the MFEs to the cloud gateway 201 pass through the underlying network of the public cloud provider (referred to herein as the “underlay”). In addition, though not shown here for simplicity, tunnels are created (through the underlay network) between each pair of the MFEs operating on the VMs 230. In some embodiments, the data message passes through additional managed forwarding elements in one of the compute VPC and the transit VPC or in both the compute and transit VPC between the illustrated MFE and the cloud gateway. In some embodiments, the intermediate MFEs implement additional intermediary logical forwarding elements.
The cloud gateway 201 can also send packets to (and receive packets from) destinations within an on-premises private datacenter 270. To send these packets, the cloud gateway 201 encapsulates the packets using its VTEP IP, so that the destination will identify the incoming packet as a logical network packet. To secure traffic between the cloud gateway 201 in the VPC 205 and the destinations in the private datacenter (e.g., VMs attached to a logical switch), the packets are sent via a VPN tunnel 204 in some embodiments. In this example, the components of the cloud gateway device 201 are not shown but are discussed in more detail in U.S. Patent publication number 2018/0062923 which is incorporated herein by reference.
Cloud gateway 201, in some embodiments, is a gateway compute node (e.g., a virtual machine) that provides access to external networks (e.g. forwards north-south traffic) to compute nodes inside the logical network. In some embodiments, cloud gateway 201 also provides services to north-south traffic received at cloud gateway 201. Services provided at the cloud gateway 201 include, in some embodiments, any one or more of a firewall service, a load balancing service, a network address translation service, intrusion detection, VPN, audit logging, service chaining (sending data messages to third-party service appliances (e.g., SVMs 202) for various middlebox processing). The services provided to traffic received at cloud gateway 201, in some embodiments, are provided by forwarding the traffic to a service compute node such as one of SVMs 202A-M executing in the Transit VPC 205. SVMs 202A-M, in some embodiments, provide a set of logical network services and a set of third party services. Having a centralized cloud gateway allows third-party appliances for north-south logical network traffic to also be centralized to the transit VPC in some embodiments leading to further cost and management overhead reductions.
As shown, each of the VMs is assigned a different workload IP (e.g., 10.1.0.2, 10.10.0.3, etc.) in the IP subnet assigned to the logical switch (e.g., 10.1.0.0/24. 10.10.0.0/24, etc.) to which it connects. The IP subnet of the logical switch is an IP subnet of the subnet assigned to the VPC in which it is implemented (e.g., 10.1.0.0/16 or 10.10.0.0/16). These IP addresses, in some embodiments, are identical to the IP addresses assigned to the VM (or the VTEP of the MFE operating on the VM) by the cloud provider. Forwarding operations performed by an MFE implementing a hybrid logical switch for such embodiments are discussed in more detail below.
In other embodiments, each of the VMs is assigned a different workload IP in a subnet of the logical network that is not the same as the logical switch subnet assigned by the cloud provider (e.g., the 192.168.1.0/24 subnet instead of the 10.1.1.0/24 subnet). When the workload application in these embodiments sends a packet, this different workload IP address will be the source IP used in the header of that packet. However, the MFEs operating on these VMs have VTEPs with different IP addresses on the VPC subnets (10.1.0.2, 10.1.0.3, 10.10.0.2, 10.10.0.3, etc.). The packets that exit the VM will thus be encapsulated using this VTEP IP address as the source IP address (after logical processing is performed by the MFE in the source VM), in order to be sent to other destinations in the VPC.
The figure also illustrates tunnels between these eight MFEs and a cloud gateway 301 in the transit VPC 305. The tunnels in some embodiments use a peering relationship between compute VPCs 310A-N and transit VPC 305 illustrated as network connections 350. These tunnels pass through the underlying network of the public cloud provider (referred to herein as the “underlay”). In addition, though not shown here for simplicity, tunnels are created (through the underlay network) between each pair of the MFEs operating on the VMs 330. These tunnels between the MFEs 340 and the cloud gateway 301 are used to forward traffic between compute nodes (e.g., VMs, containers, etc.) that are implemented by different VPCs if, for example, there is no peering between the source and destination VPCs or a centralized service provided by or through the cloud gateway 301 is required for the traffic.
The cloud gateway 301 can also send packets to (and receive packets from) destinations within an on-premises private datacenter 370. To send these packets, the cloud gateway 301 encapsulates the packets using its VTEP IP, so that the destination will identify the incoming packet as a logical network packet. To secure traffic between the cloud gateway 301 in the VPC 305 and the destinations in the private datacenter (e.g., VMs attached to a logical switch), the packets are sent via a VPN tunnel 304 in some embodiments. In this example, the gateway's connection to external networks is not shown, and is discussed in more detail in U.S. Patent publication number 2018/0062923 which is incorporated herein by reference.
Cloud gateway 301, in some embodiments, is a gateway compute node (e.g., a virtual machine) that provides access to external networks (e.g. forwards north-south traffic) to compute nodes inside the logical network. In some embodiments, cloud gateway 301 also provides services to traffic (both north-south and east-west) received at cloud gateway 301. Services provided at the cloud gateway 301 include, in some embodiments, any one or more of logical routing, a firewall service, a load balancing service, a network address translation service, intrusion detection, VPN, audit logging, service chaining (sending data messages to third-party service appliances (e.g., SVMs 302) for various middlebox processing). The services provided to traffic received at cloud gateway 301, in some embodiments, is provided by forwarding the traffic to a service compute node such as one of SVMs 302A-M executing in the Transit VPC 305. SVMs 302A-M, in some embodiments, provide a set of logical network services and a set of third party services. Having a centralized cloud gateway allows third-party appliances for north-south and east-west logical network traffic to also be centralized to the transit VPC in some embodiments leading to further cost and management overhead reductions.
Although embodiments described above and throughout the rest of the description are discussed in terms of a single transit VPC and cloud gateway, other embodiments implement a cloud gateway as a cluster of two or more cloud gateway devices in active-active or active-standby mode in a single transit VPC. Still other embodiments implement a cluster of two or more transit VPCs each hosting a cloud gateway device that are configured in one of active-active or active-standby configuration. Cloud gateways (or transit VPCs) in active-active mode are each addressable by compute nodes in the logical network and are load balanced in some embodiments or are addressed by a same multicast address. In such embodiments, MFEs are configured as described in relation to
Upon receiving the data message, the MFE performs (at 520) ingress processing. In some embodiments, the ingress processing includes a distributed firewall operation. After ingress processing, in some embodiments the MFE consults an arbiter to determine (at 530) whether the data message requires processing by a cloud gateway executing in the transit VPC. In some embodiments, the arbiter is part of the MFE. The determination, in some embodiments, is based on whether the destination of the data message is in a same VPC as the source of the data message, whether any centralized services are required for the data message, and whether a peering exists between the source and destination VPCs.
In some embodiments, a determination that a destination compute node is in a same VPC as the source compute node is based on whether a destination IP address of the data message falls within a range of IP addresses assigned to the VPC of the source compute node. Determining whether the data message requires a centralized service is based, in some embodiments, on a policy-based routing policy or other form of policy received from a network controller. The policy, in some embodiments, is specified in terms of a tuple associated with data messages (e.g., a five-tuple specifying source and destination IP address, source and destination port and protocol). In some embodiments, the determination that a data message requires a centralized service is made implicitly based on a lack of a routing entry for the destination IP address after having the routing table processed to remove routes to addresses that are not available from the VPC or that require routing through the cloud gateway in the transit VPC (as discussed in relation to
In some embodiments that implicitly determine whether a data message requires forwarding to the cloud gateway in the transit VPC (either because it needs a service or because there is no peering with a destination VPC), the MFE (or separate managed switching and routing elements as discussed below) or arbiter module performs logical switching and routing operations associated with logical switching and routing elements of the overlay network (e.g. a tenant logical network). Logical switching operations include, in some embodiments, identifying an egress logical port of the logical switching element that receives the data message from a source compute node and logically forwarding the data message to the egress logical port connected to a logical routing element. The logical routing operations, in some embodiments, include implementing a distributed logical router to identify a logical port of the logical routing element associated with the destination address of the data message. As discussed above the distributed logical router is configured, in some embodiments, to cause data messages that (1) require a service provided at the overlay cloud gateway or (2) that are destined for VPC for which no peering relationship exists from a source VPC to be routed to the cloud gateway. In some embodiments, this is equivalent to the distributed logical routing element determining that the data message should be forwarded to a centralized service routing element implemented by the overlay cloud gateway.
A data message requires processing by an overlay cloud gateway (e.g., a device implementing a logical router to provide access to external network for compute nodes in the overlay network), in some embodiments, if the data message (1) requires a centralized service provided by the cloud gateway (whether source and destination VPCs are the same or different) or (2) is destined to a VPC that is not peered to the source VPC. Conversely, a data message does not require processing by the overlay cloud gateway if (1) the data message is destined to a compute node in the same VPC and does not require a centralized service provided by the cloud gateway or (2) the data message is destined to a compute node on a VPC that is peered with the VPC of the source compute node and the data message does not require a centralized service.
If the process determines (at 530) that the data message requires processing by the cloud gateway, the process encapsulates (at 530) the data message with a header that directs the packet to a tunnel endpoint of the cloud gateway. In some embodiments, the tunnel endpoint is a Virtual Extensible Local Area Network (VXLAN) Tunnel Endpoint (VTEP). The encapsulation, in some embodiments, is a generic network virtualization encapsulation (GENEVE) header.
If the process determines (at 530) that the data message does not require processing by the cloud gateway or determines (at 530) that the data message requires processing by the cloud gateway and has encapsulated (at 540) the data message, the process forwards the data message to the destination. In some embodiments, the data message is forwarded to a forwarding element of the cloud provider and the process ends. The cloud-provider forwarding element, in some embodiments provides a further encapsulation to traverse the physical network on which the virtual private cloud is implemented.
In some embodiments, data messages encapsulated to be forwarded to the cloud gateway are received at the cloud gateway which then decapsulates the data message, provides any required services, and forwards the data message to the destination compute node. As the transit VPC on which the cloud gateway is peered with each VPC hosting compute nodes of the overlay network, in some embodiments, the forwarded data message does not require encapsulation to be forwarded from the cloud gateway to the destination compute node.
For intra-VPC traffic, first and second paths from VM1 to VM6 are depicted. The beginning of each path begins with a data message (labeled ‘1’) that is sent from the workload application using the source and destination IP addresses that are equivalent in the overlay network and the cloud provider network. In some embodiments using different overlay network IP addresses, the overlay network addresses would appear in the header of the data message from the workload application. In embodiments using containers, the container sends data message ‘1’ to an MFE (e.g. an OVS) using the IP addresses assigned by the overlay network. As described in relation to
For intra-VPC traffic that the MFE determines requires forwarding to the cloud gateway 301 based on a need for a particular service provided by the cloud gateway 301, the MFE encapsulates the data message (as data message ‘4’) with the IP address of the cloud gateway (e.g., 10.12.0.1) which is accessible based on the peering 350 between compute VPC 310A and transit VPC 305. The address of the cloud gateway 301, in this case, is both the overlay and cloud provider address, but in embodiments in which they are not the same, the MFE uses the VTEP address of the MFE to which the cloud gateway 301 attaches in the encapsulation. Data message ‘4’ is received by PCFE 690 and, similarly to data message ‘3’, is encapsulated as data message ‘5’ to traverse the underlay network to arrive at cloud gateway 301. The structure of the data messages ‘1’ through ‘5’ is shown on the right hand side of
Once the cloud gateway 301 receives data message ‘5’ it provides the service (e.g., firewall, NAT, load balancing, etc.) to/for the data message. In some embodiments providing the service includes forwarding the data message as data message ‘6’ to a service compute node (i.e., SVM 302A) that is one of a logical network service appliance or a third-party service appliance in some embodiments. Forwarding the data message to the service appliance uses the overlay encapsulation in some embodiments, while in other embodiments, the service appliance is addressed using the cloud provider logical network address. After the service is performed/provided, the service appliance returns data message ‘7’ to the cloud gateway 301 for forwarding to the destination as data message ‘8’. Data message ‘8’ in some embodiments is sent without an overlay encapsulation based on the peering relationship between transit VPC 305 and compute VPC 310A that allows addressing compute nodes in other VPCs directly. A PCFE executing on the host machine on which the cloud gateway operates will encapsulate the data message as in data messages ‘3’ and ‘5’ whether the data message is sent by the cloud gateway 301 encapsulated or unencapsulated. In some embodiments, data messages ‘6’ and ‘7’ are not sent as the service is provided by the cloud gateway itself.
Data message ‘1’ of
Data message ‘4’ is a data message sent from a VM in a first compute VPC 710A to a second compute VPC 710B for which a peering exists. MFE 740 receives a data message from an application or container on VM1 and determines that the data message does not require routing through the transit VPC 705 and that a peering relationship exists between the current compute VPC and the destination compute VPC. Based on the determination, data message ‘4’ is sent to the PCFE (not shown) using the cloud provider addresses (e.g., as for data message ‘2’ of
Once configuration data has been received (at 810) the process pushes (at 820) configuration data to compute nodes in the overlay network. In some embodiments, the configuration data pushed to the compute nodes is the configuration data received by the controller. In some embodiments, the controller converts the received configuration data into compute-node- or VPC-specific data that reflects the portion of the configuration data that is relevant to the compute node or to the VPC in which the compute node executes (e.g., that the VPC is peered with 2 additional VPCs). In some embodiments, the controller communicates the configuration information to a local agent of the compute node to be configured. This configuration data, in some embodiments, is in addition to other sets of configuration data for constructing/populating routing tables of the different MFEs.
Once the configuration information is pushed (at 820) to the compute node, the compute node receives (at 830) the configuration data. In some embodiments, the compute node operates an MFE of the overlay network. The MFE, in some embodiments, executes in a virtual machine along with a set of applications or containers for which the MFE provides forwarding services. In some embodiments, the MFE is an open vSwitch (OVS) and the configuration data is received as a set of forwarding entries for an open vSwitch (OVS). In some embodiments in which the MFE includes a routing table, the configuration includes a set of routing entries reflecting the existence of peering relationships between VPCs, the existence of VPNs, or specific flows that should be directed to a cloud gateway in a transit VPC.
In some embodiments, instead of a single MFE there is a set of managed switching and routing elements that process data messages to and from sources of data messages. The configuration data in such embodiments includes configuration data for each managed switching and routing element separately to implement the desired logical switching and routing elements, respectively, of the overlay network (e.g., a tenant logical network). In some embodiments, the compute node only executes a managed switching element with a managed routing element executing on a separate compute node.
Some embodiments place the routing entries in a routing table of a Tier-1 or Tier-0 distributed router implemented by the MFE or managed routing element. The logical network implementing Tier-1 and Tier-0 logical routers is described in U.S. Pat. No. 9,787,605 which is hereby incorporated by reference. In embodiments in which configuration data is not converted by the controller from which it is received, the compute node converts the configuration data into forwarding data (e.g., entries in a routing table, flow entries for the OVS). Some embodiments receive the configuration data as (1) policy based routing data that defines policies based on data message or flow characteristics (e.g., a five-tuple, or other header values) and (2) private routing table data that is used (i) to direct an MFE to use the private routing table when the characteristics of a data message match and (ii) to generate and populate the private routing table, respectively.
After receiving (at 830) the configuration data, the process configures (at 840) the MFE to implement the forwarding decisions based on the received configuration data. In some embodiments, configuring the MFE includes updating a routing table of the MFE (or managed routing element) with routes based on the configuration data. Routing table entries of the updated routing table may include routing entries to forward traffic (1) using the cloud provider network (e.g., sending the traffic without encapsulation by the MFE) to a destination compute node, (2) encapsulated by the MFE to use a direct tunnel between source and destination compute nodes (e.g., between MFEs to which the source and destination compute nodes connect), or (3) encapsulated to use a tunnel to a cloud gateway (e.g., a cloud gateway in a transit VPC) for the cloud gateway to forward to the destination compute node. In some embodiments, configuring the MFE includes updating flow entries of the OVS to forward traffic in any of the ways outlined above for the routing entries. The MFE is now configured and the process ends.
Many of the above-described features and applications are implemented as software processes that are specified as a set of instructions recorded on a computer readable storage medium (also referred to as computer readable medium). When these instructions are executed by one or more processing unit(s) (e.g., one or more processors, cores of processors, or other processing units), they cause the processing unit(s) to perform the actions indicated in the instructions. Examples of computer readable media include, but are not limited to, CD-ROMs, flash drives, RAM chips, hard drives, EPROMs, etc. The computer readable media does not include carrier waves and electronic signals passing wirelessly or over wired connections.
In this specification, the term “software” is meant to include firmware residing in read-only memory or applications stored in magnetic storage, which can be read into memory for processing by a processor. Also, in some embodiments, multiple software inventions can be implemented as sub-parts of a larger program while remaining distinct software inventions. In some embodiments, multiple software inventions can also be implemented as separate programs. Finally, any combination of separate programs that together implement a software invention described here is within the scope of the invention. In some embodiments, the software programs, when installed to operate on one or more electronic systems, define one or more specific machine implementations that execute and perform the operations of the software programs.
The bus 905 collectively represents all system, peripheral, and chipset buses that communicatively connect the numerous internal devices of the computer system 900. For instance, the bus 905 communicatively connects the processing unit(s) 910 with the read-only memory 930, the system memory 925, and the permanent storage device 935.
From these various memory units, the processing unit(s) 910 retrieve instructions to execute and data to process in order to execute the processes of the invention. The processing unit(s) may be a single processor or a multi-core processor in different embodiments. The read-only-memory (ROM) 930 stores static data and instructions that are needed by the processing unit(s) 910 and other modules of the computer system. The permanent storage device 935, on the other hand, is a read-and-write memory device. This device is a non-volatile memory unit that stores instructions and data even when the computer system 900 is off. Some embodiments of the invention use a mass-storage device (such as a magnetic or optical disk and its corresponding disk drive) as the permanent storage device 935.
Other embodiments use a removable storage device (such as a floppy disk, flash drive, etc.) as the permanent storage device. Like the permanent storage device 935, the system memory 925 is a read-and-write memory device. However, unlike storage device 935, the system memory is a volatile read-and-write memory, such a random access memory. The system memory stores some of the instructions and data that the processor needs at runtime. In some embodiments, the invention's processes are stored in the system memory 925, the permanent storage device 935, and/or the read-only memory 930. From these various memory units, the processing unit(s) 910 retrieve instructions to execute and data to process in order to execute the processes of some embodiments.
The bus 905 also connects to the input and output devices 940 and 945. The input devices enable the user to communicate information and select commands to the computer system. The input devices 940 include alphanumeric keyboards and pointing devices (also called “cursor control devices”). The output devices 945 display images generated by the computer system. The output devices include printers and display devices, such as cathode ray tubes (CRT) or liquid crystal displays (LCD). Some embodiments include devices such as a touchscreen that function as both input and output devices.
Finally, as shown in
Some embodiments include electronic components, such as microprocessors, storage and memory that store computer program instructions in a machine-readable or computer-readable medium (alternatively referred to as computer-readable storage media, machine-readable media, or machine-readable storage media). Some examples of such computer-readable media include RAM, ROM, read-only compact discs (CD-ROM), recordable compact discs (CD-R), rewritable compact discs (CD-RW), read-only digital versatile discs (e.g., DVD-ROM, dual-layer DVD-ROM), a variety of recordable/rewritable DVDs (e.g., DVD-RAM, DVD−RW, DVD+RW, etc.), flash memory (e.g., SD cards, mini-SD cards, micro-SD cards, etc.), magnetic and/or solid state hard drives, read-only and recordable Blu-Ray® discs, ultra-density optical discs, any other optical or magnetic media, and floppy disks. The computer-readable media may store a computer program that is executable by at least one processing unit and includes sets of instructions for performing various operations. Examples of computer programs or computer code include machine code, such as is produced by a compiler, and files including higher-level code that are executed by a computer, an electronic component, or a microprocessor using an interpreter.
While the above discussion primarily refers to microprocessor or multi-core processors that execute software, some embodiments are performed by one or more integrated circuits, such as application specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs). In some embodiments, such integrated circuits execute instructions that are stored on the circuit itself.
As used in this specification, the terms “computer”, “server”, “processor”, and “memory” all refer to electronic or other technological devices. These terms exclude people or groups of people. For the purposes of the specification, the terms display or displaying means displaying on an electronic device. As used in this specification, the terms “computer readable medium,” “computer readable media,” and “machine readable medium” are entirely restricted to tangible, physical objects that store information in a form that is readable by a computer. These terms exclude any wireless signals, wired download signals, and any other ephemeral or transitory signals.
While the invention has been described with reference to numerous specific details, one of ordinary skill in the art will recognize that the invention can be embodied in other specific forms without departing from the spirit of the invention. For instance, several figures conceptually illustrate processes. The specific operations of these processes may not be performed in the exact order shown and described. The specific operations may not be performed in one continuous series of operations, and different specific operations may be performed in different embodiments. Furthermore, the process could be implemented using several sub-processes, or as part of a larger macro process. Thus, one of ordinary skill in the art would understand that the invention is not to be limited by the foregoing illustrative details, but rather is to be defined by the appended claims.
Number | Name | Date | Kind |
---|---|---|---|
6108300 | Coile et al. | Aug 2000 | A |
6832238 | Sharma et al. | Dec 2004 | B1 |
7107360 | Phadnis et al. | Sep 2006 | B1 |
7360245 | Ramachandran et al. | Apr 2008 | B1 |
7423962 | Auterinen | Sep 2008 | B2 |
7523485 | Kwan | Apr 2009 | B1 |
7953895 | Narayanaswamy et al. | May 2011 | B1 |
8264947 | Tavares | Sep 2012 | B1 |
8296434 | Miller et al. | Oct 2012 | B1 |
8432791 | Masters | Apr 2013 | B1 |
8514868 | Hill | Aug 2013 | B2 |
8719590 | Faibish et al. | May 2014 | B1 |
8902743 | Greenberg et al. | Dec 2014 | B2 |
8958293 | Anderson | Feb 2015 | B1 |
9137209 | Brandwine et al. | Sep 2015 | B1 |
9244669 | Govindaraju et al. | Jan 2016 | B2 |
9356866 | Sivaramakrishnan et al. | May 2016 | B1 |
9413730 | Narayan et al. | Aug 2016 | B1 |
9485149 | Traina et al. | Nov 2016 | B1 |
9519782 | Aziz et al. | Dec 2016 | B2 |
9590904 | Heo et al. | Mar 2017 | B2 |
9699070 | Davie et al. | Jul 2017 | B2 |
9832118 | Miller et al. | Nov 2017 | B1 |
9871720 | Tillotson | Jan 2018 | B1 |
10135675 | Yu et al. | Nov 2018 | B2 |
10193749 | Hira et al. | Jan 2019 | B2 |
10228959 | Anderson et al. | Mar 2019 | B1 |
10326744 | Nossik et al. | Jun 2019 | B1 |
10333959 | Katrekar et al. | Jun 2019 | B2 |
10341371 | Katrekar et al. | Jul 2019 | B2 |
10348767 | Lee | Jul 2019 | B1 |
10367757 | Chandrashekhar et al. | Jul 2019 | B2 |
10397136 | Hira et al. | Aug 2019 | B2 |
10484302 | Hira et al. | Nov 2019 | B2 |
10491466 | Hira et al. | Nov 2019 | B1 |
10491516 | Ram et al. | Nov 2019 | B2 |
10567482 | Ram et al. | Feb 2020 | B2 |
10601705 | Hira et al. | Mar 2020 | B2 |
10673952 | Cohen et al. | Jun 2020 | B1 |
10778579 | Hira | Sep 2020 | B2 |
10805330 | Katrekar et al. | Oct 2020 | B2 |
10812413 | Chandrashekhar et al. | Oct 2020 | B2 |
10862753 | Hira et al. | Dec 2020 | B2 |
10924431 | Chandrashekhar et al. | Feb 2021 | B2 |
20020062217 | Fujimori | May 2002 | A1 |
20020199007 | Clayton et al. | Dec 2002 | A1 |
20070186281 | McAlister | Aug 2007 | A1 |
20070226795 | Conti et al. | Sep 2007 | A1 |
20070256073 | Troung et al. | Nov 2007 | A1 |
20080104692 | McAlister | May 2008 | A1 |
20080225888 | Valluri et al. | Sep 2008 | A1 |
20090254973 | Kwan | Oct 2009 | A1 |
20100037311 | He et al. | Feb 2010 | A1 |
20100112974 | Sahai et al. | May 2010 | A1 |
20100318609 | Lahiri et al. | Dec 2010 | A1 |
20110317703 | Dunbar et al. | Dec 2011 | A1 |
20120082063 | Fujita | Apr 2012 | A1 |
20120250682 | Vincent | Oct 2012 | A1 |
20130044636 | Koponen et al. | Feb 2013 | A1 |
20130044641 | Koponen et al. | Feb 2013 | A1 |
20130044763 | Koponen et al. | Feb 2013 | A1 |
20130058208 | Pfaff et al. | Mar 2013 | A1 |
20130058335 | Koponen et al. | Mar 2013 | A1 |
20130125230 | Koponen et al. | May 2013 | A1 |
20130198740 | Arroyo et al. | Aug 2013 | A1 |
20130263118 | Kannan et al. | Oct 2013 | A1 |
20130287026 | Davie | Oct 2013 | A1 |
20130297768 | Singh | Nov 2013 | A1 |
20130304903 | Mick et al. | Nov 2013 | A1 |
20130318219 | Kancherla | Nov 2013 | A1 |
20130346585 | Ueno | Dec 2013 | A1 |
20140010239 | Xu et al. | Jan 2014 | A1 |
20140052877 | Mao | Feb 2014 | A1 |
20140108665 | Arora et al. | Apr 2014 | A1 |
20140143853 | Onodera | May 2014 | A1 |
20140156818 | Hunt | Jun 2014 | A1 |
20140192804 | Ghanwani et al. | Jul 2014 | A1 |
20140226820 | Chopra et al. | Aug 2014 | A1 |
20140241247 | Kempf et al. | Aug 2014 | A1 |
20140245420 | Tidwell et al. | Aug 2014 | A1 |
20140280961 | Martinez et al. | Sep 2014 | A1 |
20140317677 | Vaidya et al. | Oct 2014 | A1 |
20140334495 | Stubberfield et al. | Nov 2014 | A1 |
20140337500 | Lee | Nov 2014 | A1 |
20140376560 | Senniappan et al. | Dec 2014 | A1 |
20150009995 | Gross, IV et al. | Jan 2015 | A1 |
20150016286 | Ganichev et al. | Jan 2015 | A1 |
20150016460 | Zhang et al. | Jan 2015 | A1 |
20150043383 | Farkas et al. | Feb 2015 | A1 |
20150052522 | Chanda et al. | Feb 2015 | A1 |
20150052525 | Raghu | Feb 2015 | A1 |
20150063360 | Thakkar | Mar 2015 | A1 |
20150063364 | Thakkar | Mar 2015 | A1 |
20150085870 | Narasimha et al. | Mar 2015 | A1 |
20150096011 | Watt | Apr 2015 | A1 |
20150098465 | Pete et al. | Apr 2015 | A1 |
20150103838 | Zhang et al. | Apr 2015 | A1 |
20150106804 | Chandrashekhar et al. | Apr 2015 | A1 |
20150124645 | Yadav et al. | May 2015 | A1 |
20150128245 | Brown et al. | May 2015 | A1 |
20150138973 | Kumar et al. | May 2015 | A1 |
20150139238 | Pourzandi et al. | May 2015 | A1 |
20150163137 | Kamble et al. | Jun 2015 | A1 |
20150163145 | Pettit et al. | Jun 2015 | A1 |
20150163192 | Jain et al. | Jun 2015 | A1 |
20150172075 | DeCusatis et al. | Jun 2015 | A1 |
20150172183 | DeCusatis et al. | Jun 2015 | A1 |
20150172331 | Raman | Jun 2015 | A1 |
20150263983 | Brennan et al. | Sep 2015 | A1 |
20150263992 | Kuch et al. | Sep 2015 | A1 |
20150264077 | Berger et al. | Sep 2015 | A1 |
20150271303 | Neginhal | Sep 2015 | A1 |
20150281098 | Pettit et al. | Oct 2015 | A1 |
20150281274 | Masurekar et al. | Oct 2015 | A1 |
20150295731 | Bagepalli | Oct 2015 | A1 |
20150295800 | Bala et al. | Oct 2015 | A1 |
20150304117 | Dong et al. | Oct 2015 | A1 |
20150326469 | Kern et al. | Nov 2015 | A1 |
20150339136 | Suryanarayanan et al. | Nov 2015 | A1 |
20150350059 | Chunduri et al. | Dec 2015 | A1 |
20150350101 | Sinha | Dec 2015 | A1 |
20150373012 | Bartz et al. | Dec 2015 | A1 |
20150381493 | Bansal et al. | Dec 2015 | A1 |
20160014023 | He et al. | Jan 2016 | A1 |
20160055019 | Thakkar et al. | Feb 2016 | A1 |
20160072888 | Jung et al. | Mar 2016 | A1 |
20160094364 | Subramaniyam et al. | Mar 2016 | A1 |
20160094661 | Jain et al. | Mar 2016 | A1 |
20160105488 | Thakkar et al. | Apr 2016 | A1 |
20160124742 | Rangasamy et al. | May 2016 | A1 |
20160134418 | Liu et al. | May 2016 | A1 |
20160182567 | Sood et al. | Jun 2016 | A1 |
20160191304 | Muller | Jun 2016 | A1 |
20160198003 | Luft | Jul 2016 | A1 |
20160212049 | Davie | Jul 2016 | A1 |
20160226967 | Zhang et al. | Aug 2016 | A1 |
20160274926 | Narasimhamurthy et al. | Sep 2016 | A1 |
20160308762 | Teng et al. | Oct 2016 | A1 |
20160337329 | Sood et al. | Nov 2016 | A1 |
20160352623 | Jayabalan et al. | Dec 2016 | A1 |
20160352682 | Chang et al. | Dec 2016 | A1 |
20160352747 | Khan et al. | Dec 2016 | A1 |
20160364575 | Caporal et al. | Dec 2016 | A1 |
20160380973 | Sullenberger et al. | Dec 2016 | A1 |
20170005923 | Babakian | Jan 2017 | A1 |
20170006053 | Greenberg et al. | Jan 2017 | A1 |
20170034129 | Sawant et al. | Feb 2017 | A1 |
20170034198 | Powers et al. | Feb 2017 | A1 |
20170060628 | Tarasuk-Levin et al. | Mar 2017 | A1 |
20170078248 | Bian et al. | Mar 2017 | A1 |
20170091458 | Gupta et al. | Mar 2017 | A1 |
20170091717 | Chandraghatgi et al. | Mar 2017 | A1 |
20170093646 | Chanda et al. | Mar 2017 | A1 |
20170097841 | Chang et al. | Apr 2017 | A1 |
20170099188 | Chang et al. | Apr 2017 | A1 |
20170104365 | Ghosh et al. | Apr 2017 | A1 |
20170111230 | Srinivasan et al. | Apr 2017 | A1 |
20170118115 | Tsuji | Apr 2017 | A1 |
20170126552 | Pfaff et al. | May 2017 | A1 |
20170142012 | Thakkar et al. | May 2017 | A1 |
20170149582 | Cohn et al. | May 2017 | A1 |
20170163442 | Shen et al. | Jun 2017 | A1 |
20170163599 | Shen et al. | Jun 2017 | A1 |
20170195217 | Parasmal et al. | Jul 2017 | A1 |
20170222928 | Johnsen et al. | Aug 2017 | A1 |
20170223518 | Upadhyaya et al. | Aug 2017 | A1 |
20170230241 | Neginhal | Aug 2017 | A1 |
20170279826 | Mohanty et al. | Sep 2017 | A1 |
20170289060 | Aftab et al. | Oct 2017 | A1 |
20170302529 | Agarwal et al. | Oct 2017 | A1 |
20170310580 | Caldwell et al. | Oct 2017 | A1 |
20170317972 | Bansal et al. | Nov 2017 | A1 |
20170324848 | Johnsen et al. | Nov 2017 | A1 |
20170331746 | Qiang | Nov 2017 | A1 |
20170359304 | Benny et al. | Dec 2017 | A1 |
20180006943 | Dubey | Jan 2018 | A1 |
20180007002 | Landgraf | Jan 2018 | A1 |
20180013791 | Healey et al. | Jan 2018 | A1 |
20180026873 | Cheng et al. | Jan 2018 | A1 |
20180026944 | Phillips | Jan 2018 | A1 |
20180027012 | Srinivasan et al. | Jan 2018 | A1 |
20180027079 | Ali et al. | Jan 2018 | A1 |
20180053001 | Folco et al. | Feb 2018 | A1 |
20180062880 | Yu et al. | Mar 2018 | A1 |
20180062881 | Chandrashekhar et al. | Mar 2018 | A1 |
20180062917 | Chandrashekhar et al. | Mar 2018 | A1 |
20180062923 | Katrekar | Mar 2018 | A1 |
20180062933 | Hira et al. | Mar 2018 | A1 |
20180063036 | Chandrashekhar et al. | Mar 2018 | A1 |
20180063086 | Hira et al. | Mar 2018 | A1 |
20180063087 | Hira et al. | Mar 2018 | A1 |
20180063176 | Katrekar et al. | Mar 2018 | A1 |
20180063193 | Chandrashekhar et al. | Mar 2018 | A1 |
20180077048 | Kubota et al. | Mar 2018 | A1 |
20180083923 | Bian et al. | Mar 2018 | A1 |
20180115586 | Chou et al. | Apr 2018 | A1 |
20180139123 | Qiang | May 2018 | A1 |
20180197122 | Kadt et al. | Jul 2018 | A1 |
20180336158 | Iyer et al. | Nov 2018 | A1 |
20190037033 | Khakimov et al. | Jan 2019 | A1 |
20190068493 | Ram et al. | Feb 2019 | A1 |
20190068500 | Hira | Feb 2019 | A1 |
20190068689 | Ram et al. | Feb 2019 | A1 |
20190097838 | Sahoo | Mar 2019 | A1 |
20190173757 | Hira et al. | Jun 2019 | A1 |
20190173780 | Hira et al. | Jun 2019 | A1 |
20190306185 | Katrekar et al. | Oct 2019 | A1 |
20200007497 | Jain et al. | Jan 2020 | A1 |
20200028758 | Tollet et al. | Jan 2020 | A1 |
20200067733 | Hira et al. | Feb 2020 | A1 |
20200177670 | Ram et al. | Jun 2020 | A1 |
20200351254 | Xiong et al. | Nov 2020 | A1 |
20210105208 | Hira | Apr 2021 | A1 |
Number | Date | Country |
---|---|---|
103036919 | Dec 2015 | CN |
105379227 | Mar 2016 | CN |
107534603 | Jan 2018 | CN |
1742430 | Jan 2007 | EP |
2014075731 | Apr 2014 | JP |
2015165700 | Sep 2015 | JP |
2016159113 | Oct 2016 | WO |
2018044341 | Mar 2018 | WO |
2019040720 | Feb 2019 | WO |
2019046071 | Mar 2019 | WO |
2019112704 | Jun 2019 | WO |
2020005540 | Jan 2020 | WO |
2020041074 | Feb 2020 | WO |
Entry |
---|
Firestone, Daniel, “VFP: A Virtual Switch Platform for Host SDN in the Public Cloud,” 14th USENIX Symposium on Networked Systems Design and Implementation, Mar. 27-29, 2017, 15 pages, USENIX, Boston, MA, USA. |
Non-Published commonly Owned U.S. Appl. No. 16/447,872, filed Jun. 20, 2019, 124 pages, Nicira, Inc. |
Wenjie, Zhu (Jerry), “Next Generation Service Overlay Networks,” IEEE P1903 NGSON (3GPP Draft), Aug. 22, 2014, 24 pages, IEEE. |
Author Unknown, “Network Controller,” Dec. 16, 2014, 4 pages, available at: https://web.archive.org/web/20150414112014/https://technet.microsoft.com/en-us/library/dn859239.aspx. |
Koponen, Teemu, et al., “Network Virtualization in Multi-tenant Datacenters,” Proceedings of the 11th USENIX Symposium on Networked Systems Design and Implementation (NSDI'14), Apr. 2-4, 2014, 15 pages, Seattle, WA, USA. |
Non-Published commonly Owned U.S. Appl. No. 15/831,369, filed Dec. 4, 2017, 47 pages, Nicira Inc. |
Non-Published commonly Owned U.S. Appl. No. 15/831,372, filed Dec. 4, 2017, 47 pages, Nicira Inc. |
Non-Published commonly Owned U.S. Appl. No. 16/112,597, filed Aug. 24, 2018, 38 pages, VMware, Inc. |
Non-Published commonly Owned U.S. Appl. No. 16/112,599, filed Aug. 24, 2018, 38 pages, VMware, Inc. |
Sunliang, Huang, “Future SDN-based Data Center Network,” Nov. 15, 2013, 5 pages, ZTE Corporation, available at http://wwwen.zte.com.cn/endata/magazine/ztetechnologies/2013/no6/articles/201311/120131115_412737.html. |
Non-Published commonly Owned U.S. Appl. No. 16/022,657, filed Jun. 28, 2018, 28 pages, VMware, nc. |
Non-Published commonly owned U.S. Appl. No. 16/785,620, filed Feb. 9, 2020, 52 pages, Nicira Inc. |
Black, David, et al., “An Architecture for Data Center Network Virtualization Overlays (NVO3) [draft-ietf-nvo3-arch-08],” Sep. 20, 2016, 34 pages, IETF. |
Church, Mark, “Docker Reference Architecture: Designing Scalable, Portable Docker Container Networks,” Article ID: KB000801, Jun. 20, 2017, 36 pages, retrieved from https://success.docker.com/article/networking. |
Fernando, Rex, et al., “Service Chaining using Virtual Networks with BGP,” Internet Engineering Task Force, IETF, Jul. 7, 2015, 32 pages, Internet Society (ISOC), Geneva, Switzerland, available at https://tools.ietf.org/html/draft-fm-bess-service-chaining-01. |
Lasserre, Marc, et al., “Framework for Data Center (DC) Network Virtualization,” RFC 7365, Oct. 2014, 26 pages, IETF. |
Le Bigot, Jean-Tiare, “Introduction to Linux Namespaces—Part 5: NET,” Jan. 19, 2014, 6 pages, retrieved from https://blog.yadutaf.fr/2014/01/19/introduction-to-linux-namespaces-part-5-net. |
Merkel, Dirk, “Docker: Lightweight Linux Containers for Consistent Development and Deployment,” Linux Journal, May 19, 2014, 16 pages, vol. 2014—Issue 239, Belltown Media, Houston, USA. |
Non-Published commonly Owned U.S. Appl. No. 17/020,713, filed Sep. 14, 2020, 30 pages, Nicira, Inc. |
Singla, Ankur, et al., “Architecture Documentation: OpenContrail Architecture Document,” Jan. 24, 2015, 42 pages, OpenContrail. |
Zhang, Zhe, et al., “Lark: Bringing Network Awareness to High Throughput Computing,” 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing, May 4-7, 2015, 10 pages, IEEE, Shenzhen, China. |
Non-Published commonly owned U.S. Appl. No. 17/114,322, filed Dec. 7, 2020, 50 pages, Nicira Inc. |
Non-Published commonly owned U.S. Appl. No. 17/307,983, filed May 4, 2021, 119 pages, Nicira Inc. |
Lee, Wonhyuk, et al., “Micro-Datacenter Management Architecture for Mobile Wellness Information,” 2014 International Conference on IT Convergence and Security, Oct. 28-30, 2014, 4 pages, IEEE, Beijing, China. |
Ling, Lu, et al., “Hybrid Cloud Solution Based on SDN Architecture,” Collection of Cloud Computing Industry Application Cases, Dec. 31, 2016, 20 pages, Issue 2, China Academic Journal Publishing House. [English translation of document generated from www.onlinedoctranslator.com]. |
Number | Date | Country | |
---|---|---|---|
20200067734 A1 | Feb 2020 | US |