Method and system of resiliency in cloud-delivered SD-WAN

Abstract
In one aspect, a computerized method includes the step of providing process monitor in a Gateway. The method includes the step of, with the process monitor, launching a Gateway. Daemon (GWD). The GWD runs a GWD process that implements a Network Address Translation (NAT) process. The NAT process includes receiving a set of data packets from one or more Edge devices and forwarding the set of data packets to a public Internet. The method includes the step of receiving another set of data packets from the public Internet and forwarding the other set of data packets to the one or more Edge devices. The method includes the step of launching a Network Address Translation daemon (NATD). The method includes the step of detecting that the GWD process is interrupted; moving the NAT process to the NATD.
Description
FIELD OF THE INVENTION

This application relates generally to computer networking, and more specifically to a system, article of manufacture and method of resiliency in cloud-delivered SD-WAN.


DESCRIPTION OF THE RELATED ART

Traditional methods of ensuring WAN resiliency have focused on two aspects. First, resiliency for traffic between two enterprise sites (not destined for the public Internet). Second, for subsequent flows towards the public Internet (not guaranteeing session continuity). The methods describe here provide for full resiliency for traffic destined for the public Internet including the preservation of existing flows.


BRIEF SUMMARY OF THE INVENTION

In one aspect, a computerized method includes the step of providing process monitor in a Gateway. The method includes the step of, with the process monitor, launching a Gateway Daemon (GWD). The GWD runs a GWD process that implements a Network Address Translation (NAT) process. The NAT process includes receiving a set of data packets from one or more Edge devices and forwarding the set of data packets to a public Internet. The method includes the step of receiving another set of data packets from the public Internet and forwarding the other set of data packets to the one or more Edge devices. The method includes the step of launching a Network Address Translation daemon (NATD). The method includes the step of detecting that the GWD process is interrupted; moving the NAT process to the NATD.


In another aspect, a computerized method is implemented when a public Internet flow is initiated from an Edge device connected to a Gateway system. The method includes the step of, with a GWD, looking up in a local hash table a NAT translation for a data packet's five tuple. The method includes the step of detecting that no NAT translation is extant for the data packet's five tuple. The method includes the step of creating the NAT translation for the data packet's five tuple. The method includes the step of creating returning the NAT translation for the data packet's five tuple to the Gateway system; storing the NAT translation locally in the Gateway system.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 illustrates an example process of implementing resiliency in an SD-WAN, according to some embodiments.



FIG. 2 illustrates an example of a gateway data plane running in a GWD process, according to some embodiments.



FIG. 3 illustrates an example of multiple physical or virtual instances of the gateway running and fronted by a single NATD daemon, according to some embodiments.



FIG. 4 illustrates an example of cloud traffic can be routed through an external CWS service to enable security scanning and other service insertion on the traffic before it exits to the public internet, according to some embodiments.



FIG. 5 depicts an exemplary computing system that can be configured to perform any one of the processes provided herein.



FIG. 6 conceptually illustrates an example of a single DSL connected to two gateways, according to some embodiments.





The Figures described above are a representative set, and are not exhaustive with respect to embodying the invention.


DESCRIPTION

Disclosed are a system, method, and article of manufacture for resiliency in cloud-delivered SD-WAN. The following description is presented to enable a person of ordinary skill in the art to make and use the various embodiments. Descriptions of specific devices, techniques, and applications are provided only as examples. Various modifications to the examples described herein can be readily apparent to those of ordinary skill in the art, and the general principles defined herein may be applied to other examples and applications without departing from the spirit and scope of the various embodiments.


Reference throughout this specification to “one embodiment,” “an embodiment,” ‘one example,’ or similar language means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, appearances of the phrases “in one embodiment,” “in an embodiment,” and similar language throughout, this specification may, but do not necessarily, all refer to the same embodiment.


Furthermore, the described features, structures, or characteristics of the invention may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided, such as examples of programming, software modules, user selections, network transactions, database queries, database structures, hardware modules, hardware circuits, hardware chips, etc., to provide a thorough understanding of embodiments of the invention. One skilled in the relevant art can recognize, however, that the invention may be practiced without one or more of the specific details, or with other methods, components, materials, and so forth. In other instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring aspects of the invention.


The schematic flow chart diagrams included herein are generally set forth as logical flow chart diagrams. As such, the depicted order and labeled steps are indicative of one embodiment of the presented method. Other steps and methods may be conceived that are equivalent in function, logic, or effect to one or more steps, or portions thereof, of the illustrated method. Additionally, the format and symbols employed are provided to explain the logical steps of the method and are understood not to limit the scope of the method. Although various arrow types and line types may be employed in the flow chart diagrams, and they are understood not to limit the scope of the corresponding method. Indeed, some arrows or other connectors may be used to indicate only the logical flow of the method. For instance, an arrow may indicate a waiting or monitoring period of unspecified duration between enumerated steps of the depicted method. Additionally, the order in which a particular method occurs may or may not strictly adhere to the order of the corresponding steps shown.


Definitions


Example definitions for some embodiments are now provided.


Border Gateway Protocol (BGP) can be a standardized exterior gateway protocol designed to exchange routing and reachability information among autonomous systems (AS) on the Internet.


Cloud computing can involve deploying groups of remote servers and/or software networks that allow centralized data storage and online access to computer services or resources. These groups of remote serves and/or software networks can be a collection of remote computing services.


Daemon can be a background process.


Data center, a physical location housing computing-related gear.


Dynamic tunneling is a transparent mechanism available for applications (e.g. that support the SOCKS4 or SOCKS5 client protocol).


Edge device can be a device that provides an entry point into enterprise or service provider core networks. An edge device can be software running in a virtual machine (VM) located in a branch office and/or customer premises.


Five (5) tuple refers to a set of five different values that comprise a Transmission Control Protocol/Internet Protocol (TCP/IP) connection. It includes a source IP address/port number, destination IP address/port number and the protocol in use.


Flow can be a grouping of packets that match a five (5) tuple which is a combination of Source IP Address (SIP), Destination IP Address (DIP), L4 Source Port (SPORT) and L4 Destination Port (DPORT) and the L4 protocol (PROTO).


Gateway can be a node (e.g. a router) on a computer network that serves as an access point to another network.


Internet Protocol Security (IPsec) can be a protocol suite for securing Internet Protocol (lP) communications by authenticating and encrypting each IP packet of a communication session. In IPsec tunnel mode, the entire IP packet is encrypted and authenticated. It is then encapsulated into a new IP packet with a new IP header. Tunnel mode is used to create virtual private networks for network-to-network communications (e.g. between routers to link sites), host-to-network communications (e.g. remote user access) and host-to-host communications (e.g. private chat).


Inter-process communication (IPC) can include mechanisms an operating system provides to allow the processes to manage shared data. Typically, applications can use IPC, categorized as clients and servers, where the client requests data and the server responds to client requests.


Network Address Translation (NAT) is a method of remapping one IP address space into another by modifying network address information in Internet Protocol (IP) datagram packet headers while they are in transit across a traffic routing device.


Orchestrator can include a software component that provides multi-tenant and role based centralized configuration management and visibility.


Open Shortest Path First (OSPF) can be a routing protocol for Internet Protocol (IP) networks. OSPF can use a link state routing (LSR) algorithm and falls into the group of interior gateway protocols (IGPs), operating within a single autonomous system (AS).


Software-defined networking in a wide area network (SD-WAN) a specific application of software-defined networking (SDN) technology applied to WAN connections, which are used to connect enterprise networks, including branch offices and data centers—over large geographic distances. An SD-WAN can simplify the management and operation of a WAN by decoupling the networking hardware from its control mechanism.


Tunneling protocol can allow a network, user to access or provide a network service that the underlying network does not support or provide directly.


Virtual private network (VPN) can extend a private network across a public network, such as the Internet. It can enable users to send and receive data across shared or public networks as if their computing devices were directly connected to the private network, and thus benefit from the functionality, security and management policies of the private network.


Additional example definitions are provided herein.


Examples Methods



FIG. 1 illustrates an example process of implementing resiliency in an SD-WAN, according to some embodiments. In step 102, it can be detected that a new Internet flow is initiated from an Edge device(s) connected to a Gateway. In step 104, a gateway data daemon (GWD) can look up in a local hash table to determine if a NAT translation is available for the data packet's five tuple. If no NAT translation is found, then, step 106, the GWD can query a Network Address Translation Daemon (NATD) for a new translation. If the NATD has previously translated the five (5) tuple, then, in step 108, the NATD can return the same translation that was provided before, thus ensuring session continuity. If the NATD has not previously translated the five tuple, in step 110, it can create a new translation and return it. In step 112, the new translation can be stored locally for future retrieval. When a flow is deleted on the Gateway, a signal can be sent to the NATD to delete the translation and free the associated memory in step 114.


Example Systems



FIG. 2 illustrates an example system 200 of a GWD 202 running in a GWD process 204, according to some embodiments. It is noted that GWD 202 can be a daemon. GWD process 204 can be a user-space process running in Linux. In one example, GWD can include a data plane (e.g. VeloCloud® data plane, etc.) and control plane software.


It is noted that the data plane includes the forwarding information base (FIB) and mechanisms for transmitting packets. The control plane includes the routing information base (RIB) and mechanisms for instructing Edges how to transmit packets.


In a single device case, GWD can run in a GWD process 204. GWD process 204 can receive data packets (e.g. all data packets) from the various Edge devices and forward them to the Internet, and vice versa. GWD process 204 can be interrupted for multiple reasons. For example, GWD process 204 can encounter a software fault (e.g. a crash). GWD process 204 can be restarted for troubleshooting. GWD process 204 can be restarted as part of a routine software upgrade. In these scenarios, data traffic can continue to flow uninterrupted. This can be achieved by moving the NAT process and its associated state outside the GWD context (e.g. to NATD 208). Accordingly, FIG. 2 illustrates a simplified process diagram of processes running within the Gateway 202. The process monitor (e.g. vc_procmon 210) can launch and manage two separate services independently: GWD and NATD 208. NATD 208 can be a user-space process running in Linux which contains the NAT software (e.g. VeloCloud® NAT software, etc.). NATD 208 stores its own state and communicates via IPC with GWD.


When a new Internet flow is initiated from one of the Edge devices connected to Gateway 202 the following steps can be implemented. In one step, GWD can look up in a local hash table to see if a NAT translation is available for the packet's five tuple. If no NAT translation is found, then GWD queries NATD 208 for a new translation. If NATD 208 has previously translated the five (5) tuple, then NATD can return the same translation that was provided before, ensuring session continuity. If NATD 208 has not previously translated the five tuple, it can create a new translation and return it. The new translation can be stored locally for future retrieval. When a flow is deleted on Gateway 202, a signal can be sent to NATD 208 to delete the translation and free the associated memory.



FIG. 3 illustrates an example system 300 of multiple physical or virtual instances of one more gateways 308 and 310 running and fronted by a single NATD daemon 306, according to some embodiments. This can allow for horizontal scaling of resources to provide internet connectivity from a larger number of branch devices. NATD daemon 306 can be implemented in load balancer 302. A process monitor (e.g. vc_procmon 304, 316, and 318) can launch and manage associated services (e.g. GWD 312 and 314, NATD daemon 306). In this scenario, the same NATD instance provides resiliency for NAT translations to multiple GWD instances using the same steps defined above.


It is noted that, in some embodiments, a single instance of the gateway (GWD) can have a finite number of Edge devices that can connect to it before it runs out of resources. In order to expand scale beyond this limit, a cluster of multiple gateway instances can be created and load can be distributed across those instances. These instances can share a single NATD to ensure that even if load is moved from one gateway instance to another, the session continuity is maintained.



FIG. 4 illustrates an example of cloud traffic routed through an external Cloud Web Security (CWS) service to enable security scanning and other service insertion on the traffic before it exits to the public Internet, according to some embodiments. In this scenario, the same NATD instance provides resiliency for NAT translations to multiple GWD instances using the same steps provided supra. In addition to routing traffic via the cloud gateway 410, cloud traffic can be routed through an external Cloud Web Security (CWS) 412 service to enable security scanning and other service insertion on the traffic before it exits to the public Internet. There are two mechanisms provided for connecting to the cloud service, via an aggregated IPsec tunnel through Gateway 410 and via an IPsec tunnel direct from the Edge 402 itself. By tracking the state of various connectivity points, this provides full redundancy for internet traffic even if any one of the tunnels fails.


There can be three tunnels established in the topology of FIG. 5 which enable the Edge 402 to reach the CWS 412 service. This tunnel is from the Gateway 410 to the CWS 412 service. This tunnel is established per Gateway 410 and one or more edges are assigned to this tunnel via profile, allowing edges to take advantage of the Edge-Gateway Multipath Method and ensure traffic resiliency in reaching the Gateway 410. This tunnel is established from the Edge directly to the CWS 412 service over ISP A 404. This tunnel is established from the Edge directly to the CWS 412 service over ISP B 406. Edge 402 is able to dynamically shift traffic per-packet over the optimal tunnel based on continuous measurement of tunnel state, latency, traffic priority, congestion, etc.


A variety of techniques can be used to maintain branch connectivity to data centers, cloud applications, etc. For example, in a data center topology, an edge device (e.g. an edge) can be deployed in two different ways. As a cluster, redundancy can be provided for a single data center leveraging multiple independent devices. Alternately, redundancy can be provided by deploying multiple physical edges in multiple data centers that are interconnected via routing external to a gateway. In the clustering topology, each edge in a cluster can report health statistics to a gateway at a specified period (e.g. every 30 seconds, etc.). This can enable it to make intelligent decisions about assignment and re-balancing.


Link Resiliency is now discussed. For resiliency of the individual links, multiple modes are provided. The method can include an Edge-Gateway Multipath Method where both links are considered active. In this topology, the reactivity time for blackout or brownout conditions is three hundred (300) ms and approximate bandwidth consumption on the second link is one thousand two-hundred and fifty (1250) MB per month.


Two additional modes can be provided which reduce the reactivity time but save on bandwidth consumption. The first mode provided is a pure backup mode, wherein tunnels are not established on the WAN link and ICMP probes alone are used to monitor link states. The link is still included in the link state machine tracking availability to determine availability for failover. This availability is reported as status on an Orchestrator and used to generate link up/down alerts though tunnels are not active. In this mode, usage is only twenty (20) MB per month but it may take up to two (2) seconds for the link to take over in case of blackout of the primary link and there is no brownout protection.


In a second mode, the link can be maintained in a “hot standby” mode wherein the tunnels are active however all MP control traffic is not sent across the link. In this mode, reactivity time can be seven-hundred milliseconds (700 ms) for blackout or brownout conditions and the usage is approximately two-hundred and fifty (250) MB per month.


Various cloud resiliency examples are now discussed. For cloud traffic, it can be that traffic continuity is maintained through a single peering, point due to NAT. However, the resiliency methods described above (e.g. multiple devices) can also be applicable to cloud traffic. Because sessions are translated to a given public Internet Protocol (IP) address, resiliency that utilizes multiple devices and instead resiliency behind a single NAT IP address is important. In this regard, the Gateway has the ability to provide resilient connectivity in a single or multi-device topology while preserving NAT state.


Quality of Experience Visibility is now discussed. FIG. 6 illustrates an example of a single DSL connected to two gateways, according to some embodiments. The path and link state machines will establish a set of flags to determine the eligibility of the path and link to meet scheduling criteria for the various traffic types. The path eligibility is used for path selection. The link eligibility is used for event reporting. Consider the case where there is a single DSL connected to two gateways “A” and “B”. There can be one link and two paths. Path DSL→A has 1% loss, Path DSL→B has 0% loss.


During path selection, Path DSL→A as REALTIME_VOICE_RED and Path DSL→B as REALTIME_VOICE_GREEN can be provided. Traffic to Gateway A can avoid this link while traffic to Gateway B would not. However, an event may not be generated because the issue is on the DSL→A path itself and does not appear to be local to the user's link. This abstracts the network (e.g. a Velocloud® network, etc.) problems from the user.


For a path, these flags are set based on the path statistics. For a link, these flags are set based on the statistics of the best path that is part of the link. This allows VeloCloud to report on the quality of the link irrespective of issues that may occur in the last mile of VeloCloud Gateway devices, as multiple gateways can provide for multiple data points that may be used to reflect quality.


Latency Eligibility Flags can be provided as follows:


REALTIME_VOICE_YELLOW if the average latency is >n ms.


REALTIME_VOICE_RED if the average latency is >n ms.


REALTIME_VIDEO_YELLOW if the average latency is >n ms.


REALTIME_VIDEO_RED if the average latency is >n ms.


TRANSACTIONAL_YELLOW if the average latency is >n ms.


TRANSACTIONAL_RED if the average latency is >n ms.


Jitter Eligibility Flags can be provided as follows:


REALTIME_VOICE_YELLOW if the average jitter is >n ms.


REALTIME_VOICE_RED if the average jitter is >n ms.


REALTIME_VIDEO_YELLOW if the jitter is >n ms.


REALTIME_VIDEO_RED if the jitter is >n ms.


TRANSACTIONAL_YELLOW if the jitter is >n ms.


TRANSACTIONAL_RED if the jitter is >n ms.


Loss Eligibility Flags can be provided as follows:


REALTIME_VOICE_YELLOW if the loss is >n %.


REALTIME_VOICE_RED if the loss is >n %.


REALTIME_VIDEO_YELLOW if the loss is >n %.


REALTIME_VIDEO_RED if the loss is >n %.


TRANSACTIONAL_YELLOW if the loss is >n %.


TRANSACTIONAL_RED if the less is >n %.


Path Selection methods are now discussed. During path selection, each packet can first check for the path with the lowest score that meet the jitter and loss eligibility criteria outlined for the traffic type of the flow selecting the path. For example, on the first selection of a real-time packet the check might be:


if ((jitter_flags & REALTIME_VOICE_RED)∥


(loss_flags & REALTIME_VOICE_RED))


continue;


If all the paths fail, the path with the lowest score can be chosen with the appropriate flags noted. If there are multiple eligible path the following steps can be implemented. A “fixed” path select can select the lowest score eligible path and stick to it. A “replicate” path select can send on the best scoring path for each packet and only start replicating if loss becomes an issue. A “load-balance” path select can select the best scoring path for each packet, eventually using all the eligible path if the load is high enough.


Aspects of jitter are now discussed. If an eligible path is found, the flow can select the path with jitter correction disabled. If no eligible paths are found, it can fall back to traditional path selection with jitter correction enabled. A flag can be set in the header indicating to the receive side to enable the jitter buffer. Once enabled for a flow, the jitter buffer will remain in place for the life of the flow, regardless of whether the situation clears.


Aspects of loss are now discussed. If eligible paths are found, the flow can select the path with loss correction disabled. If no eligible paths are found, it can fall back to traditional path selection with loss correction enabled. This means that loss correction state will be toggled dynamically on a per-packet basis based on the latest network conditions.


Aspects of event generation are now discussed. The link state machine can check if the above conditions are met and set/clear flags appropriately. Network events can be generated when the flags are set or cleared. The Orchestrator can display a summary chart of the quality of the link as measured and also the estimated quality once VeloCloud Error Corrections are applied. The target metrics for voice, video, transactional and bulk traffic can be measured separately and are user-configurable with recommended values. These can be used to generate a VeloCloud Quality Score (VQS) and color-coded chart as follows:


For the “before” state:


Good (“Green”): all metrics are better than the objective (obj) thresholds App. SLA met/exceeded


Fair (“Yellow”): Some or all metrics are between the objective (obj) and maximum (max) values—App. SLA is partially met


Poor (“Red”): some or all metrics have reached or exceeded the maximum (max) value—Application SLA is not met


For the “after” state:


Green: Best link meets the objective threshold or best link is yellow but can be corrected to green


Yellow: Best link does not meet the objective threshold and is yellow or best link is red but can be corrected to yellow


Red: Best link does not meet the objective threshold (is red), and cannot be corrected


An example VQS Calculation can be as follows:

VQS: Velocloud Quality Score=10*(% of time link was Green)+5*(% of time link was Yellow)+0*(% of time link was Red).


Additional Exemplary Computer Architecture and Systems



FIG. 5 depicts an exemplary computing system 500 that can be configured to perform any one of the processes provided herein. In this context, computing system 500 may include, for example, a processor, memory, storage, and I/O devices (e.g., monitor, keyboard, disk drive, Internet connection, etc.). However, computing system 500 may include circuitry or other specialized hardware for carrying out some or all aspects of the processes. In some operational settings, computing system 500 may be configured as a system that includes one or more units, each of which is configured to carry out some aspects of the processes either in software, hardware, or some combination thereof.



FIG. 5 depicts computing system 500 with a number of components that may be used to perform any of the processes described herein. The main system 502 includes a motherboard 504 having an I/O section 506, one or more central processing units (CPU) 508, and a memory section 510, which may have a flash memory card 512 related to it. The I/O section 506 can be connected to a display 514, a keyboard and/or other user input (not shown), a disk storage unit 516, and a media drive unit 518. The media drive unit 518 can read/write a computer-readable medium 520, which can contain programs 522 and/or data. Computing system 500 can include a web browser. Moreover, it is noted that computing system 500 can be configured to include additional systems in order to fulfill various functionalities. Computing system 500 can communicate with other computing devices based on various computer communication protocols such a Wi-Fi, Bluetooth® (and/or other standards for exchanging data over short distances includes those using short-wavelength radio transmissions), USB, Ethernet, cellular, an ultrasonic local area communication protocol, etc.


CONCLUSION

Although the present embodiments have been described with reference to specific example embodiments, various modifications and changes can be made to these embodiments without departing from the broader spirit and scope of the various embodiments. For example, the various devices, modules, etc. described herein can be enabled and operated using hardware circuitry, firmware, software or any combination of hardware, firmware, and software (e.g., embodied in a machine-readable medium).


In addition, it can be appreciated that the various operations, processes, and methods disclosed herein can be embodied in a machine-readable medium and/or a machine accessible medium compatible with a data processing system (e.g., a computer system), and can be performed in any order (e.g., including using means for achieving the various operations). Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense. In some embodiments, the machine-readable medium can be a non-transitory form of machine-readable medium.

Claims
  • 1. A system comprising: a first device at an edge of a branch first network;a second device operating as an access point to a second network; and
  • 2. The system of claim 1, wherein the first device is an edge device and the second device is a gateway device.
  • 3. The system of claim 1 further comprising a cloud web security service in the second network.
  • 4. The system of claim 3, wherein the cloud web security service performs security scanning for data traffic from the enterprise first network prior to the data traffic being sent to the public Internet.
  • 5. The system of claim 3, wherein the cloud web security service performs service insertion for data traffic from the enterprise first network prior to the data traffic being sent to the public Internet.
  • 6. The system of claim 3 further comprising a tunnel from the second device to the cloud web service.
  • 7. The system of claim 1, wherein the second device performs network address translation for data traffic sent from the first device to the second network.
  • 8. The system of claim 1, wherein the probe packets are ICMP probe packets.
  • 9. The system of claim 1, wherein states of the active first link and the backup second link are both included in a link state machine.
  • 10. The system of claim 9, wherein the link state machine determines availability of links for failover.
  • 11. The system of claim 10, wherein the availability is reported as a link status by a centralized configuration and management application.
  • 12. The system of claim 9, wherein the link state machine establishes a set of flags to determine eligibility of the links.
  • 13. The system of claim 1, wherein using probe packets without an established tunnel on the backup second link saves bandwidth consumption on the second link.
  • 14. The system of claim 1, wherein the first device executes in a virtual machine located in a branch office.
  • 15. The system of claim 1, wherein the first link uses a first Internet service provider and the second link uses a second Internet service provider.
  • 16. A system comprising: a first device at an edge of a branch location of an enterprise network;a second device operating as an access point to a public cloud network; andat least two links between the first device and the second device, wherein (i) a tunnel is established on an active first link of the at least two links, (ii) no tunnel is established on an inactive backup second link of the at least two links, and (iii) probe packets are used to monitor a state of the inactive backup second link in order to ensure that the second link can become active when needed to replace the first link;wherein if the first link becomes inactive, a second tunnel is established on the second link and the second link becomes the active link.
  • 17. The system of claim 16, wherein the first device is an edge device and the second device is a gateway device.
  • 18. The system of claim 16, wherein using probe packets without an established tunnel on the backup second link saves bandwidth consumption on the second link.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims priority to U.S. Provisional Application No. 62/523,477, titled and METHOD AND SYSTEM OF RESILIENCY AND VISIBILITY IN CLOUD-DELIVERED SD-WAN filed on 22 Jun. 2017. This provisional application is incorporated by reference in its entirety.

US Referenced Citations (165)
Number Name Date Kind
7003481 Banka et al. Feb 2006 B2
8111692 Ray Feb 2012 B2
8228928 Parandekar et al. Jul 2012 B2
8243589 Trost et al. Aug 2012 B1
8259566 Chen et al. Sep 2012 B2
8566452 Goodwin, III et al. Oct 2013 B1
8724456 Hong et al. May 2014 B1
8964548 Keralapura et al. Feb 2015 B1
9071607 Twitchell, Jr. Jun 2015 B2
9154327 Marino et al. Oct 2015 B1
9306949 Richard et al. Apr 2016 B1
9336040 Dong et al. May 2016 B2
9354983 Yenamandra May 2016 B1
9432245 Sorenson et al. Aug 2016 B1
9450817 Bahadur et al. Sep 2016 B1
9450852 Chen et al. Sep 2016 B1
9525564 Lee Dec 2016 B2
9665432 Kruse et al. May 2017 B2
9715401 Devine et al. Jul 2017 B2
9722815 Mukundan et al. Aug 2017 B2
9787559 Schroeder Oct 2017 B1
10135789 Mayya et al. Nov 2018 B2
10178032 Freitas Jan 2019 B1
10187289 Chen et al. Jan 2019 B1
10229017 Zou et al. Mar 2019 B1
10326830 Singh Jun 2019 B1
10348767 Lee et al. Jul 2019 B1
20020198840 Banka et al. Dec 2002 A1
20030112808 Solomon Jun 2003 A1
20030161313 Jinmei Aug 2003 A1
20030202506 Perkins et al. Oct 2003 A1
20030219030 Gubbi Nov 2003 A1
20040059831 Chu et al. Mar 2004 A1
20040068668 Lor et al. Apr 2004 A1
20050078690 DeLangis Apr 2005 A1
20060114838 Mandavilli et al. Jun 2006 A1
20060171365 Borella Aug 2006 A1
20060182034 Klinker et al. Aug 2006 A1
20060193247 Naseh et al. Aug 2006 A1
20070064604 Chen et al. Mar 2007 A1
20070091794 Filsfils et al. Apr 2007 A1
20070121486 Guichard et al. May 2007 A1
20070177511 Das et al. Aug 2007 A1
20070260746 Mirtorabi et al. Nov 2007 A1
20080049621 McGuire et al. Feb 2008 A1
20080080509 Khanna et al. Apr 2008 A1
20080095187 Jung et al. Apr 2008 A1
20080219276 Shah et al. Sep 2008 A1
20090154463 Hines et al. Jun 2009 A1
20090247204 Sennett et al. Oct 2009 A1
20100008361 Guichard et al. Jan 2010 A1
20100088440 Banks et al. Apr 2010 A1
20100118727 Draves et al. May 2010 A1
20100332657 Elyashev et al. Dec 2010 A1
20110075674 Li et al. Mar 2011 A1
20110110370 Moreno et al. May 2011 A1
20110153909 Dong Jun 2011 A1
20120008630 Ould-Brahim Jan 2012 A1
20120027013 Napierala Feb 2012 A1
20120157068 Eichen et al. Jun 2012 A1
20120173919 Patel et al. Jul 2012 A1
20120221955 Raleigh et al. Aug 2012 A1
20120250682 Vincent et al. Oct 2012 A1
20120250686 Vincent et al. Oct 2012 A1
20120300615 Kempf et al. Nov 2012 A1
20120317291 Wolfe Dec 2012 A1
20130019005 Hui et al. Jan 2013 A1
20130021968 Reznik et al. Jan 2013 A1
20130044764 Casado et al. Feb 2013 A1
20130051399 Zhang et al. Feb 2013 A1
20130124718 Griffith et al. May 2013 A1
20130124911 Griffith et al. May 2013 A1
20130124912 Griffith et al. May 2013 A1
20130128889 Mathur et al. May 2013 A1
20130173788 Song Jul 2013 A1
20130238782 Zhao et al. Sep 2013 A1
20130242718 Zhang Sep 2013 A1
20130254599 Katkar et al. Sep 2013 A1
20130258839 Wang et al. Oct 2013 A1
20130283364 Chang et al. Oct 2013 A1
20130301642 Radhakrishnan et al. Nov 2013 A1
20130329548 Nakil et al. Dec 2013 A1
20140019604 Twitchell, Jr. Jan 2014 A1
20140108665 Arora et al. Apr 2014 A1
20140156823 Liu et al. Jun 2014 A1
20140173113 Vemuri et al. Jun 2014 A1
20140219135 Li et al. Aug 2014 A1
20140223507 Xu Aug 2014 A1
20140244851 Lee Aug 2014 A1
20140317440 Biermayr et al. Oct 2014 A1
20150016249 Mukundan et al. Jan 2015 A1
20150029864 Raileanu et al. Jan 2015 A1
20150046572 Cheng Feb 2015 A1
20150096011 Watt Apr 2015 A1
20150172121 Farkas et al. Jun 2015 A1
20150188823 Williams et al. Jul 2015 A1
20150222543 Song Aug 2015 A1
20150236962 Veres et al. Aug 2015 A1
20150334696 Gu et al. Nov 2015 A1
20150350907 Timariu et al. Dec 2015 A1
20150363733 Brown Dec 2015 A1
20150372943 Hasan et al. Dec 2015 A1
20160072669 Saavedra Mar 2016 A1
20160142373 Ossipov May 2016 A1
20160164914 Madhav et al. Jun 2016 A1
20160197834 Luft Jul 2016 A1
20160197835 Luft Jul 2016 A1
20160198003 Luft Jul 2016 A1
20160210209 Verkaik et al. Jul 2016 A1
20160218947 Hughes et al. Jul 2016 A1
20160315912 Mayya et al. Oct 2016 A1
20160359738 Sullenberger et al. Dec 2016 A1
20170026283 Williams et al. Jan 2017 A1
20170034129 Sawant et al. Feb 2017 A1
20170053258 Carney et al. Feb 2017 A1
20170055131 Kong et al. Feb 2017 A1
20170064005 Lee Mar 2017 A1
20170123939 Maheshwari et al. May 2017 A1
20170126564 Mayya et al. May 2017 A1
20170134186 Mukundan et al. May 2017 A1
20170195169 Mills et al. Jul 2017 A1
20170201585 Doraiswamy et al. Jul 2017 A1
20170207976 Rovner et al. Jul 2017 A1
20170214701 Hasan Jul 2017 A1
20170223117 Messerli et al. Aug 2017 A1
20170237710 Mayya et al. Aug 2017 A1
20170257309 Appanna Sep 2017 A1
20170279717 Bethers et al. Sep 2017 A1
20170310691 Vasseur et al. Oct 2017 A1
20170317974 Masurekar et al. Nov 2017 A1
20170339070 Chang et al. Nov 2017 A1
20170364419 Lo Dec 2017 A1
20180014051 Phillips et al. Jan 2018 A1
20180034668 Mayya et al. Feb 2018 A1
20180041425 Zhang Feb 2018 A1
20180074909 Bishop et al. Mar 2018 A1
20180084081 Kuchibhotla et al. Mar 2018 A1
20180176082 Katz et al. Jun 2018 A1
20180234300 Mayya et al. Aug 2018 A1
20180270104 Zheng et al. Sep 2018 A1
20180278541 Wu et al. Sep 2018 A1
20180295529 Jen et al. Oct 2018 A1
20180302286 Mayya et al. Oct 2018 A1
20180351855 Sood et al. Dec 2018 A1
20180375744 Mayya et al. Dec 2018 A1
20190058709 Kempf et al. Feb 2019 A1
20190075083 Mayya et al. Mar 2019 A1
20190103990 Cidon et al. Apr 2019 A1
20190103991 Cidon et al. Apr 2019 A1
20190103992 Cidon et al. Apr 2019 A1
20190103993 Cidon et al. Apr 2019 A1
20190104035 Cidon et al. Apr 2019 A1
20190104049 Cidon et al. Apr 2019 A1
20190104050 Cidon et al. Apr 2019 A1
20190104051 Cidon et al. Apr 2019 A1
20190104052 Cidon et al. Apr 2019 A1
20190104053 Cidon et al. Apr 2019 A1
20190104063 Cidon et al. Apr 2019 A1
20190104064 Cidon et al. Apr 2019 A1
20190104109 Cidon et al. Apr 2019 A1
20190104111 Cidon et al. Apr 2019 A1
20190104413 Cidon et al. Apr 2019 A1
20190140889 Mayya et al. May 2019 A1
20190140890 Mayya et al. May 2019 A1
20190158605 Markuze et al. May 2019 A1
Foreign Referenced Citations (6)
Number Date Country
1912381 Apr 2008 EP
3041178 Jul 2016 EP
2012167184 Dec 2012 WO
2017083975 May 2017 WO
2019070611 Apr 2019 WO
2019094522 May 2019 WO
Non-Patent Literature Citations (7)
Entry
Petition for Post-Grant Review of U.S. Pat. No. 9,722,815, filed May 1, 2018, 106 pages.
Non-Published commonly Owned U.S. Appl. No. 15/707,124, filed Sep. 18, 2017, 24 pages, Nicira, Inc.
Non-Published commonly Owned U.S. Appl. No. 15/784,404, filed Oct. 16, 2017, 21 pages, Nicira, Inc.
Non-Published commonly Owned U.S. Appl. No. 15/811,329, filed Nov. 13, 2017, 37 pages, Nicira, Inc.
Non-Published commonly Owned U.S. Appl. No. 15/838,052, filed Dec. 11, 2017, 28 pages, Nicira, Inc.
Non-Published commonly Owned U.S. Appl. No. 15/838,355, filed Dec. 11, 2017, 29 pages, Nicira, Inc.
Mudigonda, Jayaram, et al., “NetLord: A Scalable Multi-Tenant Network Architecture for Virtualized Datacenters,” Proceedings of the ACM SIGCOMM 2011 Conference, Aug. 15-19, 2011, 12 pages, ACM, Toronto, Canada.
Related Publications (1)
Number Date Country
20180375824 A1 Dec 2018 US
Provisional Applications (1)
Number Date Country
62523477 Jun 2017 US