Dynamic SD-WAN hub cluster scaling with machine learning

Information

  • Patent Grant
  • 11979325
  • Patent Number
    11,979,325
  • Date Filed
    Wednesday, January 5, 2022
    3 years ago
  • Date Issued
    Tuesday, May 7, 2024
    9 months ago
Abstract
Some embodiments of the invention provide a method of dynamically scaling a hub cluster in a software-defined wide area network (SD-WAN) based on particular traffic statistics, the hub cluster being located in a datacenter of the SD-WAN and allowing branch sites of the SD-WAN to access resource of the datacenter by connecting to the hub cluster. A controller of the SD-WAN receives, from the hub cluster, traffic statistics centrally captured at the hub cluster. The controller then analyzes these statistics to identify traffic load fluctuations, and determines that a number of hubs in the hub cluster should be adjusted based on the identified fluctuations. The controller adjusts the number of hubs in the hub cluster based on the determination.
Description
BACKGROUND

Today, SD-WAN hub-clustering is a networking mechanism akin to server load balancing. The hub clusters reside in datacenters and allow branch sites to access resources of the datacenter by connecting to one of the hubs of the cluster in a load-balanced fashion. Determining the number of hubs to be included in the cluster, however, remains a challenge faced by network designers, as the determination is based on multiple factors, such as the number of branches and network usage patterns of those branches. As a result, network operators tend to over-provision hubs, sometimes by as much as a factor of two. The lack of an easy solution can result in unsatisfied and frustrated customers who are increasingly using cluster solutions in cloud deployments and IaaS (infrastructure as a service).


BRIEF SUMMARY

Some embodiments of the invention provide a method of dynamically scaling a hub cluster in a software-defined wide area network (SD-WAN) based on current and historical traffic statistics. The SD-WAN includes a datacenter in which the hub cluster is located, a controller, and multiple branch sites that connect to the hub cluster in order to access resources of the datacenter. In addition to providing access to resources of the datacenter, the hub cluster captures and sends traffic statistics relating to flows to and from the branch sites, and provides these traffic statistics to the controller of the SD-WAN. The controller analyzes the received traffic statistics to identify traffic load fluctuations, and determines that a number of hubs in the hub cluster should be adjusted based on the identified fluctuations. The controller then adjusts the number of hubs in the hub cluster based on the determination.


In some embodiments, each branch site includes an edge forwarding node for sending and receiving traffic flows between the branch site and other elements of the SD-WAN (e.g., between the branch site and the hub cluster). In addition to the traffic statistics received from the hub cluster, the controller in some embodiments also receives additional traffic statistics from the edge forwarding nodes at the branch sites. In some embodiments, the additional traffic statistics received from the edge forwarding nodes include more fine-grained statistics compared to the traffic statistics received from the hub cluster.


The controller, in some embodiments, includes a traffic statistics storage, a learning engine, and a decision-making and evaluation engine. In some embodiments, the traffic statistics storage stores historical traffic statistics (i.e., traffic statistics received over time). The historical traffic statistics and the real-time traffic statistics received by the controller are used as input for the learning engine, according to some embodiments. For example, the real-time traffic statistics of some embodiments are input into the learning engine to identify real-time load fluctuations (e.g., fluctuations happening at short time scales), while the historical traffic statistics in addition to the current traffic statistics are input into the learning engine to identify cyclical load fluctuations (e.g., daily morning/night fluctuations) and seasonal load fluctuations (e.g., fluctuations during peak business operations). By identifying these fluctuations, the learning engine and decision making and evaluation engine of some embodiments are able to adapt and adjust the hub cluster size in response to real-time, cyclical, and seasonal load fluctuations.


In some embodiments, the traffic statistics (both historical and real-time) are divided into two sets. The first set includes flow traffic load statistics that form the traffic parameters used by the learning engine to identify the fluctuations described above, while the second set includes flow performance statistics that form the evaluation parameters of the decision-making and evaluation engine to determine whether to adjust the hub cluster size. Flow traffic load statistics, in some embodiments, include at least the number of branch sites in the SD-WAN, the rate of flow creation per branch site, the total number of flows per branch site, flows further categorized into applications, average bandwidth per flow, total number of fat flows per branch, and average bandwidth per fat flow. The flow performance statistics in some embodiments include at least packet re-transmits per flow, packet drops per flow, and mean opinion score (MOS) of real-time flows (e.g., VoIP calls).


In order to determine that a number of hubs in the hub cluster should be adjusted based on the identified fluctuations (e.g., real-time, cyclical, or seasonal fluctuations), the decision making and evaluation engine in some embodiments uses the identified fluctuations to predict the traffic load over a certain time interval (e.g., the next three hours) and to determine the hub cluster size that can handle the predicted load with the expected flow performance (e.g., based on prior tests on sizing of the hub cluster at different traffic loads). Based on the different between the current hub cluster size and required cluster size, the decision making and evaluation engine decides whether to scale the cluster up (i.e., add hubs) or to scale the cluster down (i.e., remove hubs).


In some embodiments, in addition to deciding whether to scale the cluster up or down, the decision making and evaluation engine evaluates the learning engine's performance by comparing the predicted traffic load and the actual traffic load. This comparison, in some embodiments, can trigger a re-training of the learning engine if the engine's predictions start to deviate significantly from the actual values. In some embodiments, re-training can also be triggered after a certain time-interval to account for new data.


The preceding Summary is intended to serve as a brief introduction to some embodiments of the invention. It is not meant to be an introduction or overview of all inventive subject matter disclosed in this document. The Detailed Description that follows and the Drawings that are referred to in the Detailed Description will further describe the embodiments described in the Summary as well as other embodiments. Accordingly, to understand all the embodiments described by this document, a full review of the Summary, the Detailed Description, the Drawings, and the Claims is needed. Moreover, the claimed subject matters are not to be limited by the illustrative details in the Summary, the Detailed Description, and the Drawings.





BRIEF DESCRIPTION OF FIGURES

The novel features of the invention are set forth in the appended claims. However, for purposes of explanation, several embodiments of the invention are set forth in the following figures.



FIG. 1 conceptually illustrates an example of an SD-WAN in some embodiments.



FIG. 2 conceptually illustrates an example of the machine-learning model of some embodiments.



FIGS. 3A-3B conceptually illustrate an example of an SD-WAN before and after a machine-learning model has adjusted the hub cluster based on identified load fluctuations, according to some embodiments.



FIG. 4 illustrates a process performed by the controller in some embodiments to identify real-time load fluctuations.



FIG. 5 illustrates a process performed by the controller in some embodiments to identify cyclical load fluctuations.



FIG. 6 illustrates a process performed by the controller in some embodiments to identify seasonal load fluctuations.



FIG. 7 illustrates a process performed by the controller to evaluate the performance of the machine-learning model, in some embodiments.



FIG. 8 conceptually illustrates a computer system with which some embodiments of the invention are implemented.





DETAILED DESCRIPTION

In the following detailed description of the invention, numerous details, examples, and embodiments of the invention are set forth and described. However, it will be clear and apparent to one skilled in the art that the invention is not limited to the embodiments set forth and that the invention may be practiced without some of the specific details and examples discussed.


Some embodiments of the invention provide a method of dynamically scaling a hub cluster in a software-defined wide area network (SD-WAN) based on current and historical traffic statistics. The SD-WAN includes a datacenter in which the hub cluster is located, a controller, and multiple branch sites that connect to the hub cluster in order to access resources of the datacenter. In addition to providing access to resources of the datacenter, the hub cluster captures and sends traffic statistics relating to flows to and from the branch sites, and provides these traffic statistics to the controller of the SD-WAN. The controller analyzes the received traffic statistics to identify traffic load fluctuations, and determines that a number of hubs in the hub cluster should be adjusted based on the identified fluctuations. The controller then adjusts the number of hubs in the hub cluster based on the determination.


In some embodiments, each branch site includes an edge forwarding node for sending and receiving traffic flows between the branch site and other elements of the SD-WAN (e.g., between the branch site and the hub cluster). In addition to the traffic statistics received from the hub cluster, the controller in some embodiments also receives additional traffic statistics from the edge forwarding nodes at the branch sites. In some embodiments, the additional traffic statistics received from the edge forwarding nodes include more fine-grained statistics compared to the traffic statistics received from the hub cluster.


The controller, in some embodiments, includes a traffic statistics storage, a learning engine, and a decision-making and evaluation engine. In some embodiments, the traffic statistics storage stores historical traffic statistics (i.e., traffic statistics received over time). The historical traffic statistics and the real-time traffic statistics received by the controller are used as input for the learning engine, according to some embodiments. For example, the real-time traffic statistics of some embodiments are input into the learning engine to identify real-time load fluctuations (e.g., fluctuations happening at short time scales), while the historical traffic statistics in addition to the current traffic statistics are input into the learning engine to identify cyclical load fluctuations (e.g., daily morning/night fluctuations) and seasonal load fluctuations (e.g., fluctuations during peak business operations). By identifying these fluctuations, the learning engine and decision making and evaluation engine of some embodiments are able to adapt and adjust the hub cluster size in response to real-time, cyclical, and seasonal load fluctuations.


In some embodiments, the traffic statistics (both historical and real-time) are divided into two sets. The first set includes flow traffic load statistics that form the traffic parameters used by the learning engine to identify the fluctuations described above, while the second set includes flow performance statistics that form the evaluation parameters of the decision-making and evaluation engine to determine whether to adjust the hub cluster size. Flow traffic load statistics, in some embodiments, include at least the number of branch sites in the SD-WAN, the rate of flow creation per branch site, the total number of flows per branch site, flows further categorized into applications, average bandwidth per flow, total number of fat flows per branch, and average bandwidth per fat flow. The flow performance statistics in some embodiments include at least packet re-transmits per flow, packet drops per flow, and mean opinion score (MOS) of real-time flows (e.g., VoIP calls).


In order to determine that a number of hubs in the hub cluster should be adjusted based on the identified fluctuations (e.g., real-time, cyclical, or seasonal fluctuations), the decision making and evaluation engine in some embodiments uses the identified fluctuations to predict the traffic load over a certain time interval (e.g., the next three hours) and to determine the hub cluster size that can handle the predicted load with the expected flow performance (e.g., based on prior tests on sizing of the hub cluster at different traffic loads). Based on the different between the current hub cluster size and required cluster size, the decision making and evaluation engine decides whether to scale the cluster up (i.e., add hubs) or to scale the cluster down (i.e., remove hubs).


In some embodiments, in addition to deciding whether to scale the cluster up or down, the decision making and evaluation engine evaluates the learning engine's performance by comparing the predicted traffic load and the actual traffic load. This comparison, in some embodiments, can trigger a re-training of the learning engine if the engine's predictions start to deviate significantly from the actual values. In some embodiments, re-training can also be triggered after a certain time-interval to account for new data.



FIG. 1 illustrates an example embodiment of an SD-WAN (also referred to herein as a virtual network) for connecting multiple branch sites to each other and to a controller and at least one datacenter hub. As shown, the SD-WAN 100 includes a controller 110, three branch sites 120-124 that each include an edge forwarding node 130-134 (also referred herein as edge nodes or nodes), a cloud gateway 140, and a datacenter 150 with a hub 145.


The edge nodes in some embodiments are edge machines (e.g., virtual machines (VMs), containers, programs executing on computers, etc.) and/or standalone appliances that operate at multi-computer locations of the particular entity (e.g., at an office or datacenter of the entity) to connect the computers at their respective locations other nodes, hubs, etc. in the virtual network. In some embodiments, the nodes are clusters of nodes at each of the branch sites. In other embodiments, the edge nodes are deployed to each of the branch sites as high-availability pairs such that one edge node in the pair is the active node and the other edge node in the pair is the standby node that can take over as the active edge node in case of failover.


An example of an entity for which such a virtual network can be established includes a business entity (e.g., a corporation), a non-profit entity (e.g., a hospital, a research organization, etc.), and an education entity (e.g., a university, a college, etc.), or any other type of entity. Examples of public cloud providers include Amazon Web Services (AWS), Google Cloud Platform (GCP), Microsoft Azure, etc., while examples of entities include a company (e.g., corporation, partnership, etc.), an organization (e.g., a school, a non-profit, a government entity, etc.), etc. In other embodiments, hubs like the hub 145 can also be deployed in private cloud datacenters of a virtual WAN provider that hosts hubs to establish SD-WANs for different entities.


In the example SD-WAN 100, the hub 145 is a multi-tenant forwarding element that is deployed on the premises of the datacenter 150. The hub 145 can be used to establish secure connection links (e.g., tunnels) with edge nodes at the particular entity's multi-computer sites, such as branch sites 130-134, third party datacenters (not shown), etc. For example, the hub 145 can be used to provide access from each branch site 120-124 to each other branch site 120-124 (e.g., via the connection links 160 that terminate at the hub 145) as well as to the resources 155 of the datacenter 150. These multi-computer sites are often at different physical locations (e.g., different buildings, different cities, different states, etc.), according to some embodiments. In some embodiments, hubs can be deployed as physical nodes or virtual nodes. Additionally, hubs in some embodiments can be deployed on a cloud (e.g., as a set of virtual edges configured as a cluster).


In the SD-WAN 100, the hub 145 also provides access to the resources 155 of the datacenter 150 as mentioned above. The resources in some embodiments include a set of one or more servers (e.g., web servers, database servers, etc.) within a microservices container (e.g., a pod). Conjunctively, or alternatively, some embodiments include multiple such microservices containers, each accessible through a different set of one or more hubs of the datacenter (not shown). The resources, as well as the hubs, are within the datacenter premises, according to some embodiments. While not shown, some embodiments include multiple different SaaS datacenters, which may each be accessed via different sets of hubs, according to some embodiments. In some embodiments, the SaaS datacenters include datacenters for video conferencing SaaS providers, for middlebox (e.g., firewall) service providers, for storage service providers, etc.


Additional examples of resources accessible via the hub 145, in some embodiments, include compute machines (e.g., virtual machines and/or containers providing server operations), storage machines (e.g., database servers), and middlebox service operations (e.g., firewall services, load balancing services, encryption services, etc.). In some embodiments, the connections 160 between the branch sites and the hub 145 are secure encrypted connections that encrypt packets exchanged between the edge nodes 130-134 of the branch sites and the hub 145. Examples of secure encrypted connections used in some embodiments include VPN (virtual private network) connections, or secure IPsec (Internet Protocol security) connection.


In some embodiments, multiple secure connection links (e.g., multiple secure tunnels) can be established between an edge node and the hub 145. When multiple such links are defined between a node and a hub, each secure connection link, in some embodiments, is associated with a different physical network link between the node and an external network. For instance, to access external networks in some embodiments, a node has one or more commercial broadband Internet links (e.g., a cable mode and a fiber optic link) to access the Internet, a wireless cellular link (e.g., a 5G LTE network), etc. The collection of the edge nodes, gateway, datacenter hub, controller, and secure connections between the edge nodes, gateway, datacenter hub, and controller form the SD-WAN 100.


As mentioned above, the controller 110 communicates with each of the nodes 130-134 at the branch sites 120-124, in some embodiments, to send information such as trace monitoring rules and receive information such as trace monitoring results via the connection links 170A-170C. In some embodiments, the controller 110 also provides trace monitoring rules to, and receives trace monitoring results from, the gateway 140 via the connection link 170D and the hub 145 via the connection link 170E. While illustrated as individual connection links, the links 170A-170E are sets of multiple connection links, according to some embodiments.


In addition to the connection links 170A-170E and 160, edge nodes 132 and 134 are connected via connection link 164, while edge nodes 130 and 132 are connected to the gateway 140 via connection links 162. The gateway 140 in this example is responsible for relaying information between edge nodes (e.g., edge nodes 120 and 122, which do not share a direct connection). Also, the gateway 140 in some embodiments is used to set up direct edge-to-edge connections. In some embodiments, the gateway 140 can be used to provide the edge nodes with access to cloud resources (e.g., compute, storage, and service resources of a cloud datacenter).



FIG. 2 illustrates an example of the machine-learning model used in the embodiments described herein. The machine-learning model 200 in some embodiments is a component of the controller (e.g., controller 110). The machine-learning model includes the learning engine component 205 and the decision-making and evaluation component 210. In some embodiments, the machine learning model 200 is a time-series model that captures real-time, cyclical, and seasonal load fluctuations. The model can be a single mathematical model that captures all three fluctuations (e.g., ARIMA (autoregressive integrated moving average model) in some embodiments, or it can include a collection of different mathematical models, each of which capture the relevant fluctuation in some embodiments.


When the model includes a collection of different mathematical models, the cyclical and seasonal parameters are learned from the time-series data, and subsequently subtracted from the time-series data to capture real-time fluctuations, as will be described in further detail below. In some embodiments, the model 200 is trained on historical data and updated periodically based on new data as well as based on the model's decisions are affecting system performance.


The decision-making and evaluation engine 210, in some embodiments, uses fluctuations identified by the learning component 205 to predict traffic loads and adjust the hub cluster accordingly to account for the predicted loads. Additionally, the decision-making and evaluation engine 210 in some embodiments evaluates the performance of the machine learning model 200, and triggers re-training if necessary, as mentioned above.


In some embodiments, the machine-learning model 200 makes both real-time and historical decisions based off of the received traffic statistics. For example, using parameters like the number of branch sites, packet re-transmits or drops per flow, MOS scores as input works by having a statistical co-relation of these parameters across all of the branch sites (e.g., branch sites 120-124), in some embodiments. While one, or a handful, of branch sites reporting increased retransmits or drops may not be indicative of much, the co-relation can indicate that the increase in these parameters are indeed connected to each other and thus probably have the same root cause, according to some embodiments. As a result, the machine-learning model in some embodiments can concur that the hub cluster requires scaling.


An example of a historical decision can include using parameters like flows per second, bandwidth consumed by the flows, and fat flow bandwidth. For example, an increase of fat flows when a few branch sites perform data backup at night can indicate that the cluster needs to be scaled up at time T on days D. As another example, the number of flows per second (i.e., rate of creation of flows) can be a significant parameter that impacts the operation of large networking devices. These flows may not consume bandwidth, but the sheer volume of flows impacts the scale. By using historical data, the hub can be scaled up to account for the sudden increase in the rate of flow creations.


In a third example, a sudden onslaught of traffic for an airport's SD-WAN when several flights land around the same time can bring a sudden burst of network users, thus degrading network connectivity in an unpredictable manner. Such a model is an easy learnable model to scale hubs beforehand to avoid degradation of network connectivity.


The real-time decisions made by the machine-learning model kick in after the problem has occurred, and are meant to mitigate in time to prevent further degradation, while the historical decisions kick in before the problem occurs to avoid the problem completely. As a result, most problems are initially corrected in real-time, while historical decisions will start to fix problems over the course of operation of the network.


In some embodiments, due to the nature the historical decisions, new hubs (e.g., VMs) can be spawned ahead of time and terminated when their need has ended. While spawning new hubs can be time consuming, it does not require other resources while not in use, which works well for historical decisions allow for enough time to spawn these hubs. Real-time needs, however, are more time-constrained and the thus spawning new hubs to solve real-time problems is untenable. Thus, some embodiments have a pool of backup hubs that do not occupy any compute resources, but are fully configured and ready to launch (e.g., from a disk/storage/configuration perspective). Additional details regarding the functionality of the machine learning model 200 will be described in further detail below.



FIGS. 3A-3B illustrate an example embodiment of an SD-WAN before and after the hub cluster size has been altered. In these example, the SD-WAN 300 includes a datacenter 350, branch sites 320-324, and a controller 310. The datacenter 350 includes hub clusters 345A and 345B, respectively. Like the hub cluster 145, the hub cluster 345A/B provides the branch sites 320-324 with access to the datacenter resources 355.


As shown in FIG. 3A, the hub cluster 345A and edge nodes at the branch sites 320-324 provide traffic statistics to the learning engine 312 of the controller 310. The learning engine then identifies fluctuations for the decision making and evaluation engine 314 to use to predict traffic load for a time period and determine the necessary hub cluster size for the predicted load. The decision making and evaluation engine 314 includes a sizing tests storage 316 for storing previous hub cluster size calculations, and a load comparisons storage 318 for storing comparisons of actual and predicted loads.


After the decision-making and evaluation engine 314 performs its calculations, it adjusts the hub cluster size by scaling the cluster up or down to ensure the hub cluster can handle the predicted traffic load and is not using unnecessary bandwidth (e.g., as illustrated by the encircled 3). In some embodiments, the decision-making and evaluation engine also evaluates the performance of the machine learning model (i.e., the performance of the learning engine and decision-making and evaluation engine).


As illustrated in FIG. 3B, the hub cluster 345B includes 5 hubs compared to the 3 hubs in the hub cluster 345A as a result of the adjustment triggered by the decision-making and evaluation engine. Additionally, the hub cluster and branch sites are providing traffic statistics indicative of the actual load to the controller.


The decision-making and evaluation engine 314 then compares the actual load to the predicted load, in some embodiments, and then averages the performance of the machine-learning model over time (e.g., using data from the load comparison storage 318). When the calculations indicate significant deviations over time between the predicted loads and actual loads, model re-training is triggered, as illustrated by the encircled 3. Additional details regarding the processes performed by the machine-learning model of the controller will be described in further detail below with reference to FIGS. 4-7.



FIG. 4 illustrates a process performed by the controller in some embodiments to adjust hub cluster size based on identified real-time load fluctuations. The process 400 starts, at 405, with the controller receiving real-time traffic statistics from the hub cluster and branch sites. For example, FIG. 3A illustrates the learning engine 312 receiving traffic statistics from the edge nodes at the branch sites 320-324, as well as from the hub cluster 345A. In some embodiments, the learning engine only receives these statistics from the hub cluster and not from the branch sites.


After receiving the real-time traffic statistics at 405, the controller analyzes, at 410, the real-time traffic statistics along with historical traffic statistics using a time-series model to determine cyclical and seasonal parameters. The controller then subtracts, at 415, the cyclical and seasonal parameters from the time-series data (i.e., the real-time and historical traffic parameters) to identify potential real-time load fluctuations. Real-time load fluctuations are traffic flow fluctuations that happen at short time scales (e.g., minutes to hours), according to some embodiments. For example, in some embodiments, a burst of users connecting to an application can cause a sudden burst of traffic.


The controller next determines, at 420, whether any real-time load fluctuations have been identified. When the controller determines that no real-time load fluctuations have been identified, the process returns to 405 to receive real-time traffic statistics from the hub cluster and branch sites. Otherwise, when the controller determines at 420 that real-time load fluctuations have been identified, the process transitions to 425 to predict the traffic load for a selected time period (e.g., the next 3 hours). For example, after the learning engine 312 of the controller 310 identifies any load fluctuations, the decision-making and evaluation engine 314 uses the identified load fluctuations to predict traffic load for an upcoming time interval, in some embodiments.


Next, the controller determines, at 425, a hub cluster size that can handle the predicted traffic load with the expected flow performance. In some embodiments, this determination can be based on prior tests on sizing of the hub cluster at different traffic loads. In the example of FIG. 3A, the decision-making and evaluation engine 314 of the controller 310 makes this determination after predicting the traffic load.


After determining the needed hub cluster size at 425, the controller determines at 430 whether to scale-up the hub cluster. For example, when the needed hub cluster size determined by the decision-making and evaluation engine is larger than the current hub cluster size, a scale-up decision is triggered. When the controller determines that the hub cluster should be scaled up, the process transitions to 435 to add hubs to the cluster. The process then returns to 405 to receive real-time traffic statistics.


Alternatively, when the controller determines that the hub cluster should not be scaled up, the process transitions to 440 to determine whether to scale down the hub cluster. For example, when the determined hub cluster size for the predicted traffic load is smaller than the current hub cluster size, a scale-down decision is triggered. Accordingly, when the controller determines at 440 that the hub cluster should be scaled-down, the process transitions to 445 to remove hubs from the hub cluster. The process then transitions back to 405 to receive real-time traffic statistics.


When the controller alternately determines at 440 that the hub cluster should not be scaled-down, the process returns to 405 to receive real-time traffic statistics. In some embodiments, this occurs when the determined hub cluster size is equal to the current hub cluster size, and thus the cluster does not need to be scaled up or down.



FIG. 5 illustrates a process 500 performed by the controller in some embodiments to adjust hub cluster size based on identified cyclical load fluctuations. The process 500 starts at 505 when the controller receives real-time traffic statistics from the hub cluster and branch sites. Next, at 510, the controller analyzes the real-time traffic statistics along with historical traffic statistics using a time-series model to identify potential cyclical load fluctuations. In some embodiments, cyclical load fluctuations are traffic flow fluctuations that occur in a cyclical fashion at the time scale of business operations. For example, morning and night fluctuations can occur daily at an enterprise when users log into the enterprise network in the morning, and subsequently drop off from the network at night, according to some embodiments.


Following the analysis, the controller determines, at 515, whether any cyclical load fluctuations were identified. When the controller determines that no cyclical load fluctuations were identified (e.g., a constant load), the process transitions back to 505 to receive real-time traffic statistics. Otherwise, when the controller determines at 515 that cyclical load fluctuations have been identified, the process transitions to 520 to predict the traffic load for a selected cyclical time period. For example, in some embodiments, the controller predicts the traffic load at the time that users log into the enterprise network in the example above.


Next, the controller determines, at 525, the hub cluster size needed for the predicted traffic load. As described above, the controller can base this determination off of previous load test, in some embodiments. At 530, the controller determines whether the hub cluster should be scaled-up. For example, if the determined hub cluster size is greater than the current hub cluster size, then the hub cluster needs to be scaled-up (i.e., for the cyclical time period). When the controller determines at 530 that the hub cluster needs to be scaled-up, the process transitions to 535 to add hubs to the hub cluster. The process then returns to 505 to receive real-time traffic statistics.


Alternatively, when the controller determines at 530 that the hub cluster should not be scaled up, the process transitions to 540 to determine whether the hub cluster should be scaled down. In other words, the process determines whether the determined hub cluster size is smaller than the current hub cluster size. When the controller determines that the hub cluster should be scaled down, the process transitions to 545 to remove hubs from the hub cluster. The process then returns to 505. Otherwise, when the controller determines at 540 that the hub cluster should not be scaled down (i.e., the determined hub cluster size is equal to the current hub cluster size), the process returns to 505 without removing any hubs.



FIG. 6 illustrates a process 500 performed by the controller in some embodiments to adjust hub cluster size based on identified seasonal load fluctuations. The process 600 starts at 605 when the controller receives real-time traffic statistics from the hub cluster and branch sites. The controller then analyzes, at 610, the real-time traffic statistics along with historical traffic statistics using a time-series model to identify potential seasonal load fluctuations. In some embodiments, seasonal load fluctuations are traffic flow fluctuations that occur on a seasonal basis (i.e., on the order of months). For example, in some embodiments, the network traffic load during peak business operations periods could be much higher than normal (e.g., increased online shopping during the months of November and December).


Next, the controller determines, at 615, whether any seasonal load fluctuations were identified during the analysis. When the controller determines that no seasonal load fluctuations were identified (e.g., a constant load for seasonal periods), the process transitions back to 605 to receive real-time traffic statistics. Otherwise, when the controller determines at 615 that seasonal load fluctuations have been identified, the process transitions to 620 to predict the traffic load for a selected seasonal time period. In some embodiments, for example, the controller predicts the traffic load for shopping vendors during peak sales times each year.


At 625, the controller determines the hub cluster size needed for the predicted traffic load. In some embodiments, the controller can base this determination off of previous load test, as mentioned above. Next, at 630, the controller determines whether the hub cluster should be scaled-up. For example, if the determined hub cluster size is larger than the current hub cluster size, then the hub cluster needs to be scaled-up (i.e., for the seasonal time period). When the controller determines at 630 that the hub cluster needs to be scaled-up, the process transitions to 635 to add hubs to the hub cluster in order to scale up the cluster. The process then returns to 605 to receive real-time traffic statistics.


When the controller alternately determines at 630 that the hub cluster should not be scaled up (i.e., the determined hub cluster size is not larger than the current hub cluster size), the process transitions to 640 to determine whether the hub cluster should be scaled down. In other words, the process determines whether the determined hub cluster size is smaller than the current hub cluster size. When the controller determines that the hub cluster should be scaled down, the process transitions to 645 to remove hubs from the hub cluster. The process then returns to 605. Otherwise, when the controller determines at 640 that the hub cluster should not be scaled down (i.e., the determined hub cluster size is equal to the current hub cluster size), the process returns to 605 without removing any hubs.


In some embodiments, the processes 400-600 are performed concurrently, while in other embodiments, these processes are performed separately or in groups. In addition to the steps described for the processes 400-600, the controller of some embodiments evaluates the performance of the machine learning model by comparing the predicted traffic load to the actual traffic load (i.e., once it occurs). FIG. 7 illustrates a process performed by the controller in some embodiments to evaluate the machine learning model's performance.


The process 700 starts, at 710, by receiving data indicating actual traffic load during a particular time period (e.g., months during a seasonal load fluctuation). In some embodiments, the decision-making and evaluation engine of the machine learning model in the controller performs this process. The controller then compares, at 720, the actual traffic load with the predicted traffic load.


Next, at 730, the controller determines whether there is a significant deviation between the actual traffic load and the predicted traffic load. When the controller determines that there is no significant deviation, the process ends. Otherwise, when the controller determines at 730 that there is a significant deviation between the loads, the process transitions to 740 to calculate the average performance of the machine learning model over a particular time-interval. For example, the decision-making and evaluation engine 314 of the controller 310 includes a storage 318 for storing performance-related data.


Based on the calculations, the controller determines at 750 whether there have been significant deviations over time. When the controller determines that there have not been significant deviations over time, the process ends. Otherwise, when the controller determines that there have been significant deviations over time based on the calculations, the process transitions to 760 to trigger re-training of the machine-learning model. Following 760, the process ends. In addition to the process 700, the machine-learning model re-training in some embodiments can be triggered after a certain time-interval to account for new data.


Many of the above-described features and applications are implemented as software processes that are specified as a set of instructions recorded on a computer readable storage medium (also referred to as computer readable medium). When these instructions are executed by one or more processing unit(s) (e.g., one or more processors, cores of processors, or other processing units), they cause the processing unit(s) to perform the actions indicated in the instructions. Examples of computer readable media include, but are not limited to, CD-ROMs, flash drives, RAM chips, hard drives, EPROMs, etc. The computer readable media does not include carrier waves and electronic signals passing wirelessly or over wired connections.


In this specification, the term “software” is meant to include firmware residing in read-only memory or applications stored in magnetic storage, which can be read into memory for processing by a processor. Also, in some embodiments, multiple software inventions can be implemented as sub-parts of a larger program while remaining distinct software inventions. In some embodiments, multiple software inventions can also be implemented as separate programs. Finally, any combination of separate programs that together implement a software invention described here is within the scope of the invention. In some embodiments, the software programs, when installed to operate on one or more electronic systems, define one or more specific machine implementations that execute and perform the operations of the software programs.



FIG. 8 conceptually illustrates a computer system 800 with which some embodiments of the invention are implemented. The computer system 800 can be used to implement any of the above-described hosts, controllers, gateway and edge forwarding elements. As such, it can be used to execute any of the above described processes. This computer system includes various types of non-transitory machine readable media and interfaces for various other types of machine readable media. Computer system 800 includes a bus 805, processing unit(s) 810, a system memory 825, a read-only memory 830, a permanent storage device 835, input devices 840, and output devices 845.


The bus 805 collectively represents all system, peripheral, and chipset buses that communicatively connect the numerous internal devices of the computer system 800. For instance, the bus 805 communicatively connects the processing unit(s) 810 with the read-only memory 830, the system memory 825, and the permanent storage device 835.


From these various memory units, the processing unit(s) 810 retrieve instructions to execute and data to process in order to execute the processes of the invention. The processing unit(s) may be a single processor or a multi-core processor in different embodiments. The read-only-memory (ROM) 830 stores static data and instructions that are needed by the processing unit(s) 810 and other modules of the computer system. The permanent storage device 835, on the other hand, is a read-and-write memory device. This device is a non-volatile memory unit that stores instructions and data even when the computer system 800 is off. Some embodiments of the invention use a mass-storage device (such as a magnetic or optical disk and its corresponding disk drive) as the permanent storage device 835.


Other embodiments use a removable storage device (such as a floppy disk, flash drive, etc.) as the permanent storage device. Like the permanent storage device 835, the system memory 825 is a read-and-write memory device. However, unlike storage device 835, the system memory is a volatile read-and-write memory, such as random access memory. The system memory stores some of the instructions and data that the processor needs at runtime. In some embodiments, the invention's processes are stored in the system memory 825, the permanent storage device 835, and/or the read-only memory 830. From these various memory units, the processing unit(s) 810 retrieve instructions to execute and data to process in order to execute the processes of some embodiments.


The bus 805 also connects to the input and output devices 840 and 845. The input devices enable the user to communicate information and select commands to the computer system. The input devices 840 include alphanumeric keyboards and pointing devices (also called “cursor control devices”). The output devices 845 display images generated by the computer system. The output devices include printers and display devices, such as cathode ray tubes (CRT) or liquid crystal displays (LCD). Some embodiments include devices such as touchscreens that function as both input and output devices.


Finally, as shown in FIG. 8, bus 805 also couples computer system 800 to a network 865 through a network adapter (not shown). In this manner, the computer can be a part of a network of computers (such as a local area network (“LAN”), a wide area network (“WAN”), or an Intranet), or a network of networks (such as the Internet). Any or all components of computer system 800 may be used in conjunction with the invention.


Some embodiments include electronic components, such as microprocessors, storage and memory that store computer program instructions in a machine-readable or computer-readable medium (alternatively referred to as computer-readable storage media, machine-readable media, or machine-readable storage media). Some examples of such computer-readable media include RAM, ROM, read-only compact discs (CD-ROM), recordable compact discs (CD-R), rewritable compact discs (CD-RW), read-only digital versatile discs (e.g., DVD-ROM, dual-layer DVD-ROM), a variety of recordable/rewritable DVDs (e.g., DVD-RAM, DVD-RW, DVD+RW, etc.), flash memory (e.g., SD cards, mini-SD cards, micro-SD cards, etc.), magnetic and/or solid state hard drives, read-only and recordable Blu-Ray® discs, ultra-density optical discs, any other optical or magnetic media, and floppy disks. The computer-readable media may store a computer program that is executable by at least one processing unit and includes sets of instructions for performing various operations. Examples of computer programs or computer code include machine code, such as is produced by a compiler, and files including higher-level code that are executed by a computer, an electronic component, or a microprocessor using an interpreter.


While the above discussion primarily refers to microprocessor or multi-core processors that execute software, some embodiments are performed by one or more integrated circuits, such as application specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs). In some embodiments, such integrated circuits execute instructions that are stored on the circuit itself.


As used in this specification, the terms “computer”, “server”, “processor”, and “memory” all refer to electronic or other technological devices. These terms exclude people or groups of people. For the purposes of the specification, the terms “display” or “displaying” mean displaying on an electronic device. As used in this specification, the terms “computer readable medium,” “computer readable media,” and “machine readable medium” are entirely restricted to tangible, physical objects that store information in a form that is readable by a computer. These terms exclude any wireless signals, wired download signals, and any other ephemeral or transitory signals.


While the invention has been described with reference to numerous specific details, one of ordinary skill in the art will recognize that the invention can be embodied in other specific forms without departing from the spirit of the invention. For instance, several of the above-described embodiments deploy gateways in public cloud datacenters. However, in other embodiments, the gateways are deployed in a third party's virtual private cloud datacenters (e.g., datacenters that the third party uses to deploy cloud gateways for different entities in order to deploy virtual networks for these entities). Thus, one of ordinary skill in the art would understand that the invention is not to be limited by the foregoing illustrative details, but rather is to be defined by the appended claims.

Claims
  • 1. A method of dynamically scaling a hub cluster in a software-defined wide area network (SD-WAN) based on particular traffic statistics, wherein the hub cluster is located in a datacenter of the SD-WAN and allows a plurality of branch sites of the SD-WAN to access resources of the datacenter by connecting to the hub cluster, the method comprising: at a controller of the SD-WAN: receiving, from the hub cluster, a plurality of traffic statistics that were captured at the hub cluster;analyzing the plurality of traffic statistics to identify traffic load fluctuations, said analyzing comprising identifying repetitive load fluctuations based on the received plurality of traffic statistics and based on historical traffic statistics previously received by the controller from the hub cluster, said analyzing comprising using a set of one or more machine-trained engines to identify the load fluctuations by subtracting cyclical and seasonal parameters from time-series data derived from the received plurality of traffic statistics and the historical traffic statistics;using the identified load fluctuations to predict traffic loads, and based on the predicted traffic load, adjusting a number of hubs in the hub cluster.
  • 2. The method of claim 1, wherein the plurality of traffic statistics is a first plurality of traffic statistics, the method further comprising receiving a second plurality of traffic statistics from the plurality of branch sites, wherein the second plurality of traffic statistics were captured at each branch site in the plurality of branch sites.
  • 3. The method of claim 2, wherein each branch site in the plurality of branch sites comprises an edge forwarding node, wherein the second plurality of traffic statistics captured at each branch site in the plurality of branch sites were captured by the edge forwarding nodes at the plurality of branch sites.
  • 4. The method of claim 1, wherein the set of one or more machine-trained engines comprises (i) a learning engine for identifying real-time, cyclical, and seasonal load fluctuations based on received traffic statistics and (ii) a decision-making and evaluation engine for computing overall flow performance to determine whether to adjust the number of hubs in the hub cluster.
  • 5. The method of claim 4, wherein the plurality of received traffic statistics is a plurality of real-time traffic statistics, wherein the learning engine and decision-making and evaluation engine are trained on the historical traffic statistics previously received by the controller.
  • 6. The method of claim 5, wherein the learning engine identifies the cyclical and seasonal load fluctuations by using a mathematical model to analyze the plurality of real-time traffic statistics along with the historical traffic statistics.
  • 7. The method of claim 1, wherein using the identified traffic load fluctuations to predict traffic loads further comprises (i) predicting the traffic loads over a particular time interval and (ii) determining a number of hubs needed to handle the predicted traffic load.
  • 8. The method of claim 1, wherein adjusting the number of hubs in the hub cluster further comprises one of increasing the number of hubs in the hub cluster and decreasing the number of hubs in the hub cluster.
  • 9. The method of claim 1, wherein the traffic statistics comprise traffic load statistics and flow performance statistics.
  • 10. The method of claim 9, wherein the traffic load statistics comprise at least (i) a number of branch sites, (ii) rate of flow creation per branch site, (iii) total number of flows per branch site, (iv) flows that are further categorized by application, (v) average bandwidth per flow, (vi) total number of fat flows per branch, and (vii) average bandwidth per fat flow.
  • 11. The method of claim 9, wherein the flow performance statistics comprise at least (i) packet re-transmits per flow, (ii) packet drops per flow, and (iii) mean opinion scores (MOS) of real-time flows.
  • 12. A non-transitory machine readable medium storing a program for execution by a set of processing units, the program for dynamically scaling a hub cluster in a software-defined wide area network (SD-WAN) based on particular traffic statistics, wherein the hub cluster is located in a datacenter of the SD-WAN and allows a plurality of branch sites of the SD-WAN to access resources of the datacenter by connecting to the hub cluster, the program comprising sets of instructions for: at a controller of the SD-WAN: receiving, from the hub cluster, a plurality of traffic statistics that were captured at the hub cluster;analyzing the plurality of traffic statistics to identify traffic load fluctuations, said analyzing comprising identifying repetitive load fluctuations based on the received plurality of traffic statistics and based on historical traffic statistics previously received by the controller from the hub cluster, said analyzing comprising using a set of one or more machine-trained engines to identify the load fluctuations by subtracting cyclical and seasonal parameters from time-series data derived from the received plurality of traffic statistics and the historical traffic statistics;using the identified load fluctuations to predict traffic loads, and based on the predicted traffic load, adjusting a number of hubs in the hub cluster.
  • 13. The non-transitory machine readable medium of claim 12, wherein the plurality of traffic statistics is a first plurality of traffic statistics, the program further comprising a set of instructions for receiving a second plurality of traffic statistics from the plurality of branch sites, wherein the second plurality of traffic statistics were captured at each branch site in the plurality of branch sites.
  • 14. The non-transitory machine readable medium of claim 13, wherein each branch site in the plurality of branch sites comprises an edge forwarding node, wherein the second plurality of traffic statistics captured at each branch site in the plurality of branch sites were captured by the edge forwarding nodes at the plurality of branch sites.
  • 15. The non-transitory machine readable medium of claim 12, wherein the set of one or more machine-trained engines controller comprises (i) a learning engine for identifying real-time, cyclical, and seasonal load fluctuations based on received traffic statistics and (ii) a decision-making and evaluation engine for computing overall flow performance to determine whether to adjust the number of hubs in the hub cluster.
  • 16. The non-transitory machine readable medium of claim 15, wherein the plurality of received traffic statistics is a plurality of real-time traffic statistics, wherein the learning engine and decision-making and evaluation engine are trained on the historical traffic statistics previously received by the controller.
  • 17. The non-transitory machine readable medium of claim 16, wherein the learning engine identifies the cyclical and seasonal load fluctuations by using a mathematical model to analyze the plurality of real-time traffic statistics along with the historical traffic statistics.
  • 18. The non-transitory machine readable medium of claim 12, wherein the set of instructions for using the identified traffic load fluctuations to predict traffic loads further comprises a set of instructions for (i) predicting the traffic loads over a particular time interval and (ii) determining a number of hubs needed to handle the predicted traffic load, andwherein adjusting the number of hubs in the hub cluster further comprises one of increasing the number of hubs in the hub cluster and decreasing the number of hubs in the hub cluster.
  • 19. The non-transitory machine readable medium of claim 12, wherein the traffic statistics comprise traffic load statistics and flow performance statistics.
  • 20. The non-transitory machine readable medium of claim 19, wherein the traffic load statistics comprise at least (i) a number of branch sites, (ii) rate of flow creation per branch site, (iii) total number of flows per branch site, (iv) flows that are further categorized by application, (v) average bandwidth per flow, (vi) total number of fat flows per branch, and (vii) average bandwidth per fat flow.
  • 21. The non-transitory machine readable medium of claim 19, wherein the flow performance statistics comprise at least (i) packet re-transmits per flow, (ii) packet drops per flow, and (iii) mean opinion scores (MOS) of real-time flows.
Priority Claims (1)
Number Date Country Kind
202141003823 Jan 2021 IN national
US Referenced Citations (1003)
Number Name Date Kind
5652751 Sharony Jul 1997 A
5909553 Campbell et al. Jun 1999 A
6154465 Pickett Nov 2000 A
6157648 Voit et al. Dec 2000 A
6201810 Masuda et al. Mar 2001 B1
6363378 Conklin et al. Mar 2002 B1
6445682 Weitz Sep 2002 B1
6744775 Beshai et al. Jun 2004 B1
6976087 Westfall et al. Dec 2005 B1
7003481 Banka et al. Feb 2006 B2
7280476 Anderson Oct 2007 B2
7313629 Nucci et al. Dec 2007 B1
7320017 Kurapati et al. Jan 2008 B1
7373660 Guichard et al. May 2008 B1
7581022 Griffin et al. Aug 2009 B1
7680925 Sathyanarayana et al. Mar 2010 B2
7681236 Tamura et al. Mar 2010 B2
7751409 Carolan Jul 2010 B1
7962458 Holenstein et al. Jun 2011 B2
8094575 Vadlakonda et al. Jan 2012 B1
8094659 Arad Jan 2012 B1
8111692 Ray Feb 2012 B2
8141156 Mao et al. Mar 2012 B1
8224971 Miller et al. Jul 2012 B1
8228928 Parandekar et al. Jul 2012 B2
8243589 Trost et al. Aug 2012 B1
8259566 Chen et al. Sep 2012 B2
8274891 Averi et al. Sep 2012 B2
8301749 Finklestein et al. Oct 2012 B1
8385227 Downey Feb 2013 B1
8516129 Skene Aug 2013 B1
8566452 Goodwin et al. Oct 2013 B1
8588066 Goel et al. Nov 2013 B2
8630291 Shaffer et al. Jan 2014 B2
8661295 Khanna et al. Feb 2014 B1
8724456 Hong et al. May 2014 B1
8724503 Johnsson et al. May 2014 B2
8745177 Kazerani et al. Jun 2014 B1
8797874 Yu et al. Aug 2014 B2
8799504 Capone et al. Aug 2014 B2
8804745 Sinn Aug 2014 B1
8806482 Nagargadde et al. Aug 2014 B1
8855071 Sankaran et al. Oct 2014 B1
8856339 Mestery et al. Oct 2014 B2
8964548 Keralapura et al. Feb 2015 B1
8989199 Sella et al. Mar 2015 B1
9009217 Nagargadde et al. Apr 2015 B1
9015299 Shah Apr 2015 B1
9055000 Ghosh et al. Jun 2015 B1
9060025 Xu Jun 2015 B2
9071607 Twitchell, Jr. Jun 2015 B2
9075771 Gawali et al. Jul 2015 B1
9100329 Jiang et al. Aug 2015 B1
9135037 Petrescu-Prahova et al. Sep 2015 B1
9137334 Zhou Sep 2015 B2
9154327 Marino et al. Oct 2015 B1
9203764 Shirazipour et al. Dec 2015 B2
9225591 Beheshti-Zavareh et al. Dec 2015 B2
9306949 Richard et al. Apr 2016 B1
9323561 Ayala et al. Apr 2016 B2
9336040 Dong et al. May 2016 B2
9354983 Yenamandra et al. May 2016 B1
9356943 Lopilato et al. May 2016 B1
9379981 Zhou et al. Jun 2016 B1
9413724 Xu Aug 2016 B2
9419878 Hsiao et al. Aug 2016 B2
9432245 Sorenson et al. Aug 2016 B1
9438566 Zhang et al. Sep 2016 B2
9450817 Bahadur et al. Sep 2016 B1
9450852 Chen et al. Sep 2016 B1
9462010 Stevenson Oct 2016 B1
9467478 Khan et al. Oct 2016 B1
9485163 Fries et al. Nov 2016 B1
9521067 Michael et al. Dec 2016 B2
9525564 Lee Dec 2016 B2
9542219 Bryant et al. Jan 2017 B1
9559951 Sajassi et al. Jan 2017 B1
9563423 Pittman Feb 2017 B1
9602389 Maveli et al. Mar 2017 B1
9608917 Anderson et al. Mar 2017 B1
9608962 Chang Mar 2017 B1
9614748 Battersby et al. Apr 2017 B1
9621460 Mehta et al. Apr 2017 B2
9641551 Kariyanahalli May 2017 B1
9648547 Hart et al. May 2017 B1
9665432 Kruse et al. May 2017 B2
9686127 Ramachandran et al. Jun 2017 B2
9692714 Nair et al. Jun 2017 B1
9715401 Devine et al. Jul 2017 B2
9717021 Hughes et al. Jul 2017 B2
9722815 Mukundan et al. Aug 2017 B2
9747249 Cherian et al. Aug 2017 B2
9755965 Yadav et al. Sep 2017 B1
9787559 Schroeder Oct 2017 B1
9807004 Koley et al. Oct 2017 B2
9819540 Bahadur et al. Nov 2017 B1
9819565 Djukic et al. Nov 2017 B2
9825822 Holland Nov 2017 B1
9825911 Brandwine Nov 2017 B1
9825992 Xu Nov 2017 B2
9832128 Ashner et al. Nov 2017 B1
9832205 Santhi et al. Nov 2017 B2
9875355 Williams Jan 2018 B1
9906401 Rao Feb 2018 B1
9923826 Murgia Mar 2018 B2
9930011 Clemons, Jr. et al. Mar 2018 B1
9935829 Miller et al. Apr 2018 B1
9942787 Tillotson Apr 2018 B1
9996370 Khafizov et al. Jun 2018 B1
10038601 Becker et al. Jul 2018 B1
10057183 Salle et al. Aug 2018 B2
10057294 Xu Aug 2018 B2
10116593 Sinn et al. Oct 2018 B1
10135789 Mayya et al. Nov 2018 B2
10142226 Wu et al. Nov 2018 B1
10178032 Freitas Jan 2019 B1
10178037 Appleby et al. Jan 2019 B2
10187289 Chen et al. Jan 2019 B1
10200264 Menon et al. Feb 2019 B2
10229017 Zou et al. Mar 2019 B1
10237123 Dubey et al. Mar 2019 B2
10250498 Bales et al. Apr 2019 B1
10263832 Ghosh Apr 2019 B1
10320664 Nainar et al. Jun 2019 B2
10320691 Matthews et al. Jun 2019 B1
10326830 Singh Jun 2019 B1
10348767 Lee et al. Jul 2019 B1
10355989 Panchal et al. Jul 2019 B1
10425382 Mayya et al. Sep 2019 B2
10454708 Mibu Oct 2019 B2
10454714 Mayya et al. Oct 2019 B2
10461993 Turabi et al. Oct 2019 B2
10498652 Mayya et al. Dec 2019 B2
10511546 Singarayan et al. Dec 2019 B2
10523539 Mayya et al. Dec 2019 B2
10550093 Ojima et al. Feb 2020 B2
10554538 Spohn et al. Feb 2020 B2
10560431 Chen et al. Feb 2020 B1
10565464 Han et al. Feb 2020 B2
10567519 Mukhopadhyaya et al. Feb 2020 B1
10574482 Oréet al. Feb 2020 B2
10574528 Mayya et al. Feb 2020 B2
10594516 Cidon et al. Mar 2020 B2
10594591 Houjyo et al. Mar 2020 B2
10594659 El-Moussa et al. Mar 2020 B2
10608844 Cidon et al. Mar 2020 B2
10630505 Rubenstein et al. Apr 2020 B2
10637889 Ermagan et al. Apr 2020 B2
10666460 Cidon et al. May 2020 B2
10666497 Tahhan et al. May 2020 B2
10686625 Cidon et al. Jun 2020 B2
10693739 Naseri et al. Jun 2020 B1
10708144 Mohan et al. Jul 2020 B2
10715427 Raj et al. Jul 2020 B2
10749711 Mukundan et al. Aug 2020 B2
10778466 Cidon et al. Sep 2020 B2
10778528 Mayya et al. Sep 2020 B2
10778557 Ganichev et al. Sep 2020 B2
10805114 Cidon et al. Oct 2020 B2
10805272 Mayya et al. Oct 2020 B2
10819564 Turabi et al. Oct 2020 B2
10826775 Moreno et al. Nov 2020 B1
10841131 Cidon et al. Nov 2020 B2
10911374 Kumar et al. Feb 2021 B1
10938693 Mayya et al. Mar 2021 B2
10951529 Duan et al. Mar 2021 B2
10958479 Cidon et al. Mar 2021 B2
10959098 Cidon et al. Mar 2021 B2
10992558 Silva et al. Apr 2021 B1
10992568 Michael et al. Apr 2021 B2
10999100 Cidon et al. May 2021 B2
10999137 Cidon et al. May 2021 B2
10999165 Cidon et al. May 2021 B2
10999197 Hooda et al. May 2021 B2
11005684 Cidon May 2021 B2
11018995 Cidon et al. May 2021 B2
11044190 Ramaswamy et al. Jun 2021 B2
11050588 Mayya et al. Jun 2021 B2
11050644 Hegde et al. Jun 2021 B2
11071005 Shen et al. Jul 2021 B2
11089111 Markuze et al. Aug 2021 B2
11095612 Oswal et al. Aug 2021 B1
11102032 Cidon et al. Aug 2021 B2
11108595 Knutsen et al. Aug 2021 B2
11108851 Kurmala et al. Aug 2021 B1
11115347 Gupta et al. Sep 2021 B2
11115426 Pazhyannur et al. Sep 2021 B1
11115480 Markuze et al. Sep 2021 B2
11121962 Michael et al. Sep 2021 B2
11121985 Cidon et al. Sep 2021 B2
11128492 Sethi et al. Sep 2021 B2
11146632 Rubenstein Oct 2021 B2
11153230 Cidon et al. Oct 2021 B2
11171885 Cidon et al. Nov 2021 B2
11212140 Mukundan et al. Dec 2021 B2
11212238 Cidon et al. Dec 2021 B2
11223514 Mayya et al. Jan 2022 B2
11245641 Ramaswamy et al. Feb 2022 B2
11252079 Michael et al. Feb 2022 B2
11252105 Cidon et al. Feb 2022 B2
11252106 Cidon et al. Feb 2022 B2
11258728 Cidon et al. Feb 2022 B2
11310170 Cidon et al. Apr 2022 B2
11316743 Singh Bawa Apr 2022 B2
11323307 Mayya et al. May 2022 B2
11349722 Mayya et al. May 2022 B2
11363124 Markuze et al. Jun 2022 B2
11374904 Mayya et al. Jun 2022 B2
11375005 Rolando et al. Jun 2022 B1
11381474 Kumar et al. Jul 2022 B1
11381499 Ramaswamy et al. Jul 2022 B1
11388086 Ramaswamy et al. Jul 2022 B1
11394640 Ramaswamy et al. Jul 2022 B2
11418997 Devadoss et al. Aug 2022 B2
11438789 Devadoss et al. Sep 2022 B2
11444865 Ramaswamy et al. Sep 2022 B2
11444872 Mayya et al. Sep 2022 B2
11477127 Ramaswamy et al. Oct 2022 B2
11489720 Kempanna et al. Nov 2022 B1
11489783 Ramaswamy et al. Nov 2022 B2
11509571 Ramaswamy et al. Nov 2022 B1
11516049 Cidon et al. Nov 2022 B2
11522780 Wallace et al. Dec 2022 B1
11526434 Brooker et al. Dec 2022 B1
11533248 Mayya et al. Dec 2022 B2
11552874 Pragada et al. Jan 2023 B1
11575591 Ramaswamy et al. Feb 2023 B2
11575600 Markuze et al. Feb 2023 B2
11582144 Ramaswamy et al. Feb 2023 B2
11582298 Hood et al. Feb 2023 B2
11601356 Gandhi et al. Mar 2023 B2
11606225 Cidon et al. Mar 2023 B2
11606286 Michael et al. Mar 2023 B2
11606314 Cidon et al. Mar 2023 B2
11606712 Devadoss et al. Mar 2023 B2
11611507 Ramaswamy et al. Mar 2023 B2
11637768 Ramaswamy et al. Apr 2023 B2
11677720 Mayya et al. Jun 2023 B2
20020049687 Helsper et al. Apr 2002 A1
20020075542 Kumar et al. Jun 2002 A1
20020085488 Kobayashi Jul 2002 A1
20020087716 Mustafa Jul 2002 A1
20020152306 Tuck Oct 2002 A1
20020186682 Kawano et al. Dec 2002 A1
20020198840 Banka et al. Dec 2002 A1
20030050061 Wu et al. Mar 2003 A1
20030061269 Hathaway et al. Mar 2003 A1
20030088697 Matsuhira May 2003 A1
20030112766 Riedel et al. Jun 2003 A1
20030112808 Solomon Jun 2003 A1
20030126468 Markham Jul 2003 A1
20030161313 Jinmei et al. Aug 2003 A1
20030189919 Gupta et al. Oct 2003 A1
20030202506 Perkins et al. Oct 2003 A1
20030219030 Gubbi Nov 2003 A1
20040059831 Chu et al. Mar 2004 A1
20040068668 Lor et al. Apr 2004 A1
20040165601 Liu et al. Aug 2004 A1
20040224771 Chen et al. Nov 2004 A1
20050078690 DeLangis Apr 2005 A1
20050149604 Navada Jul 2005 A1
20050154790 Nagata et al. Jul 2005 A1
20050172161 Cruz et al. Aug 2005 A1
20050195754 Nosella Sep 2005 A1
20050210479 Andjelic Sep 2005 A1
20050265255 Kodialam et al. Dec 2005 A1
20060002291 Alicherry et al. Jan 2006 A1
20060034335 Karaoguz et al. Feb 2006 A1
20060114838 Mandavilli et al. Jun 2006 A1
20060171365 Borella Aug 2006 A1
20060182034 Klinker et al. Aug 2006 A1
20060182035 Vasseur Aug 2006 A1
20060193247 Naseh et al. Aug 2006 A1
20060193252 Naseh et al. Aug 2006 A1
20060195605 Sundarrajan et al. Aug 2006 A1
20060245414 Susai et al. Nov 2006 A1
20070050594 Augsburg et al. Mar 2007 A1
20070064604 Chen et al. Mar 2007 A1
20070064702 Bates et al. Mar 2007 A1
20070083727 Johnston et al. Apr 2007 A1
20070091794 Filsfils et al. Apr 2007 A1
20070103548 Carter May 2007 A1
20070115812 Hughes May 2007 A1
20070121486 Guichard et al. May 2007 A1
20070130325 Lesser Jun 2007 A1
20070162619 Aloni et al. Jul 2007 A1
20070162639 Chu et al. Jul 2007 A1
20070177511 Das et al. Aug 2007 A1
20070195797 Patel et al. Aug 2007 A1
20070237081 Kodialam et al. Oct 2007 A1
20070260746 Mirtorabi et al. Nov 2007 A1
20070268882 Breslau et al. Nov 2007 A1
20080002670 Bugenhagen et al. Jan 2008 A1
20080049621 McGuire et al. Feb 2008 A1
20080055241 Goldenberg et al. Mar 2008 A1
20080080509 Khanna et al. Apr 2008 A1
20080095187 Jung et al. Apr 2008 A1
20080117930 Chakareski et al. May 2008 A1
20080144532 Chamarajanagar et al. Jun 2008 A1
20080168086 Miller et al. Jul 2008 A1
20080175150 Bolt et al. Jul 2008 A1
20080181116 Kavanaugh et al. Jul 2008 A1
20080219276 Shah Sep 2008 A1
20080240121 Xiong et al. Oct 2008 A1
20080263218 Beerends et al. Oct 2008 A1
20090013210 McIntosh et al. Jan 2009 A1
20090028092 Rothschild Jan 2009 A1
20090125617 Klessig et al. May 2009 A1
20090141642 Sun Jun 2009 A1
20090154463 Hines et al. Jun 2009 A1
20090182874 Morford et al. Jul 2009 A1
20090247204 Sennett et al. Oct 2009 A1
20090268605 Campbell et al. Oct 2009 A1
20090274045 Meier et al. Nov 2009 A1
20090276657 Wetmore et al. Nov 2009 A1
20090303880 Maltz et al. Dec 2009 A1
20100008361 Guichard et al. Jan 2010 A1
20100017802 Lojewski Jan 2010 A1
20100046532 Okita Feb 2010 A1
20100061379 Parandekar et al. Mar 2010 A1
20100080129 Strahan et al. Apr 2010 A1
20100088440 Banks et al. Apr 2010 A1
20100091782 Hiscock Apr 2010 A1
20100091823 Retana et al. Apr 2010 A1
20100107162 Edwards et al. Apr 2010 A1
20100118727 Draves et al. May 2010 A1
20100118886 Saavedra May 2010 A1
20100128600 Srinivasmurthy et al. May 2010 A1
20100165985 Sharma et al. Jul 2010 A1
20100191884 Holenstein et al. Jul 2010 A1
20100223621 Joshi et al. Sep 2010 A1
20100226246 Proulx Sep 2010 A1
20100290422 Haigh et al. Nov 2010 A1
20100309841 Conte Dec 2010 A1
20100309912 Mehta et al. Dec 2010 A1
20100322255 Hao et al. Dec 2010 A1
20100332657 Elyashev et al. Dec 2010 A1
20110001604 Ludlow et al. Jan 2011 A1
20110007752 Silva et al. Jan 2011 A1
20110032939 Nozaki et al. Feb 2011 A1
20110035187 DeJori et al. Feb 2011 A1
20110040814 Higgins Feb 2011 A1
20110075674 Li et al. Mar 2011 A1
20110078783 Duan et al. Mar 2011 A1
20110107139 Middlecamp et al. May 2011 A1
20110110370 Moreno et al. May 2011 A1
20110141877 Xu et al. Jun 2011 A1
20110142041 Imai Jun 2011 A1
20110153909 Dong Jun 2011 A1
20110235509 Szymanski Sep 2011 A1
20110255397 Kadakia et al. Oct 2011 A1
20110302663 Prodan et al. Dec 2011 A1
20120008630 Ould-Brahim Jan 2012 A1
20120027013 Napierala Feb 2012 A1
20120039309 Evans et al. Feb 2012 A1
20120099601 Haddad et al. Apr 2012 A1
20120136697 Peles et al. May 2012 A1
20120140935 Kruglick Jun 2012 A1
20120157068 Eichen et al. Jun 2012 A1
20120173694 Yan et al. Jul 2012 A1
20120173919 Patel et al. Jul 2012 A1
20120182940 Taleb et al. Jul 2012 A1
20120221955 Raleigh et al. Aug 2012 A1
20120227093 Shatzkamer et al. Sep 2012 A1
20120240185 Kapoor et al. Sep 2012 A1
20120250682 Vincent et al. Oct 2012 A1
20120250686 Vincent et al. Oct 2012 A1
20120266026 Chikkalingaiah Oct 2012 A1
20120281706 Agarwal et al. Nov 2012 A1
20120287818 Corti et al. Nov 2012 A1
20120300615 Kempf et al. Nov 2012 A1
20120307659 Yamada Dec 2012 A1
20120317270 Vrbaski et al. Dec 2012 A1
20120317291 Wolfe Dec 2012 A1
20130007505 Spear Jan 2013 A1
20130019005 Hui et al. Jan 2013 A1
20130021968 Reznik et al. Jan 2013 A1
20130044764 Casado et al. Feb 2013 A1
20130051237 Ong Feb 2013 A1
20130051399 Zhang et al. Feb 2013 A1
20130054763 Merwe et al. Feb 2013 A1
20130086267 Gelenbe et al. Apr 2013 A1
20130097304 Asthana et al. Apr 2013 A1
20130103729 Cooney et al. Apr 2013 A1
20130103834 Dzerve et al. Apr 2013 A1
20130117530 Kim et al. May 2013 A1
20130124718 Griffith et al. May 2013 A1
20130124911 Griffith et al. May 2013 A1
20130124912 Griffith et al. May 2013 A1
20130128889 Mathur et al. May 2013 A1
20130142201 Kim et al. Jun 2013 A1
20130170354 Takashima et al. Jul 2013 A1
20130173768 Kundu et al. Jul 2013 A1
20130173788 Song Jul 2013 A1
20130182712 Aguayo et al. Jul 2013 A1
20130185446 Zeng et al. Jul 2013 A1
20130185729 Vasic et al. Jul 2013 A1
20130191688 Agarwal et al. Jul 2013 A1
20130223226 Narayanan et al. Aug 2013 A1
20130223454 Dunbar et al. Aug 2013 A1
20130235870 Tripathi et al. Sep 2013 A1
20130238782 Zhao et al. Sep 2013 A1
20130242718 Zhang Sep 2013 A1
20130254599 Katkar et al. Sep 2013 A1
20130258839 Wang et al. Oct 2013 A1
20130258847 Zhang et al. Oct 2013 A1
20130266015 Qu et al. Oct 2013 A1
20130266019 Qu et al. Oct 2013 A1
20130283364 Chang et al. Oct 2013 A1
20130286846 Atlas et al. Oct 2013 A1
20130297611 Moritz et al. Nov 2013 A1
20130297770 Zhang Nov 2013 A1
20130301469 Suga Nov 2013 A1
20130301642 Radhakrishnan et al. Nov 2013 A1
20130308444 Sem-Jacobsen et al. Nov 2013 A1
20130315242 Wang et al. Nov 2013 A1
20130315243 Huang et al. Nov 2013 A1
20130329548 Nakil et al. Dec 2013 A1
20130329601 Yin et al. Dec 2013 A1
20130329734 Chesla et al. Dec 2013 A1
20130346470 Obstfeld et al. Dec 2013 A1
20140016464 Shirazipour et al. Jan 2014 A1
20140019604 Twitchell, Jr. Jan 2014 A1
20140019750 Dodgson et al. Jan 2014 A1
20140040975 Raleigh et al. Feb 2014 A1
20140064283 Balus et al. Mar 2014 A1
20140071832 Johnsson et al. Mar 2014 A1
20140092907 Sridhar et al. Apr 2014 A1
20140108665 Arora et al. Apr 2014 A1
20140112171 Pasdar Apr 2014 A1
20140115584 Mudigonda et al. Apr 2014 A1
20140122559 Branson et al. May 2014 A1
20140123135 Huang et al. May 2014 A1
20140126418 Brendel et al. May 2014 A1
20140156818 Hunt Jun 2014 A1
20140156823 Liu et al. Jun 2014 A1
20140157363 Banerjee Jun 2014 A1
20140160935 Zecharia et al. Jun 2014 A1
20140164560 Ko et al. Jun 2014 A1
20140164617 Jalan et al. Jun 2014 A1
20140164718 Schaik et al. Jun 2014 A1
20140173113 Vemuri et al. Jun 2014 A1
20140173331 Martin et al. Jun 2014 A1
20140181824 Saund et al. Jun 2014 A1
20140189074 Parker Jul 2014 A1
20140208317 Nakagawa Jul 2014 A1
20140219135 Li et al. Aug 2014 A1
20140223507 Xu Aug 2014 A1
20140229210 Sharifian et al. Aug 2014 A1
20140244851 Lee Aug 2014 A1
20140258535 Zhang Sep 2014 A1
20140269690 Tu Sep 2014 A1
20140279862 Dietz et al. Sep 2014 A1
20140280499 Basavaiah et al. Sep 2014 A1
20140310282 Sprague et al. Oct 2014 A1
20140317440 Biermayr et al. Oct 2014 A1
20140321277 Lynn, Jr. et al. Oct 2014 A1
20140337500 Lee Nov 2014 A1
20140337674 Ivancic et al. Nov 2014 A1
20140341109 Cartmell et al. Nov 2014 A1
20140355441 Jain Dec 2014 A1
20140365834 Stone et al. Dec 2014 A1
20140372582 Ghanwani et al. Dec 2014 A1
20150003240 Drwiega et al. Jan 2015 A1
20150016249 Mukundan et al. Jan 2015 A1
20150029864 Raileanu et al. Jan 2015 A1
20150039744 Niazi et al. Feb 2015 A1
20150046572 Cheng et al. Feb 2015 A1
20150052247 Threefoot et al. Feb 2015 A1
20150052517 Raghu et al. Feb 2015 A1
20150056960 Egner et al. Feb 2015 A1
20150058917 Xu Feb 2015 A1
20150088942 Shah Mar 2015 A1
20150089628 Lang Mar 2015 A1
20150092603 Aguayo et al. Apr 2015 A1
20150096011 Watt Apr 2015 A1
20150100958 Banavalikar et al. Apr 2015 A1
20150106809 Reddy et al. Apr 2015 A1
20150124603 Ketheesan et al. May 2015 A1
20150134777 Onoue May 2015 A1
20150139238 Pourzandi et al. May 2015 A1
20150146539 Mehta et al. May 2015 A1
20150163152 Li Jun 2015 A1
20150169340 Haddad et al. Jun 2015 A1
20150172121 Farkas et al. Jun 2015 A1
20150172169 DeCusatis et al. Jun 2015 A1
20150188823 Williams et al. Jul 2015 A1
20150189009 Bemmel Jul 2015 A1
20150195178 Bhattacharya et al. Jul 2015 A1
20150201036 Nishiki et al. Jul 2015 A1
20150222543 Song Aug 2015 A1
20150222638 Morley Aug 2015 A1
20150236945 Michael et al. Aug 2015 A1
20150236962 Veres et al. Aug 2015 A1
20150244617 Nakil et al. Aug 2015 A1
20150249644 Xu Sep 2015 A1
20150257081 Ramanujan et al. Sep 2015 A1
20150264055 Budhani et al. Sep 2015 A1
20150271056 Chunduri et al. Sep 2015 A1
20150271104 Chikkamath et al. Sep 2015 A1
20150271303 Neginhal et al. Sep 2015 A1
20150281004 Kakadia et al. Oct 2015 A1
20150312142 Barabash et al. Oct 2015 A1
20150312760 O'Toole Oct 2015 A1
20150317169 Sinha et al. Nov 2015 A1
20150326426 Luo et al. Nov 2015 A1
20150334025 Rader Nov 2015 A1
20150334696 Gu et al. Nov 2015 A1
20150341271 Gomez Nov 2015 A1
20150349978 Wu et al. Dec 2015 A1
20150350907 Timariu et al. Dec 2015 A1
20150358232 Chen et al. Dec 2015 A1
20150358236 Roach et al. Dec 2015 A1
20150363221 Terayama et al. Dec 2015 A1
20150363733 Brown Dec 2015 A1
20150365323 Duminuco et al. Dec 2015 A1
20150372943 Hasan et al. Dec 2015 A1
20150372982 Herle et al. Dec 2015 A1
20150381407 Wang et al. Dec 2015 A1
20150381462 Choi et al. Dec 2015 A1
20150381493 Bansal et al. Dec 2015 A1
20160019317 Pawar et al. Jan 2016 A1
20160020844 Hart et al. Jan 2016 A1
20160021597 Hart et al. Jan 2016 A1
20160035183 Buchholz et al. Feb 2016 A1
20160036924 Koppolu et al. Feb 2016 A1
20160036938 Aviles et al. Feb 2016 A1
20160037434 Gopal et al. Feb 2016 A1
20160072669 Saavedra Mar 2016 A1
20160072684 Manuguri et al. Mar 2016 A1
20160080268 Anand Mar 2016 A1
20160080502 Yadav et al. Mar 2016 A1
20160105353 Cociglio Apr 2016 A1
20160105392 Thakkar et al. Apr 2016 A1
20160105471 Nunes et al. Apr 2016 A1
20160105488 Thakkar et al. Apr 2016 A1
20160117185 Fang et al. Apr 2016 A1
20160134461 Sampath et al. May 2016 A1
20160134527 Kwak et al. May 2016 A1
20160134528 Lin et al. May 2016 A1
20160134591 Liao et al. May 2016 A1
20160142373 Ossipov May 2016 A1
20160147607 Dornemann et al. May 2016 A1
20160150055 Choi May 2016 A1
20160164832 Bellagamba et al. Jun 2016 A1
20160164914 Madhav et al. Jun 2016 A1
20160173338 Wolting Jun 2016 A1
20160191363 Haraszti et al. Jun 2016 A1
20160191374 Singh et al. Jun 2016 A1
20160192403 Gupta et al. Jun 2016 A1
20160197834 Luft Jul 2016 A1
20160197835 Luft Jul 2016 A1
20160198003 Luft Jul 2016 A1
20160205071 Cooper et al. Jul 2016 A1
20160210209 Verkaik et al. Jul 2016 A1
20160212773 Kanderholm et al. Jul 2016 A1
20160218947 Hughes et al. Jul 2016 A1
20160218948 Djukic Jul 2016 A1
20160218951 Vasseur et al. Jul 2016 A1
20160234099 Jiao Aug 2016 A1
20160234161 Banerjee et al. Aug 2016 A1
20160255169 Kovvuri et al. Sep 2016 A1
20160255542 Hughes et al. Sep 2016 A1
20160261493 Li Sep 2016 A1
20160261495 Xia et al. Sep 2016 A1
20160261506 Hegde et al. Sep 2016 A1
20160261639 Xu Sep 2016 A1
20160269298 Li et al. Sep 2016 A1
20160269926 Sundaram Sep 2016 A1
20160285736 Gu Sep 2016 A1
20160299775 Madapurath et al. Oct 2016 A1
20160301471 Kunz et al. Oct 2016 A1
20160308762 Teng et al. Oct 2016 A1
20160315912 Mayya et al. Oct 2016 A1
20160323377 Einkauf et al. Nov 2016 A1
20160328159 Coddington et al. Nov 2016 A1
20160330111 Manghirmalani et al. Nov 2016 A1
20160337202 Ben-Itzhak et al. Nov 2016 A1
20160352588 Subbarayan et al. Dec 2016 A1
20160353268 Senarath et al. Dec 2016 A1
20160359738 Sullenberger et al. Dec 2016 A1
20160366187 Kamble Dec 2016 A1
20160371153 Dornemann Dec 2016 A1
20160378527 Zamir Dec 2016 A1
20160380886 Blair et al. Dec 2016 A1
20160380906 Hodique et al. Dec 2016 A1
20170005986 Bansal et al. Jan 2017 A1
20170006499 Hampel et al. Jan 2017 A1
20170012870 Blair et al. Jan 2017 A1
20170019428 Cohn Jan 2017 A1
20170024260 Chandrasekaran et al. Jan 2017 A1
20170026273 Yao et al. Jan 2017 A1
20170026283 Williams et al. Jan 2017 A1
20170026355 Mathaiyan et al. Jan 2017 A1
20170034046 Cai et al. Feb 2017 A1
20170034052 Chanda et al. Feb 2017 A1
20170034129 Sawant et al. Feb 2017 A1
20170048296 Ramalho et al. Feb 2017 A1
20170053258 Carney et al. Feb 2017 A1
20170055131 Kong et al. Feb 2017 A1
20170063674 Maskalik et al. Mar 2017 A1
20170063782 Jain et al. Mar 2017 A1
20170063783 Yong et al. Mar 2017 A1
20170063794 Jain et al. Mar 2017 A1
20170064005 Lee Mar 2017 A1
20170075710 Prasad et al. Mar 2017 A1
20170093625 Pera et al. Mar 2017 A1
20170097841 Chang et al. Apr 2017 A1
20170104653 Badea et al. Apr 2017 A1
20170104755 Arregoces et al. Apr 2017 A1
20170109212 Gaurav et al. Apr 2017 A1
20170118067 Vedula Apr 2017 A1
20170118173 Arramreddy et al. Apr 2017 A1
20170123939 Maheshwari et al. May 2017 A1
20170126475 Mahkonen May 2017 A1
20170126516 Tiagi et al. May 2017 A1
20170126564 Mayya et al. May 2017 A1
20170134186 Mukundan et al. May 2017 A1
20170134520 Abbasi et al. May 2017 A1
20170139789 Fries et al. May 2017 A1
20170142000 Cai et al. May 2017 A1
20170149637 Banikazemi et al. May 2017 A1
20170155557 Desai et al. Jun 2017 A1
20170155566 Martinsen et al. Jun 2017 A1
20170155590 Dillon et al. Jun 2017 A1
20170163473 Sadana et al. Jun 2017 A1
20170171024 Anerousis et al. Jun 2017 A1
20170171310 Gardner Jun 2017 A1
20170180220 Leckey et al. Jun 2017 A1
20170181210 Nadella et al. Jun 2017 A1
20170195161 Ruel et al. Jul 2017 A1
20170195169 Mills et al. Jul 2017 A1
20170201568 Hussam et al. Jul 2017 A1
20170201585 Doraiswamy et al. Jul 2017 A1
20170207976 Rovner et al. Jul 2017 A1
20170214545 Cheng et al. Jul 2017 A1
20170214701 Hasan Jul 2017 A1
20170223117 Messerli et al. Aug 2017 A1
20170236060 Ignatyev Aug 2017 A1
20170237710 Mayya et al. Aug 2017 A1
20170242784 Heorhiadi et al. Aug 2017 A1
20170257260 Govindan et al. Sep 2017 A1
20170257309 Appanna Sep 2017 A1
20170264496 Ao et al. Sep 2017 A1
20170279717 Bethers et al. Sep 2017 A1
20170279741 Elias et al. Sep 2017 A1
20170279803 Desai et al. Sep 2017 A1
20170280474 Vesterinen et al. Sep 2017 A1
20170288987 Pasupathy et al. Oct 2017 A1
20170289002 Ganguli et al. Oct 2017 A1
20170289027 Ratnasingham Oct 2017 A1
20170295264 Touitou et al. Oct 2017 A1
20170302501 Shi et al. Oct 2017 A1
20170302565 Ghobadi et al. Oct 2017 A1
20170310641 Jiang et al. Oct 2017 A1
20170310691 Vasseur et al. Oct 2017 A1
20170317954 Masurekar et al. Nov 2017 A1
20170317969 Masurekar et al. Nov 2017 A1
20170317974 Masurekar et al. Nov 2017 A1
20170324628 Dhanabalan Nov 2017 A1
20170337086 Zhu et al. Nov 2017 A1
20170339022 Hegde Nov 2017 A1
20170339054 Yadav et al. Nov 2017 A1
20170339070 Chang et al. Nov 2017 A1
20170346722 Smith et al. Nov 2017 A1
20170364419 Lo Dec 2017 A1
20170366445 Nemirovsky et al. Dec 2017 A1
20170366467 Martin et al. Dec 2017 A1
20170373950 Szilagyi et al. Dec 2017 A1
20170374174 Evens et al. Dec 2017 A1
20180006995 Bickhart et al. Jan 2018 A1
20180007005 Chanda et al. Jan 2018 A1
20180007123 Cheng et al. Jan 2018 A1
20180013636 Seetharamaiah et al. Jan 2018 A1
20180014051 Phillips et al. Jan 2018 A1
20180020035 Boggia et al. Jan 2018 A1
20180034668 Mayya et al. Feb 2018 A1
20180041425 Zhang Feb 2018 A1
20180062875 Tumuluru Mar 2018 A1
20180062914 Boutros et al. Mar 2018 A1
20180062917 Chandrashekhar et al. Mar 2018 A1
20180063036 Chandrashekhar et al. Mar 2018 A1
20180063193 Chandrashekhar et al. Mar 2018 A1
20180063233 Park Mar 2018 A1
20180063743 Tumuluru et al. Mar 2018 A1
20180069924 Tumuluru et al. Mar 2018 A1
20180074909 Bishop et al. Mar 2018 A1
20180077081 Lauer et al. Mar 2018 A1
20180077202 Xu Mar 2018 A1
20180084081 Kuchibhotla et al. Mar 2018 A1
20180091370 Arai Mar 2018 A1
20180097725 Wood et al. Apr 2018 A1
20180114569 Strachan et al. Apr 2018 A1
20180123910 Fitzgibbon May 2018 A1
20180123946 Ramachandran et al. May 2018 A1
20180131608 Jiang et al. May 2018 A1
20180131615 Zhang May 2018 A1
20180131720 Hobson et al. May 2018 A1
20180145899 Rao May 2018 A1
20180159796 Wang et al. Jun 2018 A1
20180159856 Gujarathi Jun 2018 A1
20180167378 Kostyukov et al. Jun 2018 A1
20180176073 Dubey et al. Jun 2018 A1
20180176082 Katz et al. Jun 2018 A1
20180176130 Banerjee et al. Jun 2018 A1
20180176252 Nimmagadda et al. Jun 2018 A1
20180181423 Gunda et al. Jun 2018 A1
20180205746 Boutnaru et al. Jul 2018 A1
20180213472 Ishii et al. Jul 2018 A1
20180219765 Michael et al. Aug 2018 A1
20180219766 Michael et al. Aug 2018 A1
20180234300 Mayya et al. Aug 2018 A1
20180248790 Tan et al. Aug 2018 A1
20180260125 Botes et al. Sep 2018 A1
20180261085 Liu Sep 2018 A1
20180262468 Kumar et al. Sep 2018 A1
20180270104 Zheng et al. Sep 2018 A1
20180278541 Wu et al. Sep 2018 A1
20180287907 Kulshreshtha et al. Oct 2018 A1
20180295101 Gehrmann Oct 2018 A1
20180295529 Jen et al. Oct 2018 A1
20180302286 Mayya et al. Oct 2018 A1
20180302321 Manthiramoorthy et al. Oct 2018 A1
20180307851 Lewis Oct 2018 A1
20180316606 Sung et al. Nov 2018 A1
20180351855 Sood et al. Dec 2018 A1
20180351862 Jeganathan et al. Dec 2018 A1
20180351863 Vairavakkalai et al. Dec 2018 A1
20180351882 Jeganathan et al. Dec 2018 A1
20180359323 Madden Dec 2018 A1
20180367445 Bajaj Dec 2018 A1
20180373558 Chang et al. Dec 2018 A1
20180375744 Mayya et al. Dec 2018 A1
20180375824 Mayya et al. Dec 2018 A1
20180375967 Pithawala et al. Dec 2018 A1
20190013883 Vargas et al. Jan 2019 A1
20190014038 Ritchie Jan 2019 A1
20190020588 Twitchell, Jr. Jan 2019 A1
20190020627 Yuan Jan 2019 A1
20190021085 Mochizuki et al. Jan 2019 A1
20190028378 Houjyo et al. Jan 2019 A1
20190028552 Johnson et al. Jan 2019 A1
20190036808 Shenoy et al. Jan 2019 A1
20190036810 Michael et al. Jan 2019 A1
20190036813 Shenoy et al. Jan 2019 A1
20190046056 Khachaturian et al. Feb 2019 A1
20190058657 Chunduri et al. Feb 2019 A1
20190058709 Kempf et al. Feb 2019 A1
20190068470 Mirsky Feb 2019 A1
20190068493 Ram et al. Feb 2019 A1
20190068500 Hira Feb 2019 A1
20190075083 Mayya et al. Mar 2019 A1
20190081894 Yousaf et al. Mar 2019 A1
20190103990 Cidon et al. Apr 2019 A1
20190103991 Cidon et al. Apr 2019 A1
20190103992 Cidon et al. Apr 2019 A1
20190103993 Cidon et al. Apr 2019 A1
20190104035 Cidon et al. Apr 2019 A1
20190104049 Cidon et al. Apr 2019 A1
20190104050 Cidon et al. Apr 2019 A1
20190104051 Cidon et al. Apr 2019 A1
20190104052 Cidon et al. Apr 2019 A1
20190104053 Cidon et al. Apr 2019 A1
20190104063 Cidon et al. Apr 2019 A1
20190104064 Cidon et al. Apr 2019 A1
20190104109 Cidon et al. Apr 2019 A1
20190104111 Cidon et al. Apr 2019 A1
20190104413 Cidon et al. Apr 2019 A1
20190109769 Jain et al. Apr 2019 A1
20190132221 Boutros et al. May 2019 A1
20190132234 Dong et al. May 2019 A1
20190132322 Song et al. May 2019 A1
20190140889 Mayya et al. May 2019 A1
20190140890 Mayya et al. May 2019 A1
20190149525 Gunda et al. May 2019 A1
20190158371 Dillon et al. May 2019 A1
20190158605 Markuze et al. May 2019 A1
20190199539 Deng et al. Jun 2019 A1
20190220703 Prakash et al. Jul 2019 A1
20190222499 Chen et al. Jul 2019 A1
20190238364 Boutros et al. Aug 2019 A1
20190238446 Barzik et al. Aug 2019 A1
20190238449 Michael et al. Aug 2019 A1
20190238450 Michael et al. Aug 2019 A1
20190238483 Marichetty et al. Aug 2019 A1
20190238497 Tourrilhes et al. Aug 2019 A1
20190268421 Markuze et al. Aug 2019 A1
20190268973 Bull et al. Aug 2019 A1
20190278631 Bernat et al. Sep 2019 A1
20190280962 Michael et al. Sep 2019 A1
20190280963 Michael et al. Sep 2019 A1
20190280964 Michael et al. Sep 2019 A1
20190288875 Shen et al. Sep 2019 A1
20190306197 Degioanni Oct 2019 A1
20190306282 Masputra et al. Oct 2019 A1
20190313278 Liu Oct 2019 A1
20190313907 Khachaturian et al. Oct 2019 A1
20190319847 Nahar et al. Oct 2019 A1
20190319881 Maskara Oct 2019 A1
20190327109 Guichard et al. Oct 2019 A1
20190334786 Dutta et al. Oct 2019 A1
20190334813 Raj et al. Oct 2019 A1
20190334820 Zhao Oct 2019 A1
20190342201 Singh Nov 2019 A1
20190342219 Liu et al. Nov 2019 A1
20190356736 Narayanaswamy et al. Nov 2019 A1
20190364099 Thakkar et al. Nov 2019 A1
20190364456 Yu Nov 2019 A1
20190372888 Michael et al. Dec 2019 A1
20190372889 Michael et al. Dec 2019 A1
20190372890 Michael et al. Dec 2019 A1
20190394081 Tahhan et al. Dec 2019 A1
20200014609 Hockett et al. Jan 2020 A1
20200014615 Michael et al. Jan 2020 A1
20200014616 Michael et al. Jan 2020 A1
20200014661 Mayya et al. Jan 2020 A1
20200014663 Chen et al. Jan 2020 A1
20200021514 Michael et al. Jan 2020 A1
20200021515 Michael et al. Jan 2020 A1
20200036624 Michael et al. Jan 2020 A1
20200044943 Bor-Yaliniz et al. Feb 2020 A1
20200044969 Hao et al. Feb 2020 A1
20200059420 Abraham Feb 2020 A1
20200059457 Raza et al. Feb 2020 A1
20200059459 Abraham et al. Feb 2020 A1
20200067831 Spraggins Feb 2020 A1
20200092207 Sipra et al. Mar 2020 A1
20200097327 Beyer et al. Mar 2020 A1
20200099625 Yigit et al. Mar 2020 A1
20200099659 Cometto et al. Mar 2020 A1
20200106696 Michael et al. Apr 2020 A1
20200106706 Mayya et al. Apr 2020 A1
20200119952 Mayya et al. Apr 2020 A1
20200127905 Mayya et al. Apr 2020 A1
20200127911 Gilson et al. Apr 2020 A1
20200153701 Mohan et al. May 2020 A1
20200153736 Liebherr et al. May 2020 A1
20200159661 Keymolen et al. May 2020 A1
20200162407 Tillotson May 2020 A1
20200169473 Rimar et al. May 2020 A1
20200177503 Hooda et al. Jun 2020 A1
20200177550 Valluri et al. Jun 2020 A1
20200177629 Hooda et al. Jun 2020 A1
20200186471 Shen et al. Jun 2020 A1
20200195557 Duan et al. Jun 2020 A1
20200204460 Schneider et al. Jun 2020 A1
20200213212 Dillon et al. Jul 2020 A1
20200213224 Cheng et al. Jul 2020 A1
20200218558 Sreenath et al. Jul 2020 A1
20200235990 Janakiraman et al. Jul 2020 A1
20200235999 Mayya et al. Jul 2020 A1
20200236046 Jain et al. Jul 2020 A1
20200241927 Yang et al. Jul 2020 A1
20200244721 S et al. Jul 2020 A1
20200252234 Ramamoorthi et al. Aug 2020 A1
20200259700 Bhalla et al. Aug 2020 A1
20200267184 Vera-Schockner Aug 2020 A1
20200267203 Jindal et al. Aug 2020 A1
20200280587 Janakiraman et al. Sep 2020 A1
20200287819 Theogaraj et al. Sep 2020 A1
20200287976 Theogaraj et al. Sep 2020 A1
20200296011 Jain et al. Sep 2020 A1
20200296026 Michael et al. Sep 2020 A1
20200301764 Thoresen et al. Sep 2020 A1
20200314006 Mackie et al. Oct 2020 A1
20200314614 Moustafa et al. Oct 2020 A1
20200322230 Natal et al. Oct 2020 A1
20200322287 Connor et al. Oct 2020 A1
20200336336 Sethi et al. Oct 2020 A1
20200344089 Motwani et al. Oct 2020 A1
20200344143 Faseela et al. Oct 2020 A1
20200344163 Gupta et al. Oct 2020 A1
20200351188 Arora et al. Nov 2020 A1
20200358878 Bansal et al. Nov 2020 A1
20200366530 Mukundan et al. Nov 2020 A1
20200366562 Mayya et al. Nov 2020 A1
20200382345 Zhao et al. Dec 2020 A1
20200382387 Pasupathy et al. Dec 2020 A1
20200403821 Dev et al. Dec 2020 A1
20200412483 Tan et al. Dec 2020 A1
20200412576 Kondapavuluru et al. Dec 2020 A1
20200413283 Shen et al. Dec 2020 A1
20210006482 Hwang et al. Jan 2021 A1
20210006490 Michael et al. Jan 2021 A1
20210021538 Meck et al. Jan 2021 A1
20210029019 Kottapalli Jan 2021 A1
20210029088 Mayya et al. Jan 2021 A1
20210036888 Makkalla et al. Feb 2021 A1
20210036987 Mishra et al. Feb 2021 A1
20210037159 Shimokawa Feb 2021 A1
20210049191 Masson et al. Feb 2021 A1
20210067372 Cidon et al. Mar 2021 A1
20210067373 Cidon et al. Mar 2021 A1
20210067374 Cidon et al. Mar 2021 A1
20210067375 Cidon et al. Mar 2021 A1
20210067407 Cidon et al. Mar 2021 A1
20210067427 Cidon et al. Mar 2021 A1
20210067442 Sundararajan et al. Mar 2021 A1
20210067461 Cidon et al. Mar 2021 A1
20210067464 Cidon et al. Mar 2021 A1
20210067467 Cidon et al. Mar 2021 A1
20210067468 Cidon et al. Mar 2021 A1
20210073001 Rogers et al. Mar 2021 A1
20210092062 Dhanabalan Mar 2021 A1
20210099360 Parsons et al. Apr 2021 A1
20210105199 C H et al. Apr 2021 A1
20210111998 Saavedra Apr 2021 A1
20210112034 Sundararajan et al. Apr 2021 A1
20210126830 R. et al. Apr 2021 A1
20210126853 Ramaswamy et al. Apr 2021 A1
20210126854 Guo et al. Apr 2021 A1
20210126860 Ramaswamy et al. Apr 2021 A1
20210144091 C H et al. May 2021 A1
20210160169 Shen et al. May 2021 A1
20210160813 Gupta et al. May 2021 A1
20210176255 Hill et al. Jun 2021 A1
20210184952 Mayya et al. Jun 2021 A1
20210184966 Ramaswamy et al. Jun 2021 A1
20210184983 Ramaswamy et al. Jun 2021 A1
20210194814 Roux et al. Jun 2021 A1
20210226880 Ramamoorthy et al. Jul 2021 A1
20210234728 Cidon et al. Jul 2021 A1
20210234775 Devadoss et al. Jul 2021 A1
20210234786 Devadoss et al. Jul 2021 A1
20210234804 Devadoss et al. Jul 2021 A1
20210234805 Devadoss et al. Jul 2021 A1
20210235312 Devadoss et al. Jul 2021 A1
20210235313 Devadoss et al. Jul 2021 A1
20210266262 Subramanian et al. Aug 2021 A1
20210279069 Salgaonkar et al. Sep 2021 A1
20210314289 Chandrashekhar et al. Oct 2021 A1
20210314385 Pande et al. Oct 2021 A1
20210328835 Mayya et al. Oct 2021 A1
20210336880 Gupta et al. Oct 2021 A1
20210377109 Shrivastava et al. Dec 2021 A1
20210377156 Michael et al. Dec 2021 A1
20210392060 Silva et al. Dec 2021 A1
20210392070 Tootaghaj et al. Dec 2021 A1
20210399920 Sundararajan et al. Dec 2021 A1
20210399978 Michael et al. Dec 2021 A9
20210400113 Markuze et al. Dec 2021 A1
20210400512 Agarwal et al. Dec 2021 A1
20210409277 Jeuk et al. Dec 2021 A1
20220006726 Michael et al. Jan 2022 A1
20220006751 Ramaswamy et al. Jan 2022 A1
20220006756 Ramaswamy et al. Jan 2022 A1
20220029902 Shemer Jan 2022 A1
20220035673 Markuze et al. Feb 2022 A1
20220038370 Vasseur et al. Feb 2022 A1
20220038557 Markuze et al. Feb 2022 A1
20220045927 Liu et al. Feb 2022 A1
20220052927 Yelahanka Raghuprasad Feb 2022 A1
20220052928 Sundararajan et al. Feb 2022 A1
20220061059 Dunsmore et al. Feb 2022 A1
20220086035 Devaraj et al. Mar 2022 A1
20220094644 Cidon et al. Mar 2022 A1
20220123961 Mukundan et al. Apr 2022 A1
20220131740 Mayya et al. Apr 2022 A1
20220131807 Srinivas et al. Apr 2022 A1
20220131898 Hooda et al. Apr 2022 A1
20220141184 Oswal et al. May 2022 A1
20220158923 Ramaswamy et al. May 2022 A1
20220158924 Ramaswamy et al. May 2022 A1
20220158926 Wennerström et al. May 2022 A1
20220166713 Markuze et al. May 2022 A1
20220191719 Roy Jun 2022 A1
20220198229 López et al. Jun 2022 A1
20220210035 Hendrickson et al. Jun 2022 A1
20220210041 Gandhi et al. Jun 2022 A1
20220210042 Gandhi et al. Jun 2022 A1
20220210122 Evin et al. Jun 2022 A1
20220217015 Vuggrala Jul 2022 A1
20220231949 Ramaswamy et al. Jul 2022 A1
20220231950 Ramaswamy et al. Jul 2022 A1
20220232411 Vijayakumar et al. Jul 2022 A1
20220294701 Mayya et al. Sep 2022 A1
20220335027 Seshadri et al. Oct 2022 A1
20220337553 Mayya et al. Oct 2022 A1
20220353152 Ramaswamy Nov 2022 A1
20220353171 Ramaswamy et al. Nov 2022 A1
20220353175 Ramaswamy et al. Nov 2022 A1
20220353182 Ramaswamy et al. Nov 2022 A1
20220353190 Ramaswamy et al. Nov 2022 A1
20220360500 Ramaswamy et al. Nov 2022 A1
20220407773 Kempanna et al. Dec 2022 A1
20220407774 Kempanna et al. Dec 2022 A1
20220407790 Kempanna et al. Dec 2022 A1
20220407820 Kempanna et al. Dec 2022 A1
20220407915 Kempanna et al. Dec 2022 A1
20230006929 Mayya et al. Jan 2023 A1
20230025586 Rolando et al. Jan 2023 A1
20230026330 Rolando et al. Jan 2023 A1
20230026865 Rolando et al. Jan 2023 A1
20230028872 Ramaswamy Jan 2023 A1
20230039869 Ramaswamy et al. Feb 2023 A1
20230041916 Zhang et al. Feb 2023 A1
20230054961 Ramaswamy et al. Feb 2023 A1
20230105680 Simlai et al. Apr 2023 A1
20230121871 Mayya et al. Apr 2023 A1
20230179445 Cidon et al. Jun 2023 A1
20230179502 Ramaswamy et al. Jun 2023 A1
20230179521 Markuze et al. Jun 2023 A1
20230179543 Cidon et al. Jun 2023 A1
Foreign Referenced Citations (50)
Number Date Country
1926809 Mar 2007 CN
102577270 Jul 2012 CN
102811165 Dec 2012 CN
104956329 Sep 2015 CN
106230650 Dec 2016 CN
106656847 May 2017 CN
106998284 Aug 2017 CN
110447209 Nov 2019 CN
111198764 May 2020 CN
1912381 Apr 2008 EP
2538637 Dec 2012 EP
2763362 Aug 2014 EP
3041178 Jul 2016 EP
3297211 Mar 2018 EP
3509256 Jul 2019 EP
3346650 Nov 2019 EP
2002368792 Dec 2002 JP
2010233126 Oct 2010 JP
2014200010 Oct 2014 JP
2017059991 Mar 2017 JP
2017524290 Aug 2017 JP
20170058201 May 2017 KR
2574350 Feb 2016 RU
03073701 Sep 2003 WO
2005071861 Aug 2005 WO
2007016834 Feb 2007 WO
2012167184 Dec 2012 WO
2015092565 Jun 2015 WO
2016061546 Apr 2016 WO
2016123314 Aug 2016 WO
2017083975 May 2017 WO
2019070611 Apr 2019 WO
2019094522 May 2019 WO
2020012491 Jan 2020 WO
2020018704 Jan 2020 WO
2020091777 May 2020 WO
2020101922 May 2020 WO
2020112345 Jun 2020 WO
2021040934 Mar 2021 WO
2021118717 Jun 2021 WO
2021150465 Jul 2021 WO
2021211906 Oct 2021 WO
2022005607 Jan 2022 WO
2022082680 Apr 2022 WO
2022154850 Jul 2022 WO
2022159156 Jul 2022 WO
2022231668 Nov 2022 WO
2022235303 Nov 2022 WO
2022265681 Dec 2022 WO
2023009159 Feb 2023 WO
Non-Patent Literature Citations (66)
Entry
VeloCloud Administration Guide. VMware SD-WAN by VeloCloud 3.3 (2019) (Year: 2019).
Barozet, Jean-Marc, “Cisco SD-WAN as a Managed Service,” BRKRST-2558, Jan. 27-31, 2020, 98 pages, Cisco, Barcelona, Spain, retrieved from https://www.ciscolive.com/c/dam/r/ciscolive/emea/docs/2020/pdf/BRKRST-2558.pdf.
Jivorasetkul, Supalerk, et al., “End-to-End Header Compression over Software-Defined Networks: a Low Latency Network Architecture,” 2012 Fourth International Conference on Intelligent Networking and Collaborative Systems, Sep. 19-21, 2012, 2 pages, IEEE, Bucharest, Romania.
Non-Published Commonly Owned U.S. Appl. No. 17/827,972, filed May 30, 2022, 30 pages, Nicira, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/850,112, filed Jun. 27, 2022, 41 pages, Nicira, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/943,147, filed Sep. 12, 2022, 42 pages, Nicira, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/967,795, filed Oct. 17, 2022, 39 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/976,784, filed Oct. 29, 2022, 55 pages, VMware, Inc.
Zakurdaev, Gieorgi, et al., “Dynamic On-Demand Virtual Extensible LAN Tunnels via Software-Defined Wide Area Networks,” 2022 IEEE 12th Annual Computing and Communication Workshop and Conference, Jan. 26-29, 2022, 6 pages, IEEE, Las Vegas, NV, USA.
Non-Published Commonly Owned U.S. Appl. No. 18/083,536, filed Dec. 18, 2022, 27 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 18/102,685, filed Jan. 28, 2023, 124 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 18/102,687, filed Jan. 28, 2023, 172 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 18/102,688, filed Jan. 28, 2023, 49 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 18/102,689, filed Jan. 28, 2023, 46 pages, VMware, Inc.
Taleb, Tarik, “D4.1 Mobile Network Cloud Component Design,” Mobile Cloud Networking, Nov. 8, 2013, 210 pages, MobileCloud Networking Consortium, retrieved from http://www.mobile-cloud-networking.eu/site/index.php?process=download&id=127&code=89d30565cd2ce087d3f8e95f9ad683066510a61f.
Valtulina, Luca, “Seamless Distributed Mobility Management (DMM) Solution in Cloud Based LTE Systems,” Master Thesis, Nov. 2013, 168 pages, University of Twente, retrieved from http://essay.utwente.nl/64411/1/Luca_Valtulina_MSc_Report_final.pdf.
Alsaeedi, Mohammed, et al., “Toward Adaptive and Scalable OpenFlow-SDN Flow Control: A Survey,” IEEE Access, Aug. 1, 2019, 34 pages, vol. 7, IEEE, retrieved from https://ieeexplore.ieee.org/document/8784036.
Alvizu, Rodolfo, et al., “SDN-Based Network Orchestration for New Dynamic Enterprise Networking Services,” 2017 19th International Conference on Transparent Optical Networks, Jul. 2-6, 2017, 4 pages, IEEE, Girona, Spain.
Barozet, Jean-Marc, “Cisco SDWAN,” Deep Dive, Dec. 2017, 185 pages, Cisco, Retreived from https://www.coursehero.com/file/71671376/Cisco-SDWAN-Deep-Divepdf/.
Bertaux, Lionel, et al., “Software Defined Networking and Virtualization for Broadband Satellite Networks, ” IEEE Communications Magazine, Mar. 18, 2015, 7 pages, vol. 53, IEEE, retrieved from https://ieeexplore.ieee.org/document/7060482.
Cox, Jacob H., et al., “Advancing Software-Defined Networks: A Survey,” IEEE Access, Oct. 12, 2017, 40 pages, vol. 5, IEEE, retrieved from https://ieeexplore.ieee.org/document/8066287.
Del Piccolo, Valentin, et al., “A Survey of Network Isolation Solutions for Multi-Tenant Data Centers,” IEEE Communications Society, Apr. 20, 2016, vol. 18, No. 4, 37 pages, IEEE.
Duan, Zhenhai, et al., “Service Overlay Networks: SLAs, QoS, and Bandwidth Provisioning,” IEEE/ACM Transactions on Networking, Dec. 2003, 14 pages, vol. 11, IEEE, New York, NY, USA.
Fortz, Bernard, et al., “Internet Traffic Engineering by Optimizing OSPF Weights,” Proceedings IEEE INFOCOM 2000, Conference on Computer Communications, Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies, Mar. 26-30, 2000, 11 pages, IEEE, Tel Aviv, Israel, Israel.
Francois, Frederic, et al., “Optimizing Secure SDN-enabled Inter-Data Centre Overlay Networks through Cognitive Routing,” 2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS), Sep. 19-21, 2016, 10 pages, IEEE, London, UK.
Guo, Xiangyi, et al., U.S. Appl. No. 62/925,193, filed Oct. 23, 2019, 26 pages.
Huang, Cancan, et al., “Modification of Q.SD-WAN,” Rapporteur Group Meeting—Doc, Study Period 2017-2020, Q4/11-DOC1 (190410), Study Group 11, Apr. 10, 2019, 19 pages, International Telecommunication Union, Geneva, Switzerland.
Lasserre, Marc, et al., “Framework for Data Center (DC) Network Virtualization,” RFC 7365, Oct. 2014, 26 pages, IETF.
Li, Shengru, et al., “Source Routing with Protocol-oblivious Forwarding (POF) to Enable Efficient e-Health Data Transfers,” 2016 IEEE International Conference on Communications (ICC), May 22-27, 2016, 6 pages, IEEE, Kuala Lumpur, Malaysia.
Lin, Weidong, et al., “Using Path Label Routing in Wide Area Software-Defined Networks with Open Flow,” 2016 International Conference on Networking and Network Applications, Jul. 2016, 6 pages, IEEE.
Long, Feng, “Research and Application of Cloud Storage Technology in University Information Service,” Chinese Excellent Masters' Theses Full-text Database, Mar. 2013, 72 pages, China Academic Journals Electronic Publishing House, China.
Michael, Nithin, et al., “HALO: Hop-by-Hop Adaptive Link-State Optimal Routing,” IEEE/ACM Transactions on Networking, Dec. 2015, 14 pages, vol. 23, No. 6, IEEE.
Ming, Gao, et al., “A Design of SD-WAN-Oriented Wide Area Network Access,” 2020 International Conference on Computer Communication and Network Security (CCNS), Aug. 21-23, 2020, 4 pages, IEEE, Xi'an, China.
Mishra, Mayank, et al., “Managing Network Reservation for Tenants in Oversubscribed Clouds,” 2013 IEEE 21st International Symposium on Modelling, Analysis and Simulation of Computer and Telecommunication Systems, Aug. 14-16, 2013, 10 pages, IEEE, San Francisco, CA, USA.
Mudigonda, Jayaram, et al., “NetLord: A Scalable Multi-Tenant Network Architecture for Virtualized Datacenters,” Proceedings of the ACM SIGCOMM 2011 Conference, Aug. 15-19, 2011, 12 pages, ACM, Toronto, Canada.
Non-Published Commonly Owned U.S. Appl. No. 17/103,614, filed Nov. 24, 2020, 38 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/143,092, filed Jan. 6, 2021, 42 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/143,094, filed Jan. 6, 2021, 42 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/194,038, filed Mar. 5, 2021, 35 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/227,016, filed Apr. 9, 2021, 37 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/227,044, filed Apr. 9, 2021, 37 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/351,327, filed Jun. 18, 2021, 48 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/351,333, filed Jun. 18, 2021, 47 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/351,340, filed Jun. 18, 2021, 48 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/351,342, filed Jun. 18, 2021, 47 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/351,345, filed Jun. 18, 2021, 48 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/384,735, filed Jul. 24, 2021, 62 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/384,736, filed Jul. 24, 2021, 63 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/384,737, filed Jul. 24, 2021, 63 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/384,738, filed Jul. 24, 2021, 62 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/510,862, filed Oct. 26, 2021, 46 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/517,639, filed Nov. 2, 2021, 46 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/517,641, filed Nov. 2, 2021, 46 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/562,890, filed Dec. 27, 2021, 36 pages, Nicira, Inc.
Non-Published Commonly Owned U.S. Appl. No. 17/572,583, filed Jan. 10, 2022, 33 pages, Nicira, Inc.
Non-Published Commonly Owned U.S. Appl. No. 15/803,964, filed Nov. 6, 2017, 15 pages, The Mode Group.
Noormohammadpour, Mohammad, et al., “DCRoute: Speeding up Inter-Datacenter Traffic Allocation while Guaranteeing Deadlines,” 2016 IEEE 23rd International Conference on High Performance Computing (HiPC), Dec. 19-22, 2016, 9 pages, IEEE, Hyderabad, India.
Ray, Saikat, et al., “Always Acyclic Distributed Path Computation,” University of Pennsylvania Department of Electrical and Systems Engineering Technical Report, May 2008, 16 pages, University of Pennsylvania ScholarlyCommons.
Sarhan, Soliman Abd Elmonsef, et al., “Data Inspection in SDN Network,” 2018 13th International Conference on Computer Engineering and Systems (ICCES), Dec. 18-19, 2018, 6 pages, IEEE, Cairo, Egypt.
Tootaghaj, Diman Zad, et al., “Homa: An Efficient Topology and Route Management Approach in SD-WAN Overlays,” IEEE INFOCOM 2020—IEEE Conference on Computer Communications, Jul. 6-9, 2020, 10 pages, IEEE, Toronto, ON, Canada.
Webb, Kevin C., et al., “Blender: Upgrading Tenant-Based Data Center Networking,” 2014 ACM/IEEE Symposium on Architectures for Networking and Communications Systems (ANCS), Oct. 20-21, 2014, 11 pages, IEEE, Marina del Rey, CA, USA.
Xie, Junfeng, et al., A Survey of Machine Learning Techniques Applied to Software Defined Networking (SDN): Research Issues and Challenges, IEEE Communications Surveys & Tutorials, Aug. 23, 2018, 38 pages, vol. 21, Issue 1, IEEE.
Yap, Kok-Kiong, et al., “Taking the Edge off with Espresso: Scale, Reliability and Programmability for Global Internet Peering,” SIGCOMM '17: Proceedings of the Conference of the ACM Special Interest Group on Data Communication, Aug. 21-25, 2017, 14 pages, Los Angeles, CA.
Funabiki, Nobuo, et al., “A Frame Aggregation Extension of Routing Algorithm for Wireless Mesh Networks,” 2014 Second International Symposium on Computing and Networking, Dec. 10-12, 2014, 5 pages, IEEE, Shizuoka, Japan.
Non-Published Commonly Owned U.S. Appl. No. 18/137,584, filed Apr. 21, 2023, 57 pages, VMware, Inc.
Non-Published Commonly Owned U.S. Appl. No. 18/197,090, filed May 14, 2023, 36 pages, Nicira, Inc.
Related Publications (1)
Number Date Country
20220239596 A1 Jul 2022 US