This application is related to U.S. application Ser. No. 15/289,755, filed Oct. 10, 2016, and entitled “ORCHESTRATION SYSTEM FOR MIGRATING USER DATA AND SERVICES BASED ON USER INFORMATION,” the contents which are explicitly incorporated herein in their entirety.
The present technology pertains to cloud and data center orchestration systems, and more specifically, orchestration of cloud and fog interactions during events such as overloads, failures, or security events.
As a result of globalization and computing mobility, users may require reliable and quick access to network data at different times and from a wide range of locations, in order to complete their tasks and business objectives. Cloud computing enables users to access data and services on “the cloud” through the Internet from anywhere in the world. Not surprisingly, the cloud has become an integral part of our lives, as it hosts and provisions a large and increasing number of services ranging from entertainment services to productivity services to infrastructure services. Service requirements and customer expectations for cloud services are similarly diverse—all varying greatly from service to service and user to user.
To meet the exceeding demands for service quality and diversity, fog computing has emerged as an extension of cloud computing. Fog computing allows services or data from the cloud to be offloaded to “the fog”. The fog can include nodes that are geographically and/or logically closer to client devices. The closer proximity of the fog to client devices can result in a reduction in latency and an increase in security and reliability for data and services hosted by the fog nodes. The fog and cloud can enable providers to balance the benefits of the cloud, such as scalability and flexibility, with the benefits of the fog, such as lower latency and better security. However, the fog-cloud architecture can also add a significant amount of complexity for providers and increase the potential points of failure for a service. In some cases, this can negatively impact service reliability and degrade the user experience.
Orchestration is the process whereby the resources of a complex network are allocated, configured, and managed. Orchestration is well known in the cloud, but is an emerging capability of fog systems. Fog orchestration poses unique challenges due to the hierarchical nature of the fog, its diverse set of resources, and widely distributed physical and logical geography. Achieving efficient and secure interaction between the cloud and levels of the fog is a particularly important capability.
In order to describe the manner in which the above-recited and other advantages and features of the disclosure can be obtained, a more particular description of the principles briefly described above will be rendered by reference to specific embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only exemplary embodiments of the disclosure and are not therefore to be considered to be limiting of its scope, the principles herein are described and explained with additional specificity and detail through the use of the accompanying drawings in which:
Various embodiments of the disclosure are discussed in detail below. While specific implementations are discussed, it should be understood that this is done for illustration purposes only. A person skilled in the relevant art will recognize that other components and configurations may be used without parting from the spirit and scope of the disclosure.
Additional features and advantages of the disclosure will be set forth in the description which follows, and in part will be obvious from the description, or can be learned by practice of the herein disclosed principles. The features and advantages of the disclosure can be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. These and other features of the disclosure will become more fully apparent from the following description and appended claims, or can be learned by the practice of the principles set forth herein.
The cloud and fog layers of a network can add complexity and points of failure to the services provisioned by the network. This can negatively impact service performance and reliability and the overall user experience. Effective and efficient orchestration mechanisms can greatly improve the service performance and reliability, as well as the user experience, in cloud and fog computing.
Disclosed herein are systems, methods, and computer-readable media for orchestrating cloud to fog interactions. The approaches set forth herein can provide effective, efficient, and intelligent orchestration of services, nodes, and workloads between cloud and fog layers in a network. Such careful orchestration can result in significant improvements in performance, reliability, and efficiency. The orchestration can be as fine-grained, dynamic, and responsive as necessary for each particular application or context.
To illustrate, in some examples, a method can involve partitioning an application into software containers. Each of the software containers can be configured to host a respective component of the application, such as a service, a function, a workload, a resource, code, etc. For example, an application suite can be divided by functions and each of the functions can be hosted on one or more specific software containers.
The method can further involve identifying nodes on respective layers of a hierarchical cloud-fog architecture for hosting the software containers. The hierarchical cloud-fog architecture can include one or more cloud layers and one or more fog layers. For example, the hierarchical cloud-fog architecture can include a cloud layer and a fog layer containing multiple sub-layers. The various layers can include a hierarchy. For example, the fog layer can include a lower sub-layer, an intermediate sub-layer, and a high sub-layer. Similarly, the cloud can include a high sub-layer and a low sub-layer. The hierarchy can be based on logical or physical proximity to a reference point, such as the cloud, the users, the client endpoints, the local area networks, etc. The hierarchy can also be based on other factors, for example, relative performance, relative bandwidth, relative resources, relative cost, etc. Finally, the hierarchy can follow the natural boundaries of an application, for example placing local, neighborhood, and regional hierarchy layers of fog nodes in a smart city, or having machine, manufacturing cell and assembly line layers of fog nodes in a smart factory.
The method can also involve deploying the software containers at the nodes on the respective hierarchical layers of the cloud-fog architecture. Each of the software containers can be deployed to a respective layer from the cloud-fog architecture, such as a cloud layer, a fog sub-layer, etc. The software containers can be deployed at respective nodes selected based on one or more specific factors, such as capacity, security, resource availability, performance, status, cost, proximity, etc. The specific factors used for mapping software containers to respective nodes can be considered individually, separately, or relative to each other for example.
The disclosed technology addresses the need in the art for orchestration of cloud and fog interactions. The present technology involves system, methods, and computer-readable media for efficiently and effectively orchestrating cloud and fog interactions during an event, such as an overload, a failure, a security event, etc.
A description of example cloud and fog network architectures, as illustrated in
The cloud 102 can provide various cloud computing services via the cloud elements 104-114, such as software as a service (SaaS) (e.g., collaboration services, email services, enterprise resource planning services, content services, communication services, etc.), infrastructure as a service (IaaS) (e.g., security services, networking services, systems management services, etc.), platform as a service (PaaS) (e.g., web services, streaming services, application development services, etc.), and other types of services such as desktop as a service (DaaS), information technology management as a service (ITaaS), managed software as a service (MSaaS), mobile backend as a service (MBaaS), etc.
The client endpoints 116 can connect with the cloud 102 to obtain one or more specific services from the cloud 102. The client endpoints 116 can communicate with elements 104-114 via one or more public networks (e.g., Internet), private networks, and/or hybrid networks (e.g., virtual private network). The client endpoints 116 can include any device with networking capabilities, such as a laptop computer, a tablet computer, a server, a desktop computer, a smartphone, a network device (e.g., an access point, a router, a switch, etc.), a smart television, a smart car, a sensor, a GPS device, a game system, a smart wearable object (e.g., smartwatch, etc.), a consumer object (e.g., Internet refrigerator, smart lighting system, etc.), a city or transportation system (e.g., traffic control, toll collection system, etc.), an internet of things (IoT) device, a camera, a network printer, a transportation system (e.g., airplane, train, motorcycle, boat, etc.), or any smart or connected object (e.g., smart home, smart building, smart retail, smart glasses, etc.), and so forth.
The fog layer 156 or “the fog” provides the computation, storage and networking capabilities of traditional cloud networks, but closer to the endpoints. The fog can thus extend the capabilities of the cloud 102 to be closer to the client endpoints 116. The fog nodes 162 can be the physical implementation of fog networks. Moreover, the fog nodes 162 can provide local or regional services and/or connectivity to the client endpoints 116. As a result, traffic and/or data can be offloaded from the cloud 102 to the fog layer 156 (e.g., via fog nodes 162). The fog layer 156 can thus provide faster services and/or connectivity to the client endpoints 116, with lower latency, as well as other advantages such as security benefits from keeping the data inside the local or regional network(s), and service resiliency in the presence of certain cloud layer or network failures.
The fog nodes 162 can include any networked computing devices, such as servers, switches, routers, controllers, cameras, access points, gateways, etc. Moreover, the fog nodes 162 can be deployed anywhere with a network connection, such as a factory floor, a power pole, alongside a railway track, in a vehicle, on an oil rig, in an airport, on an aircraft, in a shopping center, in a hospital, in a park, in a parking garage, on a street corner, in a library, etc.
In some configurations, one or more fog nodes 162 can be deployed within fog instances 158, 160. The fog instances 158, 160 can be local or regional clouds, networks, or nodes. For example, the fog instances 158, 160 can be a regional cloud or data center, a local area network, a network or cluster of fog nodes 162, etc. In some configurations, one or more fog nodes 162 can be deployed within a network, or as standalone or individual nodes, for example. Moreover, one or more of the fog nodes 162 can be interconnected with each other via links 164 in various topologies, including star, ring, mesh or hierarchical arrangements, for example.
In some cases, one or more fog nodes 162 can be mobile fog nodes. The mobile fog nodes can move to different geographic locations, logical locations or networks, and/or fog instances while maintaining connectivity with the cloud layer 154 and/or the endpoints 116. For example, a particular fog node can be placed in a vehicle, such as an aircraft or train, which can travel from one geographic location and/or logical location to a different geographic location and/or logical location. In this example, the particular fog node may connect to a particular physical and/or logical connection point with the cloud 154 while located at the starting location and switch to a different physical and/or logical connection point with the cloud 154 while located at the destination location. The particular fog node can thus move within particular clouds and/or fog instances and, therefore, serve endpoints from different locations at different times.
The levels 172-176 can represent sub-layers within the fog layer 156. Moreover, each of the levels 172-176 can include one or more nodes 162. Further, the levels 172-176 can vary based on one or more aspects, such as proximity to the cloud or client endpoints 116. For example, high level fog nodes 172 can be closer to the cloud. As used herein, the term “closer”, when referring to multiple items (e.g., a reference to an item being closer to another item), can refer to logical proximity, physical proximity, number of hops, latency of communications, performance metrics, etc.
To illustrate, in a non-limiting example, the low-level fog nodes 176 can include one or more fog nodes 162 that are closer in proximity (logical and/or physical) to the client endpoints 116 than the cloud 102, have a lower latency or faster performance of communications to the client endpoints 116 relative to the cloud 102, etc.
The intermediate level fog nodes 174 can provide a layer below the high level fog nodes 172. Thus, the intermediate level fog nodes 174 may be closer to the client endpoints 116 than the high level fog nodes 172. The low level fog nodes 176 can provide yet another layer below the high level fog nodes 172 and the intermediate level fog nodes 174. Accordingly, the low level fog nodes 176 can be closer to the client endpoints 116 than both the high level fog nodes 172 and the intermediate level fog nodes 174.
The different levels in the fog layer (i.e., levels 172-176) can provide certain advantages over the cloud 102, such as performance and security advantages. Accordingly, data, workloads, services, resources, functions, operations, etc., can be offloaded or distributed from the cloud 102 to the different levels in the fog layer 156 in order to increase performance, security, reliability, etc. Likewise, the cloud 102 can provide advantages, such as cost, resources, scalability, etc. Thus, certain aspects (e.g., data, workloads, services, resources, functions, operations, etc.) can be hosted on the cloud 102 instead of the fog layer 156. A balance of advantages can be achieved by distributing data, workloads, services, resources, functions, operations, etc., in different ways throughout the cloud 102 and the different levels 172-176 in the fog layer 156. Orchestration systems, such as orchestration system 300 shown in
Together, the cloud layer 154 and the different levels 172-176 in the fog layer 156 can allow for distribution or partitioning of an application, a service chain, a service, resources, etc. For example, as further described below with reference to
As previously noted, the application 210 can be partitioned into functions 202-208 and each of the functions 202-208 distributed on the cloud and fog architecture 170. Such partitioning can be performed, for example and without limitation, based service function chaining techniques or other partitioning techniques which enable different portions, functions, services, or aspects of an application to run on different hosts.
The functions 202-208 can be distributed across the cloud 102 and different levels 172-176 of the fog layer 156. For example, functions 202 can be hosted on the cloud 102, functions 204 can be hosted on one or more of the high level fog nodes 172, functions 206 can be hosted on one or more of the intermediate level fog nodes 174, and functions 208 can be hosted on one or more of the low level fog nodes 176.
In some examples, each function can be hosted on a separate node, container, virtual machine, etc. However, in other examples, two or more functions may be hosted on a same node, container, virtual machine, etc. The number and/or identity of functions hosted by any particular node, container, virtual machine, etc., can vary in different implementations. Such variations in the different implementations can affect or impact the various parameters of the functions 202-208 and the application suite 210 as a whole.
For example, one or more factors such as the specific partitioning and/or distribution of the functions 202-208, as well as the specific layering or hierarchical configuration (e.g., number of levels in the fog layer 156, number of nodes in a particular level, number of resources available or allocated at a particular level, the distribution of nodes within levels 172-176, the type of nodes and/or platforms at a particular level, etc.) can improve the scalability, performance, cost, security, efficiency, reliability, and/or other parameters of the respective functions 202-208 and/or the application 210 as a whole. Accordingly, the particular partitioning or distribution (e.g., the number and/or identity of functions hosted in any particular node) and/or the specific layering or hierarchical configuration can be selected or configured based on specific factor(s) and/or requirement(s), such as scalability, performance, cost, security, efficiency, reliability, location, network conditions, resource availability, etc.
Some layers or levels may be better suited than others for certain things or may confer certain benefits may be better suited for, or yield a greater impact on, certain functions. Accordingly, when configuring a particular application for a particular partitioning or distribution scheme as well as layering or hierarchical configuration, the characteristics and/or requirements associated with the specific functions 202-208, as well as the characteristics or parameters of the various layers or levels (e.g., cloud 102 and levels 172-176) can be taken into account to intelligently identify the optimal scenario or configuration for a particular application.
As one of ordinary skill in the art will recognize, the characteristics or requirements can vary between different applications. Thus, the configuration or scenario selected can be tailored for an application. Such tailoring can take into account the relative characteristics and conditions of the various layers or levels. For example, higher layers or levels, such as the cloud 102 and high level nodes 172 may generally provide, without limitation, cost, resource, and scalability benefits. In some scenarios, it can also provide other benefits such as performance, reliability, etc. On the other hand, lower layers or levels, such as the low level nodes 176 and the intermediate level nodes 174 may provide, without limitation, other benefits, such as security and performance, for example. These are general characterizations which are often applicable, but may vary in different cases. Therefore, it can be advantageous to intelligently tailor each application.
To illustrate, moving the more resource-intensive functions to higher levels or layers in the hierarchical configuration, such as the cloud 102, the high level fog nodes 172, and/or the intermediate level fog nodes 174, may provide certain benefits such as lower cost or higher performance if the higher levels or layers are equipped with faster or additional resources. On the other hand, moving the more resource-intensive functions to lower levels or layers in the hierarchical configuration, such as the intermediate level fog nodes 174 or the low level fog nodes 176, may provide certain benefits such as higher performance, better latency or reliability if such levels or layers are able to allocate adequate or comparable resources while also providing communication or bandwidth benefits resulting, for example, from fewer communication hops or bottlenecks.
In addition, the distributed or partitioned nature of applications in the orchestration configuration 200 can also result in increased efficiency, performance, security, reliability, etc., as workloads and/or functions can be serviced by different resources within the hierarchical cloud and fog architecture 170.
Having disclosed example hierarchical cloud and fog architectures and configurations, the disclosure now turns to a detailed discussion of orchestration of cloud and fog interactions in a hierarchical cloud and fog architecture.
In a fog and cloud hierarchical architecture 170, application software and/or components that may otherwise be run in the cloud 102 or cloud layer 154 can be moved to a hierarchy of fog nodes (e.g., levels 172-176) that are arranged between the cloud layer 154 and the endpoint clients 116 or endpoint “Things”. The arrangement of which software modules or components run at which layers of the cloud and fog hierarchy can be predetermined, but may also be dynamic, with the optimum location of a specific software module or component being determined by many factors, including current processor, storage or network loads on the application, latency targets, message transmission cost, node failures, security considerations, quality of service (QoS), reliability requirements, scalability, etc.
In some cases, cloud-based applications can be responsible for configuring, managing, monitoring, and load balancing some or all of the fog nodes 162 in the fog layer 156. However, given the various latency, security and availability requirements of Fog-based applications, there may be some measure of autonomy and performance scalability in the fog layer 156, which can limit the disruptions or impact in cases when, for example, the cloud 102 is unreachable, down, or overloaded.
Virtual machine and software container technologies, such as DOCKER and kernel-based virtual machine (KVM), can be implemented to manage the orchestration of resources in a cloud and fog hierarchical architecture, and further improve the versatility, performance and reliability of cloud and fog software management and orchestration.
When implementing a cloud and fog hierarchical architecture, the first step can be to partition a Cloud-Fog application, such as an application that would typically run in a single container or on a KVM system on a single host processor, into an interrelated collection or “cluster” of containers.
Partitioning can be along natural demarcation lines within the larger application, for example, cutting horizontally between the stages of a multi-step algorithm, or vertically across multiple parallel operations (e.g., for applications that support parallel execution). Well-defined inter-container communication pathways can tie the containers in a cluster together. In a simple deployment, the containers in the cluster needed to implement the entire application may run on a single host, sharing the same physical instance of an OS (Operating System) and hardware processor. The cluster of containers can be moved as a unit up or down the cloud and fog hierarchy 170 until the optimal level is found which may balance various parameters, such as the cost and/or resources used with the performance requirements. Cost considerations may naturally push cluster members up toward the cloud layer 154 where computation and storage may be cheaper, but performance requirements (e.g., latency, network bandwidth utilization, reliability, security, etc.) may push the cluster members down toward the lower fog layer 156 or the lower fog levels, such as intermediate level 174 or low level 176 for instance.
The partitioning, orchestration and management of functions between the cloud layer 154 and the levels 172-176 of the fog layer 156 may be especially carefully considered in times of highly dynamic or abnormal operation. When there is some sort of failure, natural disaster, temporary overload, or network-wide security problem, the cloud layer 154 or cloud 102 may change its mode from “overlord” to “assistant” until the problem is rectified. Containers may need to be moved between cloud and fog layers 154, 156 in response to highly dynamic network conditions. Context data consistency and integrity may also be managed. Various techniques and mechanisms can be implemented to quickly and reliably detect the need for such a change and make the transition seamlessly.
For an example of how an application may be split into a cluster of containers and orchestrated in the cloud and fog hierarchical architecture 170, consider the example shown in
To implement the multi-camera security suite 222 in a hierarchical cloud and fog model 170, the multi-camera security suite 222 can be partitioned into functions 224-234. As previously explained, such partitioning can be performed, for example and without limitation, based service function chaining techniques or other partitioning techniques which enable different portions, functions, services, or aspects of an application to run on different hosts.
In the example of multi-camera security suite 222, the “Things” or client endpoints 116 can include a network of cameras that send video streams to a multi-camera security suite application 222, which can be an analytics, storage and business intelligence application (e.g., perhaps as a single container, traditionally located in the cloud 102). The functions 224-234 can include business intelligence functions 224, video archiving 226, video security 228, video analytics 230, video decryption and compression 232, and video contrast enhancement and feature extraction 234, which makeup the multi-camera security suite 222. Other additional functions can also be included without limitation.
The functions 224-234 can be distributed across the cloud 102 and different levels 172-176 of the fog layer 156. For example, the business intelligence functions 224 can be hosted on the cloud 102. The video archiving 226 can be hosted on one or more of the high level fog nodes 172, the video security 228 and video analytics 230 functions can be hosted on one or more of the intermediate level fog nodes 174, and the video decryption and compression 232 and video contrast enhancement and feature extraction 234 functions can be hosted on one or more of the low level fog nodes 176. This example distribution is provided for illustration purposes and may vary based on one or more factors, such as resources, service requirements, network conditions, application or function characteristics, etc.
In some examples, each of functions 224-234 can be hosted on a separate node, container, virtual machine, etc. However, in other examples, two or more of the functions 224-234 may be hosted on a same node, container, virtual machine, etc. The number and/or identity of functions hosted by any particular node, container, virtual machine, etc., can vary in different implementations. In yet other examples, particularly resource intensive functions such as video analytics functions 230 may span more than one fog node in a layer, exploiting multi-node parallelism. As previously explained with reference to
The functions 224-234 can be containers assembled in a service function chain. Thus, in configuration 220, the application 222 can be partitioned into a number of containers assembled in a service function chain. In this example, the configuration 220 can be described as follows.
The lowest level containers (i.e., 232 and 234) may decrypt and decompress a video stream from a camera, and may provide contrast enhancement and feature extraction.
Next, a container 230 may include video analytics functions like pattern matching and object recognition. Another container, container 228, can take the analytics outputs and uses them to perform security functions like detecting people crossing an e-fence, or finding evidence of shoplifting. Other containers could detect the flow of customers in the videos, to help store planning and merchandising.
The next level of containers, container 226, may archive interesting video segments for later viewing. The highest level of container, container 224, can provide business intelligence functions, like alerting the owner if a security or store layout problem is discovered. If multiple cameras are in operation, multiple instances of the relevant subset of containers can be duplicated and run in parallel, still as part of the cluster.
Other applications, such as IoT applications from other verticals, like autonomous vehicles, smart cities, smart grids, and healthcare, may have their own set of natural partitioning to map their algorithms into chains of containers in a cluster.
Instead of running all containers (e.g., 224-234) in the cluster in a same cloud server or fog node, the containers may be split up and down and across the cloud and fog hierarchy. Thus, the lowest layer fog nodes 176 may run the lowest level video functions, the intermediate fog nodes 174 may run the analytics functions, the high layer fog nodes 172 may run the archival functions, and the cloud 102 may run the business intelligence.
The orchestration system can carefully manage which containers of the cluster run at which levels of the cloud and fog hierarchy 170, and may continue to tune this mapping based different measurements such as network performance. This can optimize various IoT network attributes like latency, security, network bandwidth, reliability, and cost.
The mapping of which containers in the cluster run on which layers of the cloud and fog hierarchy 170 need not be static. As the loads on the cloud 102 and fog nodes 162 fluctuate up and down, and as the performance of the application and the data complexity on which it operates varies, it is possible to dynamically move containers to other levels or layers of the hierarchy 170.
For example, if the automated orchestration system detects that latency is approaching a critical limit, some of the more computational-intensive functions may be moved one step lower in the fog layer 156, where they are presumably closer to the client endpoints 116 or “Things”, and should have shorter response times. Conversely, if other applications (with their own clusters of containers) are requesting space on some fog node that is currently fully occupied, or if an application is costing more fog resources than allocated, and if there is performance margin to spare, some containers could be moved one step up towards the cloud layer 154. This sort of movement could also be horizontal, between peer fog nodes on the same level of the hierarchy 170 to help balance the loads among them.
For example, if a single fog node is running the image processing and analytics functions 230 for multiple cameras, and is running out of resources, the containers associated with the second and subsequent cameras could be off-loaded to adjacent fog node(s).
Another reason to dynamically move containers within a cluster between fog nodes may be for fault tolerance. If a fog node, network link, or power source failure is detected, the same facilities that provided load balancing in the above paragraph could be used to automatically move the load off the failed resource to a nearby redundant resource, and non-stop operation of the application can be preserved. If an entire cloud data center in the cloud layer 154 is becoming seriously overloaded, is unreachable because of a network problem, or is in danger of a failure, the containers running on the cloud layer 154 and the orchestration controlling the entire cluster can be temporarily moved down to the highest level 172 of fog nodes, making fog an extension of, but temporarily independent from, the cloud layer 154.
The agility in software partitioning this cluster of containers scheme provides can make possible mission critical and even life critical applications in IoT, without an unacceptably high software development expense.
Other factors for moving containers within a cluster between nodes or layers in the cloud and fog hierarchy 170 include security. For example, if automatic security audits detect a node in a cluster is experiencing some sort of security compromise (e.g., physical attack, cyber attack, DDoS attack, hacking, crypto key compromise, etc.), the cluster orchestration system can instantly move load off the suspicious nodes, and isolate them for deeper investigation. Certain critical IoT applications (especially those in control of potentially dangerous IoT actuators) may have very stringent security requirements. This system can improve end-to-end security by isolating specific questionable nodes. Also, when software patches or hardware updates are required (security related or otherwise), this system can perform a rolling update process, where only specific containers in the cluster are moved to adjacent nodes, the updates are made, and the containers are then moved back.
Ideally, a container move should operate within the latency window of the worst-case fog applications—which is on the order of a few milliseconds. Thus, the algorithm can yield low latency moving of containers. The move of a container can involve pausing the operation of one container in the cluster, collect its operational context (e.g., in-flight data, intermediate results, database entries, etc.), move that context to the destination node, restore the context, and restart the container. One technique that may make this faster is to have the cluster's orchestration system, which decides when and where to move containers within a cluster, declare the possible destination nodes where each container could end up moving in advance. A “dummy container” can be made to shadow the active container on each possible destination node. Background messages can be sent between the active container and the dummy containers at the possible locations it may move to, in order to keep the slow-changing context (e.g., user databases, billing records, etc.) up-to-date throughout.
This way, when the real move is necessary, only in-flight data would have to be packaged and sent in a hurry to the new node. Hopefully, if this is done right, the application users and round-trip latency critical IoT use cases will not notice the refactoring the orchestration system is doing to continuously optimize the cloud and fog network. Pre-planning of possible destinations should the orchestration system decide to move active containers, and the two-phase process to keep the shadow containers updated can provide significant advantages.
The orchestration technologies herein can optimize use of various resources, including CPU, storage, network, etc. For example, if a special protocol conversion task is required at the Cloud-Fog boundary, the algorithm performing the special protocol conversion task can be configured into a container. The orchestration system can continuously monitor the network processing load on both the cloud 102 and fog nodes 162, and dynamically move the container to whichever side of the Cloud-Fog boundary has a lighter load. Storage can benefit from similar dynamic assignment. For example, if a video caching service is running in a container, and the network detects a focused load on a specific fog node, or a nearly exhausted storage array on the selected cloud servers, the storage container (including the compressed video it serves) can be moved down from the cloud layer 154 (e.g., cloud 102) to the fog layer 156. This process could continue to descend lower into the Fog Layer 156 until the optimal balance of resource use and performance is achieved for the instantaneous video viewing patterns of the users of the network.
As previously described, the distribution of the container elements of the cluster along the Cloud-Fog continuum (e.g., cloud and fog hierarchy 170) can be determined based on resource usage information provided by the orchestration layer (e.g., Docker/LXC mem_info, CPU usage, etc., by individual containers as fraction of total host resources). In other examples, however, the distribution of the container elements of the cluster along the Cloud-Fog continuum can be intelligently and automatically determined based on resource usage telemetry streamed from both hosted resources (e.g., containers, bare-metal servers, etc.), as well as potentially the underlying network infrastructure. This resource usage telemetry information can be obtained by a variety of channels, such as API calls by a centralized SDN (software defined network) controller, also as (type 1 or 2) metadata from the Network Service Header (NSH) of a Service Function Chain containing a variety of containerized virtual network functions (VNFs), etc.
Irrespective of the source of this resource utilization telemetry, this information can be used to specify pre-configured policies for resource utilization for the various levels of the Cloud-Fog continuum as described above. In this manner the container orchestration layer has a well-defined stratification of the Cloud-Fog continuum such that it can quickly, intelligently and automatically deploy the various container cluster members appropriately based on application needs, and redeploy them dynamically as those needs change. This well-defined resource stratification of the Cloud-Fog continuum also allows for dynamic movement of cluster member containerized workloads to ensure sustainable high availability for the cloud layer 154 by a complementary fog infrastructure (e.g., fog layer 156).
The cloud and fog hierarchy 170 can react in real-time or near real-time using these techniques. This containerization automation throughout the cloud also enables and benefits from the ACI (application-centric infrastructure) data center architecture. The container “cluster” can itself be an application within an ACI network. As the cluster is dynamically moved throughout the Cloud-Fog continuum, the Cloud-Fog orchestration system may programmatically push policy updates through the ACI (if present in the architecture). As a more monolithic application is split into a container cluster, the policy governing access of bits of the application can be transparently taught to the fabric. This policy automation can continue out into the fog layer 156 by programming other controllers present within the network.
The orchestration system herein can provide agile mapping of complex applications into multiple containers in a cluster, and automatically move containers vertically between layers of the cloud and fog hierarchy 170 to optimize the balance between efficiency and performance. The system can also automatically move containers in a cluster horizontally between peer-level fog nodes to provide load balancing and exploit parallelism, for example, or to provide redundancy for fault tolerance.
The orchestration system 300 can include one or more devices or nodes. For example, the orchestration system 300 can be a single server or a group of servers. Moreover, the orchestration system 300 can reside in any layer within the cloud layer 154 and/or the fog layer 156. In some cases, the orchestration system 300 can include multiple nodes which can be distributed within the same layer or level in the cloud and fog hierarchy 170, or different layers or levels.
The orchestration system 300 can communicate with the cloud layer 154, the fog layer 156, and/or the client endpoints 116 via a network 308. For example, the orchestration system 300 can communicate with the cloud 102 and/or one or more of the fog nodes 162 in any of levels 172-176.
The orchestration system 300 can include an orchestration module 302, which can define, partition, cluster, and/or set containers, functions, etc., and map or schedule them to one or more specific layers, levels, and/or nodes within the cloud and fog hierarchy 170. The orchestration system 300 can also include a monitoring module 304 for monitoring containers, applications, application components, functions, network conditions, layers, levels, resources, requirements, etc. For example, the monitoring module 304 can collect performance and status information from specific nodes in the cloud and fog hierarchy 170 to identify conditions or events (e.g., failures, errors, availability, overloading, security breach, etc.). The monitoring module 304 can report any data, including conditions or events, to the orchestration module 302 in order to dynamically adjust the orchestration for one or more applications. The orchestration module 302 can use the data from the monitoring module 304 to identify which containers or functions should be moved and where they should be moved.
The orchestration system 300 can include a communications module 306 for communicating with network 308 and other nodes, networks, devices, etc. The orchestration system 300 can use the communications module 306 to send and receive messages, signals, alerts, packages, and communications with other devices and networks in the hierarchical architecture 170
Having disclosed some basic system components and concepts, the disclosure now turns to the example method embodiment shown in
At step 400, the orchestration system 300 can partition an application into software containers. Each of the software containers can be configured to host one or more respective components of the application, such as a function, a feature, a service, a library, a portion of code, a data set, etc. For example, an application can be partitioned into functions, and the functions then allocated or configured on respective software containers. To illustrate, using service function chaining, an application involving ten functions can be partitioned into ten software containers where each of the software containers hosts or runs one of the ten functions. Containers can also hold the distributed data sets associated with an application. For example, static data may be in one set of containers, dynamic data and intermediate results in a second set, logs in a third set, etc.
At step 402, the orchestration system 300 can identify respective nodes on respective hierarchical layers (e.g., cloud layer 154 and fog layers 172-176) of a hierarchical cloud and fog architecture 170 for hosting the software containers on the respective hierarchical layers of the cloud and fog architecture 170. Here, the orchestration system 300 can select specific nodes or layers for hosting specific software containers. In other words, in addition to identifying the nodes, or as part of identifying the nodes, the orchestration system 300 can map or designate specific nodes to specific software containers. For example, the orchestration system 300 can map software container A to node X on fog layer Y, software container B to node W on fog layer Z, software container C to cloud 102, etc.
The identifying and mapping of nodes to containers can be based on one or more factors, such as performance, security, scalability, bandwidth, cost, resource availability, resource status, resource consumption or requirements, quality of service requirements, etc. The one or more factors can also include specific characteristics or parameters associated with the software containers, the application, and the specific components hosted at each of the software containers.
When identifying a node for a particular software container, the orchestration system 300 can compare, analyze, and/or match specific parameters of the particular software container, one or more specific nodes, and/or one or more specific layers in the hierarchical cloud and fog architecture. Specific parameters of a software container can be considered relative to parameters of other software containers. Likewise, specific parameters of a node or layer in the hierarchical cloud and fog architecture 170 can be considered relative to parameters of other nodes or layers.
For example, when identifying a node and layer to host a software container running video security functions 228 shown in
To illustrate, the orchestration system 300 can determine that the video security functions 228 have high security and/or performance requirements (or higher than other functions associated with the application 222). The orchestration system 300 may also determine that the low and intermediate level fog nodes 174, 176 provide higher security and performance than the high level fog nodes 172 or the cloud 102. The orchestration system 300 can then use this information to select or identify one or more nodes from the low level fog nodes 176 or the intermediate level fog nodes 174 for the software container associated with the video security functions 228. The orchestration system 300 can further tailor or fine tune the mapping for the software container associated with the video security functions 228 based on other considerations. For example, if the low level fog nodes 176 have limited capacity or availability or if other functions are given a higher priority to the low level fog nodes 176 based on relative parameters (e.g., performance, security, etc.), then the software container associated with the video security functions 228 can instead be mapped to the intermediate level fog nodes 174. The orchestration system 300 can thus identify a particular node from the intermediate level fog nodes 174 for the software container associated with the video security functions 228.
In some cases, nodes, functions, software containers, layers, etc., can be prioritized. For example, assume application A is partitioned by function into ten functions and corresponding software containers. The ten functions and corresponding software containers can be sorted or prioritized by one or more factors, such as performance and/or security requirements, for example. A priority can be determined for the ten functions. In some cases, multiple priorities can also be determined for the ten functions based on different factors. The priorities associated with the ten functions can be compared with specific parameters associated with the different nodes and/or layers in the hierarchical cloud and fog architecture 170. In some cases, the different nodes and/or layers of the hierarchical cloud and fog architecture 170 can also be ranked based on one or more factors. The priorities associated with the ten functions can then be compared with, or analyzed in view of, the rankings of the different nodes and/or layers. This can be used to identify and map nodes and layers to specific software containers and corresponding application components (e.g., functions).
To illustrate, an application partitioned into three functions and corresponding software containers can be analyzed to determine a relative rank or priority of the three functions based on performance demands or requirements. The ranking or prioritization can result in function A being ranked highest as having the greatest performance demands or requirements, function C ranked in the lowest as having the lowest performance demands or requirements, and function B ranked in the middle between functions A and C. The cloud and fog layers 154, 156 can also be ranked based on estimated and/or historical performance. The performance ranking can be, for example, low level fog nodes 176 ranked first (i.e., highest performance), intermediate level fog nodes 174 ranked second, high level fog nodes 172 ranked third, and the cloud 102 ranked last. The various rankings and prioritizations can be used to then map the functions to layers in the hierarchical cloud and fog architecture 170. For example, function A can be mapped to the low level fog nodes 176 based on a determination that function A is ranked highest as having the greatest performance demands or requirements, and the low level fog nodes 176 are ranked first based on performance. In some cases, a specific node within the low level fog nodes 176 can then be identified for, or mapped to, function A based on one or more factors, such as node and/or resource availability, proximity, bandwidth, cost, capacity, status, resource utilization, etc.
At step 404, the orchestration system 300 can deploy the software containers at the respective nodes on the respective hierarchical layers of the hierarchical cloud and fog architecture 170. For example, the orchestration system 300 can move, migrate, configure, install, run, and/or instantiate, the software containers on specific, selected nodes. This deployment can be performed when the application is being initially configured or setup in the hierarchical cloud and fog architecture 170, after the application has been setup in the hierarchical cloud and fog architecture 170, during operations of the application in the hierarchical cloud and fog architecture 170, etc.
In some cases, the deployment or a re-deployment of some or all of the software containers can be dynamic based on a triggering event, such as a failure, an alarm, a threshold, a performance condition, a security condition, a status, etc. For example, assume that functions 224-234 of application 222 in
The orchestration system 300 can maintain shadow containers based on specific, active containers, at specific nodes or layers for redundancy and fault tolerance. For example, the orchestration system 300 can setup a software container on a node from the low level fog nodes 176 as a backup for the video security functions 228. The shadow software container can mirror some or all of the data and/or settings from the active software container of the video security functions 228. The shadow software container can also include operational context collected from the active software container to improve the efficiency of a transition or redeployment. Accordingly, the orchestration system 300 can collect operational context and other data from active software containers and move such data to shadow containers maintained as backups.
For the sake of clarity and illustration,
The disclosure now turns to
The interfaces 502 are typically provided as modular interface cards (sometimes referred to as “line cards”). Generally, they control the sending and receiving of data packets over the network and sometimes support other peripherals used with the network device 500. Among the interfaces that may be provided are Ethernet interfaces, frame relay interfaces, cable interfaces, DSL interfaces, token ring interfaces, and the like. In addition, various very high-speed interfaces may be provided such as fast token ring interfaces, wireless interfaces, Ethernet interfaces, Gigabit Ethernet interfaces, ATM interfaces, HSSI interfaces, POS interfaces, FDDI interfaces, WIFI interfaces, 3G/4G/5G cellular interfaces, CAN BUS, LoRA, and the like. Generally, these interfaces may include ports appropriate for communication with the appropriate media. In some cases, they may also include an independent processor and, in some instances, volatile RAM. The independent processors may control such communications intensive tasks as packet switching, media control, signal processing, crypto processing, and management. By providing separate processors for the communications intensive tasks, these interfaces allow the master microprocessor 504 to efficiently perform routing computations, network diagnostics, security functions, etc.
Although the system shown in
Regardless of the network device's configuration, it may employ one or more memories or memory modules (including memory 506) configured to store program instructions for the general-purpose network operations and mechanisms for roaming, route optimization and routing functions described herein. The program instructions may control the operation of an operating system and/or one or more applications, for example. The memory or memories may also be configured to store tables such as mobility binding, registration, and association tables, etc. Memory 506 could also hold various software containers and virtualized execution environments and data.
The network device 500 can also include an application-specific integrated circuit (ASIC) 512, which can be configured to perform routing and/or switching operations. The ASIC 512 can communicate with other components in the network device 500 via the bus 510, to exchange data and signals and coordinate various types of operations by the network device 500, such as routing, switching, and/or data storage operations, for example.
The memory 620 can include multiple different types of memory with different performance characteristics. The processor 604 can include any general purpose processor and a hardware module or software module, such as module 1610, module 2612, and module 3614 stored in storage device 608, configured to control the processor 604 as well as a special-purpose processor where software instructions are incorporated into the actual processor design. The processor 604 may essentially be a completely self-contained computing system, containing multiple cores or processors, a bus, memory controller, cache, etc. A multi-core processor may be symmetric or asymmetric.
To enable user interaction with the computing device 600, an input device 622 can represent any number of input mechanisms, such as a microphone for speech, a touch-sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth. An output device 624 can also be one or more of a number of output mechanisms known to those of skill in the art. In some instances, multimodal systems can enable a user to provide multiple types of input to communicate with the computing device 600. The communications interface 626 can generally govern and manage the user input and system output. There is no restriction on operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.
Storage device 608 can be a non-volatile memory, and can be a hard disk or other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, solid state memory devices, digital versatile disks, cartridges, random access memories (RAMs) 616, read only memory (ROM) 618, and hybrids thereof.
The system 600 can include an integrated circuit 628, such as an application-specific integrated circuit (ASIC) configured to perform various operations. The integrated circuit 628 can be coupled with the bus 606 in order to communicate with other components in the system 600.
The storage device 608 can include software modules 610, 612, 614 for controlling the processor 604. Other hardware or software modules are contemplated. The storage device 608 can be connected to the system bus 606. In one aspect, a hardware module that performs a particular function can include the software component stored in a computer-readable medium in connection with the necessary hardware components, such as the processor 604, bus 606, output device 624, and so forth, to carry out the function.
It can be appreciated that example system 600 can have more than one processor 604 or be part of a group or cluster of computing devices networked together to provide greater processing capability.
For clarity of explanation, in some instances the present technology may be presented as including individual functional blocks including functional blocks comprising devices, device components, steps or routines in a method embodied in software, or combinations of hardware and software.
In some embodiments the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like. However, when mentioned, non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.
Methods according to the above-described examples can be implemented using computer-executable instructions that are stored or otherwise available from computer readable media. Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Portions of computer resources used can be accessible over a network. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.
Devices implementing methods according to these disclosures can comprise hardware, firmware and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, rackmount devices, standalone devices, and so on. Functionality described herein also can be embodied in peripherals or add-in cards. Such functionality can also be implemented on a circuit board among different chips or different processes executing in a single device, by way of further example.
The instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described in these disclosures.
Although a variety of examples and other information was used to explain aspects within the scope of the appended claims, no limitation of the claims should be implied based on particular features or arrangements in such examples, as one of ordinary skill would be able to use these examples to derive a wide variety of implementations. Further and although some subject matter may have been described in language specific to examples of structural features and/or method steps, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to these described features or acts. For example, such functionality can be distributed differently or performed in components other than those identified herein. Rather, the described features and steps are disclosed as examples of components of systems and methods within the scope of the appended claims.
Claim language reciting “at least one of” a set indicates that one member of the set or multiple members of the set satisfy the claim. For example, claim language reciting “at least one of A and B” means A, B, or A and B.
Number | Name | Date | Kind |
---|---|---|---|
5812773 | Norin | Sep 1998 | A |
5889896 | Meshinsky et al. | Mar 1999 | A |
6108782 | Fletcher et al. | Aug 2000 | A |
6178453 | Mattaway et al. | Jan 2001 | B1 |
6298153 | Oishi | Oct 2001 | B1 |
6343290 | Cossins et al. | Jan 2002 | B1 |
6643260 | Kloth et al. | Nov 2003 | B1 |
6683873 | Kwok et al. | Jan 2004 | B1 |
6721804 | Rubin et al. | Apr 2004 | B1 |
6733449 | Krishnamurthy et al. | May 2004 | B1 |
6735631 | Oehrke et al. | May 2004 | B1 |
6885670 | Regula | Apr 2005 | B1 |
6996615 | McGuire | Feb 2006 | B1 |
7054930 | Cheriton | May 2006 | B1 |
7058706 | Lyer et al. | Jun 2006 | B1 |
7062571 | Dale et al. | Jun 2006 | B1 |
7076397 | Ding et al. | Jul 2006 | B2 |
7111177 | Chauvel et al. | Sep 2006 | B1 |
7212490 | Kao et al. | May 2007 | B1 |
7277948 | Igarashi et al. | Oct 2007 | B2 |
7313667 | Pullela et al. | Dec 2007 | B1 |
7379846 | Williams et al. | May 2008 | B1 |
7480672 | Hahn et al. | Jan 2009 | B2 |
7496043 | Leong et al. | Feb 2009 | B1 |
7536476 | Alleyne | May 2009 | B1 |
7567504 | Darling et al. | Jul 2009 | B2 |
7606147 | Luft et al. | Oct 2009 | B2 |
7647594 | Togawa | Jan 2010 | B2 |
7684322 | Sand et al. | Mar 2010 | B2 |
7773510 | Back et al. | Aug 2010 | B2 |
7808897 | Mehta et al. | Oct 2010 | B1 |
7881957 | Cohen et al. | Feb 2011 | B1 |
7917647 | Cooper et al. | Mar 2011 | B2 |
8010598 | Tanimoto | Aug 2011 | B2 |
8028071 | Mahalingam et al. | Sep 2011 | B1 |
8041714 | Aymeloglu et al. | Oct 2011 | B2 |
8121117 | Amdahl et al. | Feb 2012 | B1 |
8171415 | Appleyard et al. | May 2012 | B2 |
8234377 | Cohn | Jul 2012 | B2 |
8244559 | Horvitz et al. | Aug 2012 | B2 |
8250215 | Stienhans et al. | Aug 2012 | B2 |
8280880 | Aymeloglu et al. | Oct 2012 | B1 |
8284664 | Aybay et al. | Oct 2012 | B1 |
8284776 | Petersen | Oct 2012 | B2 |
8301746 | Head et al. | Oct 2012 | B2 |
8345692 | Smith | Jan 2013 | B2 |
8406141 | Couturier et al. | Mar 2013 | B1 |
8407413 | Yucel et al. | Mar 2013 | B1 |
8448171 | Donnellan et al. | May 2013 | B2 |
8477610 | Zuo et al. | Jul 2013 | B2 |
8495252 | Lais et al. | Jul 2013 | B2 |
8495356 | Ashok et al. | Jul 2013 | B2 |
8510469 | Portolani | Aug 2013 | B2 |
8514868 | Hill | Aug 2013 | B2 |
8532108 | Li et al. | Sep 2013 | B2 |
8533687 | Greifeneder et al. | Sep 2013 | B1 |
8547974 | Guruswamy et al. | Oct 2013 | B1 |
8560639 | Murphy et al. | Oct 2013 | B2 |
8560663 | Baucke et al. | Oct 2013 | B2 |
8589543 | Dutta et al. | Nov 2013 | B2 |
8590050 | Nagpal et al. | Nov 2013 | B2 |
8611356 | Yu et al. | Dec 2013 | B2 |
8612625 | Andries et al. | Dec 2013 | B2 |
8630291 | Shaffer et al. | Jan 2014 | B2 |
8639787 | Lagergren et al. | Jan 2014 | B2 |
8656024 | Krishnan et al. | Feb 2014 | B2 |
8660129 | Brendel et al. | Feb 2014 | B1 |
8719804 | Jain | May 2014 | B2 |
8775576 | Hebert et al. | Jul 2014 | B2 |
8797867 | Chen et al. | Aug 2014 | B1 |
8805951 | Faibish et al. | Aug 2014 | B1 |
8850182 | Fritz et al. | Sep 2014 | B1 |
8856339 | Mestery et al. | Oct 2014 | B2 |
8909780 | Dickinson et al. | Dec 2014 | B1 |
8909928 | Ahmad et al. | Dec 2014 | B2 |
8918510 | Gmach et al. | Dec 2014 | B2 |
8924720 | Raghuram et al. | Dec 2014 | B2 |
8930747 | Levijarvi et al. | Jan 2015 | B2 |
8938775 | Roth et al. | Jan 2015 | B1 |
8959526 | Kansal et al. | Feb 2015 | B2 |
8977754 | Curry, Jr. et al. | Mar 2015 | B2 |
9009697 | Breiter et al. | Apr 2015 | B2 |
9015324 | Jackson | Apr 2015 | B2 |
9043439 | Bicket et al. | May 2015 | B2 |
9049115 | Rajendran et al. | Jun 2015 | B2 |
9063789 | Beaty et al. | Jun 2015 | B2 |
9065727 | Liu et al. | Jun 2015 | B1 |
9075649 | Bushman et al. | Jul 2015 | B1 |
9104334 | Madhusudana et al. | Aug 2015 | B2 |
9164795 | Vincent | Oct 2015 | B1 |
9167050 | Durazzo et al. | Oct 2015 | B2 |
9201701 | Boldyrev et al. | Dec 2015 | B2 |
9201704 | Chang et al. | Dec 2015 | B2 |
9203784 | Chang et al. | Dec 2015 | B2 |
9223634 | Chang et al. | Dec 2015 | B2 |
9244776 | Koza et al. | Jan 2016 | B2 |
9251114 | Ancin et al. | Feb 2016 | B1 |
9264478 | Hon et al. | Feb 2016 | B2 |
9313048 | Chang et al. | Apr 2016 | B2 |
9361192 | Smith et al. | Jun 2016 | B2 |
9380075 | He et al. | Jun 2016 | B2 |
9432294 | Sharma et al. | Aug 2016 | B1 |
9444744 | Sharma et al. | Sep 2016 | B1 |
9473365 | Melander et al. | Oct 2016 | B2 |
9503530 | Niedzielski | Nov 2016 | B1 |
9558078 | Farlee et al. | Jan 2017 | B2 |
9613078 | Vermeulen et al. | Apr 2017 | B2 |
9628471 | Sundaram et al. | Apr 2017 | B1 |
9632858 | Sasturkar et al. | Apr 2017 | B2 |
9658876 | Chang et al. | May 2017 | B2 |
9692802 | Bicket et al. | Jun 2017 | B2 |
9727359 | Tsirkin | Aug 2017 | B2 |
9736063 | Wan et al. | Aug 2017 | B2 |
9755858 | Bagepalli et al. | Sep 2017 | B2 |
9792245 | Raghavan et al. | Oct 2017 | B2 |
9804988 | Ayoub et al. | Oct 2017 | B1 |
9954783 | Thirumurthi et al. | Apr 2018 | B1 |
20020004900 | Patel | Jan 2002 | A1 |
20020073337 | Ioele et al. | Jun 2002 | A1 |
20020143928 | Maltz et al. | Oct 2002 | A1 |
20020166117 | Abrams et al. | Nov 2002 | A1 |
20020174216 | Shorey et al. | Nov 2002 | A1 |
20030018591 | Komisky | Jan 2003 | A1 |
20030056001 | Mate et al. | Mar 2003 | A1 |
20030228585 | Inoko et al. | Dec 2003 | A1 |
20040004941 | Malan et al. | Jan 2004 | A1 |
20040095237 | Chen et al. | May 2004 | A1 |
20040131059 | Ayyakad et al. | Jul 2004 | A1 |
20040264481 | Darling et al. | Dec 2004 | A1 |
20050060418 | Sorokopud | Mar 2005 | A1 |
20050125424 | Herriott et al. | Jun 2005 | A1 |
20060059558 | Selep et al. | Mar 2006 | A1 |
20060104286 | Cheriton | May 2006 | A1 |
20060120575 | Ahn et al. | Jun 2006 | A1 |
20060126665 | Ward et al. | Jun 2006 | A1 |
20060146825 | Hofstaedter et al. | Jul 2006 | A1 |
20060155875 | Cheriton | Jul 2006 | A1 |
20060168338 | Bruegl et al. | Jul 2006 | A1 |
20060294207 | Barsness et al. | Dec 2006 | A1 |
20070011330 | Dinker et al. | Jan 2007 | A1 |
20070174663 | Crawford et al. | Jul 2007 | A1 |
20070223487 | Kajekar et al. | Sep 2007 | A1 |
20070242830 | Conrado et al. | Oct 2007 | A1 |
20080005293 | Bhargava et al. | Jan 2008 | A1 |
20080084880 | Dharwadkar | Apr 2008 | A1 |
20080165778 | Ertemalp | Jul 2008 | A1 |
20080198752 | Fan et al. | Aug 2008 | A1 |
20080201711 | Amir Husain | Aug 2008 | A1 |
20080235755 | Blaisdell et al. | Sep 2008 | A1 |
20090006527 | Gingell, Jr. et al. | Jan 2009 | A1 |
20090010277 | Halbraich et al. | Jan 2009 | A1 |
20090019367 | Cavagnari et al. | Jan 2009 | A1 |
20090031312 | Mausolf et al. | Jan 2009 | A1 |
20090083183 | Rao et al. | Mar 2009 | A1 |
20090138763 | Arnold | May 2009 | A1 |
20090177775 | Radia et al. | Jul 2009 | A1 |
20090178058 | Stillwell, III et al. | Jul 2009 | A1 |
20090182874 | Morford et al. | Jul 2009 | A1 |
20090265468 | Annambhotla et al. | Oct 2009 | A1 |
20090265753 | Anderson et al. | Oct 2009 | A1 |
20090293056 | Ferris | Nov 2009 | A1 |
20090300608 | Ferris et al. | Dec 2009 | A1 |
20090313562 | Appleyard et al. | Dec 2009 | A1 |
20090323706 | Germain et al. | Dec 2009 | A1 |
20090328031 | Pouyadou et al. | Dec 2009 | A1 |
20100042720 | Stienhans et al. | Feb 2010 | A1 |
20100061250 | Nugent | Mar 2010 | A1 |
20100115341 | Baker et al. | May 2010 | A1 |
20100131765 | Bromley et al. | May 2010 | A1 |
20100191783 | Mason et al. | Jul 2010 | A1 |
20100192157 | Jackson et al. | Jul 2010 | A1 |
20100205601 | Abbas et al. | Aug 2010 | A1 |
20100211782 | Auradkar et al. | Aug 2010 | A1 |
20100217886 | Seren et al. | Aug 2010 | A1 |
20100293270 | Augenstein et al. | Nov 2010 | A1 |
20100318609 | Lahiri et al. | Dec 2010 | A1 |
20100325199 | Park et al. | Dec 2010 | A1 |
20100325257 | Goel et al. | Dec 2010 | A1 |
20100325441 | Laurie et al. | Dec 2010 | A1 |
20100333116 | Prahlad et al. | Dec 2010 | A1 |
20110016214 | Jackson | Jan 2011 | A1 |
20110035754 | Srinivasan | Feb 2011 | A1 |
20110055396 | Dehaan | Mar 2011 | A1 |
20110055398 | Dehaan et al. | Mar 2011 | A1 |
20110055470 | Portolani | Mar 2011 | A1 |
20110072489 | Parann-Nissany | Mar 2011 | A1 |
20110075667 | Li et al. | Mar 2011 | A1 |
20110110382 | Jabr et al. | May 2011 | A1 |
20110116443 | Yu et al. | May 2011 | A1 |
20110126099 | Anderson et al. | May 2011 | A1 |
20110138055 | Daly et al. | Jun 2011 | A1 |
20110145413 | Dawson et al. | Jun 2011 | A1 |
20110145657 | Bishop et al. | Jun 2011 | A1 |
20110173303 | Rider | Jul 2011 | A1 |
20110179186 | Li | Jul 2011 | A1 |
20110185063 | Head et al. | Jul 2011 | A1 |
20110199902 | Leavy et al. | Aug 2011 | A1 |
20110213687 | Ferris et al. | Sep 2011 | A1 |
20110213966 | Fu et al. | Sep 2011 | A1 |
20110219434 | Betz et al. | Sep 2011 | A1 |
20110231715 | Kunii et al. | Sep 2011 | A1 |
20110231899 | Pulier et al. | Sep 2011 | A1 |
20110239039 | Dieffenbach et al. | Sep 2011 | A1 |
20110252327 | Awasthi et al. | Oct 2011 | A1 |
20110261811 | Battestilli et al. | Oct 2011 | A1 |
20110261828 | Smith | Oct 2011 | A1 |
20110276675 | Singh et al. | Nov 2011 | A1 |
20110276951 | Jain | Nov 2011 | A1 |
20110295998 | Ferris et al. | Dec 2011 | A1 |
20110305149 | Scott et al. | Dec 2011 | A1 |
20110307531 | Gaponenko et al. | Dec 2011 | A1 |
20110320870 | Kenigsberg et al. | Dec 2011 | A1 |
20120005724 | Lee | Jan 2012 | A1 |
20120023418 | Frields et al. | Jan 2012 | A1 |
20120054367 | Ramakrishnan et al. | Mar 2012 | A1 |
20120072318 | Akiyama et al. | Mar 2012 | A1 |
20120072578 | Alam | Mar 2012 | A1 |
20120072581 | Tung et al. | Mar 2012 | A1 |
20120072985 | Davne et al. | Mar 2012 | A1 |
20120072992 | Arasaratnam et al. | Mar 2012 | A1 |
20120084445 | Brock et al. | Apr 2012 | A1 |
20120084782 | Chou et al. | Apr 2012 | A1 |
20120096134 | Suit | Apr 2012 | A1 |
20120102193 | Rathore et al. | Apr 2012 | A1 |
20120102199 | Hopmann et al. | Apr 2012 | A1 |
20120131174 | Ferris et al. | May 2012 | A1 |
20120137215 | Kawara | May 2012 | A1 |
20120158967 | Sedayao et al. | Jun 2012 | A1 |
20120159097 | Jennas, II et al. | Jun 2012 | A1 |
20120166649 | Watanabe et al. | Jun 2012 | A1 |
20120167094 | Suit | Jun 2012 | A1 |
20120173541 | Venkataramani | Jul 2012 | A1 |
20120173710 | Rodriguez | Jul 2012 | A1 |
20120179909 | Sagi et al. | Jul 2012 | A1 |
20120180044 | Donnellan et al. | Jul 2012 | A1 |
20120182891 | Lee et al. | Jul 2012 | A1 |
20120185632 | Lais et al. | Jul 2012 | A1 |
20120185913 | Martinez et al. | Jul 2012 | A1 |
20120192016 | Gotesdyner et al. | Jul 2012 | A1 |
20120192075 | Ebtekar et al. | Jul 2012 | A1 |
20120201135 | Ding et al. | Aug 2012 | A1 |
20120203908 | Beaty et al. | Aug 2012 | A1 |
20120204169 | Breiter et al. | Aug 2012 | A1 |
20120204187 | Breiter et al. | Aug 2012 | A1 |
20120214506 | Skaaksrud et al. | Aug 2012 | A1 |
20120222106 | Kuehl | Aug 2012 | A1 |
20120236716 | Anbazhagan et al. | Sep 2012 | A1 |
20120240113 | Hur | Sep 2012 | A1 |
20120265976 | Spiers et al. | Oct 2012 | A1 |
20120272025 | Park et al. | Oct 2012 | A1 |
20120281706 | Agarwal et al. | Nov 2012 | A1 |
20120281708 | Chauhan et al. | Nov 2012 | A1 |
20120290647 | Ellison et al. | Nov 2012 | A1 |
20120297238 | Watson et al. | Nov 2012 | A1 |
20120311106 | Morgan | Dec 2012 | A1 |
20120311568 | Jansen | Dec 2012 | A1 |
20120324092 | Brown et al. | Dec 2012 | A1 |
20120324114 | Dutta et al. | Dec 2012 | A1 |
20130003567 | Gallant et al. | Jan 2013 | A1 |
20130013248 | Brugler et al. | Jan 2013 | A1 |
20130036213 | Hasan et al. | Feb 2013 | A1 |
20130044636 | Koponen et al. | Feb 2013 | A1 |
20130066940 | Shao | Mar 2013 | A1 |
20130069950 | Adam et al. | Mar 2013 | A1 |
20130080509 | Wang | Mar 2013 | A1 |
20130080624 | Nagai et al. | Mar 2013 | A1 |
20130091557 | Gurrapu | Apr 2013 | A1 |
20130097601 | Podvratnik et al. | Apr 2013 | A1 |
20130104140 | Meng et al. | Apr 2013 | A1 |
20130111540 | Sabin | May 2013 | A1 |
20130117337 | Dunham | May 2013 | A1 |
20130124712 | Parker | May 2013 | A1 |
20130125124 | Kempf et al. | May 2013 | A1 |
20130138816 | Kuo et al. | May 2013 | A1 |
20130144978 | Jain et al. | Jun 2013 | A1 |
20130152076 | Patel | Jun 2013 | A1 |
20130152175 | Hromoko et al. | Jun 2013 | A1 |
20130159097 | Schory et al. | Jun 2013 | A1 |
20130159496 | Hamilton et al. | Jun 2013 | A1 |
20130160008 | Cawlfield et al. | Jun 2013 | A1 |
20130162753 | Hendrickson et al. | Jun 2013 | A1 |
20130169666 | Pacheco et al. | Jul 2013 | A1 |
20130179941 | McGloin et al. | Jul 2013 | A1 |
20130182712 | Aguayo et al. | Jul 2013 | A1 |
20130185413 | Beaty et al. | Jul 2013 | A1 |
20130185433 | Zhu et al. | Jul 2013 | A1 |
20130191106 | Kephart et al. | Jul 2013 | A1 |
20130198050 | Shroff et al. | Aug 2013 | A1 |
20130198374 | Zalmanovitch et al. | Aug 2013 | A1 |
20130204849 | Chacko | Aug 2013 | A1 |
20130232491 | Radhakrishnan et al. | Sep 2013 | A1 |
20130232492 | Wang | Sep 2013 | A1 |
20130246588 | Borowicz et al. | Sep 2013 | A1 |
20130250770 | Zou et al. | Sep 2013 | A1 |
20130254415 | Fullen et al. | Sep 2013 | A1 |
20130262347 | Dodson | Oct 2013 | A1 |
20130283364 | Chang et al. | Oct 2013 | A1 |
20130297769 | Chang et al. | Nov 2013 | A1 |
20130318240 | Hebert et al. | Nov 2013 | A1 |
20130318546 | Kothuri et al. | Nov 2013 | A1 |
20130339949 | Spiers et al. | Dec 2013 | A1 |
20140006481 | Frey et al. | Jan 2014 | A1 |
20140006535 | Reddy | Jan 2014 | A1 |
20140006585 | Dunbar et al. | Jan 2014 | A1 |
20140019639 | Ueno | Jan 2014 | A1 |
20140040473 | Ho et al. | Feb 2014 | A1 |
20140040883 | Tompkins | Feb 2014 | A1 |
20140052877 | Mao | Feb 2014 | A1 |
20140059310 | Du et al. | Feb 2014 | A1 |
20140074850 | Noel et al. | Mar 2014 | A1 |
20140075048 | Yuksel et al. | Mar 2014 | A1 |
20140075108 | Dong et al. | Mar 2014 | A1 |
20140075357 | Flores et al. | Mar 2014 | A1 |
20140075501 | Srinivasan et al. | Mar 2014 | A1 |
20140089727 | Cherkasova et al. | Mar 2014 | A1 |
20140098762 | Ghai et al. | Apr 2014 | A1 |
20140108985 | Scott et al. | Apr 2014 | A1 |
20140122560 | Ramey et al. | May 2014 | A1 |
20140136779 | Guha et al. | May 2014 | A1 |
20140140211 | Chandrasekaran et al. | May 2014 | A1 |
20140141720 | Princen et al. | May 2014 | A1 |
20140156557 | Zeng et al. | Jun 2014 | A1 |
20140160924 | Pfautz et al. | Jun 2014 | A1 |
20140164486 | Ravichandran et al. | Jun 2014 | A1 |
20140188825 | Muthukkaruppan et al. | Jul 2014 | A1 |
20140189095 | Lindberg et al. | Jul 2014 | A1 |
20140189125 | Amies et al. | Jul 2014 | A1 |
20140215471 | Cherkasova | Jul 2014 | A1 |
20140222953 | Karve et al. | Aug 2014 | A1 |
20140244851 | Lee | Aug 2014 | A1 |
20140245298 | Zhou et al. | Aug 2014 | A1 |
20140269266 | Filsfils et al. | Sep 2014 | A1 |
20140280805 | Sawalha | Sep 2014 | A1 |
20140282536 | Dave et al. | Sep 2014 | A1 |
20140282611 | Campbell et al. | Sep 2014 | A1 |
20140282669 | McMillan | Sep 2014 | A1 |
20140282889 | Ishaya et al. | Sep 2014 | A1 |
20140289200 | Kato | Sep 2014 | A1 |
20140297569 | Clark et al. | Oct 2014 | A1 |
20140297835 | Buys | Oct 2014 | A1 |
20140314078 | Jilani | Oct 2014 | A1 |
20140317261 | Shatzkamer et al. | Oct 2014 | A1 |
20140366155 | Chang et al. | Dec 2014 | A1 |
20140372567 | Ganesh et al. | Dec 2014 | A1 |
20150006470 | Mohan | Jan 2015 | A1 |
20150033086 | Sasturkar et al. | Jan 2015 | A1 |
20150043335 | Testicioglu et al. | Feb 2015 | A1 |
20150043576 | Dixon et al. | Feb 2015 | A1 |
20150052247 | Threefoot et al. | Feb 2015 | A1 |
20150052517 | Raghu et al. | Feb 2015 | A1 |
20150058382 | St. Laurent et al. | Feb 2015 | A1 |
20150058459 | Amendjian et al. | Feb 2015 | A1 |
20150058557 | Madhusudana et al. | Feb 2015 | A1 |
20150070516 | Shoemake et al. | Mar 2015 | A1 |
20150071285 | Kumar et al. | Mar 2015 | A1 |
20150089478 | Cheluvaraju et al. | Mar 2015 | A1 |
20150100471 | Curry, Jr. et al. | Apr 2015 | A1 |
20150106802 | Ivanov, IV et al. | Apr 2015 | A1 |
20150106805 | Melander et al. | Apr 2015 | A1 |
20150109923 | Hwang | Apr 2015 | A1 |
20150117199 | Chinnaiah Sankaran et al. | Apr 2015 | A1 |
20150117458 | Gurkan et al. | Apr 2015 | A1 |
20150120914 | Wada et al. | Apr 2015 | A1 |
20150149828 | Mukerji et al. | May 2015 | A1 |
20150178133 | Phelan et al. | Jun 2015 | A1 |
20150215819 | Bosch et al. | Jul 2015 | A1 |
20150227405 | Jan et al. | Aug 2015 | A1 |
20150242204 | Hassine et al. | Aug 2015 | A1 |
20150249709 | Teng et al. | Sep 2015 | A1 |
20150271199 | Bradley et al. | Sep 2015 | A1 |
20150280980 | Bitar | Oct 2015 | A1 |
20150281067 | Wu | Oct 2015 | A1 |
20150281113 | Siciliano et al. | Oct 2015 | A1 |
20150309908 | Pearson et al. | Oct 2015 | A1 |
20150319063 | Zourzouvillys et al. | Nov 2015 | A1 |
20150326524 | Tankala et al. | Nov 2015 | A1 |
20150339210 | Kopp et al. | Nov 2015 | A1 |
20150373108 | Fleming et al. | Dec 2015 | A1 |
20150379062 | Vermeulen et al. | Dec 2015 | A1 |
20160011925 | Kulkarni et al. | Jan 2016 | A1 |
20160013990 | Kulkarni et al. | Jan 2016 | A1 |
20160062786 | Meng et al. | Mar 2016 | A1 |
20160065417 | Sapuram et al. | Mar 2016 | A1 |
20160094398 | Choudhury et al. | Mar 2016 | A1 |
20160094480 | Kulkarni et al. | Mar 2016 | A1 |
20160094643 | Jain et al. | Mar 2016 | A1 |
20160094894 | Inayatullah et al. | Mar 2016 | A1 |
20160099847 | Melander et al. | Apr 2016 | A1 |
20160099873 | Gerö et al. | Apr 2016 | A1 |
20160103838 | Sainani et al. | Apr 2016 | A1 |
20160105393 | Thakkar et al. | Apr 2016 | A1 |
20160127184 | Bursell | May 2016 | A1 |
20160134557 | Steinder et al. | May 2016 | A1 |
20160147676 | Cha et al. | May 2016 | A1 |
20160162436 | Raghavan et al. | Jun 2016 | A1 |
20160164914 | Madhav et al. | Jun 2016 | A1 |
20160188527 | Cherian et al. | Jun 2016 | A1 |
20160234071 | Nambiar et al. | Aug 2016 | A1 |
20160239399 | Babu et al. | Aug 2016 | A1 |
20160253078 | Ebtekar et al. | Sep 2016 | A1 |
20160254968 | Ebtekar et al. | Sep 2016 | A1 |
20160261564 | Foxhoven et al. | Sep 2016 | A1 |
20160277368 | Narayanaswamy et al. | Sep 2016 | A1 |
20160292611 | Boe et al. | Oct 2016 | A1 |
20160352682 | Chang | Dec 2016 | A1 |
20160378389 | Hrischuk et al. | Dec 2016 | A1 |
20160378518 | Antony | Dec 2016 | A1 |
20170005948 | Melander et al. | Jan 2017 | A1 |
20170024260 | Chandrasekaran et al. | Jan 2017 | A1 |
20170026470 | Bhargava et al. | Jan 2017 | A1 |
20170034199 | Zaw | Feb 2017 | A1 |
20170041342 | Efremov et al. | Feb 2017 | A1 |
20170054659 | Ergin et al. | Feb 2017 | A1 |
20170063674 | Maskalik et al. | Mar 2017 | A1 |
20170097841 | Chang et al. | Apr 2017 | A1 |
20170099188 | Chang et al. | Apr 2017 | A1 |
20170104755 | Arregoces et al. | Apr 2017 | A1 |
20170126583 | Xia | May 2017 | A1 |
20170147297 | Krishnamurthy et al. | May 2017 | A1 |
20170163569 | Koganti | Jun 2017 | A1 |
20170171158 | Hoy et al. | Jun 2017 | A1 |
20170192823 | Karaje et al. | Jul 2017 | A1 |
20170264663 | Bicket et al. | Sep 2017 | A1 |
20170302521 | Lui et al. | Oct 2017 | A1 |
20170310556 | Knowles et al. | Oct 2017 | A1 |
20170317932 | Paramasivam | Nov 2017 | A1 |
20170339070 | Chang et al. | Nov 2017 | A1 |
20170371910 | Joo | Dec 2017 | A1 |
20180069885 | Patterson et al. | Mar 2018 | A1 |
20180173372 | Greenspan et al. | Jun 2018 | A1 |
20180174060 | Velez-Rojas et al. | Jun 2018 | A1 |
Number | Date | Country |
---|---|---|
101719930 | Jun 2010 | CN |
101394360 | Jul 2011 | CN |
102164091 | Aug 2011 | CN |
104320342 | Jan 2015 | CN |
105740084 | Jul 2016 | CN |
2228719 | Sep 2010 | EP |
2439637 | Apr 2012 | EP |
2645253 | Nov 2014 | EP |
10-2015-0070676 | May 2015 | KR |
394537 | Dec 2010 | TW |
WO 2009155574 | Dec 2009 | WO |
WO 2010030915 | Mar 2010 | WO |
WO 2013158707 | Oct 2013 | WO |
Entry |
---|
Dastjeredi et al. “Internet of Things, Principles and Paradigms”, Chapter 4, “Fog Computing: Principles, Architectures, and Applications”, Jun. 17, 2016 (Year: 2016). |
Bittencourtetal., “Towards Virtual Machine Migration in Fog Computing,” 2015 10th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC), Krakow, 2015, pp. 1-8. (Year: 2015). |
International Search Report and Written Opinion from the International Searching Authority, dated Sep. 18, 2017, 13 pages, for International Application No. PCT/US2017/044437, which corresponds to related matter, U.S. Appl. No. 15/289,755. |
Author Unknown, “5 Benefits of a Storage Gateway in the Cloud,” Blog, TwinStrata, Inc., Jul. 25, 2012, XP055141645, 4 pages, https://web.archive.org/web/20120725092619/http://blog.twinstrata.com/2012/07/10//5-benefits-of-a-storage-gateway-in-the-cloud. |
Author Unknown, “Joint Cisco and VMWare Solution for Optimizing Virtual Desktop Delivery: Data Center 3.0: Solutions to Accelerate Data Center Virtualization,” Cisco Systems, Inc. and VMware, Inc., Sep. 2008, 10 pages. |
Author Unknown, “Open Data Center Alliance Usage: Virtual Machine (VM) Interoperability in a Hybrid Cloud Environment Rev. 1.2,” Open Data Center Alliance, Inc., 2013, 18 pages. |
Author Unknown, “Real-Time Performance Monitoring on Juniper Networks Devices, Tips and Tools for Assessing and Analyzing Network Efficiency,” Juniper Networks, Inc., May 2010, 35 pages. |
Beyer, Steffen, “Module “Data::Locations?!”,” YAPC::Europe, London, UK,ICA, Sep. 22-24, 2000, XP002742700, 15 pages. |
Borovick, Lucinda, et al., “Architecting the Network for the Cloud,” IDC White Paper, Jan. 2011, 8 pages. |
Bosch, Greg, “Virtualization,” last modified Apr. 2012 by B. Davison, 33 pages. |
Broadcasters Audience Research Board, “What's Next,” http://Iwww.barb.co.uk/whats-next, accessed Jul. 22, 2015, 2 pages. |
Cisco Systems, Inc. “Best Practices in Deploying Cisco Nexus 1000V Series Switches on Cisco UCS B and C Series Cisco UCS Manager Servers,” Cisco White Paper, Apr. 2011, 36 pages, http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9902/white_paper_c11-558242.pdf. |
Cisco Systems, Inc., “Cisco Unified Network Services: Overcome Obstacles to Cloud-Ready Deployments,” Cisco White Paper, Jan. 2011, 6 pages. |
Cisco Systems, Inc., “Cisco Intercloud Fabric: Hybrid Cloud with Choice, Consistency, Control and Compliance,” Dec. 10, 2014, 22 pages. |
Cisco Technology, Inc., “Cisco Expands Videoscape TV Platform Into the Cloud,” Jan. 6, 2014, Las Vegas, Nevada, Press Release, 3 pages. |
CSS Corp, “Enterprise Cloud Gateway (ECG)—Policy driven framework for managing multi-cloud environments,” original published on or about Feb. 11, 2012; 1 page; http://www.css-cloud.com/platform/enterprise-cloud-qateway.php. |
Fang K., “LISP MAC-EID-TO-RLOC Mapping (LISP based L2VPN),” Network Working Group, Internet Draft, CISCO Systems, Jan. 2012, 12 pages. |
Herry, William, “Keep It Simple, Stupid: OpenStack nova-scheduler and its algorithm”, May 12, 2012, IBM, 12 pages. |
Hewlett-Packard Company, “Virtual context management on network devices”, Research Disclosure, vol. 564, No. 60, Apr. 1, 2011, Mason Publications, Hampshire, GB, Apr. 1, 2011, 524. |
Juniper Networks, Inc., “Recreating Real Application Traffic in Junosphere Lab,” Solution Brief, Dec. 2011, 3 pages. |
Kenhui, “Musings on Cloud Computing and IT-as-a-Service: [Updated for Havana] Openstack Computer for VSphere Admins, Part 2: Nova-Scheduler and DRS”, Jun. 26, 2013, Cloud Architect Musings, 12 pages. |
Kolyshkin, Kirill, “Virtualization in Linux,” Sep. 1, 2006, XP055141648, 5 pages, https://web.archive.org/web/20070120205111/http://download.openvz.org/doc/openvz-intro.pdf. |
Lerach, S.R.O., “Golem,” http://www.lerach.cz/en/products/golem, accessed Jul. 22, 2015, 2 pages. |
Linthicum, David, “VM Import could be a game changer for hybrid clouds”, InfoWorld, Dec. 23, 2010, 4 pages. |
Naik, Vijay K., et al., “Harmony: A Desktop Grid for Delivering Enterprise Computations,” Grid Computing, 2003, Fourth International Workshop on Proceedings, Nov. 17, 2003, pp. 1-11. |
Nair, Srijith K. et al., “Towards Secure Cloud Bursting, Brokerage and Aggregation,” 2012, 8 pages, www.flexiant.com. |
Nielsen, “SimMetry Audience Measurement—Technology,” http://www.nielsen-admosphere.eu/products-and-services/simmetry-audience-measurement-technology/, accessed Jul. 22, 2015, 6 pages. |
Nielsen, “Television ” http://www.nielsen.com/us/en/solutions/measurement/television.html, accessed Jul. 22, 2015, 4 pages. |
Open Stack, “Filter Scheduler,” updated Dec. 17, 2017, 5 pages, accessed on Dec. 18, 2017, https://docs.openstack.org/nova/latest/user/filter-scheduler.html. |
Rabadan, J., et al., “Operational Aspects of Proxy-ARP/ND in EVPN Networks,” BESS Worksgroup Internet Draft, draft-snr-bess-evpn-proxy-arp-nd-02, Oct. 6, 2015, 22 pages. |
Saidi, Ali, et al., “Performance Validation of Network-Intensive Workloads on a Full-System Simulator,” Interaction between Operating System and Computer Architecture Workshop, (IOSCA 2005), Austin, Texas, Oct. 2005, 10 pages. |
Shunra, “Shunra for HP Software; Enabling Confidence in Application Performance Before Deployment,” 2010, 2 pages. |
Son, Jungmin, “Automatic decision system for efficient resource selection and allocation in inter-clouds,” Jun. 2013, 35 pages. |
Wikipedia, “Filter (software)”, Wikipedia, Feb. 8, 2014, 2 pages, https://en.wikipedia.org/w/index.php?title=Filter_%28software%29&oldid=594544359. |
Wikipedia; “Pipeline (Unix)”, Wikipedia, May 4, 2014, 4 pages, https://en.wikipedia.org/w/index.php?title=Pipeline2/028Unix%29&oldid=606980114. |
Amedro, Brian, et al., “An Efficient Framework for Running Applications on Clusters, Grids and Cloud,” 2010, 17 pages. |
Author Unknown, “A Look at DeltaCloud: The Multi-Cloud API,” Feb. 17, 2012, 4 pages. |
Author Unknown, “About Deltacloud,” Apache Software Foundation, Aug. 18, 2013, 1 page. |
Author Unknown, “Architecture for Managing Clouds, A White Paper from the Open Cloud Standards Incubator,” Version 1.0.0, Document No. DSP-IS0102, Jun. 18, 2010, 57 pages. |
Author Unknown, “Cloud Infrastructure Management Interface—Common Information Model (CIMI-CIM),” Document No. DSP0264, Version 1.0.0, Dec. 14, 2012, 21 pages. |
Author Unknown, “Cloud Infrastructure Management Interface (CIMI) Primer,” Document No. DSP2027, Version 1.0.1, Sep. 12, 2012, 30 pages. |
Author Unknown, “cloudControl Documentation,” Aug. 25, 2013, 14 pages. |
Author Unknown, “Interoperable Clouds, A White Paper from the Open Cloud Standards Incubator,” Version 1.0.0, Document No. DSP-IS0101, Nov. 11, 2009, 21 pages. |
Author Unknown, “Microsoft Cloud Edge Gateway (MCE) Series Appliance,” Iron Networks, Inc., 2014, 4 pages. |
Author Unknown, “Use Cases and Interactions for Managing Clouds, A White Paper from the Open Cloud Standards Incubator,” Version 1.0.0, Document No. DSP-IS00103, Jun. 16, 2010, 75 pages. |
Author Unknown, “Apache Ambari Meetup What's New,” Hortonworks Inc., Sep. 2013, 28 pages. |
Author Unknown, “Introduction,” Apache Ambari project, Apache Software Foundation, 2014, 1 page. |
Citrix, “Citrix StoreFront 2.0” White Paper, Proof of Concept Implementation Guide, Citrix Systems, Inc., 2013, 48 pages. |
Citrix, “CloudBridge for Microsoft Azure Deployment Guide,” 30 pages. |
Citrix, “Deployment Practices and Guidelines for NetScaler 10.5 on Amazon Web Services,” White Paper, citrix.com, 2014, 14 pages. |
Gedymin, Adam, “Cloud Computing with an emphasis on Google App Engine,” Sep. 2011, 146 pages. |
Good, Nathan A., “Use Apache Deltacloud to administer multiple instances with a single API,” Dec. 17, 2012, 7 pages. |
Kunz, Thomas, et al., “OmniCloud—The Secure and Flexible Use of Cloud Storage Services,” 2014, 30 pages. |
Logan, Marcus, “Hybrid Cloud Application Architecture for Elastic Java-Based Web Applications,” F5 Deployment Guide Version 1.1, 2016, 65 pages. |
Lynch, Sean, “Monitoring cache with Claspin” Facebook Engineering, Sep. 19, 2012, 5 pages. |
Meireles, Fernando Miguel Dias, “Integrated Management of Cloud Computing Resources,” 2013-2014, 286 pages. |
Mu, Shuai, et al., “uLibCloud: Providing High Available and Uniform Accessing to Multiple Cloud Storages,” 2012 IEEE, 8 pages. |
Sun, Aobing, et al., “laaS Public Cloud Computing Platform Scheduling Model and Optimization Analysis,” Int. J. Communications, Network and System Sciences, 2011, 4, 803-811, 9 pages. |
Szymaniak, Michal, et al., “Latency-Driven Replica Placement”, vol. 47 No. 8, IPSJ Journal, Aug. 2006, 12 pages. |
Toews, Everett, “Introduction to Apache jclouds,” Apr. 7, 2014, 23 pages. |
Von Laszewski, Gregor, et al., “Design of a Dynamic Provisioning System for a Federated Cloud and Bare-metal Environment,” 2012, 8 pages. |
Ye, Xianglong, et al., “A Novel Blocks Placement Strategy for Hadoop,” 2012 IEEE/ACTS 11th International Conference on Computer and Information Science, 2012 IEEE, 5 pages. |
Al-Harbi, S.H., et al., “Adapting k-means for supervised clustering,” Jun. 2006, Applied Intelligence, vol. 24, Issue 3, pp. 219-226. |
Bohner, Shawn A., “Extending Software Change Impact Analysis into COTS Components,” 2003, IEEE, 8 pages. |
Hood, C. S., et al., “Automated Proactive Anomaly Detection,” 1997, Springer Science and Business Media Dordrecht , pp. 688-699. |
Vilalta R., et al., “An efficient approach to external cluster assessment with an application to martian topography,” Feb. 2007, 23 pages, Data Mining and Knowledge Discovery 14.1: 1-23. New York: Springer Science & Business Media. |
Number | Date | Country | |
---|---|---|---|
20180159745 A1 | Jun 2018 | US |