Large-scale clustered environments host numerous servers, sometimes on the order of thousands of servers or more. The servers may be implemented using various virtual devices such as containers, virtual machines, and the like. It may be difficult to monitor the health of the servers and manage traffic among the servers in these environments. For example, the health of a cluster of servers is determined from various factors such as individual server health, application health, and network connectivity. Conventional techniques for monitoring a group of servers and providing a network service typically involve instantiating a service provider (e.g., a monitoring service) on each application server in the cluster of servers. For clustered environments with a large number of nodes, such deployments are computationally expensive and power intensive. Thus, there is a need in the art for effective health monitoring and traffic management for large-scale clustered environments.
Various embodiments of the invention are disclosed in the following detailed description and the accompanying drawings.
The invention can be implemented in numerous ways, including as a process; an apparatus; a system; a composition of matter; a computer program product embodied on a computer readable storage medium; and/or a processor, such as a processor configured to execute instructions stored on and/or provided by a memory coupled to the processor. In this specification, these implementations, or any other form that the invention may take, may be referred to as techniques. In general, the order of the steps of disclosed processes may be altered within the scope of the invention. Unless stated otherwise, a component such as a processor or a memory described as being configured to perform a task may be implemented as a general component that is temporarily configured to perform the task at a given time or a specific component that is manufactured to perform the task. As used herein, the term ‘processor’ refers to one or more devices, circuits, and/or processing cores configured to process data, such as computer program instructions.
A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention encompasses numerous alternatives, modifications and equivalents. Numerous specific details are set forth in the following description in order to provide a thorough understanding of the invention. These details are provided for the purpose of example and the invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the invention is not unnecessarily obscured.
Conventional techniques for monitoring network health and managing network traffic includes centralized monitoring and de-centralized monitoring.
In centralized monitoring, the cluster environment includes a central controller and a local agent for each host. The central controller queries the local agent associated with a host and, in response, receives information about the health of the host from the respective local agent. However, this technique is unable to detect network connectivity issues in some instances. Suppose two hosts are each able to communicate with the central controller, but unable to communicate with each other. Under the conventional centralized monitoring scheme, the connectivity between the two hosts would not be detected because they each appear healthy to the central controller.
In decentralized monitoring, a controller is provided at each local host. A local host typically hosts a cluster or pool of servers. Using the example of load balancing, the controller is implemented as a load balancer. The load balancer load balances east-west traffic and monitors the health of the servers. East-west traffic refers to communications internal to a network such as traffic between backend application servers. Typically, the load balancer monitors not only the health of the servers provided in its respective host, but also the health of server provided on other hosts. Thus, any load balancer in this environment can detect network connectivity issues because the load balancer is aware of the health of all of the servers within the environment. For example, if there are m load balancers and n servers, then the number of checks performed is m×n. Thus, in a large-scale clustered environment, this represents a large load on the servers to respond to each query regarding its health as well as a high amount of traffic, making it impractical to use for most network traffic situations. Furthermore, generating such data is computationally expensive.
Techniques of the present disclosure provide health monitoring and traffic management in a clustered environment. In various embodiments, a service engine (e.g., a special process) provided on a host is configured to monitor its local application servers and selected application servers on other hosts. In various embodiments, the computational cost of this system is j×n, where n is the number of servers and j is a number less than the total number of application servers in the system. Thus, the computational cost of this system is less the computational cost of conventional health monitoring systems (m×n, where m is the number of service engines).
In various embodiments, a set of one or more service engines is selected from among a plurality of service engines executing on a respective plurality of hosts to perform health monitoring of an application. A plurality of instances of the application are executing on at least some of the plurality of hosts, and the set of one or more selected service engines includes a subset of the plurality of service engines. Health monitor data pertaining to the application from the selected set of one or more service engines is collected. A health condition based at least in part on the collected health monitor data is determined. The health condition is output.
In one aspect, health monitoring reflects operational characteristics of the environment and/or member nodes. This determination can be used to redirect traffic to those nodes that are more able to accommodate traffic than other nodes. Using the example of load balancers and servers, in case a load balancer loses connectively to a control plane, the service engine is configured to turn on health-monitoring for all servers deterministically to avoid black holing of traffic (e.g., in case any servers go down during the period the load balancer is not in contact with the control plane). The processes described herein are scalable and find application in large-scale clustered environments, including management of east-west traffic in such environments.
In this example, platform 100 includes a number of devices (e.g., multiple server computers or a multicore server comprising multiple processing cores). A physical device (e.g., 102, 104, 106, etc.) has hardware components and software components, and may be implemented using a device such as system 200 shown in
In some embodiments, the devices include virtual machine (VM)-based systems. In a VM-based system, a number of virtual machines (VMs) such as 118, 119, etc. are configured to execute on the same device such as 102. A VM is a software implementation of a machine (e.g., a computer) that simulates the way a physical machine executes programs. Each VM executes its own operating system. Within the operating system, services are configured to execute as they would on a standard non-VM-based system. The VMs are also referred to as “application servers” or “applications.” The part of the device's operation system that manages the VMs is referred to as the hypervisor. The hypervisor interfaces between the physical hardware and the VMs, providing a layer of abstraction for the VMs. Through its management of the VMs' sharing of the physical hardware resources, the hypervisor makes it appear as though each VM were running on its own dedicated hardware. Examples of hypervisors include VMware Workstation® and Oracle VM VirtualBox®.
In some embodiments, the devices include container-based systems. In a container-based system, multiple containers such as 118, 119, etc. are configured to execute on the same device such as 102, within the same operating system. Each container functions as a lightweight VM that can execute other programs. Examples of container-based systems include Kubernetes®, Docker®, Mesos®, etc.
In some embodiments, the devices include a combination of VM-based systems, container-based systems, and/or standard systems (e.g., non-VM-based systems and/or non-container-based systems).
In this example, clients (not shown) communicate with servers, which are configured to execute within VMs or containers to provide services (also referred to as target applications). Examples of such services include a website service, a shopping cart application, user authentication, credit card authentication, email service, file sharing service, virtual desktop, voice/video streaming, online collaboration, etc. Many other services can be implemented. A server (e.g., a web server, a credit card processing server, a database server, etc.) executes within a VM or container to provide a service, and multiple servers can execute on multiple VMs to provide the same service. For example, three separate servers execute within VMs or containers 118, 120, and 130 to provide the 1st service; three separate servers execute within VMs or containers 119, 129, and 139 to provide the 2nd service; etc.
In the example shown, inside the operating system, there are multiple modules providing different functionalities. One of the modules is a virtual switch (112, 122, 132, etc.). A physical device hardware has one or more physical ports (e.g., Ethernet ports). Network traffic (e.g., data packets) can be transmitted or received by any of the physical ports, to or from any VMs or containers. In some embodiments, the virtual switch is configured to direct traffic to and from one or more appropriate VMs or containers, such as the VM or container in which the service engine on the device is operating. In some embodiments, the virtual switch is configured to direct traffic to and from one or more VMs or containers depending on the health of the VMs or containers, relative health of the VMs and containers to each other, or overall network health.
Service engines 114, 124, and 134 (also referred to as SE1, SE2, and SE3, respectively) are instantiated on physical devices 102, 104, and 106, respectively. In some embodiments, a service engine is implemented as software executing in a virtual machine or container. The service engine is executed to provide distributed network services for processes executing on the same physical server as the service engine as well as for processes executing on different physical servers. Multiple service engines are configured to cooperate and provide the distributed network services. In some embodiments, the service engine is configured to enable appropriate service components that implement service logic. When a specific service is desired, a corresponding server is configured and invoked by the service engine to execute in a VM or container. In this example, the servers are denoted as S(i, j), where i represents the service engine identifier and j represents the service identifier. Thus, S(1, 1) indicates that the server corresponds to SE1 and the 1st service, S(2, 1) indicates that the server corresponds to SE2 and the 1st service, etc.
In various embodiments, the service engines includes an instantiation of an east-west load balancer. For example, an east-west load balancer is instantiated on each of the hosts 102, 104, 106 in the cluster. The load balancer is configured to, among other things, provide distributed load balancing, provide application maps, and support backend SSL services.
The service engine also gathers operational data for the services (e.g., numbers of open connections for the 1st service maintained by servers S(1, 1), S(2, 1), and S(3, 1), respectively; number of requests made to servers S(1, 1), S(2, 1), and S(3, 1), respectively; etc.). Note that the number of instances of servers providing a particular service can vary; for example, there can be multiple servers providing the 1st service on device 104 (in other words, multiple S(2, 1)'s can execute in multiple VMs on device 104). For example, service engine 114 monitors containers S(1,1), S(1,2), S(1,3), and selected containers of device 104 and 106. Details of how the service engine gathers operational data and interacts with controller 190 to monitor health and handle traffic are described herein in
A virtual switch such as 112 interacts with the service engines, and uses existing networking Application Programming Interfaces (APIs) (such as APIs provided by the operating system) to direct traffic and provide distributed network services for services deployed on the network. The operating system and the services implement the networking API calls (e.g., API calls to send data to or receive data from a specific socket at an Internet Protocol (IP) address). In some embodiments, the virtual switch is configured to be in-line with one or more VMs or containers and intercepts traffic designated to and from instances of the services executing on the VMs or containers. When a networking API call is invoked, traffic is intercepted by the in-line virtual switch, which directs the traffic to or from the appropriate VM on which instances of the service execute. In some embodiments, a service engine sends data to and receives data from a server via the virtual switch.
Traffic received on a physical port of a server (e.g., a communications interface such as Ethernet port 115) is sent to the virtual switch (e.g., 112). In some embodiments, the virtual switch is configured to use an API provided by the hypervisor to intercept incoming traffic designated for the target application(s) in an in-line mode, and send the traffic to an appropriate service engine. In in-line mode, packets are forwarded on without being replicated. As shown, the virtual switch passes the traffic to a service engine in the distributed network service layer (e.g., the service engine on the same physical device), which transforms the packets if needed and redirects the packets to the appropriate target application.
A controller 190 is configured to control, monitor, program, and/or provision the services, virtual machines, and/or containers. In particular, the controller is configured to control, monitor, program, and/or provision a group of service engines, and is configured to perform functions such as bringing up the service engines, downloading software onto the service engines, sending configuration information to the service engines, monitoring the service engines' operations, detecting and handling failures, collecting analytics information, and/or determining health indicator changes. The controller can be implemented as software, hardware, firmware, or any combination thereof. In this example, controller 190 determines a health condition of one or more servers and/or at overall health of at least a portion of the system with the assistance of service engines 114, 124, and 134. As will be described in greater detail below, the controller selects one or more service engines to perform health monitoring, collects health monitor data from the selected service engine(s), and determines the health condition based on the collected data.
In various embodiments, the controller is implemented by an orchestrator. The orchestrator may be configured to coordinate agents (e.g., service engines) on member devices in a system. For example, each of the devices 102-106 includes local agents that cooperate with an orchestrator to carry out various functions such as uploading, downloading, instantiating applications, health monitoring, and the like. In some instances, the orchestrator is a master that instructs agent-slaves to carry out the functions.
Optionally, platform 100 includes a central storage 180. In various embodiments, central storage 180 is configured to store information collected from one or more devices via communication channels 115, 125, and 135. The central storage 180 is accessible by controller 190 to determine health of components in the physical device(s) based on the collected information, determine overall network health, and the like. For example, the health state of the servers is published to the central storage, which published information is accessible by the service engines and the controller. For example, the shared database may be implemented by a Redis' system. Details of how storage 180 and controller 190 interact to store and/or collect health monitor data and determine a health condition based on the collected data are described herein in
In operation, controller 190 selects one or more of the service engines 114, 124, 134 to perform health monitoring of VMs/containers S(1, 1), S(1, 2), S(1, 3), S(2, 1), S(2, 2), S(2, 3), S(3, 1), S(3, 2), and/or S(3, 3), where the number of service engines selected to perform health monitoring is fewer than all of the service engines. For example, controller 190 selects SE1 and SE2 to monitor container S(1, 1). The selected service engines, SE1 and SE2 collect and/or process health data and report this information back to controller 190. In some embodiments, the reported information is stored in storage 180 and other service engines may access storage 180 to obtain this health information. In some embodiments, the reported information is sent to controller 190, and the controller 190 determines what information to make available to other service engines. Based at least in part on the information provided by SE1 and SE2, controller 190 determines a health condition of at least a portion of the clustered environment. The controller 190 then outputs the determined health condition. Details of how the controller and the service engine interact to select service engines to perform health monitoring, collect health monitor data, and determine a health condition based on the collected data are described herein in
The modules described above can be implemented as software components executing on one or more processors, as hardware components such as programmable logic devices (e.g., microprocessors, field-programmable gate arrays (FPGAs), digital signal processors (DSPs), etc.), Application Specific Integrated Circuits (ASICs) designed to perform certain functions, or a combination thereof. In some embodiments, the modules can be embodied by a form of software products which can be stored in a nonvolatile storage medium (such as optical disk, flash storage device, mobile hard disk, etc.), including a number of instructions for making a computer device (such as personal computers, servers, network equipment, etc.) implement the methods described in the embodiments of the present application. The modules may be implemented on a single device or distributed across multiple devices. The functions of the modules may be merged into one another or further split into multiple sub-modules.
Processor 202 is coupled bi-directionally with memory 210, which can include a first primary storage, typically a random access memory (RAM), and a second primary storage area, typically a read-only memory (ROM). As is well known in the art, primary storage can be used as a general storage area and as scratch-pad memory, and can also be used to store input data and processed data. Primary storage can also store programming instructions and data, in the form of data objects and text objects, in addition to other data and instructions for processes operating on processor 202. Also as is well known in the art, primary storage typically includes basic operating instructions, program code, data and objects used by the processor 202 to perform its functions (e.g., programmed instructions). For example, memory 210 can include any suitable computer-readable storage media, described below, depending on whether, for example, data access needs to be bi-directional or uni-directional. For example, processor 202 can also directly and very rapidly retrieve and store frequently needed data in a cache memory (not shown).
A removable mass storage device 212 provides additional data storage capacity for the computer system 200, and is coupled either bi-directionally (read/write) or uni-directionally (read only) to processor 202. For example, storage 212 can also include computer-readable media such as magnetic tape, flash memory, PC-CARDS, portable mass storage devices, holographic storage devices, and other storage devices. A fixed mass storage 220 can also, for example, provide additional data storage capacity. The most common example of mass storage 220 is a hard disk drive. Mass storage 212, 220 generally store additional programming instructions, data, and the like that typically are not in active use by the processor 202. It will be appreciated that the information retained within mass storage 212 and 220 can be incorporated, if needed, in standard fashion as part of memory 210 (e.g., RAM) as virtual memory.
In addition to providing processor 202 access to storage subsystems, bus 214 can also be used to provide access to other subsystems and devices. As shown, these can include a display monitor 218, a network interface 216, a keyboard 204, and a pointing device 206, as well as an auxiliary input/output device interface, a sound card, speakers, and other subsystems as needed. For example, the pointing device 206 can be a mouse, stylus, track ball, or tablet, and is useful for interacting with a graphical user interface.
The network interface 216 allows processor 202 to be coupled to another computer, computer network, or telecommunications network using a network connection as shown. For example, through the network interface 216, the processor 202 can receive information (e.g., data objects or program instructions) from another network or output information to another network in the course of performing method/process steps. Information, often represented as a sequence of instructions to be executed on a processor, can be received from and outputted to another network. An interface card or similar device and appropriate software implemented by (e.g., executed/performed on) processor 202 can be used to connect the computer system 200 to an external network and transfer data according to standard protocols. For example, various process embodiments disclosed herein can be executed on processor 202, or can be performed across a network such as the Internet, intranet networks, or local area networks, in conjunction with a remote processor that shares a portion of the processing. Additional mass storage devices (not shown) can also be connected to processor 202 through network interface 216.
An auxiliary I/O device interface (not shown) can be used in conjunction with computer system 200. The auxiliary I/O device interface can include general and customized interfaces that allow the processor 202 to send and, more typically, receive data from other devices such as microphones, touch-sensitive displays, transducer card readers, tape readers, voice or handwriting recognizers, biometrics readers, cameras, portable mass storage devices, and other computers.
In addition, various embodiments disclosed herein further relate to computer storage products with a computer readable medium that includes program code for performing various computer-implemented operations. The computer-readable medium is any data storage device that can store data which can thereafter be read by a computer system. Examples of computer-readable media include, but are not limited to, all the media mentioned above: magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROM disks; magneto-optical media such as optical disks; and specially configured hardware devices such as application-specific integrated circuits (ASICs), programmable logic devices (PLDs), and ROM and RAM devices. Examples of program code include both machine code, as produced, for example, by a compiler, or files containing higher level code (e.g., script) that can be executed using an interpreter.
The computer system shown in
At 302, one or more service engines are selected to perform health monitoring of an application. The application may be implemented by an application server executing VMs, container, etc. The service engine(s) selected to perform health monitoring can be strategically determined to reduce the load on the applications being monitored compared with conventional monitoring techniques. For example, process 300 spreads the load of health monitoring of a large number of application servers across agents (e.g., service engines) running in each physical node. In various embodiments, the service engine in the same node as the application for which health monitoring is desired is selected. Service engines in other nodes may also be selected. The number of selected service engines can be relatively small. For example, the number of service engines selected to perform health monitoring is less than the total number of service engines in the system environment. In various embodiments, the selection of the subset of service engines on the other nodes ensures that network connectivity is checked in a sampled manner from several nodes such that the report is both accurate and generated in a computationally efficient manner. In various embodiments, the set of nodes selected to perform health monitoring varies based on the service (e.g., a load balancing function such as a virtual service). For example, suppose 10 applications are deployed in a system of 100 service engines in which one service engine is local to an application and one service engine remote from the application is chosen. Using the techniques described here, only up to a threshold number (e.g., 20) service engines performs a health-check.
The selection of service engines can be based on one or more of the following factors: load, coverage, a probabilistic function, timing, and/or other factors. The factors further described herein address scalability challenges in monitoring the health of a large number of application servers. Applications may become overwhelmed when numerous service engines check the applications if the applications were required to respond individually to each request by the service engine for the health status. Here, the number and specific ones of service engines designated to perform health monitoring are selected such that the load of health monitoring is distributed among several nodes and application servers are not overwhelmed.
In various embodiments, the selection of a service engine is based on a load of the service engine. For example, a load-based decision compares a load on a service engine with a load threshold. If the load meets the threshold, the service engine is selected to perform monitoring. For example, a relatively lightly-loaded service engine (e.g., as measured against a threshold) is selected to perform monitoring. The load of the service engine used for comparison may be a current load, a load during a pre-defined time period, historical load, average load over a time period, and the like.
In another embodiment, the selection of a service engine is based on coverage of the service engine. For example, a coverage-based decision considers how many application servers the service engine is currently monitoring. If the service engine is able to accommodate monitoring of additional applications without suffering from performance degradation (e.g., the number of applications currently being monitored is less than a threshold number of applications, etc.), the service engine is assigned to monitor one or more additional applications.
In another embodiment, the selection of a service engine is based on a probabilistic function. For example, each member node (e.g., service engine) calculates a probabilistic distribution of health monitoring. The probabilistic determination may be made by a service engine about whether it will perform monitoring. The probabilistic function predicts those service engines that have more capacity to perform health monitoring at a given time and select the service engine accordingly.
A variety of probabilistic functions may be used. For example, the probability that a particular service engine is to monitor an application can be selected to achieve a desired system wide characteristic (SWC). Suppose n is a number of service engines (“SE”), p is an independent probability of health monitoring an application on a service engine, and q=1−p, i.e., a probability of not health monitoring an application on a SE.
In some embodiments, the probability p can be selected so as to arrive at different System Wide Characteristics (i.e., SWC) across all SEs as follows. SWC1, i.e., a probability of at least one service engine performing health monitoring=1−qn. SWC2, i.e., a probability of k service engines performing health monitoring=f(k)=nCk*pk*q(n-k). SWC3, i.e., a probability of at least m service engines performing health monitoring
SWC4, i.e., a probability of at least m1 service engines and at most m2 service engines performing health monitoring
The value of p, i.e., an independent probability of health monitoring an application on a service, can be determined by comparing an SWC to a probability threshold T and solving for p to a desired precision. A service engine can independently and probabilistically determine whether it needs to health monitor a service by generating a number x between 0 and 1 and health monitor the application if x is less than p.
In some embodiments, a service engine can repeatedly and independently (with respect to other service engines) determine whether to health monitor based on a computed probability p. In one aspect, while the set of service engines that health monitor can continuously vary over any period of time, the overall probability of requisite number of service engines health monitoring remains constant and a desired SWC is maintained.
In some embodiments, the system can transition from one SWC to another by instructing the service engines to re-compute p and use the newly determined p to independently determine whether to health monitor. For example, the SWC can be changed from SWC1 to SWC4 if the service being monitored is anticipated to undergo a period of instability.
At 304, health monitor data pertaining to the application is collected from the selected service engines. The health monitor data may be collected using existing networking APIs (such as APIs provided by the operating system). The operating system associated with a particular service engine implements the networking API calls (e.g., API calls to send data to or receive data from a specific socket at an IP address). For example, the health monitor data includes operational data (e.g., numbers of open connections, number of requests made to servers, etc.). In various embodiments, the health monitor data may be processed and/or filtered after collection prior to storage or transmission. In various embodiments, health monitor data is collected at pre-defined time intervals, e.g., periodic or non-periodic.
Conventionally, all service engines monitor applications simultaneously, thus duplicating work and increasing processing load. For example, suppose there are i service engines in a system and the monitoring interval is every j seconds. In this example, a service engine queries an application server being monitored every j seconds to gather health data. Typically, in conventional systems, each of the i service engines monitors the health of the same application server every j seconds. Effectively, there are i sets of health monitoring results generated every j seconds for the monitored application, the i sets often being duplicative of one another.
According to techniques of the present disclosure, monitoring is performed in a staggered fashion to minimize duplicative monitoring or optimize monitoring. In one aspect, this decreases processing load while achieving accuracy compared with conventional methods. Using the example of i service engines in a system and the monitoring interval is every j seconds, in various embodiments, each of the i service engines is assigned a random start time and each of the service engines performs monitoring every j seconds. The random start time defines when the service engine is to begin monitoring. In various embodiments, the randomization of the start times effectively staggers monitoring between service engines, which reduces duplicative monitoring.
In various embodiments, the start time for a particular service engine is determined systematically. For example, a service engine is selected periodically or after a threshold time has elapsed since the service engine was previously selected. The service engines may be selected in a round-robin fashion based on time. For example, a service engine may be assigned to perform monitoring at predetermined intervals, e.g., every second or fraction of a second. In one aspect, an application server is not overwhelmed by a large number of service engines performing health checks under this scheme because the service engines take turns performing the health checks based on time. In another aspect, time-based selection of service engine(s) ensures a more thorough connectivity check compared with relying on a single service engine to perform health monitoring. In various embodiments, these techniques provide more thorough coverage of networking paths compared with typical conventional health monitoring. In various embodiments, the collected health monitor data is stored in a shared database such as storage 180 shown in
At 306, a health condition is determined based on the collected health monitor data. The health condition may represent an operational status, state, and/or characteristics of a system such as clustered environment. The health condition indicates the health of at least a portion of the application servers. In various embodiments, the health condition is determined based on an aggregation of one or more health states of various application servers, where the aggregation may be based on data stored in a central database. For example, the data may be pushed to a central controller such as controller 190 shown in
In some embodiments, the state of the nodes is monitored and used to make subsequent determinations of which service engine(s) is to perform health monitoring such that the health of a specific application server is always checked by at least a few nodes. The health condition also finds application in determining how to auto-scale servers. For example, as the need (e.g., number of connections, amount of computation, etc.) increases, more servers are spawned to meet the need.
The health condition may be determined according to one or more policies. For example, a policy may include at least one rule for determining the health condition based on inputs. For example, the policy may be applied to received data to determine a health condition. The health condition may be a qualitative evaluation, e.g., “good,” “fair,” and “poor.” The health condition may be a quantitative evaluation, e.g., a scale from 0 to 100. By way of non-limiting example, a policy indicates that the health condition of a system is “good” if: all of the hosts are functioning, at least one host is functioning, the local service engine determines that the host is functioning and at least one other service engine determines that the host is functioning. In this example, the policy indicates that the health condition of a system is “fair” if at least one host is functioning, but fewer than a threshold of other hosts are not functioning (e.g., a percentage of total number of hosts such as 50%). The policy may indicate that the health of a condition of a system is “poor” if fewer than a threshold number of hosts are functioning (e.g., less than 50% of hosts are functioning). Other policies may specify different criteria for what is considered “good,” “fair,” or “poor.” For example, where one policy considers at least one host being functional to be “good,” another policy considers at least 75% of hosts being functional to be “good.”
In various embodiments, a more specific failure scenario may be given. For example, if a local service engine determines that the health of an application is good and a neighbor service engine determines that the health of an application is bad, this may indicate a network problem between the second service engine and the application. As another example, a delay in a response or a bad response from an application indicates misconfiguration of the application or a health problem in the system. In some embodiments, the policies are pre-determined. In some embodiments, the policies are selectable by an administrator.
At 308, the health condition is output. In various embodiments, the output includes transmitting the output to one or more other service engines, broadcasting the output, etc. In various embodiments, the output includes storing the health condition in a shared database accessible by one or more service engines. For example, the health condition is output to storage, output to a monitoring function or application to be rendered on a user interface, etc.
At 402, configuration information pertaining to a subset of application instances is received. Configuration information may include instructions regarding how and/or when to perform health monitoring. The configuration information can be specified by the manufacturer or a system administrator in a configuration file. The service engine uses this configuration information to perform health monitoring. The configuration information may be specified according to load, coverage, probabilistic function, timing, and/or other factors. For example, a relatively lightly-loaded service engine may receive configuration information instructing the service engine to perform more frequent monitoring of a particular application instance compared with a relatively heavily-loaded service engine. The selection of service engines, which forms a basis for the configuration information in various embodiments, is further described herein with respect to
At 404, health data pertaining to the subset of application instances is obtained. The health data may include information about an operational status, state, and/or characteristics of the subset of application instances being monitored (e.g., number of connections, response time from an application, response information from the application, etc.). In various embodiments, the health data is obtained as pre-defined time intervals. The pre-define time intervals may be provided with the configuration information. For example, a service engine is instructed to obtain health data beginning at a start time. The start time assigned to one service engine may be staggered with respect to another service engine. Based on the received instructions, the service engine obtains health data at the defined times.
At 406, the health data to be shared by a plurality of hosts is output. In some embodiments, the health data is output to a shared and/or distributed database such as database 180 shown in
System 500 includes four instances of an application, each instance provided in a respective one of Host 1, Host 2, Host 3, and Host 4. Instance App(1,1) is provided in host 1, instance App(2,1) is provided in host 2, instance App(3,1) is provided in host 3, and instance app (4,1) is provided in host 4. In various embodiments, each application instance is provided in a particular host having an associated virtual service. For example, instance App(1,1) is associated with service engine SE1 because they share a host. The virtual service services the respective application instance (also referred to as app). For example, each service engine (SE) performs health-monitoring for the app co-located in the same host, as well as one or more apps in at least one other host. The selection of which apps are monitored by a particular service engine may be performed according to the process described herein with respect to
The conceptual diagram shown in each of
Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, the invention is not limited to the details provided. There are many alternative ways of implementing the invention. The disclosed embodiments are illustrative and not restrictive.
This application claims priority to U.S. Provisional Patent Application No. 62/314,285 entitled SCALABLE HEALTH MONITORING IN EAST WEST DEPLOYMENTS IN LARGE SCALE CLUSTERED ENVIRONMENTS filed Mar. 28, 2016 which is incorporated herein by reference for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
5109486 | Seymour | Apr 1992 | A |
5781703 | Desai et al. | Jul 1998 | A |
6148335 | Haggard et al. | Nov 2000 | A |
6449739 | Landan | Sep 2002 | B1 |
6515968 | Combar et al. | Feb 2003 | B1 |
6714979 | Brandt et al. | Mar 2004 | B1 |
6792458 | Muret et al. | Sep 2004 | B1 |
6792460 | Oulu et al. | Sep 2004 | B2 |
6901051 | Hou et al. | May 2005 | B1 |
6996778 | Rajarajan et al. | Feb 2006 | B2 |
7076695 | McGee et al. | Jul 2006 | B2 |
7130812 | Iyer et al. | Oct 2006 | B1 |
7246159 | Aggarwal et al. | Jul 2007 | B2 |
7353272 | Robertson et al. | Apr 2008 | B2 |
7430610 | Pace et al. | Sep 2008 | B2 |
7636708 | Garcea et al. | Dec 2009 | B2 |
7933988 | Nasuto et al. | Apr 2011 | B2 |
7990847 | Leroy et al. | Aug 2011 | B1 |
8032896 | Li et al. | Oct 2011 | B1 |
8112471 | Wei et al. | Feb 2012 | B2 |
8588069 | Todd et al. | Nov 2013 | B2 |
8856797 | Siddiqui et al. | Oct 2014 | B1 |
8874725 | Ganjam et al. | Oct 2014 | B1 |
9032078 | Beerse et al. | May 2015 | B2 |
9210056 | Choudhary et al. | Dec 2015 | B1 |
9288193 | Gryb et al. | Mar 2016 | B1 |
9459980 | Arguelles | Oct 2016 | B1 |
9467476 | Shieh et al. | Oct 2016 | B1 |
9477784 | Bhave et al. | Oct 2016 | B1 |
9483286 | Basavaiah | Nov 2016 | B2 |
9495222 | Jackson | Nov 2016 | B1 |
9531614 | Nataraj et al. | Dec 2016 | B1 |
9571516 | Curcic et al. | Feb 2017 | B1 |
9608880 | Goodall | Mar 2017 | B1 |
9613120 | Kharatishvili et al. | Apr 2017 | B1 |
9626275 | Hitchcock et al. | Apr 2017 | B1 |
9674302 | Khalid et al. | Jun 2017 | B1 |
9680699 | Cohen et al. | Jun 2017 | B2 |
9692811 | Tajuddin et al. | Jun 2017 | B1 |
9697316 | Taylor et al. | Jul 2017 | B1 |
9712410 | Char et al. | Jul 2017 | B1 |
9716617 | Ahuja et al. | Jul 2017 | B1 |
9729414 | Oliveira et al. | Aug 2017 | B1 |
9749888 | Colwell et al. | Aug 2017 | B1 |
9830192 | Crouchman et al. | Nov 2017 | B1 |
9935829 | Miller et al. | Apr 2018 | B1 |
10003550 | Babcock et al. | Jun 2018 | B1 |
10212041 | Rastogi et al. | Feb 2019 | B1 |
10313211 | Rastogi et al. | Jun 2019 | B1 |
10372600 | Mathur | Aug 2019 | B2 |
10547521 | Roy et al. | Jan 2020 | B1 |
10594562 | Rastogi et al. | Mar 2020 | B1 |
10728121 | Chitalia et al. | Jul 2020 | B1 |
20020078150 | Thompson et al. | Jun 2002 | A1 |
20020198984 | Goldstein et al. | Dec 2002 | A1 |
20020198985 | Fraenkel et al. | Dec 2002 | A1 |
20030191837 | Chen | Oct 2003 | A1 |
20030236877 | Allan | Dec 2003 | A1 |
20040054680 | Kelley et al. | Mar 2004 | A1 |
20040064552 | Chong et al. | Apr 2004 | A1 |
20040103186 | Casati et al. | May 2004 | A1 |
20040243607 | Tummalapalli | Dec 2004 | A1 |
20050010578 | Doshi | Jan 2005 | A1 |
20050060574 | Klotz et al. | Mar 2005 | A1 |
20050108444 | Flauaus et al. | May 2005 | A1 |
20050120160 | Plouffe et al. | Jun 2005 | A1 |
20050172018 | Devine et al. | Aug 2005 | A1 |
20050188221 | Motsinger et al. | Aug 2005 | A1 |
20060167939 | Seidman | Jul 2006 | A1 |
20060242282 | Mullarkey | Oct 2006 | A1 |
20060271677 | Mercier | Nov 2006 | A1 |
20070226554 | Greaves et al. | Sep 2007 | A1 |
20080104230 | Nasuto et al. | May 2008 | A1 |
20080126534 | Mueller et al. | May 2008 | A1 |
20090154366 | Rossi | Jun 2009 | A1 |
20090199196 | Peracha | Aug 2009 | A1 |
20100279622 | Shuman et al. | Nov 2010 | A1 |
20110126111 | Gill et al. | May 2011 | A1 |
20110196890 | Pfeifle et al. | Aug 2011 | A1 |
20120101800 | Miao et al. | Apr 2012 | A1 |
20120110185 | Ganesan et al. | May 2012 | A1 |
20120254443 | Ueda | Oct 2012 | A1 |
20130013953 | Eck | Jan 2013 | A1 |
20130086230 | Guerra | Apr 2013 | A1 |
20130086273 | Wray et al. | Apr 2013 | A1 |
20130179289 | Calder et al. | Jul 2013 | A1 |
20130179881 | Calder et al. | Jul 2013 | A1 |
20130179894 | Calder et al. | Jul 2013 | A1 |
20130179895 | Calder et al. | Jul 2013 | A1 |
20130211559 | Lawson et al. | Aug 2013 | A1 |
20130212257 | Murase et al. | Aug 2013 | A1 |
20130343213 | Reynolds et al. | Dec 2013 | A1 |
20130346594 | Banerjee et al. | Dec 2013 | A1 |
20140006862 | Jain et al. | Jan 2014 | A1 |
20140143406 | Malhotra et al. | May 2014 | A1 |
20140173675 | Ahmed et al. | Jun 2014 | A1 |
20140215058 | Vicat-Blanc et al. | Jul 2014 | A1 |
20140215621 | Xaypanya et al. | Jul 2014 | A1 |
20140229706 | Kuesel et al. | Aug 2014 | A1 |
20140282160 | Zarpas | Sep 2014 | A1 |
20140304414 | Yengalasetti et al. | Oct 2014 | A1 |
20140344439 | Kempf et al. | Nov 2014 | A1 |
20150058265 | Padala et al. | Feb 2015 | A1 |
20150074679 | Fenoglio et al. | Mar 2015 | A1 |
20150081880 | Eaton et al. | Mar 2015 | A1 |
20150124640 | Chu et al. | May 2015 | A1 |
20150199219 | Kim et al. | Jul 2015 | A1 |
20150212829 | Kupershtok et al. | Jul 2015 | A1 |
20150288682 | Bisroev et al. | Oct 2015 | A1 |
20150293954 | Hsiao et al. | Oct 2015 | A1 |
20150295780 | Hsiao et al. | Oct 2015 | A1 |
20150295796 | Hsiao et al. | Oct 2015 | A1 |
20150358391 | Moon et al. | Dec 2015 | A1 |
20150370852 | Shastry et al. | Dec 2015 | A1 |
20160064277 | Park et al. | Mar 2016 | A1 |
20160094431 | Hall et al. | Mar 2016 | A1 |
20160094483 | Johnston et al. | Mar 2016 | A1 |
20160105335 | Choudhary et al. | Apr 2016 | A1 |
20160127204 | Ozaki et al. | May 2016 | A1 |
20160164738 | Pinski et al. | Jun 2016 | A1 |
20160182399 | Zadka et al. | Jun 2016 | A1 |
20160217022 | Velipasaoglu et al. | Jul 2016 | A1 |
20160294722 | Bhatia et al. | Oct 2016 | A1 |
20160323377 | Einkauf et al. | Nov 2016 | A1 |
20170041386 | Bhat et al. | Feb 2017 | A1 |
20170063933 | Shieh et al. | Mar 2017 | A1 |
20170093986 | Kim et al. | Mar 2017 | A1 |
20170134481 | DeCusatis et al. | May 2017 | A1 |
20170331907 | Jagannath et al. | Nov 2017 | A1 |
20180004582 | Hallenstål | Jan 2018 | A1 |
20180018244 | Yoshimura et al. | Jan 2018 | A1 |
20180041408 | Dagum et al. | Feb 2018 | A1 |
20180041470 | Schultz et al. | Feb 2018 | A1 |
20180046482 | Karve et al. | Feb 2018 | A1 |
20180088935 | Church et al. | Mar 2018 | A1 |
20180089328 | Bath et al. | Mar 2018 | A1 |
20180136931 | Hendrich et al. | May 2018 | A1 |
20180287902 | Chitalia et al. | Oct 2018 | A1 |
20180309637 | Gill et al. | Oct 2018 | A1 |
20180335946 | Wu et al. | Nov 2018 | A1 |
20190123970 | Rastogi et al. | Apr 2019 | A1 |
Entry |
---|
Author Unknown, “Autoscaler,” Compute Engine—Google Cloud Platform, Jun. 29, 2015, 6 pages, retrieved at http://web.archive.org/web/20150629041026/https://cloud.google.com/compute/docs/autoscaler/. |
Author Unknown, “Autoscaling,” Aug. 20, 2015, 4 pages, Amazon Web Services, retrieved from http://web.archive.org/web/20150820193921/https://aws.amazon.com/autoscaling/. |
Catania, V., et al., “PMT: A Tool to Monitor Performances in Distributed Systems,” Proceedings of the 3rd IEEE International Symposium on High Performance Distributed Computing, Aug. 2-5, 1994, 8 pages, San Francisco, CA, USA. |
Davis, David, “Post #8—Understanding vCenter Operations Badges,” David Davis Blog, Apr. 29, 2014, 5 pages, retrieved from http://blogs.vmware.com/management/2014/04/david-davis-on-vcenter-operations-post-8-understanding-vcenter-operations-badges.html. |
De George, Andy, “How to Scale an Application,” Jun. 16, 2015, 8 pages, Github.com. |
Liu, Feng, et al., “Monitoring of Grid Performance Based-on Agent,” 2007 2nd International Conference on Pervasive Computing and Applications, Jul. 26-27, 2007, 6 pages, IEEE, Birmingham, UK. |
Non-Published commonly owned U.S. Appl. No. 15/055,450, filed Feb. 26, 2016, 37 pages, Avi Networks. |
Non-Published commonly Owned U.S. Appl. No. 15/130,499, filed Apr. 15, 2016, 49 pages, Avi Networks. |
Sevcik, Peter, et al., “Apdex Alliance,” May 24, 2014, 5 pages, www.apdex.org. |
Wallace, Paul, et al., “Feature Brief: Stingray's Autoscaling Capability,” Brocade Community Forums, May 1, 2013, 5 pages, retrieved from http://community.brocade.com/t5/vADC-Docs/Feature-Brief-Stingray-s-Autoscaling-capability/ta-p/73843. |
Yar, Mohammed, et al., “Prediction Intervals for the Holt-Winters Forecasting Procedure,” International Journal of Forecasting, Month Unknown 1990, 11 pages, vol. 6, Issue 1, Elsevier Science Publishers B.V. |
Zhang, Xuehai, et al., “A Performance Study of Monitoring and Information Services for Distributed Systems,” Proceedings of the 12th IEEE International Symposium on High Performance Distributed Computing, Jun. 22-24, 2003, 12 pages, IEEE Computer Society, Washington, D.C., USA. |
Non-Published commonly Owned U.S. Appl. No. 16/817,604, filed Mar. 12, 2020, 52 pages, VMware, Inc. |
Number | Date | Country | |
---|---|---|---|
62314285 | Mar 2016 | US |