This invention relates to the field of systems analysis and planning, and in particular to server consolidation and virtualization through performance analysis, modeling, and prediction.
The increase in processing power and storage available in computer devices and systems has initiated a resurgence in ‘central processing systems’, where multiple tasks are serviced on a number of servers, each server being time-shared among different tasks. Simple time-sharing, however, can lead to interaction problems, particularly if one of the time-shared tasks exhibits abnormal behavior and interferes with the other tasks, or causes the computer system to shut down.
The concept of “virtual machines (VMs)” was pioneered in the mid-1960s by IBM and MIT. A ‘main-frame computer’ was partitioned into individual ‘virtual machines’, each virtual machine appearing to each user, or group of users, as an independent entity, and each virtual machine being protected from tasks running on other virtual machines. As the concept has progressed, the term “virtualization” is generally used to describe the dividing of resources of a computer into multiple execution environments, using such techniques as hardware and software partitioning, time-sharing, machine emulation, and so on.
As computers become increasingly more capable, the concept of sharing resources among multiple users is likely to increase, particularly in computing-intensive environments, such as universities, government organization, mid to large corporations, and so on. To be effective, such environments require effective management of the shared resources, to balance workloads and/or assure sufficient performance to particular tasks.
In a typical multi-computer shared environment, the management of virtualized resources generally requires satisfying particular performance expectations. In some instances, a specified service level is guaranteed in a ‘service level agreement’ (SLA), which provides at least a minimum level of service, and either limits the service to a specified performance level or provides as much performance above the specified limit as is available at the time; in other instances, all users are given equal priority, and are promised as much performance as is available at the time, without any guarantees, but with the hope of being provided an average performance that substantially exceeds an estimated minimal performance. In other situations, users acquire a particular number of ‘shares’ of the available resources, and a weighted priority scheme is used, allocating resources based on the number of acquired shares, the weighted allocation being based on either the total number of shares allocated, or the total number of shares that are currently allocated to active tasks. Such allocation rules and methods are generally referred to as ‘priority schemes’.
The effective and efficient management of a virtual system requires an effective and efficient allocation of virtual machines to actual computers to optimize the processing resources that are provided to the various users of the virtual machines and/or to optimize user satisfaction, based on each user's acquired share of the available resources. As used herein, ‘acquired share’ refers to the relative share of resources that are expected to be provided to the particular user, and may be based on a purchase of such shares, an allocation of shares based on the particular task being performed, an allocation of shares based on a status of the particular user or particular task, and/or any other scheme for prioritizing each user's allocation of available resources.
This allocation of resources, even among equal shareholders, however, is not a simple linear allocation. If two virtual machines are running on a computer, the allocation of processing to the virtual machines will not be equal to the total processing capabilities of the computer, because some processing will be consumed by the application (“virtualization technology”) that is providing the segregation of the available processing between the two tasks. As the number of virtual machines on the computer increases, the loss of available processing due to the overhead associated with providing the virtualization scheme increases, in some cases resulting in an overhead loss of 15% or more of the processing available to the virtual machines.
The aforementioned overhead loss is not solely dependent upon the number of virtual machines running on a particular computer. Other factors, such as the technology used to provide the virtualization, the utilization of resources by each task on each virtual machine, the number of physical processors on the computer, the priority associated with each task, and so on, will affect the efficiency with which the virtualization process will perform.
A variety of competing virtualization technologies are currently available, including, for example, Hyper-threaded Processor, VMware ESX, AIX Micropartition, Microsoft Virtual Machine Technology, Sun N1, HP nPar, vPar, and IBM PR/SM, each potentially producing different overhead losses under different conditions, and as the use of virtualization increases, the number of available virtualization technologies can be expected to increase. As with most competing technologies, each virtualization technology appears to have certain advantages and disadvantages, and no single virtualization technology can be said to be ideal for any and all combinations of task allocations. Selecting the appropriate virtualization technology for a given environment, or selecting the appropriate set of virtualization technologies to satisfy different groups of tasks in a given environment, is becoming increasingly difficult for managers of large computing facilities.
Satisfying the requirements for effective and efficient virtualization allocation is further complicated by the dynamic nature of most computing environments. With advances in computing technology, it is rare for a computing environment to remain static for even a year. As new computing systems are introduced and older computing systems removed from service, the allocation of resources in a virtual environment becomes increasingly complicated.
It can be expected, for example, that as virtualization becomes more prevalent, the design of future high-performance computing systems will include features targeted to virtual systems, generally to reduce the virtualization overhead for all configurations, or to minimize the virtualization overhead within a particular band of configurations. In like manner, it can be expected that current virtualization technologies will advance, and new virtualization technologies will be introduced. As can correspondingly be expected, the reallocation of virtual machines among computing systems, as existing systems are replaced by other systems and configurations having different virtualization performance characteristics, will often be a very daunting task.
It would be advantageous to provide a virtualization management system that facilitates the allocation of virtual machines among multiple computer systems. It would also be advantageous to estimate the overhead losses associated with a given allocation of virtual machines among multiple computer systems.
These advantages, and others, can be realized by an interactive virtualization management system that provides an assessment of proposed or existing virtualization schemes. A Virtual Technology Overhead Profile (VTOP) is created for each of a variety of configurations of host computer systems and virtualization technologies by measuring the overhead experienced under a variety of conditions. The multi-variate overhead profile corresponding to each target configuration being evaluated is used by the virtualization management system to determine the overhead that is to be expected on the target system, based on the particular set of conditions at the target system. Based on these overhead estimates, and the parameters of the jobs assigned to each virtual machine on each target system, the resultant overall performance of the target system for meeting the performance criteria of each of the jobs in each virtual machine is determined, and over-committed virtual machines and computer systems are identified.
The invention is explained in further detail, and by way of example, with reference to the accompanying drawings wherein:
Throughout the drawings, the same reference numerals indicate similar or corresponding features or functions. The drawings are included for illustrative purposes and are not intended to limit the scope of the invention.
In the following description, for purposes of explanation rather than limitation, specific details are set forth such as the particular architecture, interfaces, techniques, etc., in order to provide a thorough understanding of the concepts of the invention. However, it will be apparent to those skilled in the art that the present invention may be practiced in other embodiments, which depart from these specific details. In like manner, the text of this description is directed to the example embodiments as illustrated in the Figures, and is not intended to limit the claimed invention beyond the limits expressly included in the claims. For purposes of simplicity and clarity, detailed descriptions of well-known devices, circuits, and methods are omitted so as not to obscure the description of the present invention with unnecessary detail.
This invention addresses the performance aspect of an actual or planned virtualization embodiment/environment, and provides an assessment of the virtualization environment to assist users in providing an embodiment that effectively achieves performance and business goals. These goals may be explicit or assumed; for example, in many virtualization environments, services are provided in accordance with a Service Level Agreement (SLA), wherein the provider of the virtual machine services contracts to provide a minimum performance level.
When a virtual machine is created on a host computer, the performance of that virtual machine will be dependent upon a variety of factors. The primary factors affecting the performance of a virtual machine include, for example, the processing power of the host computer, the number of other virtual machines on the host, the total load/utilization placed on the host by the virtual machines, and so on. The dependency on the number and utilization of the virtual machines is often a function of the particular virtualization technology that is used to provide the virtualization services. Other factors will also affect the performance of virtual machines on a host computer, such as the priority scheme used for allocating the processing capabilities among the virtual machines, the configuration of the computer system in terms of the number of processors, the number of threads per processor, size of memory, and so on.
The load of a virtual machine is generally defined in terms of ‘utilization’. For a given computer system, the amount of resources (typically percentage of CPU time) used by a task in a non-shared environment defines its ‘native’ utilization. If a task is slowed down because of insufficient computer resources, the resources are being over-utilized. Preferably, sufficient resources should be provided so as never to delay the task; that is, in a preferred system, the resources should always be under-utilized. Because the utilization of a task generally varies during the execution of the task, avoiding the introduction of any delay during the execution of the task generally requires that the system should be configured to be substantially under-utilized relative to the overall/average utilization of a task, to avoid delays during peak utilization periods.
The difference between a job's performance/utilization on a single dedicated host and the job's performance on a nominally-equivalent virtual machine is generally defined as the “virtualization overhead”. This overhead arises due to the time required to switch physical resources between virtual machines, the control of the switching between machines, losses introduced by interrupting ongoing processes, and so on.
A number of performance assessment schemes have been proposed, generally using analytic models to capture the dependency of the virtual machine on each of the factors that are considered relevant to the achievable performance of a given virtualization technology. Such analytic models, however, are difficult to develop, and require extensive testing to determine their validity/accuracy. Additionally, when modifications are made to an existing virtualization technology, or a new virtualization technology is introduced, there will generally be a substantial delay before new analytic models can be created.
In accordance with this invention, empirical models, based on the performance of actual virtualization embodiments, are used to estimate the performance that can be expected to be achieved on other virtualization embodiments. A collection of performance measurements are determined for a given virtualization technology on a given host computer for a variety of conditions. These measures are used to determine the overhead introduced for each of the variety of conditions, thereby forming a multi-variate overhead profile for the given virtualization technology on the given host computer system. As new or modified virtualization technologies and computer systems are introduced, the performance of the new configuration is measured under the aforementioned variety of conditions, and a new overhead profile is created for this new combination of virtualization technology and host computer system.
The measurement of overhead for the variety of conditions/operational parameters form a locus of points in a multi-dimensional space, one of the dimensions corresponding to the measured overhead and each of the other dimensions corresponding to a varied operational parameter, such as the number of virtual machines, the total task (or job) utilization, the CPU speed, the number of cores per CPU, the number of threads per CPU, and so on. Assuming substantial continuity in each dimension, this locus of measured points defines a surface in the multi-dimensional space.
Using these multi-variate overhead profiles, a performance modeling system selects the profile corresponding to the virtualization technology and host computer system of the system being modeled (the ‘target’ system), then determines the point on the multi-dimensional surface corresponding to the overhead produced by the particular condition/set of operational parameters of the particular target system. Because this performance modeling system determines the overhead based on a multi-dimensional profile that is created by measuring the overhead of a particular system configuration under a variety of conditions, this performance modeling system can be used to determine the overhead on any system configuration for which a profile exists or can be created.
The creation of a test scenario for measuring overhead at different utilization levels can be facilitated by creating a test script that includes an input parameter that controls the number of CPU cycles consumed by the script, then executing this script for each of a desired amount of utilization, based on the computer system's speed.
As can be seen, the example overhead measurements of
Other segments of the overhead profile are created for other conditions 101 with the same host “Dell PE 1850” and same virtualization technology “VMware ESX”. For example, the measurements of
One would expect that if a target system under evaluation is operated under the same conditions as one of the measured points defining the overhead profile, its overhead would be similar to the previously measured overhead. For example, one would expect that a target system comprising a Dell PE 1850 running five virtual machines under VMware ESX at 1256 MHz with one core per CPU, one thread per core, and 30% utilization would experience about 5% overhead, as illustrated by the dashed lines in
In like manner, a target system comprising the same host computer and virtualization technology as illustrated in
As noted above, the overhead profile will generally exhibit a non-linear dependency on the condition values. As such, the aforementioned n-dimensional interpolation is preferably a non-linear interpolation, typically using a polynomial function. The degree of the polynomial will be dependent upon the observed degree of non-linearity and an error estimate associated with the determined coefficients of the polynomial. For example, if a curve fitting program provides coefficients of a second-degree polynomial with a lower error factor than a third-degree polynomial, the simpler second-degree polynomial would preferably be used.
In a preferred embodiment, in lieu of maintaining a record of all measured points in the n-dimensional space and performing interpolations between measured points in the vicinity of the target system's coordinate in the n-dimensional space, the overhead dependency on one or more of the dimensions can be represented by the coefficients of a polynomial. For example, if a particular condition is known to provide a substantially linear dependency, this dependency can be defined by the slope and intercept values of a linear equation (a 1-degree polynomial). If these coefficients vary as a function of the other condition values, different slope and intercept values can be defined for each of the other condition values, or a multi-variate polynomial can be defined to represent these dependencies.
In like manner, CPU speed is generally known to provide a determinable scaling of performance. In a preferred embodiment, multiple sets of measurements are not performed to measure the overhead at different CPU speeds. All measurements for a given overhead profile are performed at a fixed CPU speed, and the overhead on a system with a different CPU speed is determined as discussed above without regard to CPU speed, and then scaled appropriately. The appropriate scaling factor can be found using a variety of known techniques, or a variety of published results. For example, Standard Performance Evaluation Corporation (SPEC) is a non-profit corporation that establishes, maintains, and endorses a standardized set of benchmarks for high-performance computers, and provides validated measures that can be used to effect this scaling based on CPU speeds.
In this example, a user interface 240 is provided to enable the definition of an allocation of virtual machines to computer systems 250. This definition is based on the workload that needs to be supported 210 and the computer systems that are available to provide this support 220. The allocation will also typically be defined in view of performance criteria associated with each of the workloads 210, as the hosts to which jobs having significant performance criteria are allocated will generally be kept substantially underutilized, if possible.
As each job is added to the computer system, the total allocated utilization is increased by a corresponding amount of native utilization associated with the tasks of the job. If the native utilization of a task had been measured on a different computer system than the system to which it is being allocated, the native utilization is scaled appropriately, using the aforementioned standard scaling factors associated with high-performance computer systems.
In a preferred embodiment, jobs can be defined as a collection of tasks, as indicated by the “+” icon at each job; when the job is added to a virtual machine, the individual tasks are displayed. As each job is allocated to a virtual machine, its symbol in the jobs window is dimmed to indicate the allocation. After all of the jobs are allocated, the VM-Computer allocation (250 in
The VM-Computer allocation 250 identifies the configuration of each computer in terms of virtualization technology employed, and defines the operational parameters specific to the configured computer system. The allocation of virtual machines to the defined computer configuration defines the other operational parameters, including the number of virtual machines assigned and the total utilization by all of the defined jobs on these virtual machines.
A set of previously created overhead profiles (VTOPs—Virtual Technology Overhead Profile) 230, discussed above, are made available to a performance modeling engine 270 for determining the overhead associated with each of configured host computer systems defined in the VM-computer allocation 250. A VTOP selector 260 is configured to select an overhead profile corresponding to the configuration of each target computer from among the set of available VTOPs 230. In the example embodiment, wherein an overhead profile is stored for each measured computer system and virtual technology combination, the set of VTOPs 230 is searched to find a matching computer system and virtualization technology for each target computer system. If the overhead profiles are collected and stored using a different set of configuration parameters, to reduce, for example, the number of dimensions of the profile by removing discrete parameters for which interpolation is not performed, the selector 260 finds the overhead profile corresponding to these configuration parameters.
Having identified the appropriate overhead profile, the performance modeling engine 270 uses the values of the operational parameters (number of virtual machines, utilization, etc.) of each target computer system to determine the overhead that is to be expected on each target computer system operating with these parameter values. As discussed above, any of a variety of techniques can be used to determine the specific value of the overhead, including multi-dimensional interpolation and/or the use of equations determined by curve-fitting to the measured overhead values.
After determining the overhead incurred on each computer system, the impact of this overhead on the performance of each of the individual tasks is determined. The achievable utilization by each task is affected by the utilization demanded by all of the other competing tasks on the computer. The determined overhead can be considered to be the utilization attributed to a hypothetical overhead task that is also one of these other competing tasks that affect the task's achievable utilization.
As taught by Ethan Bolker and Yiping Ding, in “VIRTUAL PERFORMANCE WON'T DO: CAPACITY PLANNING FOR VITUAL SYSTEMS” [CMG Proceedings, V1, pp 39-49, December 2005], the achievable utilization by each task in an equal-priority scheme can be estimated as:
where Vi is the achievable utilization by task i, Ui is the native utilization of task i, and UT is the total utilization, including each task's native utilization and the determined overhead utilization. Appropriate modifications to this equation can be made for other priority schemes using conventional queuing theory models.
Other techniques for estimating the impact of overhead utilization on each device's performance, common in the art, may also be used. In like manner, the impact of the determined overhead on other measures of performance may be assessed in addition to or in lieu of the achievable utilization. For example, in the above referenced article of Bolker and Ding, techniques are provided for determining the impact of virtualization overhead on the response time of each task. The impact on other measures, such as creation or aggravation of bottlenecks, additional memory usage, and so on, can also be determined using techniques common in the art.
The determined overhead and resultant performance 280 is provided to a report generator 290 that provides a variety of reports via the user interface 240. Of particular significance, the report generator 290 compares the criteria 210 associated with each task, typically expressed in a Service Level Agreement (SLA), with the determined performance, and identifies any task that does not satisfy these criteria.
In a preferred embodiment, the user interface allows a user to investigate details associated with a given report. For example, in
The report generator 290 also provides composite printed reports that identify the allocations on each computer system, the determined overhead, the achievable utilization for each task, and so on.
In a preferred embodiment, the report generator 290 is also configured to assess the allocation to provide recommended changes, based on accepted ‘good practice’ techniques. In like manner, the system may be configured to automate all or some of the virtual machine allocations, using such good practice techniques. For example, an accepted good practice technique is to equalize the utilization allocated to each computer system. To automate the allocation, the workload demand and the total available computer system resources may be used to determine an average utilization per computer system, normalized for different computer systems' performance ratings, and tasks allocated to each computer system until the allocated utilization is approximately equal to the desired average.
Known allocation algorithms, such as the greedy algorithm, may also be used to allocate the jobs in order of utilization demand, each to a different computer system until all systems have one job, then the allocation continues by allocating each next-smaller job to the computer system having the least total allocated utilization. If the jobs have different priority ratings, the jobs are allocated in priority order, and the determination of the next-computer to receive the next-smaller job is based on a weighted measure of utilization, such that the allocation of a higher priority job increases this measure by a larger weight than a lower priority job.
As noted above, because the virtualization planning system of this invention determines virtualization overhead and resultant task performance based on sample measurements of a virtualization technology on particular computer configurations under a variety of conditions, its use is applicable for analyzing virtualization strategies on any given virtualization technology for which such sample measures are available, or can be obtained.
Because the principles of this invention can be applied to any virtualization technology, it would be advantageous for virtualization technology vendors, or third party vendors, to create overhead profiles for a variety of computer systems of the class commonly used to provide virtual services. In like manner, it would be advantageous for vendors of such computer systems to create overhead profiles for a variety of virtualization systems, as an incentive for purchasing their systems for large-scale computer complexes, particularly those with required capacity to meet the Service Level Agreement (SLA).
The foregoing merely illustrates the principles of the invention. It will thus be appreciated that those skilled in the art will be able to devise various arrangements which, although not explicitly described or shown herein, embody the principles of the invention and are thus within its spirit and scope. For example, in the above description, the actual native utilization is assumed to be known; however, any of a number of techniques may be used to determine/estimate native utilizations based on measured virtual utilizations, including the technique disclosed in the aforementioned article by Ethan Bolker and Yiping Ding.
Similarly, the above description discloses the use of multi-dimensional interpolation to determine/estimate the overhead from a given VTOP; however, any of a number of techniques may be used to determine/estimate the overhead based on measured overhead values.
For example, the measured overheads under the different conditions may be used to train a neural network for each configuration, wherein each set of conditions is provided as input and the corresponding measured output is provided as the desired output from the neural network. In this embodiment, each neural network corresponds to a VTOP for each configuration. After the VTOPs are trained using the measured overhead values for the variety of conditions for each configuration, the overhead of a target system can be determined/estimated by providing the conditions of the target system to the VTOP neural network corresponding to the configuration of the target system.
In like manner, although each VTOP is presented above as being discrete for each configuration, one of skill in the art will recognize that a plurality of VTOPs may be included in a single data structure that uses the particular configuration of the computer system as an input parameter for accessing the appropriate overhead values for a given configuration of the target system.
These and other system configuration and optimization features will be evident to one of ordinary skill in the art in view of this disclosure, and are included within the scope of the following claims.
In interpreting these claims, it should be understood that:
a) the word “comprising” does not exclude the presence of other elements or acts than those listed in a given claim;
b) the word “a” or “an” preceding an element does not exclude the presence of a plurality of such elements;
c) any reference signs in the claims do not limit their scope;
d) several “means” may be represented by the same item or hardware or software implemented structure or function;
e) each of the disclosed elements may be comprised of hardware portions (e.g., including discrete and integrated electronic circuitry), software portions (e.g., computer programming), and any combination thereof;
f) hardware portions may be comprised of one or both of analog and digital portions;
g) any of the disclosed devices or portions thereof may be combined together or separated into further portions unless specifically stated otherwise;
h) no specific sequence of acts is intended to be required unless specifically indicated; and
i) the term “plurality of” an element includes two or more of the claimed element, and does not imply any particular range of number of elements; that is, a plurality of elements can be as few as two elements, and can include an immeasurable number of elements.
This application is a continuation of U.S. patent application Ser. No. 12/196,044, entitled “Virtualization Planning System,” filed Aug. 21, 2008, which claims the benefit of U.S. Provisional Patent Application 60/968,025, filed 24 Aug. 2007, both of which are incorporated by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
7818145 | Wood et al. | Oct 2010 | B2 |
20080288941 | Adams et al. | Nov 2008 | A1 |
Entry |
---|
Wang, et al., “Capacity and Performance Overhead in Dynamic Resource Allocation to Virtual Containers”, Integrated Network Management, 2007; 10th IFIP/IEEE International Symposium, May 1, 2007. |
Desai, “Profiling and Load Distribution for Virtual Machines”, [online] Aug. 8, 2006. |
Salsberg, et al., “It May be Virtual—But the Overhead Is Not”, Measure It [online], Mar. 2007. |
Ajiro, et al., “Measuring Resource Management Overhead for server Virtualization”, IPSI SIG technical Report; Aug. 20, 2006. |
Bolker, et al., “Virtual Performance Won't Do: Capacity Planning for Virtual Systems”, CMG Proceedings, V1; Dec. 2005. |
Number | Date | Country | |
---|---|---|---|
20120324451 A1 | Dec 2012 | US |
Number | Date | Country | |
---|---|---|---|
60968025 | Aug 2007 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12196044 | Aug 2008 | US |
Child | 13595372 | US |