US9037717B2 - Virtual machine demand estimation - Google Patents

Virtual machine demand estimation Download PDF

Info

Publication number
US9037717B2
US9037717B2 US12/563,445 US56344509A US9037717B2 US 9037717 B2 US9037717 B2 US 9037717B2 US 56344509 A US56344509 A US 56344509A US 9037717 B2 US9037717 B2 US 9037717B2
Authority
US
United States
Prior art keywords
resource
entity
resources
demand
cpu
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US12/563,445
Other versions
US20110072138A1 (en
Inventor
Isci Canturk
James E. Hanson
Jeffrey O. Kephart
Malgorzata Steinder
Ian N. Whalley
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US12/563,445 priority Critical patent/US9037717B2/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: STEINDER, MALGORZATA, CANTURK, ISCI, HANSON, JAMES E., KEPHART, JEFFREY O., WHALLEY, IAN N.
Publication of US20110072138A1 publication Critical patent/US20110072138A1/en
Application granted granted Critical
Publication of US9037717B2 publication Critical patent/US9037717B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5011Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/50Indexing scheme relating to G06F9/50
    • G06F2209/504Resource capping
    • Y02B60/142
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Definitions

  • aspects of the present invention are directed to a method and a system for virtual machine CPU demand estimation for autonomous resource and power management.
  • a virtualized environment the physical hardware is decoupled from the operating system and application software via a hypervisor layer.
  • resource consolidation is a commonly applied technology that is enabled by computing hardware virtualization.
  • the hypervisor helps multiple virtual machines (VMs), which may run different operating systems on different virtual hardware, share the same physical resources.
  • VMs virtual machines
  • Some solutions aim to address the problem of identifying the resource requirements of a VM. These solutions rely on determining resource needs from within a virtual machine by inspecting application and operating system behavior within the VM to track certain indicative behavior such as detecting the occurrences of an idle loop within a virtual machine.
  • information about the VM CPU load is determined based on its interrupt response time. While this technique provides indirect measures on whether a VM is receiving its resource requirements, it is generally intrusive and does not provide a direct measure of a VM's actual resource demand.
  • a method for use in a system in which computational entities are distributed across physical computing resources to place the entities on the resources includes estimating actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity, computing a best allocation of the resources to the entities from the estimated actual resource demand for each entity and distributing the resources to the entities in accordance with the computed best allocation.
  • a method for estimating actual resource demand for running entities includes obtaining a map of the entities relative to resources, obtaining resource usage metrics for each entity and resource, obtaining resource capacities and computing the estimated actual resource demand based on the map of the entities relative to the resources, the resource usage metrics and the resource capacities.
  • a system in which computational entities are distributed across physical computing resources to place the entities on the resources.
  • the system includes a processor and a computer readable medium coupled to the processor having executable instructions stored thereon, which, when executed instruct the processor to estimate actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity, to compute a best allocation of the resources to the entities from the estimated actual resource demand for each entity and to distribute the resources to the entities in accordance with the computed best allocation.
  • FIG. 1 is a schematic diagram of a resource consolidated environment, and its operation without the employment of demand estimation as disclosed in this invention
  • FIG. 2 is a flowchart illustrating a method to place entities on resources in accordance with embodiments of the invention
  • FIG. 3 is a table illustrating an exemplary resource consolidated environment in which virtual machines (VMs) are consolidated on a single host;
  • VMs virtual machines
  • FIG. 4 is a flow chart illustrating an analytical derivation of VM CPU demand estimation
  • FIG. 5 is a schematic illustration of a virtualized environment in accordance with embodiments of the present invention.
  • FIG. 6 is a snapshot of evaluation results of demand estimation tests.
  • an exemplary system includes a virtualized cluster with 3 hosts, each with total CPU capacity of 100%, and 8 virtual machines (VMs).
  • Dynamic placement aims to balance resources for use by the VMs such that each host's utilization does not, for example, exceed 85%.
  • Dynamic placement also employs power management such that if the total resource demand by the VMs can be satisfied with fewer than 3 hosts, the remaining hosts are powered down. When the demand rises, the hosts are powered back on to satisfy the demand.
  • a dynamic resource manager uses resource usage as a measure of demand to respond in the absence of the demand estimation of the invention and an iterative process including 5 separate decisions that converges to the placement of resources.
  • desired resource demand of the VMs is captured initially and guides dynamic resource balancing actions in a single operation. That is, the resource management actions performed in steps T 2 -T 5 of FIG. 1 are realized in a single operation with demand estimation. This is achieved using scheduling statistics commonly available to operating systems and hypervisors to derive reliable estimates for actual VM resource demand.
  • a method for use in a system in which computational entities are distributed across physical computing resources to place the entities on the resources includes estimating actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity 200 , such as a hypervisor, computing a best allocation of the resources to the entities from the estimated actual resource demand for each entity 220 , and distributing the resources to the entities in accordance with the computed best allocation 230 .
  • the estimating of operation 200 may include obtaining a map of the entities relative to the resources 201 , obtaining resource usage metrics for each entity and resource 202 , obtaining resource capacities 203 and computing the estimated actual resource demand based on the map of the entities relative to the resources, the resource usage metrics and the resource capacities 210 .
  • the resource capacities may be potentially time-varying due to employment of dynamic power/performance management techniques (such as dynamic voltage and frequency scaling, dynamic memory sizing), resource reservations, dynamically-varying system resource requirements, and other contributors.
  • the computing of the estimated resource demand of operation 210 may further include computing estimated actual resource demand specific to CPU resources 211 , computing estimated actual resource demand specific to memory resources 212 , computing estimated actual resource demand specific to storage I/O resources 213 and computing estimated actual resource demand specific to network I/O resources 214 .
  • the computing of the estimated actual resource demand specific to CPU resources 211 may include obtaining CPU resource usage metrics for each entity and resource, obtaining CPU resource capacities for each resource and computing actual resource demand specific to CPU resources for each entity from the CPU resource usage metrics and the CPU resource capacities.
  • the CPU resource usage metrics may include CPU cycles during which the entity was actually granted the use of the CPU resources, CPU cycles during which the entity was ready to use the CPU resources, CPU cycles during which the entity was waiting on a resource other than the CPU and CPU cycles during which the entity was idle.
  • the CPU resource usage metrics may further include CPU cycles during which the resource performed system-related operations.
  • the CPU resource capacities may include available CPU cycles aggregated over all available logical processing units enclosed by the resource.
  • the computing of the estimated actual resource demand specific to memory resources 212 may include obtaining memory resource usage metrics for each entity and resource, obtaining memory resource capacities for each entity and resource and computing actual resource demand specific to memory resources for each entity from the memory resource usage metrics and the memory resource capacities.
  • the memory resource usage metrics may include an amount of memory that is active for each entity, an amount of memory that is mapped for each entity, an amount of memory that is swapped for each entity, an amount of memory that is shared with other entities and an amount of reserved memory.
  • the memory resource capacities may include available and configured memory for each resource.
  • the computing of the estimated actual resource demand specific to storage I/O resources 213 may include obtaining storage resource usage metrics for each entity and resource, obtaining storage resource capacities for each entity and resource and computing actual resource demand specific to storage I/O resources for each entity from the storage resource usage metrics and the storage resource capacities.
  • the storage resource usage metrics may include storage read and write rates for each entity, storage read and write requests for each entity, storage read and write latencies for each entity and resource and storage device read and write queue lengths for each resource.
  • the storage resource capacities may include storage bandwidth for each resource.
  • the computing of the estimated actual resource demand specific to network I/O resources 214 may include obtaining network resource usage metrics for each entity and resource, obtaining network resource capacities for each entity and resource and computing actual resource demand specific to network I/O resources for each entity from the network resource usage metrics and the network resource capacities.
  • the network resource usage metrics include network packet transfer and receive rates for each entity and resource, packet transfer and receive latencies for each entity and transmit and receive queue lengths and latencies for each resource.
  • the network resource capacities may include available network bandwidth for each resource and configured network bandwidth for each entity.
  • the methods described above may be embodied in a system, in which computational entities are distributed across physical computing resources to place the entities on the resources.
  • the system may include a processor and a computer readable medium coupled to the processor having executable instructions stored thereon. When executed, the executable instructions instruct the processor to estimate actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity, to compute a best allocation of the resources to the entities from the estimated actual resource demand for each entity and to distribute the resources to the entities in accordance with the computed best allocation.
  • the scheduling metrics may include CPU Used, CPU Wait, CPU Ready and CPU System metrics.
  • the CPU Used metric refers to a time spent while the CPU is in use.
  • the CPU Wait metric refers to a time spent waiting for some resource other than the CPU.
  • the CPU Ready metric refers to a time spent waiting for the CPU to be available.
  • the CPU System metric refers to a time spent in the hypervisor/kernel.
  • the metrics can be defined for each VM, as well as each physical processor.
  • the demand estimation method then uses the statistics gathered for each VM over a sampling period T that is smaller than the average decision making period used by a dynamic resource balancing mechanism. It is understood that while demand estimation relates to computing resource management as described herein, it may also be applied to memory, bandwidth and/or other similar computing resources.
  • the CPU Wait metric describes a time spent waiting on some other resource such as a disk or a network input/output (I/O) unit. During wait time, the VM cannot progress even if there is an available CPU resource since this resource is effectively blocked.
  • the CPU Ready metric captures a time during which the VM was ready to run on the CPU, but had to relinquish the processor to some other VM or task computing for the same resource. During ready time, the VM could actually make useful progress had it been given more processing resources.
  • the CPU System metric represents a time spent in lower-level operations such as hypervisor or kernel tasks and may be an important component in VM CPU accounting.
  • Insights relevant to demand estimation are drawn from these metrics. For example, for an idle VM, used and ready times are close to 0 seconds, while wait time is around 100%. For a completely CPU-bound VM, the sum of used and ready times is expected to by approximately 100%. In general, the sum of all the four metrics is expected to approximate total elapsed time for a VM.
  • the CPU demand estimate for a VM follows from these insights. The VM is expected to be in one of the four states described by the metrics and the distribution of VM CPU time to these states depends on an amount of physical resources available to the VM and each VM's actual desired demand.
  • FIG. 3 shows an example where 10 VMs are consolidated on a single host.
  • the host has 2 physical CPUs aggregating to a total CPU capacity of 200%.
  • Each VM has a CPU demand of 80%.
  • t represents an arbitrary time quantum during which only a pair of the VMs can run on the two physical cores. For simplicity, we assume the VMs get their turns to run in a pairwise round-robin fashion.
  • the VM load is constructed such that each VM wants to perform CPU operations worth 4t time, followed by a 1t idle period. This load pattern is repeated with this 5t period to achieve the 80% CPU demand over a sampling period T>>5t.
  • the three different states are represented with U (CPU Used), W (CPU Wait) and R (CPU Ready).
  • the columns of the figure represent the timeslices in t, and the whole chart from left to right shows the execution timeline of the 10 VMs.
  • VM 1 and VM 2 are ready to run.
  • the first two VMs are scheduled on the physical processors and, therefore, they are in the U state.
  • the other VMs stay in the R state, as they are ready to run, but there are no available CPUs.
  • VM 1 and VM 2 relinquish the CPUs, which are then allotted to VM 3 and VM 4 , which go into the U state.
  • VM 1 and VM 2 run for the fourth t interval. After this, they have completed their CPU-bound operation worth 4t CPU time and they remain idle for 1t. This is indicated with the W state. That is, at this point VM 1 and VM 2 are not competing for the CPU resources.
  • the following VMs enter the U state to perform their fourth t interval of CPU-bound execution and consecutively go into W state.
  • the VMs go back into R state after the W state when they start competing for resources again after idling for the 1t interval.
  • the whole 10VM execution pattern shows a fundamental period of 20t. Therefore, a sampling interval T>>20t shall capture a similar distribution of states as shown in the highlighted 20t region.
  • an estimated actual demand for a VM can be represented as:
  • This equation shows an approach to VM CPU demand estimation, which relies on the use of scheduling parameters for a hypervisor.
  • An additional “system” component is also included in the above equation. This represents a small overhead component for the time spent in local hypervisor resource management processes that is not attributed to the U, R and W states. While the equation shows the conceptual derivation of VM CPU demand, an analytical representation can be derived as shown in FIG. 4 .
  • a VM is assumed to be in a “u” state, where the VM gets to use the CPU for that tick, a “w” state, where the VM is waiting for an I/O wakeup, as an example, or an “r” state where the VM is not given access to the CPU but could have used it had it been.
  • N u the number of “u” state seconds
  • N w the number of “w” state seconds
  • N r the number of “r” state seconds
  • the CPU scheduling algorithm is treated as a random process in which, at each tick of the clock, the CPU is given to something other than the VM with a fixed probability.
  • demand estimation may be implemented and validated in an embodiment of a virtualized environment using, for example, one or more hypervisors 500 , such as VMware hypervisors, and VMs 520 A-C, such as Linux VMs, which are distributed across hosts 510 , 511 .
  • a virtualization manager 530 collects host/VM CPU usage stats in an aggregator 531 and stored in a virtual appliance 540 .
  • This exemplary validation prototype is implemented as a virtual appliance application programming interface (API) client that gathers the described VM CPU statistics from the hypervisors and as an integrated dynamic resource balancing engine that performs resource management and dynamic placement.
  • API application programming interface
  • demand estimation can be carried out by agents in individual hypervisors or as part of a general virtualization management endpoint. Experiments similar to the case shown in FIG. 1 have been conducted in the prototype implementation of FIG. 5 and indicate that demand estimation relatively accurately captures VM resource demand and that demand estimation eliminates the iterative progress described in FIG. 1 . Demand estimation therefore enables a dynamic placement engine to perform correct resource balancing actions in a single operation.
  • FIG. 6 shows an actual snapshot of evaluations for an experimental case with 2 hosts and 5 VMs.
  • VMs are configured to have different levels of demand as shown in Table 2, so that an effectiveness of demand estimation can be verified for accuracy in a general/realistic scenario.
  • the top highlighted regions show the observed CPU usage and the demand estimation, which tracks actual VM CPU demand relatively closely.
  • the lower area shows the impact of demand estimation in dynamic resource balancing.
  • the placement engine identifies and performs three live migrations to satisfy the actual demand of all VMs in the cluster in a single operation.
  • FIG. 6 corresponds to the same initial state of the scenario described in FIG. 1 , where all VMs reside in the first host.
  • the estimator identifies the actual demand of each VM with reasonable accuracy.
  • dynamic placement performs all the necessary balancing actions immediately.
  • all the intermediate steps of FIG. 1 are merged to a single correct action.
  • Table 3 shows the timeline of the actual placement decisions applied in this example with and without demand estimation. Without demand estimation, dynamic placement progresses in an iterative fashion, similar to FIG. 1 . In contrast, with demand estimation, all of the necessary balancing actions are completed in a single iteration.
  • relative priorities of VMs can be changed, resource caps on VMs can be introduced, resource guarantees for certain VMs can be provided.
  • the demand estimate approach performs in a similar fashion as demonstrated above.
  • the individual VM performances however show different characteristics as dictated by the applied resource constraints.
  • demand estimation can also relate to heterogeneous resources.
  • the demand estimate for a VM is seen to be less than 100% (minus an empirically-determined threshold for resilience against measurement fluctuations) of host capacity, this demand can be translated into another host by carrying over absolute resource requirements.
  • the case where the VM demand is seen to be close to 100% can be handled as a special case.
  • the VM is known to desire at least as much as its current host's capacity. Therefore, for another host with lower capacity than the VM's current host, it is known that the VM demand exceeds 100% of that host's capacity. This can be represented as either a percentage greater than 100% or by using absolute resource demand measures.
  • the actual demand of the VM can be estimated within a range. It is known that, the demand can be at least as much as the original host's resource capacity. This capacity translates into a percentage that is lower than 100% in the other (bigger-capacity) host. This measure constitutes a lower bound on VM demand. As the new measured demand cannot exceed capacity of the other host, the VM demand can be only as high as 100% of the other host's capacity. Thus, a demand of 100% constitutes an upper bound for conservatively estimating such a VM's demand on another, bigger-capacity host. For performance-efficient placement, this conservative estimate relatively cautiously assumes the upper bound of the possible range of VM resource demand.
  • the improved demand estimates provided by the proposed invention can be employed for other purposes.
  • a first additional purpose for which the invention may be employed is to improve the effectiveness of forecasting approaches that predict future resource demand of VMs. Without demand estimation, a forecasting approach that aims to discern trends and patterns from past resource usage history is exposed to spurious and potentially sharp changes in these historical measurements. These spurious shifts in past resource usage history occur because the patterns for each VM are susceptible to both other VMs' dynamically-varying application characteristics and the dynamic placement decisions as they work to correct any resource overcommitment.
  • a forecasting approach that takes into account the actual demand created by a VM is exposed to a representative view of the actual VM behavior that is only dependent on the same VM's own application characteristics. Such a forecasting approach is not misdirected by inter-VM influences on resource usage and ongoing dynamic placement decisions.
  • a second additional purpose for which the invention may be employed is to provide administrators directly with information about the evolution of demand over time, rather than feeding the demand estimate directly to an automated resource allocator.
  • the estimated actual demand of each VM (or application) over time could be plotted alongside the more traditional CPU utilization plot in the Graphical User Interface of a system administrator tool.
  • the GUI could also display a forecast of future demand, computed from the time series of estimated demands.
  • the future demand forecast would be more accurate and informative than a future CPU utilization forecast.

Abstract

A method for use in a system in which computational entities are distributed across physical computing resources to place the entities on the resources includes estimating actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity, computing a best allocation of the resources to the entities from the estimated actual resource demand for each entity and distributing the resources to the entities in accordance with the computed best allocation.

Description

BACKGROUND
Aspects of the present invention are directed to a method and a system for virtual machine CPU demand estimation for autonomous resource and power management.
In a virtualized environment, the physical hardware is decoupled from the operating system and application software via a hypervisor layer. In such environments, resource consolidation is a commonly applied technology that is enabled by computing hardware virtualization. The hypervisor helps multiple virtual machines (VMs), which may run different operating systems on different virtual hardware, share the same physical resources.
Current virtualization technologies incorporate management schemes that present a unified resource view for multiple virtualized hosts and provide methods for dynamic resource balancing among these hosts by, e.g., migrating VMs between resources. Such resource balancing requires explicit knowledge of each VM's resource demand to commit resources efficiently. One key challenge in these processes is the determination of the actual desired resource demand of each VM. In a case where a host is undercommitted and has more resources than the total demand produced by its occupant VMs, each VM's resource usage tracks demand closely. However, in the case where a host is overcommitted and had less resources than the total demand produced by its occupant VMs, VM resource usage deviates from demand. This latter case is far more important in terms of resource balancing and may lead to VM performance degradation. In overcommitment, the existing solutions have drawbacks in that they either use resource usage, such as each VM's CPU usage percent, as an indicator of demand or they may rely on legacy implementations tightly coupled with the underlying virtualization technology for specific implementations.
Generally, it is seen that some of the existing techniques either explicitly rely on CPU usage or introduce disruptive hooks to the overlaying virtualized applications for resource management and capacity planning. Some of these proposed techniques argue for the necessity of application-level intervention to distinguish demand from usage. While other work also argues that demand does not necessarily follow usage in both native and virtualized systems and points out some of the indicators of resource contention, they do not provide an estimation technique. Other work that proposes virtual machine performance modeling and estimation techniques rely on tracking virtual machine resource usage/demand and develop forecasting or application-level performance modeling techniques. These studies are not applicable for arbitrary black-box virtualization management (that is, virtualization management done without any monitoring hooks inside the VMs, where VMs are treated as black boxes) cases, and do not provide a general representation of actual VM resource demand. Some solutions aim to address the problem of identifying the resource requirements of a VM. These solutions rely on determining resource needs from within a virtual machine by inspecting application and operating system behavior within the VM to track certain indicative behavior such as detecting the occurrences of an idle loop within a virtual machine. In an approach for hypervisor load balancing, information about the VM CPU load is determined based on its interrupt response time. While this technique provides indirect measures on whether a VM is receiving its resource requirements, it is generally intrusive and does not provide a direct measure of a VM's actual resource demand.
SUMMARY
In accordance with an aspect of the invention, a method for use in a system in which computational entities are distributed across physical computing resources to place the entities on the resources is provided. The method includes estimating actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity, computing a best allocation of the resources to the entities from the estimated actual resource demand for each entity and distributing the resources to the entities in accordance with the computed best allocation.
In accordance with another aspect of the invention, a method for estimating actual resource demand for running entities is provided and includes obtaining a map of the entities relative to resources, obtaining resource usage metrics for each entity and resource, obtaining resource capacities and computing the estimated actual resource demand based on the map of the entities relative to the resources, the resource usage metrics and the resource capacities.
In accordance with another aspect of the invention, a system, in which computational entities are distributed across physical computing resources to place the entities on the resources, is provided. The system includes a processor and a computer readable medium coupled to the processor having executable instructions stored thereon, which, when executed instruct the processor to estimate actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity, to compute a best allocation of the resources to the entities from the estimated actual resource demand for each entity and to distribute the resources to the entities in accordance with the computed best allocation.
BRIEF DESCRIPTIONS OF THE SEVERAL VIEWS OF THE DRAWINGS
The subject matter regarded as the invention is particularly pointed out and distinctly claimed in the claims at the conclusion of the specification. The foregoing and other aspects, features, and advantages of the invention are apparent from the following detailed description taken in conjunction with the accompanying drawings in which:
FIG. 1 is a schematic diagram of a resource consolidated environment, and its operation without the employment of demand estimation as disclosed in this invention;
FIG. 2 is a flowchart illustrating a method to place entities on resources in accordance with embodiments of the invention;
FIG. 3 is a table illustrating an exemplary resource consolidated environment in which virtual machines (VMs) are consolidated on a single host;
FIG. 4 is a flow chart illustrating an analytical derivation of VM CPU demand estimation;
FIG. 5 is a schematic illustration of a virtualized environment in accordance with embodiments of the present invention; and
FIG. 6 is a snapshot of evaluation results of demand estimation tests.
DETAILED DESCRIPTION
With reference to FIG. 1, an exemplary system includes a virtualized cluster with 3 hosts, each with total CPU capacity of 100%, and 8 virtual machines (VMs). Dynamic placement aims to balance resources for use by the VMs such that each host's utilization does not, for example, exceed 85%. Dynamic placement also employs power management such that if the total resource demand by the VMs can be satisfied with fewer than 3 hosts, the remaining hosts are powered down. When the demand rises, the hosts are powered back on to satisfy the demand. In this example, the VMs are initially close to idle (each demanding 10% CPU) and, they fit within a single host (8×10%=80%<85%), they all reside in a single host with the other two machines powered down. When each VM's demand increases to 25% a dynamic resource manager uses resource usage as a measure of demand to respond in the absence of the demand estimation of the invention and an iterative process including 5 separate decisions that converges to the placement of resources.
In the present disclosure embodiments, desired resource demand of the VMs is captured initially and guides dynamic resource balancing actions in a single operation. That is, the resource management actions performed in steps T2-T5 of FIG. 1 are realized in a single operation with demand estimation. This is achieved using scheduling statistics commonly available to operating systems and hypervisors to derive reliable estimates for actual VM resource demand.
With reference to FIG. 2, a method for use in a system in which computational entities are distributed across physical computing resources to place the entities on the resources is provided as described above. This includes estimating actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity 200, such as a hypervisor, computing a best allocation of the resources to the entities from the estimated actual resource demand for each entity 220, and distributing the resources to the entities in accordance with the computed best allocation 230.
The estimating of operation 200 may include obtaining a map of the entities relative to the resources 201, obtaining resource usage metrics for each entity and resource 202, obtaining resource capacities 203 and computing the estimated actual resource demand based on the map of the entities relative to the resources, the resource usage metrics and the resource capacities 210. The resource capacities may be potentially time-varying due to employment of dynamic power/performance management techniques (such as dynamic voltage and frequency scaling, dynamic memory sizing), resource reservations, dynamically-varying system resource requirements, and other contributors.
The computing of the estimated resource demand of operation 210 may further include computing estimated actual resource demand specific to CPU resources 211, computing estimated actual resource demand specific to memory resources 212, computing estimated actual resource demand specific to storage I/O resources 213 and computing estimated actual resource demand specific to network I/O resources 214.
In accordance with embodiments of the invention, the computing of the estimated actual resource demand specific to CPU resources 211 may include obtaining CPU resource usage metrics for each entity and resource, obtaining CPU resource capacities for each resource and computing actual resource demand specific to CPU resources for each entity from the CPU resource usage metrics and the CPU resource capacities. Here, The CPU resource usage metrics may include CPU cycles during which the entity was actually granted the use of the CPU resources, CPU cycles during which the entity was ready to use the CPU resources, CPU cycles during which the entity was waiting on a resource other than the CPU and CPU cycles during which the entity was idle. The CPU resource usage metrics may further include CPU cycles during which the resource performed system-related operations. The CPU resource capacities may include available CPU cycles aggregated over all available logical processing units enclosed by the resource.
In accordance with further embodiments of the invention, the computing of the estimated actual resource demand specific to memory resources 212 may include obtaining memory resource usage metrics for each entity and resource, obtaining memory resource capacities for each entity and resource and computing actual resource demand specific to memory resources for each entity from the memory resource usage metrics and the memory resource capacities. Here, the memory resource usage metrics may include an amount of memory that is active for each entity, an amount of memory that is mapped for each entity, an amount of memory that is swapped for each entity, an amount of memory that is shared with other entities and an amount of reserved memory. The memory resource capacities may include available and configured memory for each resource.
In accordance with further embodiments of the invention, the computing of the estimated actual resource demand specific to storage I/O resources 213 may include obtaining storage resource usage metrics for each entity and resource, obtaining storage resource capacities for each entity and resource and computing actual resource demand specific to storage I/O resources for each entity from the storage resource usage metrics and the storage resource capacities. Here, the storage resource usage metrics may include storage read and write rates for each entity, storage read and write requests for each entity, storage read and write latencies for each entity and resource and storage device read and write queue lengths for each resource. The storage resource capacities may include storage bandwidth for each resource.
In accordance with further embodiments of the invention, the computing of the estimated actual resource demand specific to network I/O resources 214 may include obtaining network resource usage metrics for each entity and resource, obtaining network resource capacities for each entity and resource and computing actual resource demand specific to network I/O resources for each entity from the network resource usage metrics and the network resource capacities. Here, the network resource usage metrics include network packet transfer and receive rates for each entity and resource, packet transfer and receive latencies for each entity and transmit and receive queue lengths and latencies for each resource. The network resource capacities may include available network bandwidth for each resource and configured network bandwidth for each entity.
The methods described above may be embodied in a system, in which computational entities are distributed across physical computing resources to place the entities on the resources. The system may include a processor and a computer readable medium coupled to the processor having executable instructions stored thereon. When executed, the executable instructions instruct the processor to estimate actual resource demand for each entity on each resource based on application resource usage data collected from a data source external from the entity, to compute a best allocation of the resources to the entities from the estimated actual resource demand for each entity and to distribute the resources to the entities in accordance with the computed best allocation.
Demand estimation of aspects of the present invention generally relies on basic scheduling metrics. As shown in Table 1 below, the scheduling metrics may include CPU Used, CPU Wait, CPU Ready and CPU System metrics. The CPU Used metric refers to a time spent while the CPU is in use. The CPU Wait metric refers to a time spent waiting for some resource other than the CPU. The CPU Ready metric refers to a time spent waiting for the CPU to be available. The CPU System metric refers to a time spent in the hypervisor/kernel.
TABLE 1
CPU scheduling statistics used in demand estimation:
CPU Used Time spent while using the CPU
CPU Wait Time spent waiting for some resource other than the CPU
CPU Ready Time spent waiting for the CPU to be available
CPU System Time Spent in the hypervisor/kernel
The metrics can be defined for each VM, as well as each physical processor. The demand estimation method then uses the statistics gathered for each VM over a sampling period T that is smaller than the average decision making period used by a dynamic resource balancing mechanism. It is understood that while demand estimation relates to computing resource management as described herein, it may also be applied to memory, bandwidth and/or other similar computing resources.
The CPU Used metric tracks an amount of time a VM was actually running on a physical processor. The amount of time reported in these metrics may be considered as a portion of the sampling period T. For example, for a case in which T=20 seconds, if the CPU Used metric is found to be 10 seconds, this translates to 50% CPU usage for the VM. The CPU Wait metric describes a time spent waiting on some other resource such as a disk or a network input/output (I/O) unit. During wait time, the VM cannot progress even if there is an available CPU resource since this resource is effectively blocked. The CPU Ready metric captures a time during which the VM was ready to run on the CPU, but had to relinquish the processor to some other VM or task computing for the same resource. During ready time, the VM could actually make useful progress had it been given more processing resources. The CPU System metric represents a time spent in lower-level operations such as hypervisor or kernel tasks and may be an important component in VM CPU accounting.
Insights relevant to demand estimation are drawn from these metrics. For example, for an idle VM, used and ready times are close to 0 seconds, while wait time is around 100%. For a completely CPU-bound VM, the sum of used and ready times is expected to by approximately 100%. In general, the sum of all the four metrics is expected to approximate total elapsed time for a VM. The CPU demand estimate for a VM follows from these insights. The VM is expected to be in one of the four states described by the metrics and the distribution of VM CPU time to these states depends on an amount of physical resources available to the VM and each VM's actual desired demand.
With reference to FIG. 3, a hypothetical example demonstrates how the VM time is distributed during the lifetime of a VM. System time is omitted in the description for clarity. FIG. 3 shows an example where 10 VMs are consolidated on a single host. The host has 2 physical CPUs aggregating to a total CPU capacity of 200%. Each VM has a CPU demand of 80%. In the example, t represents an arbitrary time quantum during which only a pair of the VMs can run on the two physical cores. For simplicity, we assume the VMs get their turns to run in a pairwise round-robin fashion. That is, at t=1, VM1 and VM2 run, while the remaining are eight VMs wait their turn; at t=2, VM3 and VM4 acquire the cores, while the others wait; and so on. In addition the VM load is constructed such that each VM wants to perform CPU operations worth 4t time, followed by a 1t idle period. This load pattern is repeated with this 5t period to achieve the 80% CPU demand over a sampling period T>>5t. The three different states are represented with U (CPU Used), W (CPU Wait) and R (CPU Ready). The columns of the figure represent the timeslices in t, and the whole chart from left to right shows the execution timeline of the 10 VMs.
At t=1, all VMs are ready to run. The first two VMs are scheduled on the physical processors and, therefore, they are in the U state. The other VMs stay in the R state, as they are ready to run, but there are no available CPUs. At t=2, VM1 and VM2 relinquish the CPUs, which are then allotted to VM3 and VM4, which go into the U state. The same pattern continues until t=16. At this point, VM1 and VM2 run for the fourth t interval. After this, they have completed their CPU-bound operation worth 4t CPU time and they remain idle for 1t. This is indicated with the W state. That is, at this point VM1 and VM2 are not competing for the CPU resources. The following VMs enter the U state to perform their fourth t interval of CPU-bound execution and consecutively go into W state. The VMs go back into R state after the W state when they start competing for resources again after idling for the 1t interval. As the highlighted region shows, the whole 10VM execution pattern shows a fundamental period of 20t. Therefore, a sampling interval T>>20t shall capture a similar distribution of states as shown in the highlighted 20t region.
Focusing on the representative 20t region, each VM shows a state distribution of {U=4t, W=1t, R=15t}. While the amount spent in R is generally dependent on the amount of competition between VMs, W and U can be expected to be moderately consistent regardless of the level of overcommitment. As such, it is seen that the demand of the VMs is captured by a ratio of how much the VMs are in the U and W states and that a ratio of the CPU Used time (4t) to the sum of CPU Used and CPU Wait times (5t) represents the actual CPU demand of the VMs (80%). Thus, an estimated actual demand for a VM can be represented as:
CPU_Demand Used ( + System ) Used + Wait ( + System ) = Used ( + System ) Stats_Period - Ready
This equation shows an approach to VM CPU demand estimation, which relies on the use of scheduling parameters for a hypervisor. An additional “system” component is also included in the above equation. This represents a small overhead component for the time spent in local hypervisor resource management processes that is not attributed to the U, R and W states. While the equation shows the conceptual derivation of VM CPU demand, an analytical representation can be derived as shown in FIG. 4.
As shown in FIG. 4, at operation 400, at each tick of a CPU clock a VM is assumed to be in a “u” state, where the VM gets to use the CPU for that tick, a “w” state, where the VM is waiting for an I/O wakeup, as an example, or an “r” state where the VM is not given access to the CPU but could have used it had it been. Next, at operation 410, counts for each state over an interval of N ticks are defined as Nu (the number of “u” state seconds), Nw (the number of “w” state seconds) and Nr (the number of “r” state seconds). It is then assumed, at operation 420, that the VM's behavior is statistically stationary, such that for over a large number of ticks, where N>>1, the following quantities are well defined and measurable: U=Nu/N=Fraction of “u” seconds, W=Nw/N=Fraction of “w” seconds, and R=Nr/N=Fraction of “r” seconds.
If, at operation 430, it is determined that there is no contention for the CPU, the VM gets access to the CPU resources whenever it needs such access and it is known that R=O. Here, baseline quantities of U=U0, W=W0 and R=R0=0 are defined with the goal being to calculate U0 given values of U, W and R measured when there is contention for the CPU. When there is contention for the CPU, at operation 440, the CPU scheduling algorithm is treated as a random process in which, at each tick of the clock, the CPU is given to something other than the VM with a fixed probability. In this way, it may be represented that, if the VM would have been in state “u” if there were no contention, and if the CPU were to be given to something else for that clock tick, then the VM is placed in the “r” state and its program counter would not advance. Alternatively, if the VM is in the “w” state when the CPU is given to something else, the VM is unaffected. This substitution leads to the desired equations:
U 0 = U U + W = U 1 - R
With reference to FIG. 5, demand estimation may be implemented and validated in an embodiment of a virtualized environment using, for example, one or more hypervisors 500, such as VMware hypervisors, and VMs 520A-C, such as Linux VMs, which are distributed across hosts 510, 511. A virtualization manager 530 collects host/VM CPU usage stats in an aggregator 531 and stored in a virtual appliance 540. This exemplary validation prototype is implemented as a virtual appliance application programming interface (API) client that gathers the described VM CPU statistics from the hypervisors and as an integrated dynamic resource balancing engine that performs resource management and dynamic placement. In other embodiments, demand estimation can be carried out by agents in individual hypervisors or as part of a general virtualization management endpoint. Experiments similar to the case shown in FIG. 1 have been conducted in the prototype implementation of FIG. 5 and indicate that demand estimation relatively accurately captures VM resource demand and that demand estimation eliminates the iterative progress described in FIG. 1. Demand estimation therefore enables a dynamic placement engine to perform correct resource balancing actions in a single operation.
FIG. 6 shows an actual snapshot of evaluations for an experimental case with 2 hosts and 5 VMs. In FIG. 5, VMs are configured to have different levels of demand as shown in Table 2, so that an effectiveness of demand estimation can be verified for accuracy in a general/realistic scenario.
TABLE 2
CPU demand configurations for a case in which five VMs
are used in demand estimator validation:
CPU Demand
VM1
100%
VM2
80%
VM3 60%
VM4 40%
VM5
20%
In FIG. 6, the top highlighted regions show the observed CPU usage and the demand estimation, which tracks actual VM CPU demand relatively closely. The lower area shows the impact of demand estimation in dynamic resource balancing. The placement engine identifies and performs three live migrations to satisfy the actual demand of all VMs in the cluster in a single operation.
The case shown in FIG. 6 corresponds to the same initial state of the scenario described in FIG. 1, where all VMs reside in the first host. In this case, the estimator identifies the actual demand of each VM with reasonable accuracy. As a result, dynamic placement performs all the necessary balancing actions immediately. In other words, all the intermediate steps of FIG. 1 are merged to a single correct action.
Table 3 shows the timeline of the actual placement decisions applied in this example with and without demand estimation. Without demand estimation, dynamic placement progresses in an iterative fashion, similar to FIG. 1. In contrast, with demand estimation, all of the necessary balancing actions are completed in a single iteration.
TABLE 3
Timeline of dynamic placement actions, with and without
demand estimation:
Timeline Without Demand Estimation With Demand Estimation
Iteration
1 Migrate VM5 → Host2 Migrate VM5, VM4 &
VM3 → Host2
Iteration
2 Migrate VM4 → Host2 ALL DEMANDS MET
Iteration
3 Migrate VM3 → Host2
Iteration
4 ALL DEMANDS MET
In accordance with further embodiments of the invention, relative priorities of VMs can be changed, resource caps on VMs can be introduced, resource guarantees for certain VMs can be provided. In each case, the demand estimate approach performs in a similar fashion as demonstrated above. The individual VM performances however show different characteristics as dictated by the applied resource constraints.
While the discussed prototype focuses on homogeneous resources, demand estimation can also relate to heterogeneous resources. In the case of heterogeneous host configurations, if the demand estimate for a VM is seen to be less than 100% (minus an empirically-determined threshold for resilience against measurement fluctuations) of host capacity, this demand can be translated into another host by carrying over absolute resource requirements. The case where the VM demand is seen to be close to 100% can be handled as a special case. In this case, the VM is known to desire at least as much as its current host's capacity. Therefore, for another host with lower capacity than the VM's current host, it is known that the VM demand exceeds 100% of that host's capacity. This can be represented as either a percentage greater than 100% or by using absolute resource demand measures. For another host with capacity higher than the VM's current host, the actual demand of the VM can be estimated within a range. It is known that, the demand can be at least as much as the original host's resource capacity. This capacity translates into a percentage that is lower than 100% in the other (bigger-capacity) host. This measure constitutes a lower bound on VM demand. As the new measured demand cannot exceed capacity of the other host, the VM demand can be only as high as 100% of the other host's capacity. Thus, a demand of 100% constitutes an upper bound for conservatively estimating such a VM's demand on another, bigger-capacity host. For performance-efficient placement, this conservative estimate relatively cautiously assumes the upper bound of the possible range of VM resource demand.
In addition to improving the power/performance efficiency of virtualized environments, the improved demand estimates provided by the proposed invention can be employed for other purposes.
A first additional purpose for which the invention may be employed is to improve the effectiveness of forecasting approaches that predict future resource demand of VMs. Without demand estimation, a forecasting approach that aims to discern trends and patterns from past resource usage history is exposed to spurious and potentially sharp changes in these historical measurements. These spurious shifts in past resource usage history occur because the patterns for each VM are susceptible to both other VMs' dynamically-varying application characteristics and the dynamic placement decisions as they work to correct any resource overcommitment.
In contrast, a forecasting approach that takes into account the actual demand created by a VM is exposed to a representative view of the actual VM behavior that is only dependent on the same VM's own application characteristics. Such a forecasting approach is not misdirected by inter-VM influences on resource usage and ongoing dynamic placement decisions.
A second additional purpose for which the invention may be employed is to provide administrators directly with information about the evolution of demand over time, rather than feeding the demand estimate directly to an automated resource allocator. For example, the estimated actual demand of each VM (or application) over time could be plotted alongside the more traditional CPU utilization plot in the Graphical User Interface of a system administrator tool. If used in conjunction with forecasting (see, in FIG. 2, operation 240 of forecasting future demand based on the estimated actual resource demand), the GUI could also display a forecast of future demand, computed from the time series of estimated demands. As noted above, the future demand forecast would be more accurate and informative than a future CPU utilization forecast.
While the disclosure has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from the scope of the disclosure. In particular, the disclosure can be applied to a more general set of computational entities than VMs, such as applications or processes. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the disclosure without departing from the essential scope thereof. Therefore, it is intended that the disclosure not be limited to the particular exemplary embodiment disclosed as the best mode contemplated for carrying out this disclosure, but that the disclosure will include all embodiments falling within the scope of the appended claims.

Claims (22)

What is claimed is:
1. A method for use in a system in which computational entities are distributed across physical computing resources, the system including a processor and a non-transitory computer readable medium coupled to the processor and having executable instructions stored thereon, which, when executed instruct the processor to execute the method, which comprises:
tracking use time during which each entity runs on each resource, wait time during which each entity waits on a disk or input/output (I/O) unit, ready time during which each entity is ready to run on each resource but relinquishes the resource to another entity and system time for each entity;
calculating an estimated actual resource demand as a sum of use and system times divided by a sum of use, wait and system times and as the sum of use and system times divided by a result of ready time subtracted from a predefined statistics period;
computing a best allocation of the resources to the entities from the estimated actual resource demand for each entity such that, upon distribution of the resources to the entities, the resources will substantially satisfy the estimated actual resource demand for each entity; and
distributing the resources to the entities in accordance with the computed best allocation,
wherein the system time is representative of time spent in a local hypervisor resource management process.
2. The method according to claim 1, wherein the estimating comprises:
obtaining a map of the entities relative to the resources;
obtaining resource usage metrics for each entity and resource;
obtaining resource capacities; and
computing the estimated actual resource demand based on the map of the entities relative to the resources, the resource usage metrics and the resource capacities.
3. The method according to claim 2, wherein the resource capacities are time-varying.
4. The method according to claim 2, wherein the computing of the estimated actual resource demand comprises:
computing estimated actual resource demand specific to CPU resources;
computing estimated actual resource demand specific to memory resources;
computing estimated actual resource demand specific to storage I/O resources; and
computing estimated actual resource demand specific to network I/O resources.
5. The method according to claim 4, wherein the computing of the estimated actual resource demand specific to CPU resources comprises:
obtaining CPU resource usage metrics for each entity and resource;
obtaining CPU resource capacities for each resource; and
computing actual resource demand specific to CPU resources for each entity from the CPU resource usage metrics and the CPU resource capacities.
6. The method according to claim 5, wherein the CPU resource usage metrics comprise:
CPU cycles during which the entity was actually granted the use of the CPU resources;
CPU cycles during which the entity was ready to use the CPU resources;
CPU cycles during which the entity was waiting on a resource other than the CPU; and
CPU cycles during which the entity was idle.
7. The method according to claim 6, wherein the CPU resource usage metrics further comprise CPU cycles during which the resource performed system-related operations.
8. The method according to claim 5, wherein the CPU resource capacities comprise available CPU cycles aggregated over all available logical processing units enclosed by the resource.
9. The method according to claim 4, wherein the computing of the estimated actual resource demand specific to memory resources comprises:
obtaining memory resource usage metrics for each entity and resource;
obtaining memory resource capacities for each entity and resource; and
computing actual resource demand specific to memory resources for each entity from the memory resource usage metrics and the memory resource capacities.
10. The method according to claim 9, wherein the memory resource usage metrics comprise:
an amount of memory that is active for each entity;
an amount of memory that is mapped for each entity;
an amount of memory that is swapped for each entity;
an amount of memory that is shared with other entities; and
an amount of reserved memory.
11. The method according to claim 9, wherein the memory resource capacities comprise available and configured memory for each resource.
12. The method according to claim 4, wherein the computing of the estimated actual resource demand specific to storage I/O resources comprises:
obtaining storage resource usage metrics for each entity and resource;
obtaining storage resource capacities for each entity and resource; and
computing actual resource demand specific to storage I/O resources for each entity from the storage resource usage metrics and the storage resource capacities.
13. The method according to claim 12, wherein the storage resource usage metrics comprise:
storage read and write rates for each entity;
storage read and write requests for each entity;
storage read and write latencies for each entity and resource; and
storage device queue read and write latencies for each resource.
14. The method according to claim 12, wherein the storage resource capacities comprise storage bandwidth for each resource.
15. The method according to claim 4, wherein the computing of the estimated actual resource demand specific to network I/O resources comprises:
obtaining network resource usage metrics for each entity and resource;
obtaining network resource capacities for each entity and resource; and
computing actual resource demand specific to network I/O resources for each entity from the network resource usage metrics and the network resource capacities.
16. The method according to claim 15, wherein the network resource usage metrics comprise:
network packet transfer and receive rates for each entity and resource;
packet transfer and receive latencies for each entity; and
transmit and receive queue lengths and latencies for each resource.
17. The method according to claim 15, wherein the network resource capacities comprise:
available network bandwidth for each resource; and
configured network bandwidth for each entity.
18. The method according to claim 2, further comprising forecasting future demand based on the estimated actual resource demand.
19. A method for use in a system including a processor and a non-transitory computer readable medium coupled to the processor and having executable instructions stored thereon, which, when executed instruct the processor to estimate actual resource demand for running entities, the method comprising:
obtaining a map of the entities relative to a number of resources;
obtaining resource usage metrics for each entity and resource;
obtaining resource capacities; and,
based on the map of the entities relative to the resources, the resource usage metrics and the resource capacities;
tracking use time during which each entity runs on each resource, wait time during which each entity waits on a disk or input/output (I/O) unit, ready time during which each entity is ready to run on each resource but relinquishes the resource to another entity and system time for each entity;
calculating an estimated actual resource demand for each entity on each resource as a sum of use and system times divided by a sum of use, wait and system times and as the sum of use and system times divided by a result of ready time subtracted from a predefined statistics period,
computing a best allocation of the resources to the entities from the estimated actual resource demand for each entity such that, upon distribution of the resources to the entities, a fewest possible number of the resources will substantially satisfy the estimated actual resource demand for each entity without exceeding an established utilization target, which is established as a percentage of total capacity for each resource, the percentage being less than 100% of the total capacity, and
distributing the fewest possible number of the resources to the entities in accordance with the computed best allocation in a single distribution iteration and powering down a remaining number of the resources,
wherein the system time is representative of time spent in a local hypervisor resource management process.
20. The method according to claim 19, wherein the estimating actual resource demand comprises:
computing estimated actual resource demand specific to CPU resources;
computing estimated actual resource demand specific to memory resources;
computing estimated actual resource demand specific to storage I/O resources; and
computing estimated actual resource demand specific to network I/O resources.
21. The method according to claim 19, further comprising forecasting future demand based on the estimated actual resource demand.
22. A system, in which computational entities are distributed across a number of physical computing resources to place the entities on the resources, the system including a processor and a non-transitory computer readable medium coupled to the processor and having executable instructions stored thereon, which, when executed instruct the processor to:
establish a utilization target for each resource as a percentage of total capacity for each resource, the percentage being less than 100% of the total capacity,
track use time during which each entity runs on each resource, wait time during which each entity waits on a disk or input/output (I/O) unit, ready time during which each entity is ready to run on each resource but relinquishes the resource to another entity and system time for each entity,
calculate an estimated actual resource demand for each entity on each resource as a sum of use and system times divided by a sum of use, wait and system times and as the sum of use and system times divided by a result of ready time subtracted from a predefined statistics period,
compute a best allocation of the resources to the entities from the estimated actual resource demand for each entity, such that, upon distribution of the resources to the entities, a fewest number of the resources will substantially satisfy the estimated actual resource demand for each entity without exceeding the established utilization target,
distribute the fewest number of the resources to the entities in accordance with the computed best allocation in a single distribution iteration and
power down a remaining number of the resources,
wherein the system time is representative of time spent in a local hypervisor resource management process.
US12/563,445 2009-09-21 2009-09-21 Virtual machine demand estimation Expired - Fee Related US9037717B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/563,445 US9037717B2 (en) 2009-09-21 2009-09-21 Virtual machine demand estimation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/563,445 US9037717B2 (en) 2009-09-21 2009-09-21 Virtual machine demand estimation

Publications (2)

Publication Number Publication Date
US20110072138A1 US20110072138A1 (en) 2011-03-24
US9037717B2 true US9037717B2 (en) 2015-05-19

Family

ID=43757575

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/563,445 Expired - Fee Related US9037717B2 (en) 2009-09-21 2009-09-21 Virtual machine demand estimation

Country Status (1)

Country Link
US (1) US9037717B2 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150106245A1 (en) * 2013-10-16 2015-04-16 Vmware, Inc. Dynamic unit resource usage price calibrator for a virtual data center
US10838735B2 (en) 2018-12-07 2020-11-17 Nutanix, Inc. Systems and methods for selecting a target host for migration of a virtual machine
US10884779B2 (en) 2018-12-07 2021-01-05 Nutanix, Inc. Systems and methods for selecting virtual machines to be migrated
US11016889B1 (en) 2019-12-13 2021-05-25 Seagate Technology Llc Storage device with enhanced time to ready performance
US11797327B2 (en) * 2013-05-03 2023-10-24 Vmware, Inc. Dynamic virtual machine sizing

Families Citing this family (120)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8028090B2 (en) 2008-11-17 2011-09-27 Amazon Technologies, Inc. Request routing utilizing client location information
US7991910B2 (en) 2008-11-17 2011-08-02 Amazon Technologies, Inc. Updating routing information based on client location
US7970820B1 (en) 2008-03-31 2011-06-28 Amazon Technologies, Inc. Locality based content distribution
US8606996B2 (en) 2008-03-31 2013-12-10 Amazon Technologies, Inc. Cache optimization
US8321568B2 (en) 2008-03-31 2012-11-27 Amazon Technologies, Inc. Content management
US8156243B2 (en) 2008-03-31 2012-04-10 Amazon Technologies, Inc. Request routing
US8533293B1 (en) 2008-03-31 2013-09-10 Amazon Technologies, Inc. Client side cache management
US8447831B1 (en) 2008-03-31 2013-05-21 Amazon Technologies, Inc. Incentive driven content delivery
US8601090B1 (en) 2008-03-31 2013-12-03 Amazon Technologies, Inc. Network resource identification
US7962597B2 (en) 2008-03-31 2011-06-14 Amazon Technologies, Inc. Request routing based on class
US9912740B2 (en) 2008-06-30 2018-03-06 Amazon Technologies, Inc. Latency measurement in resource requests
US7925782B2 (en) 2008-06-30 2011-04-12 Amazon Technologies, Inc. Request routing using network computing components
US9407681B1 (en) 2010-09-28 2016-08-02 Amazon Technologies, Inc. Latency measurement in resource requests
US8065417B1 (en) 2008-11-17 2011-11-22 Amazon Technologies, Inc. Service provider registration by a content broker
US8732309B1 (en) 2008-11-17 2014-05-20 Amazon Technologies, Inc. Request routing utilizing cost information
US8122098B1 (en) 2008-11-17 2012-02-21 Amazon Technologies, Inc. Managing content delivery network service providers by a content broker
US8060616B1 (en) 2008-11-17 2011-11-15 Amazon Technologies, Inc. Managing CDN registration by a storage provider
US8073940B1 (en) 2008-11-17 2011-12-06 Amazon Technologies, Inc. Managing content delivery network service providers
US8521880B1 (en) 2008-11-17 2013-08-27 Amazon Technologies, Inc. Managing content delivery network service providers
US8521851B1 (en) 2009-03-27 2013-08-27 Amazon Technologies, Inc. DNS query processing using resource identifiers specifying an application broker
US8412823B1 (en) 2009-03-27 2013-04-02 Amazon Technologies, Inc. Managing tracking information entries in resource cache components
US8688837B1 (en) 2009-03-27 2014-04-01 Amazon Technologies, Inc. Dynamically translating resource identifiers for request routing using popularity information
US8756341B1 (en) 2009-03-27 2014-06-17 Amazon Technologies, Inc. Request routing utilizing popularity information
US8782236B1 (en) 2009-06-16 2014-07-15 Amazon Technologies, Inc. Managing resources using resource expiration data
US8397073B1 (en) 2009-09-04 2013-03-12 Amazon Technologies, Inc. Managing secure content in a content delivery network
US8433771B1 (en) 2009-10-02 2013-04-30 Amazon Technologies, Inc. Distribution network with forward resource propagation
US8504686B2 (en) * 2009-11-02 2013-08-06 InMon Corp. Method and apparatus for combining data associated with hardware resources and network traffic
US8490087B2 (en) * 2009-12-02 2013-07-16 International Business Machines Corporation System and method for transforming legacy desktop environments to a virtualized desktop model
US9495338B1 (en) 2010-01-28 2016-11-15 Amazon Technologies, Inc. Content distribution network
US8140735B2 (en) * 2010-02-17 2012-03-20 Novell, Inc. Techniques for dynamic disk personalization
US9245246B2 (en) * 2010-04-22 2016-01-26 International Business Machines Corporation Capacity over-commit management in resource provisioning environments
DE102010029209B4 (en) * 2010-05-21 2014-06-18 Offis E.V. A method for dynamically distributing one or more services in a network of a plurality of computers
US8631406B2 (en) * 2010-06-30 2014-01-14 Sap Ag Distributed cloud computing architecture
US8707300B2 (en) 2010-07-26 2014-04-22 Microsoft Corporation Workload interference estimation and performance optimization
US8577992B1 (en) 2010-09-28 2013-11-05 Amazon Technologies, Inc. Request routing management based on network components
US10958501B1 (en) 2010-09-28 2021-03-23 Amazon Technologies, Inc. Request routing information based on client IP groupings
US9003035B1 (en) 2010-09-28 2015-04-07 Amazon Technologies, Inc. Point of presence management in request routing
US9712484B1 (en) 2010-09-28 2017-07-18 Amazon Technologies, Inc. Managing request routing information utilizing client identifiers
US8819283B2 (en) 2010-09-28 2014-08-26 Amazon Technologies, Inc. Request routing in a networked environment
US8468247B1 (en) 2010-09-28 2013-06-18 Amazon Technologies, Inc. Point of presence management in request routing
US10097398B1 (en) 2010-09-28 2018-10-09 Amazon Technologies, Inc. Point of presence management in request routing
US8452874B2 (en) 2010-11-22 2013-05-28 Amazon Technologies, Inc. Request routing processing
US10467042B1 (en) 2011-04-27 2019-11-05 Amazon Technologies, Inc. Optimized deployment based upon customer locality
US8843933B1 (en) 2011-05-25 2014-09-23 Vmware, Inc. System and method for managing a virtualized computing environment
US8843924B2 (en) * 2011-06-17 2014-09-23 International Business Machines Corporation Identification of over-constrained virtual machines
US8966084B2 (en) 2011-06-17 2015-02-24 International Business Machines Corporation Virtual machine load balancing
US8949428B2 (en) 2011-06-17 2015-02-03 International Business Machines Corporation Virtual machine load balancing
US9250944B2 (en) * 2011-08-30 2016-02-02 International Business Machines Corporation Selection of virtual machines from pools of pre-provisioned virtual machines in a networked computing environment
TW201324357A (en) * 2011-12-01 2013-06-16 Univ Tunghai Green energy management of virtual machine cluster
US8904009B1 (en) 2012-02-10 2014-12-02 Amazon Technologies, Inc. Dynamic content delivery
US10021179B1 (en) 2012-02-21 2018-07-10 Amazon Technologies, Inc. Local resource delivery network
US10353738B2 (en) 2012-03-21 2019-07-16 International Business Machines Corporation Resource allocation based on social networking trends in a networked computing environment
US9146763B1 (en) 2012-03-28 2015-09-29 Google Inc. Measuring virtual machine metrics
US10623408B1 (en) 2012-04-02 2020-04-14 Amazon Technologies, Inc. Context sensitive object management
US9244742B2 (en) 2012-05-31 2016-01-26 Vmware, Inc. Distributed demand-based storage quality of service management using resource pooling
US9154551B1 (en) 2012-06-11 2015-10-06 Amazon Technologies, Inc. Processing DNS queries to identify pre-processing information
US8959383B2 (en) 2012-06-13 2015-02-17 Microsoft Corporation Failover estimation using contradiction
WO2013190649A1 (en) * 2012-06-20 2013-12-27 富士通株式会社 Information processing method and device related to virtual-disk migration
US8930948B2 (en) * 2012-06-21 2015-01-06 Vmware, Inc. Opportunistically proactive resource management using spare capacity
US9043787B2 (en) * 2012-07-13 2015-05-26 Ca, Inc. System and method for automated assignment of virtual machines and physical machines to hosts
US20140019964A1 (en) * 2012-07-13 2014-01-16 Douglas M. Neuse System and method for automated assignment of virtual machines and physical machines to hosts using interval analysis
US9525659B1 (en) 2012-09-04 2016-12-20 Amazon Technologies, Inc. Request routing utilizing point of presence load information
US9323577B2 (en) * 2012-09-20 2016-04-26 Amazon Technologies, Inc. Automated profiling of resource usage
US11321118B2 (en) 2012-10-24 2022-05-03 Messageone, Inc. System and method for controlled sharing of consumable resources in a computer cluster
US9491114B2 (en) * 2012-10-24 2016-11-08 Messageone, Inc. System and method for optimizing resource utilization in a clustered or cloud environment
US10205698B1 (en) 2012-12-19 2019-02-12 Amazon Technologies, Inc. Source-dependent address resolution
US10114719B2 (en) * 2013-02-21 2018-10-30 International Business Machines Corporation Estimating power usage in a computing environment
US9471394B2 (en) 2013-03-13 2016-10-18 Cloubrain, Inc. Feedback system for optimizing the allocation of resources in a data center
US9294391B1 (en) 2013-06-04 2016-03-22 Amazon Technologies, Inc. Managing network computing components utilizing request routing
US9830236B2 (en) * 2013-06-05 2017-11-28 Vmware, Inc. System and method for assigning memory reserved for high availability failover to virtual machines
US10002059B2 (en) 2013-06-13 2018-06-19 Vmware, Inc. System and method for assigning memory available for high availability failover to virtual machines
WO2014198001A1 (en) * 2013-06-14 2014-12-18 Cirba Inc System and method for determining capacity in computer environments using demand profiles
US9727355B2 (en) * 2013-08-23 2017-08-08 Vmware, Inc. Virtual Hadoop manager
US9686207B2 (en) * 2014-01-29 2017-06-20 Vmware, Inc. Application service level objective aware demand estimation
US20150244645A1 (en) * 2014-02-26 2015-08-27 Ca, Inc. Intelligent infrastructure capacity management
US9298492B2 (en) * 2014-03-05 2016-03-29 Ca, Inc. System and method for modifying allocated resources
US10146567B2 (en) 2014-11-20 2018-12-04 Red Hat Israel, Ltd. Optimizing virtual machine allocation to cluster hosts
US10091096B1 (en) 2014-12-18 2018-10-02 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10097448B1 (en) 2014-12-18 2018-10-09 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10033627B1 (en) 2014-12-18 2018-07-24 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
TW201624277A (en) * 2014-12-31 2016-07-01 萬國商業機器公司 Method of facilitating live migration of virtual machines
US10225326B1 (en) 2015-03-23 2019-03-05 Amazon Technologies, Inc. Point of presence based data uploading
US9887932B1 (en) 2015-03-30 2018-02-06 Amazon Technologies, Inc. Traffic surge management for points of presence
US9819567B1 (en) 2015-03-30 2017-11-14 Amazon Technologies, Inc. Traffic surge management for points of presence
US9887931B1 (en) 2015-03-30 2018-02-06 Amazon Technologies, Inc. Traffic surge management for points of presence
US9832141B1 (en) 2015-05-13 2017-11-28 Amazon Technologies, Inc. Routing based request correlation
US10616179B1 (en) 2015-06-25 2020-04-07 Amazon Technologies, Inc. Selective routing of domain name system (DNS) requests
US10079715B1 (en) * 2015-07-16 2018-09-18 VCE IP Holding Company LLC Methods, systems and computer readable mediums for performing metadata-driven data collection
US10417099B1 (en) * 2015-07-30 2019-09-17 EMC IP Holding Company LLC Resilient backups for large Hyper-V cluster shared volume environments
US10097566B1 (en) 2015-07-31 2018-10-09 Amazon Technologies, Inc. Identifying targets of network attacks
US9794281B1 (en) 2015-09-24 2017-10-17 Amazon Technologies, Inc. Identifying sources of network attacks
US9774619B1 (en) 2015-09-24 2017-09-26 Amazon Technologies, Inc. Mitigating network attacks
US9742795B1 (en) 2015-09-24 2017-08-22 Amazon Technologies, Inc. Mitigating network attacks
US10270878B1 (en) 2015-11-10 2019-04-23 Amazon Technologies, Inc. Routing for origin-facing points of presence
US10049051B1 (en) 2015-12-11 2018-08-14 Amazon Technologies, Inc. Reserved cache space in content delivery networks
US10257307B1 (en) 2015-12-11 2019-04-09 Amazon Technologies, Inc. Reserved cache space in content delivery networks
US10348639B2 (en) 2015-12-18 2019-07-09 Amazon Technologies, Inc. Use of virtual endpoints to improve data transmission rates
CN106959889A (en) * 2016-01-11 2017-07-18 阿里巴巴集团控股有限公司 A kind of method and apparatus of server resource adjustment
US10387210B2 (en) 2016-04-04 2019-08-20 International Business Machines Corporation Resource schedule optimization
US10075551B1 (en) 2016-06-06 2018-09-11 Amazon Technologies, Inc. Request management for hierarchical cache
US10110694B1 (en) 2016-06-29 2018-10-23 Amazon Technologies, Inc. Adaptive transfer rate for retrieving content from a server
US9992086B1 (en) 2016-08-23 2018-06-05 Amazon Technologies, Inc. External health checking of virtual private cloud network environments
US10033691B1 (en) 2016-08-24 2018-07-24 Amazon Technologies, Inc. Adaptive resolution of domain name requests in virtual private cloud network environments
US10469513B2 (en) 2016-10-05 2019-11-05 Amazon Technologies, Inc. Encrypted network addresses
US10382535B2 (en) 2016-10-05 2019-08-13 Vmware, Inc. Pairwise comparison for load balancing
AU2017345796A1 (en) * 2016-10-21 2019-05-23 DataRobot, Inc. Systems for predictive data analytics, and related methods and apparatus
US10831549B1 (en) 2016-12-27 2020-11-10 Amazon Technologies, Inc. Multi-region request-driven code execution system
US10372499B1 (en) 2016-12-27 2019-08-06 Amazon Technologies, Inc. Efficient region selection system for executing request-driven code
US10938884B1 (en) 2017-01-30 2021-03-02 Amazon Technologies, Inc. Origin server cloaking using virtual private cloud network environments
US10942760B2 (en) 2017-02-03 2021-03-09 Microsoft Technology Licensing, Llc Predictive rightsizing for virtual machines in cloud computing systems
US10423455B2 (en) * 2017-02-03 2019-09-24 Microsoft Technology Licensing, Llc Method for deploying virtual machines in cloud computing systems based on predicted lifetime
US10296367B2 (en) 2017-02-03 2019-05-21 Microsoft Technology Licensing, Llc Resource management for virtual machines in cloud computing systems
US10503613B1 (en) 2017-04-21 2019-12-10 Amazon Technologies, Inc. Efficient serving of resources during server unavailability
US11075987B1 (en) 2017-06-12 2021-07-27 Amazon Technologies, Inc. Load estimating content delivery network
US10447648B2 (en) 2017-06-19 2019-10-15 Amazon Technologies, Inc. Assignment of a POP to a DNS resolver based on volume of communications over a link between client devices and the POP
US10742593B1 (en) 2017-09-25 2020-08-11 Amazon Technologies, Inc. Hybrid content request routing system
US10592578B1 (en) 2018-03-07 2020-03-17 Amazon Technologies, Inc. Predictive content push-enabled content delivery network
US10862852B1 (en) 2018-11-16 2020-12-08 Amazon Technologies, Inc. Resolution of domain name requests in heterogeneous network environments
US11025747B1 (en) 2018-12-12 2021-06-01 Amazon Technologies, Inc. Content request pattern-based routing system
US11023287B2 (en) * 2019-03-27 2021-06-01 International Business Machines Corporation Cloud data center with reduced energy consumption

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6215772B1 (en) * 1997-11-26 2001-04-10 International Business Machines Corporation Dynamic parameter estimation for efficient transport of HPR data on IP
US20060064687A1 (en) 2004-09-23 2006-03-23 Jan Dostert Thread-level resource usage measurement
US20070006218A1 (en) 2005-06-29 2007-01-04 Microsoft Corporation Model-based virtual system provisioning
US20070162720A1 (en) 2006-01-12 2007-07-12 International Business Machines Corporation Apparatus and method for autonomically adjusting one or more computer program configuration settings when resources in a logical partition change
US20070300239A1 (en) * 2006-06-23 2007-12-27 International Business Machines Corporation Dynamic application instance placement in data center environments
US20080222025A1 (en) 2005-01-12 2008-09-11 International Business Machines Corporation Automatically distributing a bid request for a grid job to multiple grid providers and analyzing responses to select a winning grid provider
US20080288941A1 (en) 2007-05-14 2008-11-20 Vmware, Inc. Adaptive dynamic selection and application of multiple virtualization techniques
US20080301473A1 (en) 2007-05-29 2008-12-04 International Business Machines Corporation Method and system for hypervisor based power management
US20090006878A1 (en) * 2007-06-28 2009-01-01 International Business Machines Corporation Method and system for monitoring system processes energy consumption
US20090113422A1 (en) 2007-10-31 2009-04-30 Toshimitsu Kani Dynamic allocation of virtual machine devices
US20090164356A1 (en) 2007-10-12 2009-06-25 Alexander Bakman Method, system and apparatus for calculating chargeback for virtualized computing resources
US7716446B1 (en) * 2006-04-27 2010-05-11 Vmware, Inc. System and method for cooperative virtual machine memory scheduling
US20100131959A1 (en) * 2008-11-26 2010-05-27 Spiers Adam Z Proactive application workload management
US20100191854A1 (en) * 2009-01-26 2010-07-29 Vmware, Inc. Process demand prediction for distributed power and resource management

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6215772B1 (en) * 1997-11-26 2001-04-10 International Business Machines Corporation Dynamic parameter estimation for efficient transport of HPR data on IP
US20060064687A1 (en) 2004-09-23 2006-03-23 Jan Dostert Thread-level resource usage measurement
US20080222025A1 (en) 2005-01-12 2008-09-11 International Business Machines Corporation Automatically distributing a bid request for a grid job to multiple grid providers and analyzing responses to select a winning grid provider
US20070006218A1 (en) 2005-06-29 2007-01-04 Microsoft Corporation Model-based virtual system provisioning
US20070162720A1 (en) 2006-01-12 2007-07-12 International Business Machines Corporation Apparatus and method for autonomically adjusting one or more computer program configuration settings when resources in a logical partition change
US7716446B1 (en) * 2006-04-27 2010-05-11 Vmware, Inc. System and method for cooperative virtual machine memory scheduling
US20070300239A1 (en) * 2006-06-23 2007-12-27 International Business Machines Corporation Dynamic application instance placement in data center environments
US20080288941A1 (en) 2007-05-14 2008-11-20 Vmware, Inc. Adaptive dynamic selection and application of multiple virtualization techniques
US20080301473A1 (en) 2007-05-29 2008-12-04 International Business Machines Corporation Method and system for hypervisor based power management
US20090006878A1 (en) * 2007-06-28 2009-01-01 International Business Machines Corporation Method and system for monitoring system processes energy consumption
US20090164356A1 (en) 2007-10-12 2009-06-25 Alexander Bakman Method, system and apparatus for calculating chargeback for virtualized computing resources
US20090113422A1 (en) 2007-10-31 2009-04-30 Toshimitsu Kani Dynamic allocation of virtual machine devices
US20100131959A1 (en) * 2008-11-26 2010-05-27 Spiers Adam Z Proactive application workload management
US20100191854A1 (en) * 2009-01-26 2010-07-29 Vmware, Inc. Process demand prediction for distributed power and resource management

Non-Patent Citations (7)

* Cited by examiner, † Cited by third party
Title
Andrew Tanenbaum, Structured computer organization; (2nd ed.), Prentice-Hall, Inc., Upper Saddle River, NJ, 1984. *
Black-box and Gray-box Strategies for Virtual Machine Migration. Timothy Wood, Prashant Shenoy, Arun Venkataramani, and Mazin Yousif. Fourth Symposium on Networked Systems Design and Implementation (NSDI), Apr. 11-13, 2007. *
Chapin,Steve J. et al., "Support for Implementing Scheduling Algorithms Using Messiahs," pp. 1-27, 1994.
Frank Vahid, "The Softening of Hardware," Computer, vol. 36, No. 4, pp. 27-34, Apr. 2003, doi:10.1109/MC.2003.1193225. *
Hongzhang Shan et al., "Job Superscheduler Architecture and Performance in Computational Grid Environments," SC'03, pp. 1-15.
IEEE article "Dynamic Placement of Virtual Machines for Managing SLA Violations" (May 21-25, 2007) to Bobroff et al. ("Bobroff"). *
Steve McConnell, "Who Needs Software Engineering?," IEEE Software, vol. 18, No. 1, pp. 5-8, Jan./Feb. 2001, doi:10.1109/MS.2001.903148. *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11797327B2 (en) * 2013-05-03 2023-10-24 Vmware, Inc. Dynamic virtual machine sizing
US20150106245A1 (en) * 2013-10-16 2015-04-16 Vmware, Inc. Dynamic unit resource usage price calibrator for a virtual data center
US9524516B2 (en) * 2013-10-16 2016-12-20 Vmware, Inc. Dynamic unit resource usage price calibrator for a virtual data center
US10838735B2 (en) 2018-12-07 2020-11-17 Nutanix, Inc. Systems and methods for selecting a target host for migration of a virtual machine
US10884779B2 (en) 2018-12-07 2021-01-05 Nutanix, Inc. Systems and methods for selecting virtual machines to be migrated
US11016889B1 (en) 2019-12-13 2021-05-25 Seagate Technology Llc Storage device with enhanced time to ready performance

Also Published As

Publication number Publication date
US20110072138A1 (en) 2011-03-24

Similar Documents

Publication Publication Date Title
US9037717B2 (en) Virtual machine demand estimation
US10896055B2 (en) Capacity risk management for virtual machines
Grandl et al. Multi-resource packing for cluster schedulers
Mars et al. Contention aware execution: online contention detection and response
US8180604B2 (en) Optimizing a prediction of resource usage of multiple applications in a virtual environment
Chen et al. Cloudscope: Diagnosing and managing performance interference in multi-tenant clouds
US8131519B2 (en) Accuracy in a prediction of resource usage of an application in a virtual environment
JP5885920B2 (en) Virtual CPU based frequency and voltage control
US8145455B2 (en) Predicting resource usage of an application in a virtual environment
US8145456B2 (en) Optimizing a prediction of resource usage of an application in a virtual environment
Peschlow et al. A flexible dynamic partitioning algorithm for optimistic distributed simulation
Stavrinides et al. Energy-aware scheduling of real-time workflow applications in clouds utilizing DVFS and approximate computations
Rameshan et al. Hubbub-scale: Towards reliable elastic scaling under multi-tenancy
Lu et al. Predictive VM consolidation on multiple resources: Beyond load balancing
Gottschlag et al. Automatic core specialization for AVX-512 applications
US20190102233A1 (en) Method for power optimization in virtualized environments and system implementing the same
Pons et al. Cloud white: Detecting and estimating qos degradation of latency-critical workloads in the public cloud
Joshi et al. Sherlock: Lightweight detection of performance interference in containerized cloud services
Hlavacs et al. Predicting web service levels during vm live migrations
Pfitscher et al. Customer-oriented diagnosis of memory provisioning for iaas clouds
HoseinyFarahabady et al. Data-intensive workload consolidation in serverless (Lambda/FaaS) platforms
Jiang et al. Resource allocation in contending virtualized environments through VM performance modeling and feedback
Ghosh et al. Symbiotic scheduling for shared caches in multi-core systems using memory footprint signature
Zhang et al. LVMCI: Efficient and effective VM live migration selection scheme in virtualized data centers
Hauser et al. Predictability of resource intensive big data and hpc jobs in cloud data centres

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CANTURK, ISCI;HANSON, JAMES E.;KEPHART, JEFFREY O.;AND OTHERS;SIGNING DATES FROM 20090918 TO 20090921;REEL/FRAME:023258/0603

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20190519