System and method for scheduling resources within a compute environment using a scheduler process with reservation mask function

Information

  • Patent Grant
  • 12124878
  • Patent Number
    12,124,878
  • Date Filed
    Thursday, March 17, 2022
    2 years ago
  • Date Issued
    Tuesday, October 22, 2024
    a month ago
Abstract
A system and method of dynamically controlling a reservation of resources within a cluster environment to maximize a response time are disclosed. The method embodiment of the invention includes receiving from a requestor a request for a reservation of resources in the cluster environment, reserving a first group of resources, evaluating resources within the cluster environment to determine if the response time can be improved and if the response time can be improved, then canceling the reservation for the first group of resources and reserving a second group of resources to process the request at the improved response time.
Description
RELATED APPLICATIONS

The present application is related to U.S. patent application Ser. No. 10/530,583 (now U.S. Pat. No. 7,620,706); U.S. patent application Ser. No. 10/530,582 (now U.S. Pat. No. 7,971,204); U.S. patent application Ser. No. 10/530,577 (now U.S. Pat. No. 9,268,607); U.S. patent application Ser. No. 10/530,576 (now U.S. Pat. No. 9,176,785); U.S. patent application Ser. No. 10/589,339 (now U.S. Pat. No. 7,490,325); U.S. patent application Ser. No. 10/530,578 (now U.S. Pat. No. 8,151,103); U.S. patent application Ser. No. 10/530,580 (now U.S. Pat. No. 9,558,042), and U.S. patent application Ser. No. 10/530,575 (now U.S. Pat. No. 8,108,869). The content of each of these cases is incorporated herein by reference.


BACKGROUND OF THE INVENTION
1. Field of the Invention

The present invention relates to reservations in a cluster or more specifically to a system and method of providing a self-optimizing reservation in time of compute resources.


2. Introduction

The present invention relates to a system and method of allocation resources in the context of a grid or cluster of computers. Grid computing may be defined as coordinated resource sharing and problem solving in dynamic, multi-institutional collaborations. Many computing projects require much more computational power and resources than a single computer may provide. Networked computers with peripheral resources such as printers, scanners, I/O devices, storage disks, scientific devices and instruments, etc. may need to be coordinated and utilized to complete a task.


Grid/cluster resource management generally describes the process of identifying requirements, matching resources to applications, allocating those resources, and scheduling and monitoring grid resources over time in order to run grid applications as efficiently as possible. Each project will utilize a different set of resources and thus is typically unique. In addition to the challenge of allocating resources for a particular job, grid administrators also have difficulty obtaining a clear understanding of the resources available, the current status of the grid and available resources, and real-time competing needs of various users. One aspect of this process is the ability to reserve resources for a job. A cluster manager will seek to reserve a set of resources to enable the cluster to process a job at a promised quality of service.


General background information on clusters and grids may be found in several publications. See, e.g., Grid Resource Management, State of the Art and Future Trends, Jarek Nabrzyski, Jennifer M. Schopf, and Jan Weglarz, Kluwer Academic Publishers, 2004; and Beowulf Cluster Computing with Linux, edited by William Gropp, Ewing Lusk, and Thomas Sterling, Massachusetts Institute of Technology, 2003.


It is generally understood herein that the terms grid and cluster are interchangeable in that there is no specific definition of either. In general, a grid will comprise a plurality of clusters as will be shown in FIG. 1A. Several general challenges exist when attempting to maximize resources in a grid. First, there are typically multiple layers of grid and cluster schedulers. A grid 100 generally comprises a group of clusters or a group of networked computers. The definition of a grid is very flexible and may mean a number of different configurations of computers. The introduction here is meant to be general given the variety of configurations that are possible. A grid scheduler 102 communicates with a plurality of cluster schedulers 104A, 104B and 104C. Each of these cluster schedulers communicates with a respective resource manager 106A, 106B or 106C. Each resource manager communicates with a respective series of compute resources shown as nodes 108A, 108B, 108C in cluster 110, nodes 108D, 108E, 108F in cluster 112 and nodes 108G, 108H, 108I in cluster 114.


Local schedulers (which may refer to either the cluster schedulers 104 or the resource managers 106) are closer to the specific resources 108 and may not allow grid schedulers 102 direct access to the resources. Examples of compute resources include data storage devices such as hard drives and computer processors. The grid level scheduler 102 typically does not own or control the actual resources. Therefore, jobs are submitted from the high level grid-scheduler 102 to a local set of resources with no more permissions that then user would have. This reduces efficiencies and can render the reservation process more difficult.


The heterogeneous nature of the shared resources also causes a reduction in efficiency. Without dedicated access to a resource, the grid level scheduler 102 is challenged with the high degree of variance and unpredictability in the capacity of the resources available for use. Most resources are shared among users and projects and each project varies from the other. The performance goals for projects differ. Grid resources are used to improve performance of an application but the resource owners and users have different performance goals: from optimizing the performance for a single application to getting the best system throughput or minimizing response time. Local policies may also play a role in performance.


Within a given cluster, there is only a concept of resource management in space. An administrator can partition a cluster and identify a set of resources to be dedicated to a particular purpose and another set of resources can be dedicated to another purpose. In this regard, the resources are reserved in advance to process the job. There is currently no ability to identify a set of resources over a time frame for a purpose. By being constrained in space, the nodes 108A, 108B, 108C, if they need maintenance or for administrators to perform work or provisioning on the nodes, have to be taken out of the system, fragmented permanently or partitioned permanently for special purposes or policies. If the administrator wants to dedicate them to particular users, organizations or groups, the prior art method of resource management in space causes too much management overhead requiring a constant adjustment the configuration of the cluster environment and also losses in efficiency with the fragmentation associated with meeting particular policies.


To manage the jobs submissions, a cluster scheduler will employ reservations to insure that jobs will have the resources necessary for processing. FIG. 1B illustrates a cluster/node diagram for a cluster 124 with nodes 120. Time is along the X axis. An access control list 114 (ACL) to the cluster is static, meaning that the ACL is based on the credentials of the person, group, account, class or quality of service making the request or job submission to the cluster. The ACL 114 determines what jobs get assigned to the cluster 110 via a reservation 112 shown as spanning into two nodes of the cluster. Either the job can be allocated to the cluster or it can't and the decision is determined based on who submits the job at submission time. The deficiency with this approach is that there are situations in which organizations would like to make resources available but only in such a way as to balance or meet certain performance goals. Particularly, groups may want to establish a constant expansion factor and make that available to all users or they may want to make a certain subset of users that are key people in an organization and want to give them special services but only when their response time drops below a certain threshold. Given the prior art model, companies are unable to have the flexibility over their cluster resources.


To improve the management of cluster resources, what is needed in the art is a method for a scheduler, a cluster scheduler or cluster workload management system to manage resources in a dimensional addition to space. Furthermore, given the complexity of the cluster environment, what is needed is more power and flexibility in the reservations process.


SUMMARY OF THE INVENTION

Additional features and advantages of the invention will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention. The features and advantages of the invention may be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. These and other features of the present invention will become more fully apparent from the following description and appended claims, or may be learned by the practice of the invention as set forth herein.


The invention includes systems, methods and computer-readable media embodiments. The method aspect of the invention comprises a method of dynamically controlling a reservation of resources within a compute environment to maximize a response time. The compute environment may be a cluster, grid or any environment of a plurality of compute devices. The method comprises receiving from a requestor a request for a reservation of resources in the cluster environment, reserving a first group of resources and evaluating resources within the cluster environment to determine if the response time can be improved. If the response time can be improved, the method comprises canceling the reservation for the first group of resources and reserving a second group of resources to process the request at the improved response time.





BRIEF DESCRIPTION OF THE DRAWINGS

In order to describe the manner in which the above-recited and other advantages and features of the invention can be obtained, a more particular description of the invention briefly described above will be rendered by reference to specific embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only typical embodiments of the invention and are not therefore to be considered to be limiting of its scope, the invention will be described and explained with additional specificity and detail through the use of the accompanying drawings in which:



FIG. 1A illustrates generally a grid scheduler, cluster scheduler, and resource managers interacting with compute nodes;



FIG. 1B illustrates an access control list (ACL) interacting with a group of nodes;



FIG. 2A illustrates a method embodiment of the invention;



FIG. 2B illustrates another method embodiment of the invention;



FIG. 2C illustrates yet another method embodiment of the invention;



FIG. 2D illustrates yet another method embodiment of the invention;



FIG. 2E illustrates yet another method embodiment of the invention;



FIG. 3A illustrates a reservation sandbox;



FIG. 3B illustrates aspects of a reservation sandbox;



FIG. 4 illustrates a rollback reservation;



FIG. 5 illustrates another aspect of the rollback reservation;



FIG. 6 illustrates a dynamic access control list;



FIG. 7 illustrates an interface for creating a reservation;



FIG. 8 illustrates a dynamic reservation migration process; and



FIG. 9 illustrates a method embodiment of the invention.





DETAILED DESCRIPTION OF THE INVENTION

Various embodiments of the invention are discussed in detail below. While specific implementations are discussed, it should be understood that this is done for illustration purposes only. A person skilled in the relevant art will recognize that other components and configurations may be used without parting from the spirit and scope of the invention.


The present invention relates to resource reservations in the context of a cluster environment. The cluster may be operated by a hosting facility, hosting center, a virtual hosting center, data center, grid, cluster and/or utility-based computing environments. Software modules and components operate within a computing environment to manage the reservations of resources. The “system” embodiment of the invention may comprise a computing device that includes the necessary hardware and software components to enable a workload manager or a software module performing the steps of the invention. Such a computing device may include such known hardware elements as one or more central processors, random access memory (RAM), read-only memory (ROM), storage devices such as hard disks, communication means such as a modem or a card to enable networking with other computing devices, a bus that provides data transmission between various hardware components, a keyboard, a display, an operating system and so forth. There is no restriction that the particular system embodiment of the invention have any specific hardware components and any known or future developed hardware configurations are contemplated as within the scope of the invention when the computing device operates as is claimed.


An advance reservation is the mechanism by which the present invention guarantees the availability of a set of resources at a particular time. With an advanced reservation a site has an ability to actually specify how the scheduler should manage resources in both space and time. Every reservation consists of three major components, a list of resources, a timeframe (a start and an end time during which it is active), and an access control list (ACL). These elements are subject to a set of rules. The ACL acts as a doorway determining who or what can actually utilize the resources of the cluster. It is the job of the cluster scheduler to make certain that the ACL is not violated during the reservation's lifetime (i.e., its timeframe) on the resources listed. The ACL governs access by the various users to the resources. The ACL does this by determining which of the jobs, various groups, accounts, jobs with special service levels, jobs with requests for specific resource types or attributes and many different aspects of requests can actually come in and utilize the resources. With the ability to say that these resources are reserved, the scheduler can then enforce true guarantees and can enforce policies and enable dynamic administrative tasks to occur. The system greatly increases in efficiency because there is no need to partition the resources as was previously necessary and the administrative overhead is reduced it terms of staff time because things can be automated and scheduled ahead of time and reserved.


As an example of a reservation, a reservation may specify that node002 is reserved for user John Doe on Friday. The scheduler will thus be constrained to make certain that only John Doe's jobs can use node002 at any time on Friday. Advance reservation technology enables many features including backfill, deadline based scheduling, QOS support, and meta scheduling.


There are several reservation concepts that will be introduced as aspects of the invention. These include dynamic reservations, co-allocating reservation resources of different types, reservations that self-optimize in time, reservations that self-optimization in space, reservations rollbacks and reservation masks. Each of these will be introduced and explained.


Dynamic reservations are reservations that are able to be modified once they are created. FIG. 2A illustrates a dynamic reservation. Attributes of a reservation may change based on a feedback mechanism that adds intelligence as to ideal characteristics of the reservation and how it should be applied as the context of its environment or an entities needs change. One example of a dynamic reservation is a reservation that provides for a guarantee of resources for a project unless that project is not using the resources it has been given. A job associated with a reservation begins in a cluster environment (202). At a given portion of time into processing the job on compute resources, the system receives compute resource usage feedback relative to the job (204). For example, a dynamic reservation policy may apply which says that if the project does not use more than 25% of what it is guaranteed by the time that 50% of its time has expired, then, based on the feedback, the system dynamically modifies the reservation of resources to more closely match the job (206). In other words, the reservation dynamically adjust itself to reserve X % fewer resources for this project, thus freeing up unused resource for others to use.


Another dynamic reservation may perform the following step: if usage of resources provided by a reservation is above 90% with fewer than 10 minutes left in the reservation then the reservation will attempt to add 10% more time to the end of the reservation to help ensure the project is able to complete. In summary, it is the ability for a reservation to receive manual or automatic feedback to an existing reservation in order to have it more accurately match any given needs, whether those be of the submitting entity, the community of users, administrators, etc. The dynamic reservation improves the state of the art by allowing the ACL to the reservation to have a dynamic aspect instead of simply being based on who the requestor is. The reservation can be based on a current level of service or response time being delivered to the requestor.


Another example of a dynamic reservation is consider a user submitting a job and the reservation may need an ACL that requires that the only job that can access these resources are those that have a queue time that is currently exceeded two hours. If the job has sat in the queue for two hours it will then access the additional resources to prevent the queue time for the user from increasing significantly beyond this time frame. You can also key the dynamic reservation off of utilization, off of an expansion factor and other performance metrics of the j ob.


The ACL and scheduler are able to monitor all aspects of the request by looking at the current job inside the queue and how long it has sat there and what the response time target is. It is preferable, although not required, that the scheduler itself determines whether all requirements of the ACL are satisfied. If the requirements are satisfied, the scheduler releases the resources that are available to the job.


The benefits of this model is it makes it significantly easier for a site to balance or provide guaranteed levels of service or constant levels of service for key players or the general populace. By setting aside certain resources and only making them available to the jobs which threaten to violate their quality of service targets it increases the probability of satisfying it.


Another reservation type is a self optimizing reservation in time. This is shown in FIG. 2B. In many cases, people will request resources and request that they be available at a particular time. For example, a person is doing a demonstration and it happens to be from 2:00 pm to 4:00 pm. In many other cases, people will simply have a deadline or simply want processing as early as possible. With a self-optimizing in time reservation, the scheduler is actually able to lock in a set of resources for a particular request and then over time evaluate the cluster resources and determine if it can actually improve on it and improve on the reservation in such a way as to guarantee that it does not lose the resources that it has already made available.


The method aspect of the invention relates to a method of dynamically controlling a reservation of resources within a cluster environment to maximize a response time for processing the reservation. The method comprises receiving from a requestor a request for a reservation of resources in the cluster environment (210), reserving a first group of resources and guaranteeing to the requestor a response time to process the request (212), evaluating resources within the cluster environment to determine if the response time can be improved (214) and determining whether the response time can be improved (216). If the response time can be improved, the system cancels the reservation for the first group of resources and reserves a second group of resources to process the request at the improved response time (218). If the response time cannot be improved, then the system continues to evaluate the resources according to step 214.


The reservation for the first group of resources and the reservation for the second group of resources can overlap in time and/or in terms of the physical resources, or other resources such as software, or license rights, etc. that are reserved. With self-optimizing reservations in time, a particular request may come in request resources that meet the following criteria but the requester prefers resources that meet a more increasingly strict criteria. The scheduler, in finding the reservation, may be able to satisfy the required criteria but not necessarily satisfy all the preferred criteria. Over time, the scheduler, once it has established a reservation that meets the minimum criteria, it can continue to look at newly freed up resources and determine if it can, to a larger and larger extent, satisfy the preferred resource needs as well.


The self optimizing reservation technology is also useful to work around resource failures in the case of a reservation that has already had reserved all the resources it needs and it has a node failure. Other types of resources failures may also be monitored and reservations modified to meet the original promised quality or service or response time to the requestor. For example, one the requestor submits a request for a reservation of resources, and the system promises a certain response time and reserves a group of resources, the system will monitor those resources for failure. If a component of the group of resources fails, such as a node, or a router, or memory or a CPU, the system will seek to modify the reservation by identifying replacement resources that can be included in the reservation such that the promised quality of service can be met. The system can actually continue to locate resources and reallocate resources that are still up and running and be able to satisfy the time frame it originally promised by excluding the failed node and picking up a newly available compute node. This may be performed by modifying the original reservation, or canceling the original reservation and reserving a second group of resources that excludes the failed resource and includes the reallocated working resource such that the requestor maintains his or her quality of service.


The determination of whether the response time can be improved may includes a comparison of a cost of canceling the first group of resources and reserving the second group of resources with the improved response time gained from meeting at least one of the preferred criteria. In this regard, a threshold value may be established that indicates when overall resource efficiency may be improved in that enough of the preferred criteria may be met to overcome the cost of canceling a reservation and establishing a the new reservation of resources.



FIG. 4 illustrates the roll-back reservation mask 402. An important concept in this type of reservation is that it stays ahead of the current time by either a fixed or dynamic amount of time 408. A method embodiment of this reservation is shown in one aspect in FIG. 2B. As the self-optimization registration slides forward at a fixed (or dynamic) distance in the future (230), it analyses reservations, jobs, objects or other resources 404 (232) and determines whether the level of service may be improved (234). If yes, the method comprises creating a new reservation and making the associated changes (236) and then leaves them 406 behind to be processed when the current time catches up.


A self optimizing reservation will only slide forward barring resource failure of the actual compute resources. It does this by, when it makes a query to determine what resources are available, as part of its algorithm, it determines that it has availability to both free resources and the resources it already has reserved. In such a case in then goes and analyzes it, looks at resources that were recently freed by other workload and other reservations that completed early which is actually quite common in a cluster environment, and if it can find that it can improve the level of service delivered to the request or it will actually create the new reservation and will remove the old reservation and adjust things as needed. A self optimizing reservation therefore has the ability to improve any given attribute of service to the submitting entity, community of users, administrators, etc.


Another aspect of the self-optimizing reservation in time is illustrated in FIG. 2D. A reservation is created for a job on a cluster environment (210). Self optimizing may consist of improving response time, starting the reservation earlier if possible, adding additional resources as available, using fewer resources when fewer resources are required. Another example of a self-optimizing reservation is if an organization were trying to guarantee a specific number of compute nodes, say 32 for instance, and at first look the first time that 32 nodes can be available is in three days, but if they were available earlier the using organization would want them earlier. Other reservations and cluster resources are monitored (212). As time goes on, if another reservation ends its usage earlier and now the 32 node reservation can be set up in only 2 days. The system determines whether an optimization of the reservation exits (214). If yes, that an optimization exists, the reservation self-optimizes (216) and changes its start date ahead by one day. If no optimization can occur, the system continues to monitor the cluster resources (212). Yet another example of a self-optimizing reservation is that a group reserves 8 nodes for 4 hours, but would really prefer to get more nodes to get the work done faster. Due to another reservation concluding early, this reservation can now self-optimize and run the jobs on 32 nodes for just one hour and the using group is done four times faster.


Another reservation is the self-terminating reservation. FIG. 2C illustrates this reservation. A self-terminating reservation is a reservation that can cancel itself if certain criteria take place. As a reservation time begins for a job (220), the system monitors for jobs and cluster resources (222). An example of a self-terminating reservation is a reservation that uses an event policy to check that if after 30 minutes no jobs have been submitted against the reservation, or if utilization of the assigned resources is below x then the reservation will cancel itself, thus making those resources available to be used by others. Thus, if monitored events justify terminating the reservation (224), then the reservation terminates itself (226). If the monitored events do not justify canceling the reservation, the system continues to monitor events (222).



FIG. 3A illustrates a standing reservation. In cluster 302, there are standing reservations shown as 304A, 304B and 304C. These reservations show resources allocated and reserved on a periodic basis. These are consuming reservations meaning that cluster resources will be consumed by the reservation.


Another embodiment of reservation is something called a reservation mask, which allows a site to create “sandboxes” in which other guarantees can be made. The most common aspects of this reservation are for grid environments and personal reservation environments. In a grid environment, a remote entity will be requesting resources and will want to use these resources on an autonomous cluster for the autonomous cluster to participate. In many cases it will want to constrain when and where the entities can reserve or utilize resources. One way of doing that is via the reservation mask.



FIG. 3B illustrates the reservation mask shown as creating sandboxes 306A, 306B, 306C in cluster 310 and allows the autonomous cluster to state that only a specific subset of resources can be used by these remote requesters during a specific subset of times. When a requester asks for resources, the scheduler will only report and return resources available within this reservation, after which point the remote entity desires it, he can actually make a consumption reservation and that reservation is guaranteed to be within the reservation mask space. The consumption reservations 312A, 312B, 312C, 312D are shown within the reservation masks.


In cluster 310 the reservation masks operate differently from consuming reservations in that they are enabled to allow personal reservations to be created within the space that is reserved. ACL's are independent inside of a sandbox reservation or a reservation mask in that one can also exclude other requesters out of those spaces so they dedicated for these particular users.


The benefits of this approach include preventing local job starvation, and providing a high level of control to the cluster manager in that he or she can determine exactly when, where, how much and who can use these resources even though he doesn't necessarily know who the requesters are or the combination or quantity of resources they will request. The administrator can determine when, how and where requestors will participate in these grids. A valuable use is in the space of personal reservations which typically involves a local user given the authority to reserve a block of resources for a rigid time frame. Again, with a personal reservation mask, the requests are limited to only allow resource reservation within the mask time frame and mask resource set, providing again the administrator the ability to constrain exactly when and exactly where and exactly how much of resources individual users can reserve for a rigid time frame. The individual user is not known ahead of time but it is known to the system, it is a standard local cluster user.


The reservation masks 306A, 306B and 306C define periodic, personal reservation masks where other reservations in a cluster 310 may be created, i.e., outside the defined boxes. These are provisioning or policy-based reservations in contrast to consuming reservations. In this regard, the resources in this type of reservation are not specifically allocated but the time and space defined by the reservation mask cannot be reserved for other jobs. Reservation masks enable the system to be able to control the fact that resources are available for specific purposes, during specific time frames. The time frames may be either single time frames or repeating time frames to dedicate the resources to meet project needs, policies, guarantees of service, administrative needs, demonstration needs, etc. This type of reservation insures that reservations are managed and scheduled in time as well as space. Boxes 308A, 308B, 308C and 308D represent non-personal reservation masks. They have the freedom to be placed anywhere in cluster including overlapping some or all of the reservation masks 306A, 306B, 306C. Overlapping is allowed when the personal reservation mask was setup with a global ACL. A global ACL is an ACL that anyone can use. It is wide open in the sense that anyone can take advantage of the resources within that space. To prevent the possibility of an overlap of a reservation mask by a non-personal reservation, the administrator can set an ACL to constrain it is so that only personal consumption reservations are inside. These personal consumption reservations are shown as boxes 312B, 312A, 312C, 312D which are constrained to be within the personal reservation masks 306A, 306B, 306C. The 308A, 308B, 308C and 308D reservations, if allowed, can go anywhere within the cluster 310 including overlapping the other personal reservation masks. The result is the creation of a “sandbox” where only personal reservations can go without in any way constraining the behavior of the scheduler to schedule other requests.


Another reservation type is the reservation roll-back shown in FIG. 4. This reservation has particular application for enforcing policies or allowing support for service level guarantees in service level agreements. A level of service guarantee allows a site or cluster to guarantee that a particular consumer or organization or type of credential is guaranteed a certain quantity of resources within a certain amount of time. The standard way to provide those guarantees would be to dedicate a block of resources that satisfy the needs and would be statically and rigidly partitioned so that no one else could access it. The request of that organization could not extend beyond the bounds of the dedicated block.


With the present invention regarding the reservation roll-back, an administrator can create a reservation 402 which enforces its policy and continues to float in time a certain distance 408 ahead of the current time. Typically the rectangular area of the reservation has a height that corresponds to guaranteed throughput when processing jobs and the horizontal distance that corresponds to the length in time of the reservation. The reservation 402 may correspond to a certain amount of time according to a service level agreement, such as 3 or 4 months for example. The reservation 402 may extend into infinity as well if there is no defined ending time. The reservation 402 is a provisioning reservation and maintains the time offset 402 to the current time.


To illustrate the reservation roll-back, consider a service level agreement with a company to have twenty resources available within one hour of the request for the resources and that they can make the request anytime. The time offset 408 can then be set to one hour and the company will never will they wait more than one hour to get up to twenty resources. The reservation 402 monitors the resources and when a request is made for resources, consumption reservations 404 are allocated and left behind 406 as the roll-back reservation maintains its offset.


An implementation with reservation rollback would allow a site to set up basically a floating reservation that extends from one hour in the future until a time further in the future, such as 4 or 8 hours in the future, and continues to slide forward in time. The reservation 402 will only allow jobs from this organization can drop down requests or reserve host resources underneath the reservation. As time moves forward, the reservation slides forward in time so it always maintains a constant distance in the future allowing these guarantees 404 to be created and maintained 406 on the cluster.


The time offset 408 may be static or dynamic. A static offset 408 will maintain a constant offset time, such as one hour into the future. The static offset will likely be set by a service level agreement wherein a company requests that the resources become available within an hour. The offset 408 may also by dynamic. There may be requests in the service level agreement where under a given event or set of events, the offset would change wherein the reservation slides closer or farther away from the current time to provide a guarantee of resources within ½ (instead of 1 hour) or 2 hours in the future. There are a variety of ways to vary the offset. One can be to simply cancel the current sliding reservation and create a new reservation at a different offset. Another way would be to maintain the current reservation but slide it closer or farther away from the current time. The factors that adjust the dynamic nature of the offset may be based on company requests, the nature and use of the cluster resources, the time the request is made, historical information, and so forth. For example, if the request for resources is made at midnight on a Friday night, perhaps instead of the 1 hour availability of resources, the hosting center analyzes the cluster resources and the time of the request and determines that it can deliver the resources in ½. The company may want a flexible offset where if the request is made during a block of time such as between 3-4:30 pm (near the end of the work day) that the offset be shorted so that the job can be processed sooner. The modifications to the offset may be automatic based on a feedback loop of information or may be adjustable by an administrator.


The reservation rollback policy mask is stackable allowing multiple different types of service or service level agreements to be simultaneously satisfied and share a collection of resources. This feature is illustrated in FIG. 5. A reservation 502 is shown and can generally be considered as an aggregation of requests from various masks 504, 506, 508510. These are aggregated into one space 502 which will then allow reservations to be created on a first come first serve basis, or based on other factors. If these reservation masks 504, 506, 508 and 510 are stacked with individual offsets from the current time (not shown), the administrator can allow the masks to be partitioned among consumers. A useful component of this stackable approach is the capability to have an enveloping reservation 502 created with a total quantity of resource and rollback time offset 408 and a duration to the end of the SLA. Once that reservation space is established or paid for, as a service, the hosting center sub-partitions the space using reservation to provide service guarantees, response time guarantees, quantity or resources guarantees taking advantage of the stacking capability.


A company may therefore establish the enveloping reservation 502 and request from the hosting center that they partition the space according to various organizations within the enveloping reservation 502. This eliminates the need for a large entity to have its own group of clusters of computer.


As mentioned above, the present application is related to U.S. patent application Ser. No. 10/530,583, which was incorporated herein by reference. The following paragraphs, modified for formatting, are from that application.


A system and method are disclosed for dynamically reserving resources within a cluster environment. The method embodiment of the invention comprises receiving a request for resources in the cluster environment, monitoring events after receiving the request for resources and based on the monitored events, dynamically modifying at least one of the request for resources and the cluster environment.


The present invention relates to reservations in a compute environment and more specifically to a system and method of providing advanced reservations to resources within a compute environment such as a cluster.


There are challenges in the complex process of managing the consumption of resources within a compute environment such as a grid, compute farm or cluster of computers. Grid computing may be defined as coordinated resource sharing and problem solving in dynamic, multi-institutional collaborations. Many computing projects require much more computational power and resources than a single computer may provide. Networked computers with peripheral resources such as printers, scanners, I/O devices, storage disks, scientific devices and instruments, etc. may need to be coordinated and utilized to complete a task. The term compute resource generally refers to computer processors, network bandwidth, and any of these peripheral resources as well. A compute farm may comprise a plurality of computers coordinated for such purposes of handling Internet traffic. The web search website Google® had a compute farm used to process its network traffic and Internet searches.


Grid/cluster resource management generally describes the process of identifying requirements, matching resources to applications, allocating those resources, and scheduling and monitoring grid resources over time in order to run grid applications or jobs submitted to the compute environment as efficiently as possible. Each project or job will utilize a different set of resources and thus is typically unique. For example, a job may utilize computer processors and disk space, while another job may require a large amount of network bandwidth and a particular operating system. In addition to the challenge of allocating resources for a particular job or a request for resources, administrators also have difficulty obtaining a clear understanding of the resources available, the current status of the compute environment and available resources, and real-time competing needs of various users. One aspect of this process is the ability to reserve resources for a job. A cluster manager will seek to reserve a set of resources to enable the cluster to process a job at a promised quality of service.


General background information on clusters and grids may be found in several publications. See, e.g., Grid Resource Management, State of the Art and Future Trends, Jarek Nabrzyski, Jennifer M. Schopf, and Jan Weglarz, Kluwer Academic Publishers, 2004; and Beowulf Cluster Computing with Linux, edited by William Gropp, Ewing Lusk, and Thomas Sterling, Massachusetts Institute of Technology, 2003.


It is generally understood herein that the terms grid and cluster are interchangeable, although they have different connotations. For example, when a grid is referred to as receiving a request for resources and the request is processed in a particular way, the same method may also apply to other compute environments such as a cluster or a compute farm. A cluster is generally defined as a collection of compute nodes organized for accomplishing a task or a set of tasks. In general, a grid will comprise a plurality of clusters as will be shown in FIG. 1A. Several general challenges exist when attempting to maximize resources in a grid. First, there are typically multiple layers of grid and cluster schedulers. A grid 100 generally comprises a group of clusters or a group of networked computers. The definition of a grid is very flexible and may mean a number of different configurations of computers. The introduction here is meant to be general given the variety of configurations that are possible. A grid scheduler 102 communicates with a plurality of cluster schedulers 104A, 104B and 104C. Each of these cluster schedulers communicates with a respective resource manager 106A, 106B or 106C. Each resource manager communicates with a respective series of compute resources shown as nodes 108A, 108B, 108C in cluster 110, nodes 108D, 108E, 108F in cluster 112 and nodes 108G, 108H, 108I in cluster 114.


Local schedulers (which may refer to either the cluster schedulers 104 or the resource managers 106) are doser to the specific resources 108 and may not allow grid schedulers 102 direct access to the resources. The grid level scheduler 102 typically does not own or control the actual resources. Therefore, jobs are submitted from the high level grid-scheduler 102 to a local set of resources with no more permissions that then user would have. This reduces efficiencies and can render the reservation process more difficult.


The heterogeneous nature of the shared compute resources also causes a reduction in efficiency. Without dedicated access to a resource, the grid level scheduler 102 is challenged with the high degree of variance and unpredictability in the capacity of the resources available for use. Most resources are shared among users and projects and each project varies from the other. The performance goals for projects differ. Grid resources are used to improve performance of an application but the resource owners and users have different performance goals: from optimizing the performance for a single application to getting the best system throughput or minimizing response time. Local policies may also play a role in performance.


Within a given cluster, there is only a concept of resource management in space. An administrator can partition a cluster and identify a set of resources to be dedicated to a particular purpose and another set of resources can be dedicated to another purpose. In this regard, the resources are reserved in advance to process the job. There is currently no ability to identify a set of resources over a time frame for a purpose. By being constrained in space, the nodes 108A, 108B, 108C, if they need maintenance or for administrators to perform work or provisioning on the nodes, have to be taken out of the system, fragmented permanently or partitioned permanently for special purposes or policies. If the administrator wants to dedicate them to particular users, organizations or groups, the prior art method of resource management in space causes too much management overhead requiring a constant adjustment the configuration of the cluster environment and also losses in efficiency with the fragmentation associated with meeting particular policies.


To manage the jobs submissions or requests for resources within a cluster, a cluster scheduler will employ reservations to insure that jobs will have the resources necessary for processing. FIG. 1B illustrates a cluster/node diagram for a cluster 124 with nodes 120. Time is along the X axis. An access control list 114 (ACL) to the cluster is static, meaning that the ACL is based on the credentials of the person, group, account, class or quality of service making the request or job submission to the cluster. The ACL 114 determines what jobs get assigned to the cluster 110 via a reservation 112 shown as spanning into two nodes of the cluster. Either the job can be allocated to the cluster or it can't and the decision is determined based on who submits the job at submission time. The deficiency with this approach is that there are situations in which organizations would like to make resources available but only in such a way as to balance or meet certain performance goals. Particularly, groups may want to establish a constant expansion factor and make that available to all users or they may want to make a certain subset of users that are key people in an organization and want to give them special services but only when their response time drops below a certain threshold. Given the prior art model, companies are unable to have the flexibility over their cluster resources.


To improve the management of compute resources, what is needed in the art is a method for a scheduler, such as a grid scheduler, a cluster scheduler or cluster workload management system to manage resources more efficiently. Furthermore, given the complexity of the cluster environment, what is needed is more power and flexibility in the reservations process.


The invention relates to systems, methods and computer-readable media for dynamically modifying either compute resources or a reservation for compute resources within a compute environment such as a grid or a cluster. In one aspect of the invention, a method of dynamically modifying resources within a compute environment comprises receiving a request for resources in the compute environment, monitoring events after receiving the request for resources and based on the monitored events, dynamically modifying at least one of the request for resources and the compute environment.


The invention enables an improved matching between a reservation and jobs submitted for processing in the compute environment. A benefit of the present invention is that the compute environment and the reservation or jobs submitted under the reservation will achieve a better fit. The closer the fit between jobs, reservations and the compute resources provides increased efficiency of the resources.


The present invention relates to reservations of resources within the context of a compute environment. One example of a compute environment is a cluster. The cluster may be, for example, a group of computing devices operated by a hosting facility, a hosting center, a virtual hosting center, a data center, grid and/or utility-based computing environments. Every reservation consists of three major components: a set of resources, a timeframe, and an access control list (ACL). Additionally, a reservation may also have a number of optional attributes controlling its behavior and interaction with other aspects of scheduling. A reservation's ACL specifies which jobs can use the reservation. Only jobs which meet one or more of a reservation's access criteria are allowed to use the reserved resources during the reservation timeframe. The reservation access criteria comprises, in one example, at least following: users, groups, accounts, classes, quality of service (QOS) and job duration. A job may be any venue or end of consumption of resource for any broad purpose, whether it be for a batch system, direct volume access or other service provisioning.


A workload manager, or scheduler, will govern access to the compute environment by receiving requests for reservations of resources and creating reservations for processing jobs. A workload manager functions by manipulating five primary, elementary objects. These are jobs, nodes, reservations, QOS structures, and policies. In addition to these, multiple minor elementary objects and composite objects are also utilized. These objects are also defined in a scheduling dictionary.


A workload manager may operate on a single computing device or multiple computing devices to manage the workload of a compute environment The “system” embodiment of the invention may comprise a computing device that includes the necessary hardware and software components to enable a workload manager or a software module performing the steps of the invention. Such a computing device may include such known hardware elements as one or more central processors, random access memory (RAM), read-only memory (ROM), storage devices such as hard disks, communication means such as a modem or a card to enable networking with other computing devices, a bus that provides data transmission between various hardware components, a keyboard, a display, an operating system and so forth. There is no restriction that the particular system embodiment of the invention have any specific hardware components and any known or future developed hardware configurations are contemplated as within the scope of the invention when the computing device operates as is claimed.


Job information is provided to the workload manager scheduler from a resource manager such as Loadleveler, the Portable Batch System (PBS), Wiki or Platform's LSF products. Those of skill in the art will be familiar with each of these software products and their variations. Job attributes include ownership of the job, job state, amount and type of resources required by the job, required criteria (I need this job finished in 1 hour), preferred criteria (I would like this job to complete in ½ hour) and a wallclock limit, indicating how long the resources are required. A job consists of one or more requirements each of which requests a number of resources of a given type. For example, a job may consist of two requirements, the first asking for ‘1 IBM node with at least 512 MB of RAM’ and the second asking for ‘24 IBM nodes with at least 128 MB of RAM’. Each requirement consists of one or more tasks where a task is defined as the minimal independent unit of resources. A task is a collection of elementary resources which must be allocated together within a single node. For example, a task may consist of one processor, 512 MB or memory, and 2 GB of local disk. A task may also be just a single processor. In symmetric multiprocessor (SMP) environments, however, users may wish to tie one or more processors together with a certain amount of memory and/or other resources. A key aspect of a task is that the resources associated with the task must be allocated as an atomic unit, without spanning node boundaries. A task requesting 2 processors cannot be satisfied by allocating 2 uni-processor nodes, nor can a task requesting 1 processor and 1 GB of memory be satisfied by allocating 1 processor on one node and memory on another.


A job requirement (or req) consists of a request for a single type of resources. Each requirement consists of the following components: (1) a task definition is a specification of the elementary resources which compose an individual task; (2) resource constraints provide a specification of conditions which must be met in order for resource matching to occur. Only resources from nodes which meet all resource constraints may be allocated to the job requirement; (3) a task count relates to the number of task instances required by the requirement; (4) a task List is a list of nodes on which the task instances have been located; and (5) requirement statistics are statistics tracking resource utilization.


As far as the workload manager is concerned, a node is a collection of resources with a particular set of associated attributes. In most cases, it fits nicely with the canonical world view of a node such as a PC cluster node or an SP node. In these cases, a node is defined as one or more CPU's, memory, and possibly other compute resources such as local disk, swap, network adapters, software licenses, etc. Additionally, this node will described by various attributes such as an architecture type or operating system. Nodes range in size from small uni-processor PC's to large SMP systems where a single node may consist of hundreds of CPU's and massive amounts of memory.


Information about nodes is provided to the scheduler chiefly by the resource manager. Attributes include node state, configured and available resources (i.e., processors, memory, swap, etc.), run classes supported, etc.


Policies are generally specified via a configuration file and serve to control how and when jobs start. Policies include, but are not limited to, job prioritization, fairness policies, fairshare configuration policies, and scheduling policies. Jobs, nodes, and reservations all deal with the abstract concept of a resource. A resource in the workload manager world is one of the following: (1) processors which are specified with a simple count value; (2) memory such as real memory or ‘RAM’ is specified in megabytes (MB); (3) swap which is virtual memory or ‘swap’ is specified in megabytes (MB); and (4) disk space such as a local disk is specified in megabytes (MB) or gigabytes (GB). In addition to these elementary resource types, there are two higher level resource concepts used within workload manager. These are the task and the processor equivalent (PE).


In a workload manager, jobs or reservations that request resources make such a request in terms of tasks typically using a task count and a task definition. By default, a task maps directly to a single processor within a job and maps to a fill node within reservations. In all cases, this default definition can be overridden by specifying a new task definition. Within both jobs and reservations, depending on task definition, it is possible to have multiple tasks from the same job mapped to the same node. For example, a job requesting 4 tasks using the default task definition of 1 processor per task, can be satisfied by two dual processor nodes.


The concept of the PE arose out of the need to translate multi-resource consumption requests into a scalar value. It is not an elementary resource, but rather, a derived resource metric. It is a measure of the actual impact of a set of requested resources by a job on the total resources available system wide. It is calculated as:

PE=MAX(ProcsRequestedByJob/TotalConfiguredProcs, MemoryRequestedByJob/TotalConfiguredMemory, DiskRequestedByJob/TotalConfiguredDisk, SwapRequestedByJob/TotalConfiguredSwap)*TotalConfiguredProcs


For example, say a job requested 20% of the total processors and 50% of the total memory of a 128 processor MPP system. Only two such jobs could be supported by this system. The job is essentially using 50% of all available resources since the system can only be scheduled to its most constrained resource, in this case memory. The processor equivalents for this job should be 50% of the PE=64.


A further example will be instructive. Assume a homogeneous 100 node system with 4 processors and 1 GB of memory per node. A job is submitted requesting 2 processors and 768 MB of memory. The PE for this job would be calculated as:

PE=MAX(2/(100*4),768/(100*1024))*(100*4)=3.


This result makes sense since the job would be consuming ¾ of the memory on a 4 processor node. The calculation works equally well on homogeneous or heterogeneous systems, uni-processor or large way SMP systems.


A class (or queue) is a logical container object which can be used to implicitly or explicitly apply policies to jobs. In most cases, a class is defined and configured within the resource manager and associated with one or more of the attributes or constraints shown in Table 1 below.









TABLE 1







Attributes of a Class








Attribute
Description





Default Job
A queue may be associated with a default job duration,


Attributes
default size, or default resource requirements


Host
A queue may constrain job execution to a


Constraints
particular set of hosts


Job
A queue may constrain the attributes of jobs which may


Constraints
submitted including setting limits such as max wallclock



time, minimum number of processors, etc.


Access List
A queue may constrain who may submit jobs into it



based on user lists, group lists, etc.


Special
A queue may associate special privileges with jobs


Access
including adjusted job priority.









As stated previously, most resource managers allow full class configuration within the resource manager. Where additional class configuration is required, the CLASSCFG parameter may be used. The workload manager tracks class usage as a consumable resource allowing sites to limit the number of jobs using a particular class. This is done by monitoring class initiators which may be considered to be a ticket to run in a particular class. Any compute node may simultaneously support several types of classes and any number of initiators of each type. By default, nodes will have a one-to-one mapping between class initiators and configured processors. For every job task run on the node, one class initiator of the appropriate type is consumed. For example, a three processor job submitted to the class batch will consume three batch class initiators on the nodes where it is run.


Using queues as consumable resources allows sites to specify various policies by adjusting the class initiator to node mapping. For example, a site running serial jobs may want to allow a particular 8 processor node to run any combination of batch and special jobs subject to the following constraints:

    • only 8 jobs of any type allowed simultaneously
    • no more than 4 special jobs allowed simultaneously


To enable this policy, the site may set the node's MAXJOB policy to 8 and configure the node with 4 special class initiators and 8 batch class initiators. Note that in virtually all cases jobs have a one-to-one correspondence between processors requested and class initiators required. However, this is not a requirement and, with special configuration sites may choose to associate job tasks with arbitrary combinations of class initiator requirements.


In displaying class initiator status, workload manager signifies the type and number of class initiators available using the format [<CLASSNAME>:<CLASSCOUNT>]. This is most commonly seen in the output of node status commands indicating the number of configured and available class initiators, or in job status commands when displaying class initiator requirements.


Nodes can also be configured to support various arbitrary resources. Information about such resources can be specified using the NODECFG parameter. For example, a node may be configured to have “256 MB RAM, 4 processors, 1 GB Swap, and 2 tape drives”.


We next turn to the concept of reservations. There are several types of reservations which sites typically deal with. The first, administrative reservations, are typically one-time reservations created for special purposes and projects. These reservations are created using a command that sets a reservation. These reservations provide an integrated mechanism to allow graceful management of unexpected system maintenance, temporary projects, and time critical demonstrations. This command allows an administrator to select a particular set of resources or just specify the quantity of resources needed. For example, an administrator could use a regular expression to request a reservation be created on the nodes ‘blue0[1-9]’ or could simply request that the reservation locate the needed resources by specifying a quantity based request such as ‘TASKS==20’.


Another type of reservation is called a standing reservation. This is shown in FIG. 2A. A standing reservation is useful for recurring needs for a particular type of resource distribution. For example, a site could use a standing reservation to reserve a subset of its compute resources for quick turnaround jobs during business hours on Monday thru Friday. Standing reservations are created and configured by specifying parameters in a configuration file.


As shown in FIG. 3A, the compute environment 302 includes standing reservations shown as 304A, 304B and 304C. These reservations show resources allocated and reserved on a periodic basis. These are, for example, consuming reservations meaning that cluster resources will be consumed by the reservation. These reservations are specific to a user or a group of users and allow the reserved resources to be also customized specific to the workload submitted by these users or groups. For example, one aspect of the invention is that a user may have access to reservation 304A and not only submit jobs to the reserved resources but request, perhaps for optimization or to meet preferred criteria as opposed to required criteria, that the resources within the reservation be modified by virtual partitioning or some other means to accommodate the particular submitted job. In this regard, this embodiment of the invention enables the user to submit and perhaps request modification or optimization within the reserved resources for that particular job. There may be an extra charge or debit of an account of credits for the modification of the reserved resources. The modification of resources within the reservation according to the particular job may also be performed based on a number of factors discussed herein, such as criteria, class, quality of service, policies etc.


Standing reservations build upon the capabilities of advance reservations to enable a site to enforce advanced usage policies in an efficient manner. Standing reservations provide a superset of the capabilities typically found in a batch queuing system's class or queue architecture. For example, queues can be used to allow only particular types of jobs access to certain compute resources. Also, some batch systems allow these queues to be configured so that they only allow this access during certain times of the day or week. Standing reservations allow these same capabilities but with greater flexibility and efficiency than is typically found in a normal queue management system.


Standing Reservations provide a mechanism by which a site can dedicate a particular block of resources for a special use on a regular daily or weekly basis. For example, node X could be dedicated to running jobs only from users in the accounting group every Friday from 4 to 10 PM. A standing reservation is a powerful means of controlling access to resources and controlling turnaround of jobs.


Another embodiment of reservation is something called a reservation mask, which allows a site to create “sandboxes” in which other guarantees can be made. The most common aspects of this reservation are for grid environments and personal reservation environments. In a grid environment, a remote entity will be requesting resources and will want to use these resources on an autonomous cluster for the autonomous cluster to participate. In many cases it will want to constrain when and where the entities can reserve or utilize resources. One way of doing that is via the reservation mask.



FIG. 3B illustrates the reservation mask shown as creating sandboxes 306A, 306B, 306C in compute environment 310 and allows the autonomous cluster to state that only a specific subset of resources can be used by these remote requesters during a specific subset of times. When a requester asks for resources, the scheduler will only report and return resources available within this reservation, after which point the remote entity desires it, it can actually make a consumption reservation and that reservation is guaranteed to be within the reservation mask space. The consumption reservations 312A, 312B, 312C, 312D are shown within the reservation masks.


Another concept related to reservations is the personal reservation and/or the personal reservation mask. In compute environment 310, the reservation masks operate differently from consuming reservations in that they are enabled to allow personal reservations to be created within the space that is reserved. ACL's are independent inside of a sandbox reservation or a reservation mask in that you can also exclude other requesters out of those spaces so they're dedicated for these particular users.


One benefit of the personal reservation approach includes preventing local job starvation, and providing a high level of control to the cluster manager in that he or she can determine exactly when, where, how much and who can use these resources even though he doesn't necessarily know who the requesters are or the combination or quantity of resources they will request. The administrator can determine when, how and where requestors will participate in these clusters or grids. A valuable use is in the space of personal reservations which typically involves a local user given the authority to reserve a block of resources for a rigid time frame. Again, with a personal reservation mask, the requests are limited to only allow resource reservation within the mask time frame and mask resource set, providing again the administrator the ability to constrain exactly when and exactly where and exactly how much of resources individual users can reserve for a rigid time frame. The individual user is not known ahead of time but it is known to the system, it is a standard local cluster user.


The reservation masks 306A, 306B and 306C define periodic, personal reservation masks where other reservations in the compute environment 310 may be created, i.e., outside the defined boxes. These are provisioning or policy-based reservations in contrast to consuming reservations. In this regard, the resources in this type of reservation are not specifically allocated but the time and space defined by the reservation mask cannot be reserved for other jobs. Reservation masks enable the system to be able to control the fact that resources are available for specific purposes, during specific time frames. The time frames may be either single time frames or repeating time frames to dedicate the resources to meet project needs, policies, guarantees of service, administrative needs, demonstration needs, etc. This type of reservation insures that reservations are managed and scheduled in time as well as space. Boxes 308A, 308B, 308C and 308D represent non-personal reservation masks. They have the freedom to be placed anywhere in cluster including overlapping some or all of the reservation masks 306A, 306B, 306C. Overlapping is allowed when the personal reservation mask was setup with a global ACL. To prevent the possibility of an overlap of a reservation mask by a non-personal reservation, the administrator can set an ACL to constrain it is so that only personal consumption reservations are inside. These personal consumption reservations are shown as boxes 312B, 312A, 312C, 312D which are constrained to be within the personal reservation masks 306A, 306B, 306C. The 308A, 308B, 308C and 308D reservations, if allowed, can go anywhere within the cluster 310 including overlapping the other personal reservation masks. The result is the creation of a “sandbox” where only personal reservations can go without in any way constraining the behavior of the scheduler to schedule other requests.


All reservations possess a start and an end time which define the reservation's active time. During this active time, the resources within the reservation may only be used as specified by the reservation ACL. This active time may be specified as either a start/end pair or a start/duration pair. Reservations exist and are visible from the time they are created until the active time ends at which point they are automatically removed.


For a reservation to be useful, it must be able to limit who or what can access the resources it has reserved. This is handled by way of an access control list, or ACL. With reservations, ACL's can be based on credentials, resources requested, or performance metrics. In particular, with a standing reservation, the attributes userlist, grouplist, accountlist, classlist, qoslist, jobattrlist, proclimit, timelimit and others may be specified.



FIG. 6 illustrates an aspect of the present invention that allows the ACL 606 for the reservation 604 to have a dynamic aspect instead of simply being based on who the requester is. The ACL decision-making process is based at least in part on the current level of service or response time that is being delivered to the requester. To illustrate the operation of the ACL 606, assume that a user 608 submits a job 614 to a queue 610 and that the ACL 606 reports that the only job that can access these resources 602 are those that have a queue time that currently exceeds two hours. The resources 602 are shown with resources N on the y axis and time on the x axis. If the job 614 has sat in the queue 610 for two hours it will then access the additional resources to prevent the queue time for the user 608 from increasing significantly beyond this time frame. The decision to allocate these additional resources can be keyed off of utilization of an expansion factor and other performance metrics of the job. For example, the reservation 604 may be expanded or contracted or migrated to cover a new set of resources.


Whether or not an ACL 606 is satisfied is typically and preferably determined the scheduler 104A. However, there is no restriction in the principle of the invention regarding where or on what node in the network the process of making these allocation of resource decisions occurs. The scheduler 104A is able to monitor all aspects of the request by looking at the current job 614 inside the queue 610 and how long it has sat there and what the response time target is and the scheduler itself determines whether all requirements of the ACL 606 are satisfied. If requirements are satisfied, it releases the resources that are available to the job 614. A job 614 that is located in the queue and the scheduler communicating with the scheduler 104A. If resources are allocated, the job 614 is taken from the queue 610 and inserted into the reservation 614 in the cluster 602.


An example benefit of this model is that it makes it significantly easier for a site to balance or provide guaranteed levels of service or constant levels of service for key players or the general populace. By setting aside certain resources and only making them available to the jobs which threaten to violate their quality of service targets, the system increases the probability of satisfying targets.


When specifying which resources to reserve, the administrator has a number of options. These options allow control over how many resources are reserved and where they are reserved at. The following reservation attributes allow the administrator to define resources.


An important aspect of reservations is the idea of a task. The scheduler uses the task concept extensively for its job and reservation management. A task is simply an atomic collection of resources, such as processors, memory, or local disk, which must be found on the same node. For example, if a task requires 4 processors and 2 GB of memory, the scheduler must find all processors AND memory on the same node; it cannot allocate 3 processors and 1 GB on one node and 1 processor and 1 GB of memory on another node to satisfy this task. Tasks constrain how the scheduler must collect resources for use in a standing reservation, however, they do not constrain the way in which the scheduler makes these cumulative resources available to jobs. A job can use the resources covered by an accessible reservation in whatever way it needs. If reservation X allocated 6 tasks with 2 processors and 512 MB of memory each, it could support job Y which requires 10 tasks of 1 processor and 128 MB of memory or job Z which requires 2 tasks of 4 processors and 1 GB of memory each. The task constraints used to acquire a reservation's resources are completely transparent to a job requesting use of these resources. Using the task description, the taskcount attribute defines how many tasks must be allocated to satisfy the reservation request. To create a reservation, a taskcount and/or a hostlist may be specified.


A hostlist constrains the set of resource which are available to a reservation. If no taskcount is specified, the reservation will attempt to reserve one task on each of the listed resources. If a taskcount is specified which requests fewer resources than listed in the hostlist, the scheduler will reserve only the number of tasks from the hostlist specified by the taskcount attribute. If a taskcount is specified which requests more resources than listed in the hostlist, the scheduler will reserve the hostlist nodes first and then seek additional resources outside of this list.


Reservation flags allow specification of special reservation attributes or behaviors. Supported flags are listed in table 2 below.










TABLE 2





Flag Name
Description







BESTEFFORT
N/A


BYNAME
reservation will only allow access to jobs which meet reservation



ACL's and explicitly request the resources of this reservation using



the job ADVRES flag


IGNRSV
request will ignore existing resource reservations allowing the



reservation to be forced onto available resources even if this



conflicts with other reservations.


OWNERPREEMPT
job's by the reservation owner are allowed to preempt non-owner



jobs using reservation resources


PREEMPTEE
Preempts a job or other object


SINGLEUSE
reservation is automatically removed after completion of the first job



to use the reserved resources


SPACEFLEX
reservation is allowed to adjust resources allocated over time in an



attempt to optimize resource utilization


TIMEFLEX
reservation is allowed to adjust the reserved timeframe in an attempt



to optimize resource utilization









Reservations must explicitly request the ability to float for optimization purposes by using a flag such as the SPACEFLEX flag. The reservations may be established and then identified as self-optimizing in either space or time. If the reservation is flagged as such, then after the reservation is created, conditions within the compute environment may be monitored to provide feedback on where optimization may occur. If so justified, a reservation may migrate to a new time or migrate to a new set of resources that are more optimal than the original reservation.



FIG. 7 illustrates a reservation creation window that includes the use of the flags in Table 2. A user Scott input reservation information in a variety of fields for name, partition, node features and floating reservation. Each of these input fields includes a drop-down menu to enable the selection of options easy. An access control list input field allows the user to input an account, class/queue, user, group and QoS information. Resources may be assigned and searched and tasks created and reservation flags set, such as best effort, single use, preemptee, time flex, by name, owner preempt, space flex, exclusive and force. These flags set parameters that may cause the reservation to be optimized such as in time or space where it migrates to a new time or over new resources based on monitored events or other feedback.


A reservation time-frame may also be input such as one, daily, weekly, with start and end times for the reservation. Menu drop down calendars and docks are available for easily enabling the user to view and graphically input and select the various timeframe parameters. Event triggers may also be input wherein the user can create one or more triggers associated with the reservation. As generally shown in FIG. 7, the use of a graphical interface simplifies the reservation-creation process for the administrator or user.



FIG. 8 illustrates a particular instance where the user has identified the time-flex and space-flex flags within the reservation. A window 800 identifies three reservations 802 for 96 nodes, 804 for 128 nodes and 806 for 256 nodes. The height of each of these reservations generally relates to resources reserved, such as a number of processors reserved or processors and disk space. The X-axis represents time. Reservation 808 represents a reservation in the future that will in a position to receive submitted jobs. Assume that reservation 806 which was scheduled to end at time T2 has finished early at time T1. Also assume that reservation 808 is flagged for time flex and space flex. In this case, based on the monitored event that reservation 806 has ended early, the system would cause reservation 808 to migrate in time (and space in this example) to position 810. This represents a movement of the reservation to a new time and a new set of resources. If reservation 804 ends early, and reservation 808 migrates to position 820, that would represent a migration in time (to an earlier time) but not in space. This would be enabled by the time-flex flag being set wherein the migration would seek to create a new reservation at the earliest time possible and/or according to available resources. The new time may be based on criteria to minimize the time for the reservation or to maximize usage of the overall resources or better performance of the compute environment.


Next, assume that reservation 808 is for 128 processors and reservation 806 is for 256 processors and reservation 808 is flagged for space flex. If reservation 806 ends are time T1 instead of time T2, then reservation 808 may migrate to position 812 to a reservation of 256 processors. The time frame of the starting and ending time may be the same but the reservation has migrated in space and thus been optimized.


In another aspect of reservation migration, assume that reservation 808 is set but that a node or a group of nodes that are part of the reservation go down or are projected to fail as represented by 818. In this regard, reservation 808 may be enabled to migrate as shown by 816 and 808 to cover new resources but to accommodate for the nodes that are no longer available.


Standing reservations allow resources to be dedicated for particular uses. This dedication can be configured to be permanent or periodic, recurring at a regular time of day and/or time of week. There is extensive applicability of standing reservations for everything from daily dedicated job runs to improved use of resources on weekends. All standing reservation attributes are specified via a parameter using available attributes


In addition to standing and administrative reservations, a workload manager according to the invention can also create priority reservations. These reservations are used to allow the benefits of out-of-order execution (such as is available with a backfill feature) without the side effect of job starvation. Starvation can occur in any system where the potential exists for a job to be overlooked by the scheduler for an indefinite period. In the case of backfill, small jobs may continue to be run on available resources as they become available while a large job sits in the queue never able to find enough nodes available simultaneously to run on. To avoid such situations, priority reservations are created for high priority jobs which cannot run immediately. When making these reservations, the scheduler determines the earliest time the job could start, and then reserves these resources for use by this job at that future time. By default, only the highest priority job will receive a priority reservation. However, this behavior is configurable via a reservation depth policy. The workload manager's default behavior of only reserving the highest priority job allows backfill to be used in a form known as liberal backfill. This liberal backfill tends to maximize system utilization and minimize overall average job turnaround time. However, it does lead to the potential of some lower priority jobs being indirectly delayed and may lead to greater variance in job turnaround time. A reservation depth parameter can be set to a very large value, essentially enabling what is called conservative backfill where every job which cannot run is given a reservation. Most sites prefer the liberal backfill approach associated with the default reservation depth 1 or select a slightly higher value. It is important to note that to prevent starvation in conjunction with reservations, monotonically increasing priority factors such as queuetime or job x-factor should be enabled.


Another important consequence of backfill and reservation depth is its affect on job priority. In the workload manager, all jobs are preferably prioritized. Backfill allows jobs to be run out of order and thus, to some extent, job priority to be ignored. This effect, known as ‘priority dilution’ can cause many site policies implemented via workload manager prioritization policies to be ineffective. Setting the reservation depth parameter to a higher value will give job priority ‘more teeth’ at the cost of slightly lower system utilization. This lower utilization results from the constraints of these additional reservations, decreasing the scheduler's freedom and its ability to find additional optimizing schedules. Anecdotal evidence indicates that these utilization losses are fairly minor, rarely exceeding 8%.


In addition to the reservation depth parameter, sites also have the ability to control how reservations are maintained. The workload manager's dynamic job prioritization allows sites to prioritize jobs so that their priority order can change over time. It is possible that one job can be at the top of the priority queue for a time, and then get bypassed by another job submitted later. A reservation policy parameter allows a site to determine what how existing reservations should be handled when new reservations are made. The value “highest” will cause that all jobs which have ever received a priority reservation will maintain that reservation until they run even if other jobs later bypass them in priority value. The value of the parameter “current highest” will cause that only the current top <RESERVATIONDEPTH> priority jobs will receive reservations. If a job had a reservation but has been bypassed in priority by another job so that it no longer qualifies as being among the top <RESERVATIONDEPTH> jobs, it will lose its reservation. Finally, the value “never” indicates that no priority reservations will be made.


QOS-based reservation depths can be enabled via the reservation QOS list parameter. This parameter allows varying reservation depths to be associated with different sets of job QoS's. For example, the following configuration will create two reservation depth groupings:

    • - - -
    • RESERVATIONDEPTH[0] 8
    • RESERVATIONQOSLIST[0] highprio interactive debug
    • RESERVATIONDEPTH[1] 2
    • RESERVATIONQOSLIST[1] batch
    • - - -


This example will cause that the top 8 jobs belonging to the aggregate group of highprio, interactive, and debug QoS jobs will receive priority reservations. Additionally, the top 2 batch QoS jobs will also receive priority reservations. Use of this feature allows sites to maintain high throughput for important jobs by guaranteeing a significant proportion of these jobs are making progress toward starting through use of the priority reservation. The following are example default values for some of these parameters: RESERVATIONDEPTH[DEFAULT]=1; RESERVATIONQOSLIST[DEFAULT]=ALL.


This allows one job with the highest priority to get a reservation. These values can be overwritten by modifying the default policy.


A final reservation policy is in place to handle a number of real-world issues. Occasionally when a reservation becomes active and a job attempts to start, various resource manager race conditions or corrupt state situations will prevent the job from starting. By default, the workload manager assumes the resource manager is corrupt, releases the reservation, and attempts to re-create the reservation after a short timeout. However, in the interval between the reservation release and the re-creation timeout, other priority reservations may allocate the newly available resources, reserving them before the original reservation gets an opportunity to reallocate them. Thus, when the original job reservation is re-established, its original resource may be unavailable and the resulting new reservation may be delayed several hours from the earlier start time. The parameter reservation retry time allows a site that is experiencing frequent resource manager race conditions and/or corruption situations to tell the workload manager to hold on to the reserved resource for a period of time in an attempt to allow the resource manager to correct its state.


Next we discuss the use of partitions. Partitions are a logical construct which divide available resources and any single resource (i.e., compute node) may only belong to a single partition. Often, natural hardware or resource manager bounds delimit partitions such as in the case of disjoint networks and diverse processor configurations within a cluster. For example, a cluster may consist of 256 nodes containing four 64 port switches. This cluster may receive excellent interprocess communication speeds for parallel job tasks located within the same switch but sub-stellar performance for tasks which span switches. To handle this, the site may choose to create four partitions, allowing jobs to run within any of the four partitions but not span them.


While partitions do have value, it is important to note that within the workload manager, the standing reservation facility provides significantly improved flexibility and should be used in the vast majority of politically motivated cases where partitions may be required under other resource management systems. Standing reservations provide time flexibility, improved access control features, and more extended resource specification options. Also, another workload manager facility called node sets allows intelligent aggregation of resources to improve per job node allocation decisions. In cases where system partitioning is considered for such reasons, node sets may be able to provide a better solution.


An important aspect of partitions over standing reservations and node sets is the ability to specify partition specific policies, limits, priorities, and scheduling algorithms although this feature is rarely required. An example of this need may be a cluster consisting of 48 nodes owned by the Astronomy Department and 16 nodes owned by the Mathematics Department. Each department may be willing to allow sharing of resources but wants to specify how their partition will be used. As mentioned earlier, many of the workload manager's scheduling policies may be specified on a per partition basis allowing each department to control the scheduling goals within their partition.


The partition associated with each node should be specified as indicated in the node location section. With this done, partition access lists may be specified on a per job or per QOS basis to constrain which resources a job may have access to. By default, QOS's and jobs allow global partition access. Note that by default, a job may only utilize resources within a single partition.


If no partition is specified, the workload manager creates one partition per resource manager into which all resources corresponding to that resource manager are placed. This partition may be given the same name as the resource manager. A partition preferably does not span multiple resource managers. In addition to these resource manager partitions, a pseudo-partition named [ALL] is created which contains the aggregate resources of all partitions. While the resource manager partitions are real partitions containing resources not explicitly assigned to other partitions, the [ALL] partition is only a convenience object and is not a real partition; thus it cannot be requested by jobs or included in configuration ACL's.


Node-to-partition mappings are established using a node configuration parameter as shown in this example:

    • NODECFG[node001] PARTITION=astronomy
    • NODECFG[node002] PARTITION=astronomy
    • . . .
    • NODECFG[node049] PARTITION=math
    • . . .


By default, the workload manager only allows the creation of 4 partitions total. Two of these partitions, DEFAULT, and [ALL], are used internally, leaving only two additional partition definition slots available. If more partitions will be needed, the maximum partition count should be adjusted. Increasing the maximum number of partitions can be managed.


Determining who can use which partition is specified using *CFG parameters (for example, these parameters may be defined as: usercfg, groupcfg, accountcfg, quoscfg, classcfg and systemcfg). These parameters allow both a partition access list and default partition to be selected on a credential or system wide basis using the PLIST and PDEF keywords. By default, the access associated with any given job is the logical or of all partition access lists assigned to the job's credentials. Assume a site with two partitions: general and test. The site management would like everybody to use the general partition by default. However, one user, Steve, needs to perform the majority of his work on the test partition. Two special groups, staff and mgmt will also need access to use the test partition from time to time but will perform most of their work in the general partition. The example configuration below will enable the needed user and group access and defaults for this site.

    • SYSCFG[base] PLIST=
    • USERCFG[DEFAULT] PLIST=general
    • USERCFG[steve] PLIST=general:test PDEF=test
    • GROUPCFG[staff] PLIST=general:test PDEF=general
    • GROUPCFG[mgmt] PLIST=general:test PDEF=general


By default, the system partition access list allows global access to all partitions. If using logically or based partition access lists, the system partition list should be explicitly constrained using the SYSCFG parameter. While using a logical or approach allows sites to add access to certain jobs, some sites prefer to work the other way around. In these cases, access is granted by default and certain credentials are then restricted from access various partitions. To use this model, a system partition list must be specified. See the example below:

    • SYSCFG[base] PLIST=general,test&
    • USERCFG[demo] PLIST=test&
    • GROUP[staff] PLIST=general&


In the above example, note the ampersand (‘&’). This character, which can be located anywhere in the PLIST line, indicates that the specified partition list should be logically AND'd with other partition access lists. In this case, the configuration will limit jobs from user demo to running in partition test and jobs from group staff to running in partition general. All other jobs will be allowed to run in either partition. When using and based partition access lists, the base system access list must be specified with SYSCFG.


Users may request to use any partition they have access to on a per job basis. This is accomplished using the resource manager extensions, since most native batch systems do not support the partition concept. For example, on a PBS system, a job submitted by a member of the group staff could request that the job run in the test partition by adding the line ‘#PBS-W×=PARTITION:test’ to the command file. Special jobs may be allowed to span the resources of multiple partitions if desired by associating the job with a QOS which has the flag ‘SPAN’ set.


The disclosure now continues to discuss reservations further. An advance reservation is the mechanism by which the present invention guarantees the availability of a set of resources at a particular time. With an advanced reservation a site now has an ability to actually specify how the scheduler should manage resources in both space and time. Every reservation consists of three major components, a list of resources, a timeframe (a start and an end time during which it is active), and the ACL. These elements are subject to a set of rules. The ACL acts as a doorway determining who or what can actually utilize the resources of the cluster. It is the job of the cluster scheduler to make certain that the ACL is not violated during the reservation's lifetime (i.e., its timeframe) on the resources listed. The ACL governs access by the various users to the resources. The ACL does this by determining which of the jobs, various groups, accounts, jobs with special service levels, jobs with requests for specific resource types or attributes and many different aspects of requests can actually come in and utilize the resources. With the ability to say that these resources are reserved, the scheduler can then enforce true guarantees and can enforce policies and enable dynamic administrative tasks to occur. The system greatly increases in efficiency because there is no need to partition the resources as was previously necessary and the administrative overhead is reduced it terms of staff time because things can be automated and scheduled ahead of time and reserved.


As an example of a reservation, a reservation may specify that node 002 is reserved for user John Doe on Friday. The scheduler will thus be constrained to make certain that only John Doe's jobs can use node 002 at any time on Friday. Advance reservation technology enables many features including backfill, deadline based scheduling, QOS support, and meta scheduling.


There are several reservation concepts that will be introduced as aspects of the invention. These include dynamic reservations, co-allocating reservation resources of different types, reservations that self-optimize in time, reservations that self-optimization in space, reservations rollbacks and reservation masks. The present invention relates to a system and method of providing dynamic reservations in a compute environment. Dynamic reservations are reservations that are able to be modified once they are created. The workload manager allows dynamic modification of most scheduling parameters allowing new scheduling policies, algorithms, constraints, and permissions to be set at any time. For example, a reservation may be expanded or contracted after a job is submitted to more closely match the reservation to the workload. Changes made via client commands are preferably temporary and will be overridden by values specified in a config files the next time the workload manager is shutdown and restarted.


Various commands may be used manually or automatically to control reservations. Examples of such commands and their function are illustrated in Table 3:











TABLE 3







mdiag
-r
display summarized reservation information




and any unexpected state


mrsvctl

reservation control


mrsvctl
-r
remove reservations


mrsvctl
-c
create an administrative reservation


showres

display information regarding location




and state of reservations










FIG. 9 illustrates a method of dynamically modifying a request, a reservation or the compute environment. Attributes of a reservation may change based on a feedback mechanism that adds intelligence as to ideal characteristics of the reservation and how it should be applied as the context of its environment or an entities needs change. One example of a dynamic reservation is a reservation that provides for a guarantee of resources for a request unless no jobs that consume resources are submitted under the request or if the user is not using the reserved resources. In other words, if no jobs are submitted on reserved resources or the job that is submitted does not need all of the reserved resources.


The example method in FIG. 9 can relate to the scenario where a job has or has not yet been submitted to reserved compute resources. The method comprises receiving a request for resources within the compute environment (902) and monitoring events after receiving the request for resources (904). Based on the monitored events, the method comprises dynamically modifying at least one of the request for resources, a reservation and the compute environment (906). The compute environment may be a computer farm, a cluster, a grid, an on-demand computing center and the like.


The request for resources may be a request for consumption of resources such as processor time and network bandwidth. The request may also be for provisioning resources such as available licenses for particular software or operating systems. The request may also be for such things as a request to process a batch job or for direct volume access, or a request for a virtual private cluster.


The monitored events may further mean monitoring events related to the compute environment. Events that may be identified include, but are not limited to, new resources that become available because other jobs finish early, compute nodes that go down and are unavailable, other jobs submitted to the compute environment. In this regard, the monitoring may include jobs submitted by an administrator, other users or the requestor. For example, if the requestor never submits a job within a reservation made according to the request, then the method may modify the reservation by shrinking the reservation or reduce the reserved amount of resources for efficiency. The request or the reservation may also be canceled if no jobs are submitted or based on other criteria.


A job submitted may also be one of a reservation, an object that monitors policy, an object that monitors credentials, an object that monitors node states and an object that monitors the compute environment. If the compute environment is dynamically modified according to the monitored events, the modification may be performed to satisfy the request for resources or preferences within the request. The modifications to the compute environment may also be constrained within the reservation space.


Examples of modifications that may be done to the compute environment include but are not limited to modifying a node or nodes, modifying at least one operating system or other software, installing end-user applications, dynamically partitioning node resources and adjusting network configurations. Once a job has been submitted, the compute resources may be dynamically modified to more adequately process the job or more efficiently process the job. For example, if it is foreseen that the job will end early, the system may shorten the reservation of time for the resources to free-up migration of other reservations in that time and space. Another example may exist where if a reservation is partly consumed by a job, but based on monitored events, the remaining reserved resources, say 128 nodes, could be expanded to 256 nodes such that the job may finish early. In that case, the reservation from the current time would be dynamically modified to include additional resources.


The modifications to a request, a reservation or a compute environment may be based on a policy. For example, a dynamic reservation policy may apply which says that if the project does not use more than 25% of what it is guaranteed by the time that 50% of its time has expired, then, based on the feedback, the system dynamically modifies the reservation of resources to more closely match the job (606). In other words, the reservation dynamically adjust itself to reserve X % fewer resources for this project, thus freeing up unused resource for others to use.


If the party submitting the request for resources has not submitted a job for processing after a predetermined amount of time, then the request for resources or the job submitted to the reservation may be cancelled. This is illustrated more with reference to FIG. 2C which illustrates this reservation. A self-terminating reservation is a reservation that can cancel itself if certain criteria take place. A reservation of compute resources is created (220) and the system monitors events associated with the reservation (222). The system determines whether the monitored events justifies canceling the reservation or jobs submitted according to the reservation (224). If no, there is no justification to terminate, then the system continues to monitor events in step (222). If, however, conditions justify terminating one of the reservation or a job, then the reservation terminates itself or a job is cancelled (226).


An example of a self-terminating reservation is a reservation that uses an event policy to check that if after 30 minutes no jobs have been submitted against the reservation, or if utilization of the assigned resources is below x % then the reservation will cancel itself, thus making those resources available to be used by others. Another example is if a job is submitted to the reserved cluster resources, but to process the job would require the use of compute resources beyond the reservation time or the reserved cluster resources, then the job may be canceled and notification provided to the submitted regarding the reasons for the cancellation. Options may then be provided to the submitter for modifying the reservation, or modifying the job and so forth to enable the job to be resubmitted under modified circumstances that may enable the job to be processed.


Based on the monitored events in the cluster environment, modifying the request for resources may involve dynamically modifying the compute environment or modifying the compute environment to more adequately process jobs submitted within the reservation.


Preferably, the option of extending the reservation to accommodate the job is subject to pre-established policies that are either required or preferred. One example of presenting these types of offers includes presenting the submitter the option of extending the reservation according to a pricing plan that would meet the preferred policies. This pricing plan may include options to pay for extended time, extended or modified resources, licenses, other provisioning options and so forth. Any combination of job or resource modification is envisioned. In this regard, the reservation of resources could migrate to “cover” a new set of resources that may meet a preferred criteria, an increased payment plan, or some other threshold. The migration of a reservation may be in both space (compute resources) and time (such as, for example, to move the start time of the reservation to as soon as possible). The migration in space may be for the purpose of increasing the performance for the overall compute environment or may be for optimizing the time of completion for a job or jobs. The migration may be for any other reason such as to modify the resources used because of a node failure or a projected maintenance of other failure of a resource. The system may also present a user with the option of allowing jobs running within a personal reservation to complete although the job is projected to run beyond the window of time for the reservation of resources.


As mentioned above, the option of extending or modifying a reservation may be based on pre-established policies that govern whether a reservation may be modified and to what extent it may be modified. There are preferably thresholds established in time and space governing the modifications.


The request for resources in a compute environment may include a request for a reservation of resources for a window of time in which at least one user can submit personal reservations. A personal reservation is a non-administrator reservation submitted by an individual user or a group of users that are not considered administrators. The personal reservation may be submitted by an administrator but is of a non-administrative nature. The window of time may also be a request for cluster resources for a periodic window of time, such as daily, weekly, monthly, quarterly and so on. Then, if the system receives a personal reservation for the use of compute resources within the window of time, the system provides access to the reserved cluster resources for the personal reservation to process submitted jobs. If the processing personal reservation exceeds the window of time for the reservation of compute resources, then the system may cancel and lock out the personal reservation from access to the cluster resources. Before canceling and locking out the personal reservation, the system may present to a user who submitted the personal reservation an option of allowing the personal reservation to complete although it is beyond the window of time for their reservation of compute resources. If a job submitted under the personal reservation would exceed the personal reservation, then the system may extend the personal reservation to meet the needs of the job or perform some other modification. A consumption job submitted may exceed the window of time allowed for the reservation and thus the system may never start the consumption job in the first place.


Charging for resource use and reservation is also an aspect of the present invention. The system may also charge the requestor a specific rate for reserved resources and a different rate for consumed resources. Yet a different rate may be charged for reserved resources that are never used.


The user/requestor may be charged for use of the resources in the cluster environment in a variety of ways. For example, the user may be charged for reserved resources at one rate, and another rate for reserved and consumed resources.


Within a reservation, the system may provide a modification of the compute resources within the reservation space. For example, the system may optimize the use of resources within that reservation to meet needs and preferences of particular jobs submitted under that reservation.


Another dynamic reservation may perform the following step: if usage of resources provided by a reservation is above 90% with fewer than 10 minutes left in the reservation then the reservation will attempt to add 10% more time to the end of the reservation to help ensure the project is able to complete. In summary, it is the ability for a reservation to receive manual or automatic feedback to an existing reservation in order to have it more accurately match any given needs, whether those be of the submitting entity, the community of users, administrators, etc. The dynamic reservation improves the state of the art by allowing the ACL to the reservation to have a dynamic aspect instead of simply being based on who the requestor is. The reservation can be based on a current level of service or response time being delivered to the requestor.


The ACL and scheduler are able to monitor all aspects of the request by looking at the current job inside the queue and how long it has sat there and what the response time target is. It is preferable, although not required, that the scheduler itself determines whether all requirements of the ACL are satisfied. If the requirements are satisfied, the scheduler releases the resources that are available to the job.


The benefits of this model is it makes it significantly easier for a site to balance or provide guaranteed levels of service or constant levels of service for key players or the general populace. By setting aside certain resources and only making them available to the jobs which threaten to violate their quality of service targets it increases the probability of satisfying it.


As can be appreciated, the methods described above for managing a compute environment provide marked improvements in how resources are reserved and how those reservations are managed in connection with the compute environment to maximize efficiency for both the user and the compute environment.


Embodiments within the scope of the present invention may also include computer-readable media for carrying or having computer-executable instructions or data structures stored thereon. Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer. By way of example, and not limitation, such computer-readable media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code means in the form of computer-executable instructions or data structures. When information is transferred or provided over a network or another communications connection (either hardwired, wireless, or combination thereof) to a computer, the computer properly views the connection as a computer-readable medium. Thus, any such connection is properly termed a computer-readable medium. Combinations of the above should also be included within the scope of the computer-readable media.


Computer-executable instructions include, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Computer-executable instructions also include program modules that are executed by computers in stand-alone or network environments. Generally, program modules include routines, programs, objects, components, and data structures, etc. that perform particular tasks or implement particular abstract data types. Computer-executable instructions, associated data structures, and program modules represent examples of the program code means for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represents examples of corresponding acts for implementing the functions described in such steps.


Those of skill in the art will appreciate that other embodiments of the invention may be practiced in network computing environments with many types of computer system configurations, including personal computers, hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, and the like. Embodiments may also be practiced in distributed computing environments where tasks are performed by local and remote processing devices that are linked (either by hardwired links, wireless links, or by a combination thereof) through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.


Although the above description may contain specific details, they should not be construed as limiting the claims in any way. Other configurations of the described embodiments of the invention are part of the scope of this invention. Accordingly, the appended claims and their legal equivalents should only define the invention, rather than any specific examples given.

Claims
  • 1. A method for use within a computerized scheduler process of a commonly administrated multi-node compute environment, the method comprising: receiving a first request for processing one or more workload items, the processing of the one or more workload items requiring one or more resources;causing reserving of resources sufficient to meet the required one or more resources associated with processing the first request, the causing of the reserving of the resources comprising creating at least one mask, each of the at least one mask constraining the resources to a specific subset of resources during a specific subset of times, and guaranteeing that any subsequent request for resource consumption is within the specific subset of resources and the specific subset of times of the at least one mask;causing preservation of the reserved resources for at least a period of time based on the at least one mask;causing initiation of processing of the one or more workload items; andmonitoring one or more conditions within the commonly administrated multi-node compute environment to determine that the processing was prevented from starting;wherein the at least period of time comprises a period of time in which the reserved resources must be preserved; andwherein the causing of the preservation of the reserved resources for the at least period of time comprises causing the preservation of the reserved resources at least until a completion of an attempt to remedy a condition that caused the prevention of the processing from starting, based at least on a parameter reservation retry time from a computerized user process, the parameter reservation retry time instructing the computerized scheduler process to cause the preservation of the reserved resources for at least the period of time to prevent the reserved resources from being reallocated for use with one or more other workload items during the at least period of time to prevent at least one race condition during the at least period of time.
  • 2. The method of claim 1, further comprising receiving a second request for processing the one or more other workload items, the processing of the one or more other workload items requiring one or more resources; wherein at least the causing of the preservation of the reserved resources for the at least period of time until the completion of the attempt to remedy the condition that caused the prevention of the processing from starting, prevents the one or more race conditions from occurring between the first request and the second request.
  • 3. The method of claim 1, further comprising, based at least on the determination that the processing was prevented from starting, causing cancelation, after the period of time, of the reserving of the resources sufficient to meet the required one or more resources associated with processing the first request, the cancellation enabling the resources to be allocated to at least one other request.
  • 4. The method of claim 3, further comprising: disposing the first request in a queue, the disposing the first request in the queue occurring after the receiving the first request for the processing of the one or more workload items and the causing of the reserving of the resources; andbased at least on the cancelation of the reserving of the resources sufficient to meet the required one or more resources associated with processing the first request, causing the first request to be processed.
  • 5. The method of claim 3, wherein the resources sufficient to meet the required one or more resources associated with processing the first request comprise resources not sufficient to meet one or more preferred criteria.
  • 6. The method of claim 5, wherein the resources sufficient to meet the required one or more resources associated with processing the first request comprise resources sufficient to meet the one or more preferred criteria for the at least one other request.
  • 7. The method of claim 1, further comprising: dynamically modifying the reserved resources in response to determining that the one or more workload items will end early.
  • 8. A computer-readable storage apparatus comprising a non-transitory storage medium storing instructions for controlling a computerized device to schedule resources within a compute environment using a computerized scheduler process, the compute environment comprising a plurality of compute nodes under common administrative control, the instructions configured to, when executed by a processor apparatus, cause the computerized device to: receive a first request for processing one or more workload items, the processing of the one or more workload items requiring one or more resources, the first request having been generated by a computerized user process;create a self-optimizing reservation of resources that is sufficient to meet the required one or more resources associated with processing the first request, the self-optimizing reservation automatically (i) reserving processing resources or memory resources within the plurality of compute nodes, and (ii) starting at a first time; wherein the creation of the self-optimizing reservation of the resources comprises utilization of a mask function in which one or more time periods and a specific subset of resources are not available to other workload items besides the one or more workload items, the self-optimizing reservation being created within the mask function;monitor one or more conditions within the compute environment;utilize the self-optimizing reservation to automatically: (i) determine, based at least on the monitoring indicating that a number of the compute nodes guaranteed by the self-optimizing reservation are available earlier than previously determined, that the reservation can be started at a second time earlier than the first time, each of the second time and the first time being within at least one of personal reservation mask or non-personal reservation mask, wherein the at least one of the personal reservation mask or the non-personal reservation mask comprises a personal reservation mask and a non-personal reservation mask, the non-personal reservation mask being able to overlap with the personal reservation mask based on the personal reservation mask being setup with a global access control list;(ii) enable the self-optimizing reservation to start at the second time or a time thereafter earlier than the first time, the enablement causing completion of the processing of the one or more workload items quicker than when being started at the first time; andprocess the one or more workload items based on the self-optimizing reservation.
  • 9. The computer-readable storage apparatus of claim 8, wherein the instructions are further configured to, when executed by the processor apparatus, cause the computerized device to: cause placement of at least one of the one or more workload items in a queue, the placement of the at least one of the one or more workload items in the queue occurring after the receipt of the first request and causing the reserving of resources;cause placement of at least one workload item associated with one or more other requests in the queue; andcause servicing of the at least one workload item associated with the one or more other requests in the queue by the scheduler process upon the processing of at least one of the one or more workload items having been started.
  • 10. The computer-readable storage apparatus of claim 8, wherein: the first request comprises data indicative of a priority classification for processing of the one or more workload items of the first request; andthe instructions are further configured to, when executed by the processor apparatus, cause the computerized device to:cause placement of the one or more workload items of the first request in a queue, the placement occurring after the receipt of the first request and the creating of the self-optimizing reservation of the resources; andcause processing of the one or more queued workload items, the causing of the processing being performed ahead of workload of one or more other requests in the queue based at least on the data indicative of the priority classification.
  • 11. The non-transitory computer-readable storage apparatus of claim 8, wherein the instructions are further configured to, when executed on the computerized device, rollback the reservation, the rollback of the reservation comprising setup of a moving reservation that extends from a first future time offset from a then-current time to a second future time offset, and continues to slide forward in time, thereby allowing guarantees to be created and maintained for at least one of the plurality of compute nodes.
  • 12. A computerized scheduler apparatus configured for use within a managed cluster of compute nodes, the computerized scheduler apparatus comprising: at least one data processor;a network interface in data communication with the at least one data processor; anda non-transitory computer-readable storage apparatus in data communication with the at least one data processor and comprising a storage medium storing instructions for controlling scheduling of resources within the managed cluster, the instructions configured to, when executed on the at least one data processor;receive a first request for processing workload, the processing of the workload requiring at least one resource, the first request having been received via at least the network interface;cause queueing of the workload within a scheduler queue of a scheduler process;cause resources of at least one of the compute nodes that at least meet the required at least one resource for processing the workload to be reserved exclusively for use by the workload at least until a processing status of the workload has been determined;monitor the processing status within the managed cluster to determine that processing of the workload has failed due to one or more corruptions or errors; andbased at least on the determination that the processing of the workload has failed; cause:release of the resources of the at least one of the compute nodes reserved exclusively for use by the workload for use by second workload associated with one or more other requests; andfurther processing of the first request by the scheduler process:wherein a prescribed subset of all of the compute nodes of the managed cluster comprises a subset selected according to at least one logical masking or policy using two of more periodic reservation masks each defining separate logical boundaries in which one or more reservations for the resources can be created during a specific time period and where available resources are to be used for specific purposes, the two or more reservation masks indicating two or more repeating time periods, respectively.
  • 13. The computerized scheduler apparatus of claim 12, wherein the causation of the queueing of the workload within the scheduler queue of the scheduler process comprises queueing the workload within the scheduler queue relative to a workload of the one or more other requests based at least on priority data associated with at least the first request.
  • 14. The computerized scheduler apparatus of claim 12, wherein the instructions are further configured to, when executed on the at least one data processor, cause application of at least one policy mechanism to the first request and the one or more other requests, the at least one policy mechanism configured to determine whether one or more policies are met by each of the first request and the one or more other requests prior to allowing the queueing.
  • 15. The computerized scheduler apparatus of claim 14, wherein the one or more policies comprise one or more prioritization policies for workload associated with the first request and the one or more other requests, respectively.
  • 16. The computerized scheduler apparatus of claim 12, wherein the causation of the resources of the at least one of the compute nodes that at least meet the required at least one resource for the processing of the workload to be reserved exclusively for use by the workload comprises identification of at least two of the compute nodes each having a respective portion of the required at least one resource for processing the workload available simultaneously, the respective portions collectively at least meeting the required at least one resource.
  • 17. The computerized scheduler apparatus of claim 12, wherein the causation of the resources of the at least one of the compute nodes that at least meet the required at least one resource for the processing the workload to be reserved exclusively for use by the workload comprises: identification of at least one of the compute nodes having the required at least one resource for processing the workload available from a prescribed subset of all of the compute nodes of the managed cluster; andallocating at least the available required at least one resource of the identified at least one compute node to the workload on an exclusive basis.
  • 18. The computerized scheduler apparatus of claim 12, wherein the reservation mask comprises a personal reservation mask.
  • 19. A computerized scheduler apparatus configured for use within a managed cluster of compute nodes, the computerized scheduler apparatus comprising: at least one data processor;a network interface in data communication with the at least one data processor; and
  • 20. The computerized scheduler apparatus of claim 19, wherein: the first request comprises data indicative of a priority classification for processing of the at least one workload item by the cluster; andthe queueing of the at least one workload item within the scheduler queue of the scheduler process comprises queueing of the at least one workload item based on the data indicative of priority.
  • 21. The computerized scheduler apparatus of claim 19, wherein: the determination that at least a portion of the further processing of the at least one workload item has failed comprises determination, via a resource manager process in data communication with the scheduler process, that one or more resources required to support the at least one workload item are not available.
  • 22. The computerized scheduler apparatus of claim 19, wherein: the determination that at least a portion of the further processing of the at least one workload item has failed comprises determination, via a resource manager process in data communication with the scheduler process, that one or more resources required to support the at least one workload item are not available; andthe instructions are configured to, when executed on the at least one data processor:cause the determination that at least the portion of the further processing of the at least one workload item has failed only after a prescribed time delay has expired since the determination via the resource manager process that the one or more resources required to support the at least one workload item are not available.
  • 23. The computerized scheduler apparatus of claim 19, wherein the instructions are configured to, when executed on the at least one data processor, predict a node failure.
  • 24. A computerized scheduler apparatus configured for use within a managed cluster of compute nodes, the computerized scheduler apparatus comprising: at least one data processor;a network interface in data communication with the at least one data processor; anda non-transitory computer-readable storage apparatus in data communication with the at least one data processor and comprising a storage medium storing instructions for controlling scheduling of resources within the managed cluster, the instructions configured to, when executed on the at least one data processor:receive a first request for processing at least one workload item, the processing of the at least one workload item requiring at least one resource, the first request having been (i) initiated by a computerized user process in data communication with a compute environment within which the computerized scheduler apparatus operates, and (ii) received via at least the network interface;cause queueing of the at least one workload item within a scheduler queue of a scheduler process, the queueing comprising utilization of at least one class initiator-to-node mapping, the at least one class initiator-to-node mapping enabling at least one of the compute nodes to run at least two different types of jobs simultaneously, subject to one or more constraints, the one or more constraints comprising 1) a maximum number of total jobs allowed to run on the at least one of the compute nodes and 2) a maximum number of at least one of the at least two different types of jobs allowed to run on the at least one of the compute nodes;cause selection of the at least one workload item from the scheduler queue for further processing;based on the selection, cause creation of a reservation of resources within a first periodic reservation mask of at least two periodic reservation masks, the resources to be utilized for the further processing;cause initiation of processing of the selected at least one workload item;monitor a processing status of the selected at least one workload item to algorithmically determine whether at least a portion of the further processing of the at least one workload item has failed; andbased on the failure of the at least portion of the further processing, causing migration of the reservation of the resources to be within a second periodic reservation mask of at least two periodic reservation masks, the second periodic reservation mask having a repeating time frame from the first periodic reservation mask, the further processing of the at least one workload item being successfully completed based on the migration and use of the second periodic reservation mask.
PRIORITY CLAIM

The present application is a continuation of U.S. patent application Ser. No. 17/088,954, filed Nov. 4, 2020, which is a continuation of U.S. patent application Ser. No. 13/855,241, filed Apr. 2, 2013 (now U.S. Pat. No. 10,871,999), which is a continuation of U.S. patent application Ser. No. 10/530,581, filed Aug. 11, 2006 (now U.S. Pat. No. 8,413,155), which is a national stage application of PCT/US2005/008296, filed Mar. 11, 2005, which claims priority to U.S. Provisional Application No. 60/552,653, filed Mar. 13, 2004, the contents of which are incorporated herein by reference in their entirety.

US Referenced Citations (2019)
Number Name Date Kind
4215406 Gomola et al. Jul 1980 A
4412288 Herman Oct 1983 A
4525780 Bratt et al. Jun 1985 A
4532893 Day et al. Aug 1985 A
4542458 Kitajima Sep 1985 A
4553202 Trufyn Nov 1985 A
4677614 Circo Jun 1987 A
4850891 Walkup et al. Jul 1989 A
4852001 Tsushima et al. Jul 1989 A
4943932 Lark et al. Jul 1990 A
4975840 DeTore Dec 1990 A
4992958 Kageyama Feb 1991 A
5012409 Fletcher Apr 1991 A
5056070 Shibuya Oct 1991 A
5084696 Guscott Jan 1992 A
5132625 Shaland Jul 1992 A
5146561 Carey et al. Sep 1992 A
5168441 Onarheim Dec 1992 A
5175800 Galis et al. Dec 1992 A
5233533 Edstrom Aug 1993 A
5257374 Hammer et al. Oct 1993 A
5274809 Iwasaki Dec 1993 A
5276877 Friedrich Jan 1994 A
5299115 Fields et al. Mar 1994 A
5307496 Ichinose et al. Apr 1994 A
5325526 Cameron et al. Jun 1994 A
5349682 Rosenberry Sep 1994 A
5355508 Kan Oct 1994 A
5377332 Entwistle et al. Dec 1994 A
5408663 Miller Apr 1995 A
5451936 Yang et al. Sep 1995 A
5465328 Dievendorff Nov 1995 A
5469566 Hohenstein Nov 1995 A
5473773 Aman et al. Dec 1995 A
5477546 Shibata Dec 1995 A
5495533 Linehan et al. Feb 1996 A
5504894 Ferguson et al. Apr 1996 A
5542000 Semba Jul 1996 A
5550970 Cline et al. Aug 1996 A
5594901 Andoh Jan 1997 A
5594908 Hyatt Jan 1997 A
5598536 Slaughter et al. Jan 1997 A
5600844 Shaw et al. Feb 1997 A
5623641 Kadoyashiki Apr 1997 A
5623672 Popat Apr 1997 A
5651006 Fujino et al. Jul 1997 A
5652841 Nemirovsky et al. Jul 1997 A
5666293 Metz Sep 1997 A
5675739 Eilert et al. Oct 1997 A
5701451 Rogers et al. Dec 1997 A
5729754 Estes Mar 1998 A
5732077 Whitehead Mar 1998 A
5734818 Kern et al. Mar 1998 A
5737009 Payton Apr 1998 A
5745694 Egawa Apr 1998 A
5752022 Chiu May 1998 A
5752030 Konno et al. May 1998 A
5757771 Li May 1998 A
5761433 Billings Jun 1998 A
5761475 Yung Jun 1998 A
5761484 Agarwal et al. Jun 1998 A
5765146 Wolf Jun 1998 A
5774660 Brendel et al. Jun 1998 A
5774668 Choquier et al. Jun 1998 A
5781187 Gephardt et al. Jul 1998 A
5781624 Mitra et al. Jul 1998 A
5787000 Lilly Jul 1998 A
5787459 Stallmo et al. Jul 1998 A
5799174 Muntz et al. Aug 1998 A
5801985 Roohparvar et al. Sep 1998 A
5826080 Dworzecki Oct 1998 A
5826082 Bishop et al. Oct 1998 A
5826236 Narimatsu et al. Oct 1998 A
5826239 Du et al. Oct 1998 A
5828743 Pinnell Oct 1998 A
5828888 Kozaki et al. Oct 1998 A
5832517 Knutsen, II Nov 1998 A
5854887 Kindell et al. Dec 1998 A
5862478 Cutler, Jr. et al. Jan 1999 A
5867382 McLaughlin Feb 1999 A
5874789 Su Feb 1999 A
5881238 Aman et al. Mar 1999 A
5901048 Hu May 1999 A
5908468 Hartmann Jun 1999 A
5911143 Deinhart et al. Jun 1999 A
5913921 Tosey Jun 1999 A
5918017 Attanasio et al. Jun 1999 A
5920545 Raesaenen et al. Jul 1999 A
5920863 McKeehan et al. Jul 1999 A
5926798 Carter Jul 1999 A
5930167 Lee et al. Jul 1999 A
5933417 Rottoo Aug 1999 A
5935293 Detering et al. Aug 1999 A
5950190 Yeager Sep 1999 A
5956715 Glasser Sep 1999 A
5958003 Preining et al. Sep 1999 A
5961599 Kalavade et al. Oct 1999 A
5963911 Walker Oct 1999 A
5968176 Nessett et al. Oct 1999 A
5971804 Gallagher et al. Oct 1999 A
5978356 Elwalid et al. Nov 1999 A
5987611 Freund Nov 1999 A
6003061 Jones et al. Dec 1999 A
6006192 Cheng et al. Dec 1999 A
6012052 Altschuler et al. Jan 2000 A
6021425 Waldron, III et al. Feb 2000 A
6032224 Blumenau Feb 2000 A
6052707 D'Souza Apr 2000 A
6055618 Thorson Apr 2000 A
6058416 Mukherjee et al. May 2000 A
6067545 Wolff May 2000 A
6076174 Freund Jun 2000 A
6078953 Vaid et al. Jun 2000 A
6079863 Furukawa Jun 2000 A
6085238 Yuasa et al. Jul 2000 A
6088718 Altschuler et al. Jul 2000 A
6092178 Jindal et al. Jul 2000 A
6094712 Follett Jul 2000 A
6097882 Mogul Aug 2000 A
6098090 Burns Aug 2000 A
6101508 Wolff Aug 2000 A
6105117 Ripley Aug 2000 A
6108662 Hoskins et al. Aug 2000 A
6115382 Abe Sep 2000 A
6122664 Boukobza Sep 2000 A
6141214 Ahn Oct 2000 A
6151598 Shaw et al. Nov 2000 A
6154778 Koistinen et al. Nov 2000 A
6161170 Burger et al. Dec 2000 A
6167445 Gai et al. Dec 2000 A
6175869 Ahuja et al. Jan 2001 B1
6181699 Crinion et al. Jan 2001 B1
6182139 Brendel et al. Jan 2001 B1
6182142 Win et al. Jan 2001 B1
6185272 Hiraoglu Feb 2001 B1
6185575 Orcutt Feb 2001 B1
6185601 Wolff Feb 2001 B1
6189111 Alexander Feb 2001 B1
6192414 Horn Feb 2001 B1
6195678 Komuro Feb 2001 B1
6198741 Yoshizawa et al. Mar 2001 B1
6201611 Carter et al. Mar 2001 B1
6202080 Lu et al. Mar 2001 B1
6205465 Schoening et al. Mar 2001 B1
6210275 Olsen Apr 2001 B1
6212542 Kahle et al. Apr 2001 B1
6223202 Bayeh Apr 2001 B1
6226677 Slemmer May 2001 B1
6226788 Schoening et al. May 2001 B1
6240453 Chang May 2001 B1
6247056 Chou et al. Jun 2001 B1
6252878 Locklear, Jr. Jun 2001 B1
6253230 Couland et al. Jun 2001 B1
6256704 Hlava Jul 2001 B1
6259675 Honda Jul 2001 B1
6263359 Fong et al. Jul 2001 B1
6266667 Olsson Jul 2001 B1
6269398 Leong Jul 2001 B1
6278712 Takihiro et al. Aug 2001 B1
6282561 Jones et al. Aug 2001 B1
6289382 Bowman-Amuah Sep 2001 B1
6298352 Kannan et al. Oct 2001 B1
6304549 Srinivasan Oct 2001 B1
6314114 Coyle et al. Nov 2001 B1
6314487 Hahn et al. Nov 2001 B1
6314501 Gulick et al. Nov 2001 B1
6314555 Ndumu et al. Nov 2001 B1
6317787 Boyd et al. Nov 2001 B1
6324279 Kalmanek, Jr. et al. Nov 2001 B1
6327364 Shaffer et al. Dec 2001 B1
6330008 Razdow et al. Dec 2001 B1
6330562 Boden et al. Dec 2001 B1
6330583 Reiffin Dec 2001 B1
6330605 Christensen et al. Dec 2001 B1
6333936 Johansson et al. Dec 2001 B1
6334114 Jacobs et al. Dec 2001 B1
6338085 Ramaswamy Jan 2002 B1
6338112 Wipfel et al. Jan 2002 B1
6339717 Baumgartl et al. Jan 2002 B1
6343311 Nishida et al. Jan 2002 B1
6343488 Hackfort Feb 2002 B1
6345287 Fong et al. Feb 2002 B1
6345294 O'Toole et al. Feb 2002 B1
6349295 Tedesco Feb 2002 B1
6351775 Yu Feb 2002 B1
6353844 Bitar et al. Mar 2002 B1
6363434 Eytchison Mar 2002 B1
6363488 Ginter et al. Mar 2002 B1
6366945 Fong et al. Apr 2002 B1
6370154 Wickham Apr 2002 B1
6370584 Bestavros et al. Apr 2002 B1
6373841 Goh et al. Apr 2002 B1
6374254 Cochran et al. Apr 2002 B1
6374297 Wolf et al. Apr 2002 B1
6384842 DeKoning May 2002 B1
6385302 Antonucci et al. May 2002 B1
6392989 Jardetzky et al. May 2002 B1
6393569 Orenshteyn May 2002 B1
6393581 Friedman et al. May 2002 B1
6400996 Hoffberg et al. Jun 2002 B1
6401133 York Jun 2002 B1
6404768 Basak et al. Jun 2002 B1
6405212 Samu Jun 2002 B1
6405234 Ventrone Jun 2002 B2
6418459 Gulick Jul 2002 B1
6434568 Bowman-Amuah Aug 2002 B1
6438125 Brothers Aug 2002 B1
6438134 Chow et al. Aug 2002 B1
6438553 Yamada Aug 2002 B1
6438594 Bowman-Amuah Aug 2002 B1
6438652 Jordan et al. Aug 2002 B1
6442137 Yu et al. Aug 2002 B1
6445968 Jalla Sep 2002 B1
6446192 Narasimhan et al. Sep 2002 B1
6446206 Feldbaum Sep 2002 B1
6452809 Jackson et al. Sep 2002 B1
6452924 Golden et al. Sep 2002 B1
6453349 Kano et al. Sep 2002 B1
6453383 Stoddard et al. Sep 2002 B1
6460082 Lumelsky et al. Oct 2002 B1
6463454 Lumelsky et al. Oct 2002 B1
6464261 Dybevik et al. Oct 2002 B1
6466935 Stuart Oct 2002 B1
6466965 Chessell et al. Oct 2002 B1
6466980 Lumelsky et al. Oct 2002 B1
6477575 Koeppel Nov 2002 B1
6477580 Bowman-Amuah Nov 2002 B1
6487390 Virine et al. Nov 2002 B1
6490432 Wegener et al. Dec 2002 B1
6496566 Posthuma Dec 2002 B1
6496866 Attanasio et al. Dec 2002 B2
6496872 Katz et al. Dec 2002 B1
6502135 Munger et al. Dec 2002 B1
6505228 Schoening et al. Jan 2003 B1
6507586 Satran et al. Jan 2003 B1
6519571 Guheen et al. Feb 2003 B1
6520591 Jun et al. Feb 2003 B1
6526442 Stupek, Jr. et al. Feb 2003 B1
6529499 Doshi et al. Mar 2003 B1
6529932 Dadiomov et al. Mar 2003 B1
6538994 Horspool Mar 2003 B1
6549940 Allen et al. Apr 2003 B1
6556952 Magro Apr 2003 B1
6557073 Fujiwara Apr 2003 B1
6564261 Gudjonsson et al. May 2003 B1
6571215 Mahapatro May 2003 B1
6571391 Acharya et al. May 2003 B1
6574238 Thrysoe Jun 2003 B1
6574632 Fox et al. Jun 2003 B2
6578005 Lesaint Jun 2003 B1
6578068 Bowman-Amuah Jun 2003 B1
6584489 Jones et al. Jun 2003 B1
6584499 Jantz et al. Jun 2003 B1
6587469 Bragg Jul 2003 B1
6587938 Eilert et al. Jul 2003 B1
6590587 Wichelman et al. Jul 2003 B1
6594718 Ebner Jul 2003 B1
6600898 Bonet et al. Jul 2003 B1
6601234 Bowman-Amuah Jul 2003 B1
6606660 Bowman-Amuah Aug 2003 B1
6618820 Krum Sep 2003 B1
6622168 Datta Sep 2003 B1
6626077 Gilbert Sep 2003 B1
6628649 Raj et al. Sep 2003 B1
6629081 Cornelius et al. Sep 2003 B1
6629148 Ahmed et al. Sep 2003 B1
6633544 Rexford et al. Oct 2003 B1
6636853 Stephens, Jr. Oct 2003 B1
6640145 Hoffberg et al. Oct 2003 B2
6640238 Bowman-Amuah Oct 2003 B1
6640248 Jorgensen Oct 2003 B1
6651098 Carroll et al. Nov 2003 B1
6651125 Maergner Nov 2003 B2
6661671 Franke et al. Dec 2003 B1
6661787 O'Connell et al. Dec 2003 B1
6662202 Krusche et al. Dec 2003 B1
6662219 Nishanov et al. Dec 2003 B1
6668304 Satran et al. Dec 2003 B1
6678065 Hikawa Jan 2004 B1
6687257 Balasubramanian Feb 2004 B1
6690400 Moayyad et al. Feb 2004 B1
6690647 Tang et al. Feb 2004 B1
6701318 Fox et al. Mar 2004 B2
6704489 Kurauchi Mar 2004 B1
6708220 Olin Mar 2004 B1
6711691 Howard et al. Mar 2004 B1
6714778 Nykanen et al. Mar 2004 B2
6724733 Schuba et al. Apr 2004 B1
6725456 Bruno et al. Apr 2004 B1
6735188 Becker et al. May 2004 B1
6735630 Gelvin et al. May 2004 B1
6735716 Podanoffsky May 2004 B1
6738736 Bond May 2004 B1
6738974 Nageswaran May 2004 B1
6745221 Ronca Jun 2004 B1
6745246 Erimli et al. Jun 2004 B1
6745262 Benhase Jun 2004 B1
6748559 Pfister Jun 2004 B1
6754892 Johnson Jun 2004 B1
6757723 O'Toole et al. Jun 2004 B1
6757897 Shi Jun 2004 B1
6760306 Pan et al. Jul 2004 B1
6763519 McColl et al. Jul 2004 B1
6763520 Seeds Jul 2004 B1
6766389 Hayter et al. Jul 2004 B2
6771661 Chawla et al. Aug 2004 B1
6772211 Lu et al. Aug 2004 B2
6775701 Pan et al. Aug 2004 B1
6779016 Aziz et al. Aug 2004 B1
6781990 Puri et al. Aug 2004 B1
6782408 Chandra Aug 2004 B1
6785724 Drainville et al. Aug 2004 B1
6785794 Chase et al. Aug 2004 B2
6813676 Henry et al. Nov 2004 B1
6816750 Klaas Nov 2004 B1
6816903 Rakoshitz et al. Nov 2004 B1
6816905 Sheets et al. Nov 2004 B1
6823377 Wu et al. Nov 2004 B1
6826607 Gelvin et al. Nov 2004 B1
6829206 Watanabe Dec 2004 B1
6829762 Arimilli et al. Dec 2004 B2
6832251 Gelvin et al. Dec 2004 B1
6836806 Raciborski et al. Dec 2004 B1
6842430 Melnik Jan 2005 B1
6850966 Matsuura et al. Feb 2005 B2
6857020 Chaar et al. Feb 2005 B1
6857026 Cain Feb 2005 B1
6857938 Smith et al. Feb 2005 B1
6859831 Gelvin et al. Feb 2005 B1
6859927 Moody et al. Feb 2005 B2
6862451 Alard Mar 2005 B1
6862606 Major et al. Mar 2005 B1
6868097 Soda et al. Mar 2005 B1
6874031 Corbeil Mar 2005 B2
6882718 Smith Apr 2005 B1
6894792 Abe May 2005 B1
6904460 Raciborski et al. Jun 2005 B1
6912533 Hornick Jun 2005 B1
6922664 Fernandez et al. Jul 2005 B1
6925431 Papaefstathiou Aug 2005 B1
6928471 Pabari et al. Aug 2005 B2
6931640 Asano et al. Aug 2005 B2
6934702 Faybishenko et al. Aug 2005 B2
6938256 Deng et al. Aug 2005 B2
6947982 McGann et al. Sep 2005 B1
6948171 Dan et al. Sep 2005 B2
6950821 Faybishenko et al. Sep 2005 B2
6950833 Costello et al. Sep 2005 B2
6952828 Greene Oct 2005 B2
6954784 Aiken, Jr. et al. Oct 2005 B2
6963917 Callis et al. Nov 2005 B1
6963926 Robinson Nov 2005 B1
6963948 Gulick Nov 2005 B1
6965930 Arrowood et al. Nov 2005 B1
6966033 Gasser et al. Nov 2005 B1
6968323 Bansal et al. Nov 2005 B1
6971098 Khare et al. Nov 2005 B2
6975609 Khaleghi et al. Dec 2005 B1
6977939 Joy et al. Dec 2005 B2
6978310 Rodriguez et al. Dec 2005 B1
6978447 Okmianski Dec 2005 B1
6985461 Singh Jan 2006 B2
6985937 Keshav et al. Jan 2006 B1
6988170 Barroso et al. Jan 2006 B2
6990063 Lenoski et al. Jan 2006 B1
6990616 Botton-Dascal Jan 2006 B1
6990677 Pietraszak et al. Jan 2006 B1
6996821 Butterworth Feb 2006 B1
6996822 Willen Feb 2006 B1
7003414 Wichelman et al. Feb 2006 B1
7006881 Hoffberg et al. Feb 2006 B1
7013303 Faybishenko et al. Mar 2006 B2
7013322 Lahr Mar 2006 B2
7017186 Day Mar 2006 B2
7020695 Kundu et al. Mar 2006 B1
7020701 Gelvin et al. Mar 2006 B1
7020719 Grove et al. Mar 2006 B1
7032119 Fung Apr 2006 B2
7034686 Matsumura Apr 2006 B2
7035230 Shaffer et al. Apr 2006 B1
7035240 Balakrishnan et al. Apr 2006 B1
7035854 Hsiao et al. Apr 2006 B2
7035911 Lowery et al. Apr 2006 B2
7043605 Suzuki May 2006 B2
7058070 Tran et al. Jun 2006 B2
7058716 Sundaresan et al. Jun 2006 B1
7058949 Willen Jun 2006 B1
7058951 Bril et al. Jun 2006 B2
7065579 Traversat et al. Jun 2006 B2
7065764 Prael et al. Jun 2006 B1
7072807 Brown et al. Jul 2006 B2
7076717 Grossman, IV et al. Jul 2006 B2
7080078 Slaughter et al. Jul 2006 B1
7080283 Songer et al. Jul 2006 B1
7080285 Kosugi Jul 2006 B2
7080378 Noland et al. Jul 2006 B1
7082606 Wood et al. Jul 2006 B2
7085825 Pishevar et al. Aug 2006 B1
7085837 Kimbrel et al. Aug 2006 B2
7085893 Krissell et al. Aug 2006 B2
7089294 Baskey et al. Aug 2006 B1
7093256 Bloks Aug 2006 B2
7095738 Desanti Aug 2006 B1
7099933 Wallace et al. Aug 2006 B1
7100192 Igawa et al. Aug 2006 B1
7102996 Amdahl et al. Sep 2006 B1
7103625 Hipp et al. Sep 2006 B1
7103664 Novaes et al. Sep 2006 B1
7107578 Alpem Sep 2006 B1
7107589 Tal Sep 2006 B1
7117208 Tamayo et al. Oct 2006 B2
7117273 O'Toole et al. Oct 2006 B1
7119591 Lin Oct 2006 B1
7124289 Suorsa Oct 2006 B1
7124410 Berg et al. Oct 2006 B2
7126913 Patel et al. Oct 2006 B1
7127613 Pabla et al. Oct 2006 B2
7127633 Olson et al. Oct 2006 B1
7136927 Traversat et al. Nov 2006 B2
7140020 McCarthy et al. Nov 2006 B2
7143088 Green et al. Nov 2006 B2
7143153 Black et al. Nov 2006 B1
7143168 DiBiasio et al. Nov 2006 B1
7145995 Oltmanns et al. Dec 2006 B2
7146233 Aziz et al. Dec 2006 B2
7146353 Garg et al. Dec 2006 B2
7146416 Yoo et al. Dec 2006 B1
7150044 Hoefelmeyer et al. Dec 2006 B2
7154621 Rodriguez Dec 2006 B2
7155478 Ims et al. Dec 2006 B2
7155502 Galloway et al. Dec 2006 B1
7165107 Pouyoul et al. Jan 2007 B2
7165120 Giles et al. Jan 2007 B1
7167920 Traversat et al. Jan 2007 B2
7168049 Day Jan 2007 B2
7170315 Bakker et al. Jan 2007 B2
7171415 Kan et al. Jan 2007 B2
7171476 Maeda et al. Jan 2007 B2
7171491 O'Toole et al. Jan 2007 B1
7171593 Whittaker Jan 2007 B1
7177823 Lam et al. Feb 2007 B2
7180866 Chartre et al. Feb 2007 B1
7185046 Ferstl et al. Feb 2007 B2
7185073 Gai et al. Feb 2007 B1
7185077 O'Toole et al. Feb 2007 B1
7188145 Lowery et al. Mar 2007 B2
7188174 Rolia et al. Mar 2007 B2
7191244 Jennings et al. Mar 2007 B2
7197071 Weigand Mar 2007 B1
7197549 Salama et al. Mar 2007 B1
7197559 Goldstein et al. Mar 2007 B2
7197561 Lovy et al. Mar 2007 B1
7197565 Abdelaziz et al. Mar 2007 B2
7200716 Aiello Apr 2007 B1
7203063 Bash et al. Apr 2007 B2
7203746 Harrop Apr 2007 B1
7203753 Yeager et al. Apr 2007 B2
7206819 Schmidt Apr 2007 B2
7206841 Traversat et al. Apr 2007 B2
7206934 Pabla et al. Apr 2007 B2
7213047 Yeager et et al. May 2007 B2
7213050 Shaffer May 2007 B1
7213062 Raciborski et al. May 2007 B1
7213065 Watt May 2007 B2
7216173 Clayton et al. May 2007 B2
7222187 Yeager et al. May 2007 B2
7222343 Heyrman et al. May 2007 B2
7225249 Barry et al. May 2007 B1
7225442 Dutta et al. May 2007 B2
7228348 Farley Jun 2007 B1
7228350 Hong et al. Jun 2007 B2
7231445 Aweya et al. Jun 2007 B1
7233569 Swallow Jun 2007 B1
7233669 Swallow Jun 2007 B2
7236915 Algieri et al. Jun 2007 B2
7237243 Sutton et al. Jun 2007 B2
7242501 Ishimoto Jul 2007 B2
7243351 Kundu Jul 2007 B2
7249179 Romero et al. Jul 2007 B1
7251222 Chen et al. Jul 2007 B2
7251688 Leighton et al. Jul 2007 B2
7254608 Yeager et al. Aug 2007 B2
7257655 Burney et al. Aug 2007 B1
7260846 Day Aug 2007 B2
7263288 Islam Aug 2007 B1
7263560 Abdelaziz et al. Aug 2007 B2
7263596 Wideman Aug 2007 B1
7274705 Chang et al. Sep 2007 B2
7275018 Abu-El-Zeet et al. Sep 2007 B2
7275102 Yeager et al. Sep 2007 B2
7275249 Miller et al. Sep 2007 B1
7278008 Case et al. Oct 2007 B1
7278142 Bandhole et al. Oct 2007 B2
7278582 Siegel et al. Oct 2007 B1
7281045 Aggarwal et al. Oct 2007 B2
7283838 Lu Oct 2007 B2
7284109 Paxie et al. Oct 2007 B1
7289619 Vivadelli et al. Oct 2007 B2
7289985 Zeng et al. Oct 2007 B2
7293092 Sukegawa Nov 2007 B2
7296268 Darling et al. Nov 2007 B2
7299294 Bruck et al. Nov 2007 B1
7305464 Phillipi et al. Dec 2007 B2
7308496 Yeager et al. Dec 2007 B2
7308687 Trossman et al. Dec 2007 B2
7310319 Awsienko et al. Dec 2007 B2
7313793 Traut et al. Dec 2007 B2
7315887 Liang et al. Jan 2008 B1
7320025 Steinberg et al. Jan 2008 B1
7324555 Chen et al. Jan 2008 B1
7325050 O'Connor et al. Jan 2008 B2
7328243 Yeager et al. Feb 2008 B2
7328264 Babka Feb 2008 B2
7328406 Kalinoski et al. Feb 2008 B2
7334108 Case et al. Feb 2008 B1
7334230 Chung et al. Feb 2008 B2
7337333 O'Conner et al. Feb 2008 B2
7337446 Sankaranarayan et al. Feb 2008 B2
7340500 Traversat et al. Mar 2008 B2
7340578 Khanzode Mar 2008 B1
7340777 Szor Mar 2008 B1
7343467 Brown et al. Mar 2008 B2
7349348 Johnson et al. Mar 2008 B1
7350186 Coleman et al. Mar 2008 B2
7353276 Bain et al. Apr 2008 B2
7353362 Georgiou et al. Apr 2008 B2
7353495 Somogyi Apr 2008 B2
7356655 Allen et al. Apr 2008 B2
7356770 Jackson Apr 2008 B1
7363346 Groner et al. Apr 2008 B2
7366101 Varier et al. Apr 2008 B1
7366719 Shaw Apr 2008 B2
7370092 Aderton et al. May 2008 B2
7373391 Iinuma May 2008 B2
7373524 Motsinger et al. May 2008 B2
7376693 Neiman et al. May 2008 B2
7380039 Miloushev et al. May 2008 B2
7382154 Ramos et al. Jun 2008 B2
7383433 Yeager et al. Jun 2008 B2
7386586 Headley et al. Jun 2008 B1
7386611 Dias et al. Jun 2008 B2
7386850 Mullen Jun 2008 B2
7386888 Liang et al. Jun 2008 B2
7389310 Bhagwan et al. Jun 2008 B1
7392325 Grove et al. Jun 2008 B2
7392360 Aharoni Jun 2008 B1
7395536 Verbeke et al. Jul 2008 B2
7395537 Brown Jul 2008 B1
7398216 Barnett et al. Jul 2008 B2
7398471 Rambacher Jul 2008 B1
7398525 Leymann Jul 2008 B2
7401114 Block et al. Jul 2008 B1
7401152 Traversal et al. Jul 2008 B2
7401153 Traversal et al. Jul 2008 B2
7401355 Supnik et al. Jul 2008 B2
7403994 Vogl et al. Jul 2008 B1
7409433 Lowery et al. Aug 2008 B2
7412492 Waldspurger Aug 2008 B1
7412703 Cleary et al. Aug 2008 B2
7415709 Hipp et al. Aug 2008 B2
7418518 Grove et al. Aug 2008 B2
7418534 Hayter et al. Aug 2008 B2
7421402 Chang et al. Sep 2008 B2
7421500 Talwar et al. Sep 2008 B2
7423971 Mohaban et al. Sep 2008 B1
7426489 Van Soestbergen et al. Sep 2008 B2
7426546 Breiter et al. Sep 2008 B2
7428540 Coates et al. Sep 2008 B1
7433304 Galloway et al. Oct 2008 B1
7437460 Chidambaran et al. Oct 2008 B2
7437540 Paolucci et al. Oct 2008 B2
7437730 Goyal Oct 2008 B2
7441261 Slater et al. Oct 2008 B2
7447147 Nguyen et al. Nov 2008 B2
7447197 Terrell et al. Nov 2008 B2
7451197 Davis Nov 2008 B2
7451199 Kandefer et al. Nov 2008 B2
7451201 Alex et al. Nov 2008 B2
7454467 Girouard et al. Nov 2008 B2
7461134 Ambrose Dec 2008 B2
7463587 Rajsic et al. Dec 2008 B2
7464159 Luoffo et al. Dec 2008 B2
7464160 Iszlai et al. Dec 2008 B2
7466712 Makishima et al. Dec 2008 B2
7466810 Quon et al. Dec 2008 B1
7467225 Anerousis et al. Dec 2008 B2
7467306 Cartes et al. Dec 2008 B2
7467358 Kang et al. Dec 2008 B2
7475419 Basu et al. Jan 2009 B1
7483945 Blumofe Jan 2009 B2
7484008 Gelvin et al. Jan 2009 B1
7484225 Hugly et al. Jan 2009 B2
7487254 Walsh et al. Feb 2009 B2
7487509 Hugly et al. Feb 2009 B2
7492720 Pruthi et al. Feb 2009 B2
7496494 Altman Feb 2009 B2
7502747 Pardo et al. Mar 2009 B1
7502884 Shah et al. Mar 2009 B1
7503045 Aziz et al. Mar 2009 B1
7505463 Schuba Mar 2009 B2
7512649 Faybishenko et al. Mar 2009 B2
7512894 Hintermeister Mar 2009 B1
7516208 Kerrison Apr 2009 B1
7516221 Souder et al. Apr 2009 B2
7516455 Matheson et al. Apr 2009 B2
7519677 Lowery et al. Apr 2009 B2
7519843 Buterbaugh et al. Apr 2009 B1
7526479 Zenz Apr 2009 B2
7529835 Agronow et al. May 2009 B1
7533141 Nadgi et al. May 2009 B2
7533161 Hugly et al. May 2009 B2
7533172 Traversal et al. May 2009 B2
7533385 Barnes May 2009 B1
7536541 Isaacson May 2009 B2
7543052 Klein Jun 2009 B1
7546553 Bozak et al. Jun 2009 B2
7551614 Teisberg et al. Jun 2009 B2
7554930 Gaddis et al. Jun 2009 B2
7555666 Brundridge et al. Jun 2009 B2
7562143 Fellenstein et al. Jul 2009 B2
7568199 Bozak et al. Jul 2009 B2
7570943 Sorvari et al. Aug 2009 B2
7571438 Jones et al. Aug 2009 B2
7574523 Traversal et al. Aug 2009 B2
7577722 Khandekar et al. Aug 2009 B1
7577834 Traversal et al. Aug 2009 B1
7577959 Nguyen et al. Aug 2009 B2
7580382 Amis et al. Aug 2009 B1
7580919 Hannel Aug 2009 B1
7583607 Steele et al. Sep 2009 B2
7583661 Chaudhuri Sep 2009 B2
7584239 Yan Sep 2009 B1
7584274 Bond et al. Sep 2009 B2
7586841 Vasseur Sep 2009 B2
7590746 Slater et al. Sep 2009 B2
7590747 Coates et al. Sep 2009 B2
7594011 Chandra Sep 2009 B2
7594015 Bozak et al. Sep 2009 B2
7596144 Pong Sep 2009 B2
7596784 Abrams et al. Sep 2009 B2
7599360 Edsall et al. Oct 2009 B2
7606225 Xie et al. Oct 2009 B2
7606245 Ma et al. Oct 2009 B2
7610266 Cascaval Oct 2009 B2
7610289 Muret et al. Oct 2009 B2
7613796 Harvey et al. Nov 2009 B2
7616646 Ma et al. Nov 2009 B1
7620057 Aloni et al. Nov 2009 B1
7620635 Hornick Nov 2009 B2
7620706 Jackson Nov 2009 B2
7624118 Schipunov et al. Nov 2009 B2
7624194 Kakivaya et al. Nov 2009 B2
7627691 Buchsbaum et al. Dec 2009 B1
7631066 Schatz et al. Dec 2009 B1
7631307 Wang et al. Dec 2009 B2
7640353 Shen et al. Dec 2009 B2
7640547 Neiman et al. Dec 2009 B2
7644215 Wallace et al. Jan 2010 B2
7657535 Moyaux et al. Feb 2010 B2
7657597 Arora et al. Feb 2010 B2
7657626 Zwicky Feb 2010 B1
7657677 Huang et al. Feb 2010 B2
7657756 Hall Feb 2010 B2
7657779 Kaminsky Feb 2010 B2
7660887 Reedy et al. Feb 2010 B2
7660922 Harriman Feb 2010 B2
7664110 Lovett et al. Feb 2010 B1
7665090 Tormasov et al. Feb 2010 B1
7668809 Kelly et al. Feb 2010 B1
7673164 Agarwal Mar 2010 B1
7680933 Fatula, Jr. Mar 2010 B2
7685281 Saraiya et al. Mar 2010 B1
7685599 Kanai et al. Mar 2010 B2
7685602 Tran et al. Mar 2010 B1
7689661 Lowery et al. Mar 2010 B2
7693976 Perry et al. Apr 2010 B2
7693993 Sheets et al. Apr 2010 B2
7694076 Lowery et al. Apr 2010 B2
7694305 Karlsson et al. Apr 2010 B2
7698386 Amidon et al. Apr 2010 B2
7698398 Lai Apr 2010 B1
7698430 Jackson Apr 2010 B2
7701948 Rabie et al. Apr 2010 B2
7702779 Gupta et al. Apr 2010 B1
7707088 Schmelzer Apr 2010 B2
7707185 Czezatke Apr 2010 B1
7710936 Morales Barroso May 2010 B2
7711652 Schmelzer May 2010 B2
7716193 Krishnamoorthy May 2010 B2
7716334 Rao et al. May 2010 B2
7719834 Miyamoto et al. May 2010 B2
7721125 Fung May 2010 B2
7725583 Jackson May 2010 B2
7730220 Hasha et al. Jun 2010 B2
7730262 Lowery et al. Jun 2010 B2
7730488 Ilzuka et al. Jun 2010 B2
7739308 Baffier et al. Jun 2010 B2
7739541 Rao et al. Jun 2010 B1
7742425 El-Damhougy Jun 2010 B2
7742476 Branda et al. Jun 2010 B2
7743147 Suorsa et al. Jun 2010 B2
7747451 Keohane et al. Jun 2010 B2
RE41440 Briscoe et al. Jul 2010 E
7751433 Dollo et al. Jul 2010 B2
7752258 Lewin et al. Jul 2010 B2
7752624 Crawford, Jr. et al. Jul 2010 B2
7756658 Kulkarni et al. Jul 2010 B2
7757033 Mehrotra Jul 2010 B1
7757236 Singh Jul 2010 B1
7760720 Pullela et al. Jul 2010 B2
7761557 Fellenstein et al. Jul 2010 B2
7761687 Blumrich et al. Jul 2010 B2
7765288 Bainbridge et al. Jul 2010 B2
7765299 Romero Jul 2010 B2
7769620 Fernandez et al. Aug 2010 B1
7769803 Birdwell et al. Aug 2010 B2
7770120 Baudisch Aug 2010 B2
7774331 Barth et al. Aug 2010 B2
7774495 Pabla et al. Aug 2010 B2
7778234 Cooke et al. Aug 2010 B2
7782813 Wheeler et al. Aug 2010 B2
7783777 Pabla et al. Aug 2010 B1
7783786 Lauterbach Aug 2010 B1
7783910 Felter et al. Aug 2010 B2
7788403 Darugar et al. Aug 2010 B2
7788477 Huang et al. Aug 2010 B1
7791894 Bechtolsheim Sep 2010 B2
7792113 Foschiano et al. Sep 2010 B1
7793288 Sameske Sep 2010 B2
7796399 Clayton et al. Sep 2010 B2
7796619 Feldmann et al. Sep 2010 B1
7797367 Gelvin et al. Sep 2010 B1
7797393 Qiu et al. Sep 2010 B2
7801132 Ofek et al. Sep 2010 B2
7802017 Uemura et al. Sep 2010 B2
7805448 Andrzejak et al. Sep 2010 B2
7805575 Agarwal et al. Sep 2010 B1
7810090 Gebhart Oct 2010 B2
7813822 Hoffberg Oct 2010 B1
7827361 Karlsson et al. Nov 2010 B1
7830820 Duke et al. Nov 2010 B2
7831839 Hatakeyama Nov 2010 B2
7840353 Ouksel et al. Nov 2010 B2
7840703 Arimilli et al. Nov 2010 B2
7840810 Eastham Nov 2010 B2
7844687 Gelvin et al. Nov 2010 B1
7844787 Ranganathan et al. Nov 2010 B2
7848262 El-Damhougy Dec 2010 B2
7849139 Wolfson et al. Dec 2010 B2
7849140 Abdel-Aziz et al. Dec 2010 B2
7853880 Porter Dec 2010 B2
7860999 Subramanian Dec 2010 B1
7865614 Lu et al. Jan 2011 B2
7886023 Johnson Feb 2011 B1
7889675 Mack-Crane et al. Feb 2011 B2
7890571 Kriegsman Feb 2011 B1
7890701 Lowery et al. Feb 2011 B2
7891004 Gelvin et al. Feb 2011 B1
RE42262 Stephens, Jr. Mar 2011 E
7899047 Cabrera et al. Mar 2011 B2
7899864 Margulis Mar 2011 B2
7900206 Joshi et al. Mar 2011 B1
7904569 Gelvin et al. Mar 2011 B1
7921169 Jacobs Apr 2011 B2
7925795 Tamir et al. Apr 2011 B2
7930397 Midgley Apr 2011 B2
7934005 Fascenda Apr 2011 B2
7958262 Hasha et al. Jun 2011 B2
7970830 Staggs Jun 2011 B2
7970929 Mahalingaiah Jun 2011 B1
7971204 Jackson Jun 2011 B2
7975032 Lowery et al. Jul 2011 B2
7975035 Popescu et al. Jul 2011 B2
7975110 Spaur et al. Jul 2011 B1
7984137 O'Toole, Jr. et al. Jul 2011 B2
7984183 Andersen et al. Jul 2011 B2
7991817 Dehon et al. Aug 2011 B2
7991922 Hayter et al. Aug 2011 B2
7992151 Warrier et al. Aug 2011 B2
7992983 Nanjo Aug 2011 B2
7995501 Jetcheva et al. Aug 2011 B2
7996458 Nielsen Aug 2011 B2
7996510 Vicente Aug 2011 B2
8000288 Wheeler et al. Aug 2011 B2
8014408 Habetha et al. Sep 2011 B2
8018860 Cook Sep 2011 B1
8019832 De Sousa et al. Sep 2011 B2
8032634 Eppstein Oct 2011 B1
8037202 Yeager et al. Oct 2011 B2
8037475 Jackson Oct 2011 B1
8041773 Abu-Ghazaleh et al. Oct 2011 B2
8055788 Chan et al. Nov 2011 B1
8060552 Hinni et al. Nov 2011 B2
8060619 Saulpaugh Nov 2011 B1
8060760 Shetty et al. Nov 2011 B2
8060775 Sharma et al. Nov 2011 B1
8073978 Sengupta et al. Dec 2011 B2
8078708 Wang et al. Dec 2011 B1
8079118 Gelvin et al. Dec 2011 B2
8082400 Chang et al. Dec 2011 B1
8090880 Hasha et al. Jan 2012 B2
8095600 Hasha et al. Jan 2012 B2
8095601 Hasha et al. Jan 2012 B2
8103543 Zwicky Jan 2012 B1
8108455 Yeager et al. Jan 2012 B2
8108508 Goh et al. Jan 2012 B1
8108512 Howard et al. Jan 2012 B2
8108930 Hoefelmeyer et al. Jan 2012 B2
8122269 Houlihan et al. Feb 2012 B2
8132034 Lambert et al. Mar 2012 B2
8135812 Lowery et al. Mar 2012 B2
8140658 Gelvin et al. Mar 2012 B1
8151103 Jackson Apr 2012 B2
8155113 Agarwal Apr 2012 B1
8156362 Branover et al. Apr 2012 B2
8160077 Traversat et al. Apr 2012 B2
8161391 McClelland et al. Apr 2012 B2
8165120 Maruccia et al. Apr 2012 B2
8166063 Andersen et al. Apr 2012 B2
8166204 Basu et al. Apr 2012 B2
8170040 Konda May 2012 B2
8171136 Petite May 2012 B2
8176189 Traversat et al. May 2012 B2
8176490 Jackson May 2012 B1
8180996 Fullerton et al. May 2012 B2
8185776 Gentes et al. May 2012 B1
8189612 Lemaire et al. May 2012 B2
8194659 Ban Jun 2012 B2
8196133 Kakumani et al. Jun 2012 B2
8199636 Rouyer et al. Jun 2012 B1
8204992 Arora et al. Jun 2012 B2
8205044 Lowery et al. Jun 2012 B2
8205103 Kazama et al. Jun 2012 B2
8205210 Cleary et al. Jun 2012 B2
8244671 Chen et al. Aug 2012 B2
8260893 Bandhole et al. Sep 2012 B1
8261349 Peng Sep 2012 B2
8266321 Johnston-Watt et al. Sep 2012 B2
8271628 Lowery et al. Sep 2012 B2
8271980 Jackson Sep 2012 B2
8275881 Fellenstein et al. Sep 2012 B2
8302100 Deng et al. Oct 2012 B2
8321048 Coss et al. Nov 2012 B1
8346591 Fellenstein et al. Jan 2013 B2
8346908 Vanyukhin et al. Jan 2013 B1
8359397 Traversal et al. Jan 2013 B2
8370898 Jackson Feb 2013 B1
8379425 Fukuoka et al. Feb 2013 B2
8380846 Abu-Ghazaleh et al. Feb 2013 B1
8386622 Jacobson Feb 2013 B2
8392515 Kakivaya et al. Mar 2013 B2
8396757 Fellenstein et al. Mar 2013 B2
8397092 Karnowski Mar 2013 B2
8402540 Kapoor et al. Mar 2013 B2
8407428 Cheriton et al. Mar 2013 B2
8413155 Jackson Apr 2013 B2
8417715 Bruckhaus et al. Apr 2013 B1
8417813 Kakivaya et al. Apr 2013 B2
8429396 Trivedi Apr 2013 B1
8458333 Stoica et al. Jun 2013 B1
8463867 Robertson et al. Jun 2013 B2
8464250 Ansel Jun 2013 B1
8484382 Das et al. Jul 2013 B2
8495201 Klincewicz Jul 2013 B2
8504663 Lowery et al. Aug 2013 B2
8504791 Cheriton et al. Aug 2013 B2
8516470 van Rietschote Aug 2013 B1
8544017 Prael et al. Sep 2013 B1
8554920 Chen et al. Oct 2013 B2
8560639 Murphy et al. Oct 2013 B2
8572326 Lowery et al. Oct 2013 B2
RE44610 Krakirian et al. Nov 2013 E
8578130 DeSota et al. Nov 2013 B2
8584129 Czajkowski Nov 2013 B1
8589517 Hoefelmeyer et al. Nov 2013 B2
8599863 Davis Dec 2013 B2
8601595 Gelvin et al. Dec 2013 B2
8606800 Lagad et al. Dec 2013 B2
8615602 Li et al. Dec 2013 B2
8626820 Levy Jan 2014 B1
8631130 Jackson Jan 2014 B2
8684802 Gross et al. Apr 2014 B1
8701121 Saffre Apr 2014 B2
8726278 Shawver et al. May 2014 B1
8737410 Davis May 2014 B2
8738860 Griffin et al. May 2014 B1
8745275 Ikeya et al. Jun 2014 B2
8745302 Davis et al. Jun 2014 B2
8782120 Jackson Jul 2014 B2
8782231 Jackson Jul 2014 B2
8782321 Harriman et al. Jul 2014 B2
8782654 Jackson Jul 2014 B2
8812400 Faraboschi et al. Aug 2014 B2
8824485 Biswas et al. Sep 2014 B2
8826270 Lewis Sep 2014 B1
8854831 Arnouse Oct 2014 B2
8863143 Jackson Oct 2014 B2
8903964 Breslin Dec 2014 B2
8924560 Pang Dec 2014 B2
8930536 Jackson Jan 2015 B2
8954584 Subbarayan et al. Feb 2015 B1
9008079 Davis et al. Apr 2015 B2
9038078 Jackson May 2015 B2
9054990 Davis Jun 2015 B2
9060060 Lobig Jun 2015 B2
9069611 Jackson Jun 2015 B2
9069929 Borland Jun 2015 B2
9075655 Davis et al. Jul 2015 B2
9075657 Jackson Jul 2015 B2
9077654 Davis Jul 2015 B2
9092594 Borland Jul 2015 B2
9112813 Jackson Aug 2015 B2
9116755 Jackson Aug 2015 B2
9128767 Jackson Sep 2015 B2
9152455 Jackson Oct 2015 B2
9176785 Jackson Nov 2015 B2
9231886 Jackson Jan 2016 B2
9258276 Dalal et al. Feb 2016 B2
9262225 Davis Feb 2016 B2
9268607 Jackson Feb 2016 B2
9288147 Kern Mar 2016 B2
9304896 Chandra et al. Apr 2016 B2
9311269 Davis Apr 2016 B2
9367802 Arndt et al. Jun 2016 B2
9405584 Davis Aug 2016 B2
9413687 Jackson Aug 2016 B2
9438515 McCormick Sep 2016 B2
9450875 Tong Sep 2016 B1
9454403 Davis Sep 2016 B2
9465771 Davis et al. Oct 2016 B2
9479463 Davis Oct 2016 B2
9491064 Jackson Nov 2016 B2
9509552 Davis Nov 2016 B2
9575805 Jackson Feb 2017 B2
9585281 Schnell Feb 2017 B2
9602573 Abu-Ghazaleh et al. Mar 2017 B1
9619296 Jackson Apr 2017 B2
9648102 Davis et al. May 2017 B1
9680770 Davis Jun 2017 B2
9749326 Davis Aug 2017 B2
9778959 Jackson Oct 2017 B2
9785479 Jackson Oct 2017 B2
9792249 Borland Oct 2017 B2
9825860 Hu Nov 2017 B2
9866477 Davis Jan 2018 B2
9876735 Davis Jan 2018 B2
9886322 Jackson Feb 2018 B2
9929976 Davis Mar 2018 B2
9959140 Jackson May 2018 B2
9959141 Jackson May 2018 B2
9961013 Jackson May 2018 B2
9965442 Borland May 2018 B2
9977763 Davis May 2018 B2
9979672 Jackson May 2018 B2
10021806 Schnell Jul 2018 B2
10050970 Davis Aug 2018 B2
10135731 Davis Nov 2018 B2
10140245 Davis et al. Nov 2018 B2
10212092 Dalal et al. Feb 2019 B2
10277531 Jackson Apr 2019 B2
10311014 Dalton Jun 2019 B2
10333862 Jackson Jun 2019 B2
10379909 Jackson Aug 2019 B2
10445146 Jackson Oct 2019 B2
10445148 Jackson Oct 2019 B2
10585704 Jackson Mar 2020 B2
10608949 Jackson Mar 2020 B2
10733028 Jackson Aug 2020 B2
10735505 Abu-Ghazaleh et al. Aug 2020 B2
10871999 Jackson Dec 2020 B2
10951487 Jackson Mar 2021 B2
10977090 Jackson Apr 2021 B2
11132277 Dalton Sep 2021 B2
11134022 Jackson Sep 2021 B2
11144355 Jackson Oct 2021 B2
11356385 Jackson Jun 2022 B2
11467883 Jackson Oct 2022 B2
11494235 Jackson Nov 2022 B2
11496415 Jackson Nov 2022 B2
11522811 Jackson Dec 2022 B2
11522952 Abu-Ghazaleh Dec 2022 B2
11526304 Davis et al. Dec 2022 B2
11533274 Jackson Dec 2022 B2
11537434 Jackson Dec 2022 B2
11537435 Jackson Dec 2022 B2
11630704 Jackson Apr 2023 B2
11650857 Jackson May 2023 B2
11652706 Jackson May 2023 B2
11656907 Jackson May 2023 B2
11658916 Jackson May 2023 B2
11709709 Jackson Jul 2023 B2
11720290 Davis Aug 2023 B2
11762694 Jackson Sep 2023 B2
11765101 Jackson Sep 2023 B2
11831564 Jackson Nov 2023 B2
11861404 Jackson Jan 2024 B2
11886915 Jackson Jan 2024 B2
11960937 Jackson Apr 2024 B2
12008405 Jackson Jun 2024 B2
12009996 Jackson Jun 2024 B2
20010010605 Aoki Aug 2001 A1
20010015733 Sklar Aug 2001 A1
20010023431 Horiguchi Sep 2001 A1
20010032109 Gonyea Oct 2001 A1
20010034752 Kremien Oct 2001 A1
20010037311 McCoy et al. Nov 2001 A1
20010044667 Nakano Nov 2001 A1
20010044759 Kutsumi Nov 2001 A1
20010046227 Matsuhira et al. Nov 2001 A1
20010051929 Suzuki Dec 2001 A1
20010052016 Skene et al. Dec 2001 A1
20010052108 Bowman-Amuah Dec 2001 A1
20020002578 Yamashita Jan 2002 A1
20020002636 Vange et al. Jan 2002 A1
20020004833 Tonouchi Jan 2002 A1
20020004912 Fung Jan 2002 A1
20020007389 Jones et al. Jan 2002 A1
20020010783 Primak et al. Jan 2002 A1
20020016809 Foulger Feb 2002 A1
20020018481 Mor et al. Feb 2002 A1
20020031364 Suzuki et al. Mar 2002 A1
20020032716 Nagato Mar 2002 A1
20020035605 McDowell Mar 2002 A1
20020040391 Chaiken et al. Apr 2002 A1
20020049608 Hartsell et al. Apr 2002 A1
20020052909 Seeds May 2002 A1
20020052961 Yoshimine et al. May 2002 A1
20020053006 Kawamoto May 2002 A1
20020059094 Hosea et al. May 2002 A1
20020059274 Hartsell et al. May 2002 A1
20020062377 Hillman et al. May 2002 A1
20020062451 Scheidt et al. May 2002 A1
20020062465 Goto May 2002 A1
20020065864 Hartsell et al. May 2002 A1
20020083299 Van Huben et al. Jun 2002 A1
20020083352 Fujimoto et al. Jun 2002 A1
20020087611 Tanaka et al. Jul 2002 A1
20020087699 Karagiannis et al. Jul 2002 A1
20020090075 Gabriel Jul 2002 A1
20020091786 Yamaguchi et al. Jul 2002 A1
20020093915 Larson Jul 2002 A1
20020097732 Worster et al. Jul 2002 A1
20020099842 Jennings et al. Jul 2002 A1
20020103681 Tolis Aug 2002 A1
20020103886 Rawson, III Aug 2002 A1
20020107903 Richter et al. Aug 2002 A1
20020107962 Richter et al. Aug 2002 A1
20020116234 Nagasawa Aug 2002 A1
20020116721 Dobes et al. Aug 2002 A1
20020120741 Webb et al. Aug 2002 A1
20020124128 Qiu Sep 2002 A1
20020129160 Habelha Sep 2002 A1
20020129274 Baskey et al. Sep 2002 A1
20020133537 Lau et al. Sep 2002 A1
20020133821 Shteyn Sep 2002 A1
20020137565 Blanco Sep 2002 A1
20020138459 Mandal Sep 2002 A1
20020138635 Redlich et al. Sep 2002 A1
20020138679 Koning Sep 2002 A1
20020143855 Traversat Oct 2002 A1
20020143944 Traversat et al. Oct 2002 A1
20020147663 Walker et al. Oct 2002 A1
20020147771 Traversal et al. Oct 2002 A1
20020147810 Traversal et al. Oct 2002 A1
20020151271 Tatsuji Oct 2002 A1
20020152299 Traversal et al. Oct 2002 A1
20020152305 Jackson et al. Oct 2002 A1
20020156675 Pedone Oct 2002 A1
20020156699 Gray et al. Oct 2002 A1
20020156891 Ulrich et al. Oct 2002 A1
20020156893 Pouyoul et al. Oct 2002 A1
20020156904 Gullotta et al. Oct 2002 A1
20020156984 Padovano Oct 2002 A1
20020159452 Foster et al. Oct 2002 A1
20020161869 Griffin et al. Oct 2002 A1
20020161917 Shapiro et al. Oct 2002 A1
20020166110 Powell Nov 2002 A1
20020166117 Abrams et al. Nov 2002 A1
20020172205 Tagore-Brage et al. Nov 2002 A1
20020173984 Robertson et al. Nov 2002 A1
20020174165 Kawaguchi Nov 2002 A1
20020174227 Hartsell et al. Nov 2002 A1
20020184129 Arena Dec 2002 A1
20020184310 Traversal et al. Dec 2002 A1
20020184311 Traversal et al. Dec 2002 A1
20020184357 Traversal et al. Dec 2002 A1
20020184358 Traversal et al. Dec 2002 A1
20020186656 Vu Dec 2002 A1
20020188657 Traversal et al. Dec 2002 A1
20020194242 Chandrasekaran Dec 2002 A1
20020194384 Habelha Dec 2002 A1
20020194412 Bottom Dec 2002 A1
20020196611 Ho et al. Dec 2002 A1
20020196734 Tanaka et al. Dec 2002 A1
20020198734 Greene et al. Dec 2002 A1
20020198923 Hayes Dec 2002 A1
20030004772 Dutta et al. Jan 2003 A1
20030005130 Cheng Jan 2003 A1
20030005162 Habelha Jan 2003 A1
20030007493 Oi et al. Jan 2003 A1
20030009506 Bril et al. Jan 2003 A1
20030014503 Legout et al. Jan 2003 A1
20030014524 Tormasov Jan 2003 A1
20030014539 Reznick Jan 2003 A1
20030014613 Soni Jan 2003 A1
20030018573 Comas Jan 2003 A1
20030018766 Duvvuru Jan 2003 A1
20030018803 El Batt et al. Jan 2003 A1
20030028443 Ellis Feb 2003 A1
20030028585 Yeager et al. Feb 2003 A1
20030028642 Agarwal et al. Feb 2003 A1
20030028645 Romagnoli Feb 2003 A1
20030028656 Babka Feb 2003 A1
20030033547 Larson et al. Feb 2003 A1
20030036820 Yellepeddy et al. Feb 2003 A1
20030039213 Holtzman Feb 2003 A1
20030039246 Guo et al. Feb 2003 A1
20030041141 Abdelaziz et al. Feb 2003 A1
20030041238 French Feb 2003 A1
20030041266 Ke et al. Feb 2003 A1
20030041308 Ganesan et al. Feb 2003 A1
20030046330 Hayes Mar 2003 A1
20030050924 Faybishenko et al. Mar 2003 A1
20030050959 Faybishenko et al. Mar 2003 A1
20030050989 Marinescu et al. Mar 2003 A1
20030051127 Miwa Mar 2003 A1
20030055894 Yeager et al. Mar 2003 A1
20030055898 Yeager et al. Mar 2003 A1
20030058277 Bowman-Amuah Mar 2003 A1
20030061260 Rajkumar Mar 2003 A1
20030061261 Greene Mar 2003 A1
20030061262 Hahn et al. Mar 2003 A1
20030065703 Aborn Apr 2003 A1
20030065784 Herrod Apr 2003 A1
20030069828 Blazey Apr 2003 A1
20030069918 Lu et al. Apr 2003 A1
20030069949 Chan et al. Apr 2003 A1
20030072263 Peterson Apr 2003 A1
20030074090 Becka Apr 2003 A1
20030076832 Ni Apr 2003 A1
20030081938 Nishimura May 2003 A1
20030084435 Messer May 2003 A1
20030088457 Keil et al. May 2003 A1
20030093255 Freyensee et al. May 2003 A1
20030093624 Arimilli et al. May 2003 A1
20030093647 Mogi May 2003 A1
20030097284 Shinozaki May 2003 A1
20030097429 Wu et al. May 2003 A1
20030097439 Strayer et al. May 2003 A1
20030101084 Perez May 2003 A1
20030103413 Mandava et al. Jun 2003 A1
20030105655 Kimbrel et al. Jun 2003 A1
20030105721 Ginter et al. Jun 2003 A1
20030110262 Hasan et al. Jun 2003 A1
20030112792 Cranor et al. Jun 2003 A1
20030115562 Martin Jun 2003 A1
20030120472 Lind Jun 2003 A1
20030120701 Pulsipher et al. Jun 2003 A1
20030120704 Tran et al. Jun 2003 A1
20030120710 Pulsipher et al. Jun 2003 A1
20030120780 Zhu Jun 2003 A1
20030126013 Shand Jul 2003 A1
20030126200 Wolff Jul 2003 A1
20030126202 Watt Jul 2003 A1
20030126265 Aziz et al. Jul 2003 A1
20030126283 Prakash et al. Jul 2003 A1
20030131043 Berg et al. Jul 2003 A1
20030131209 Lee Jul 2003 A1
20030135509 Davis Jul 2003 A1
20030135615 Wyatt Jul 2003 A1
20030135621 Romagnoli Jul 2003 A1
20030140190 Mahony et al. Jul 2003 A1
20030144894 Robertson et al. Jul 2003 A1
20030149685 Trossman et al. Aug 2003 A1
20030154112 Neiman et al. Aug 2003 A1
20030158884 Alford Aug 2003 A1
20030158940 Leigh Aug 2003 A1
20030159083 Fukuhara et al. Aug 2003 A1
20030169269 Sasaki et al. Sep 2003 A1
20030172191 Williams Sep 2003 A1
20030177050 Crampton Sep 2003 A1
20030177121 Moona et al. Sep 2003 A1
20030177239 Shinohara Sep 2003 A1
20030177334 King et al. Sep 2003 A1
20030182421 Faybishenko et al. Sep 2003 A1
20030182425 Kurakake Sep 2003 A1
20030182429 Jagels Sep 2003 A1
20030182496 Yoo Sep 2003 A1
20030185229 Shachar et al. Oct 2003 A1
20030187907 Ito Oct 2003 A1
20030188083 Kumar et al. Oct 2003 A1
20030191795 Bernardin et al. Oct 2003 A1
20030191857 Terrell et al. Oct 2003 A1
20030193402 Post et al. Oct 2003 A1
20030195931 Dauger Oct 2003 A1
20030200109 Honda et al. Oct 2003 A1
20030200258 Hayashi Oct 2003 A1
20030202520 Witkowski et al. Oct 2003 A1
20030202709 Simard et al. Oct 2003 A1
20030204709 Rich Oct 2003 A1
20030204773 Petersen et al. Oct 2003 A1
20030204786 Dinker Oct 2003 A1
20030210694 Jayaraman et al. Nov 2003 A1
20030212738 Wookey et al. Nov 2003 A1
20030212792 Raymond Nov 2003 A1
20030216927 Sridhar Nov 2003 A1
20030216951 Ginis et al. Nov 2003 A1
20030217129 Knittel et al. Nov 2003 A1
20030218627 Gusler Nov 2003 A1
20030227934 White Dec 2003 A1
20030231624 Alappat et al. Dec 2003 A1
20030231647 Petrovykh Dec 2003 A1
20030233378 Butler et al. Dec 2003 A1
20030233446 Earl Dec 2003 A1
20030236745 Hartsell Dec 2003 A1
20030236854 Rom Dec 2003 A1
20030236880 Srivastava Dec 2003 A1
20040003077 Bantz et al. Jan 2004 A1
20040003086 Parham et al. Jan 2004 A1
20040009751 Michaelis Jan 2004 A1
20040010544 Slater et al. Jan 2004 A1
20040010550 Gopinath Jan 2004 A1
20040010592 Carver et al. Jan 2004 A1
20040011761 Hensley Jan 2004 A1
20040013113 Singh et al. Jan 2004 A1
20040015579 Cooper et al. Jan 2004 A1
20040015973 Skovira Jan 2004 A1
20040017806 Yazdy et al. Jan 2004 A1
20040017808 Forbes et al. Jan 2004 A1
20040021678 Ullah Feb 2004 A1
20040024853 Cates Feb 2004 A1
20040030741 Wolton et al. Feb 2004 A1
20040030743 Hugly et al. Feb 2004 A1
20040030794 Hugly et al. Feb 2004 A1
20040030938 Barr et al. Feb 2004 A1
20040034873 Zenoni Feb 2004 A1
20040039815 Evans et al. Feb 2004 A1
20040042487 Ossman Mar 2004 A1
20040043755 Shimooka Mar 2004 A1
20040044718 Ferstl et al. Mar 2004 A1
20040044727 Abdelaziz et al. Mar 2004 A1
20040054630 Ginter et al. Mar 2004 A1
20040054777 Ackaouy et al. Mar 2004 A1
20040054780 Romero Mar 2004 A1
20040054807 Harvey et al. Mar 2004 A1
20040054999 Willen Mar 2004 A1
20040064511 Abdel-Aziz et al. Apr 2004 A1
20040064512 Arora et al. Apr 2004 A1
20040064568 Arora et al. Apr 2004 A1
20040064817 Shibayama et al. Apr 2004 A1
20040066782 Nassar Apr 2004 A1
20040068411 Scanlan Apr 2004 A1
20040068676 Larson et al. Apr 2004 A1
20040068730 Miller et al. Apr 2004 A1
20040071147 Roadknight et al. Apr 2004 A1
20040073650 Nakamura Apr 2004 A1
20040073854 Windl Apr 2004 A1
20040073908 Benejam et al. Apr 2004 A1
20040081148 Yamada Apr 2004 A1
20040083287 Gao et al. Apr 2004 A1
20040088347 Yeager et al. May 2004 A1
20040088348 Yeager et al. May 2004 A1
20040088369 Yeager et al. May 2004 A1
20040098391 Robertson et al. May 2004 A1
20040098424 Seidenberg May 2004 A1
20040098447 Verbeke et al. May 2004 A1
20040103078 Smedberg et al. May 2004 A1
20040103305 Ginter et al. May 2004 A1
20040103339 Chalasani et al. May 2004 A1
20040103413 Mandava et al. May 2004 A1
20040107123 Haffner Jun 2004 A1
20040107273 Biran et al. Jun 2004 A1
20040107281 Bose et al. Jun 2004 A1
20040109428 Krishnamurthy Jun 2004 A1
20040111307 Demsky et al. Jun 2004 A1
20040111612 Choi et al. Jun 2004 A1
20040117610 Hensley Jun 2004 A1
20040117768 Chang et al. Jun 2004 A1
20040121777 Schwarz et al. Jun 2004 A1
20040122970 Kawaguchi et al. Jun 2004 A1
20040128495 Hensley Jul 2004 A1
20040128670 Robinson et al. Jul 2004 A1
20040133620 Habelha Jul 2004 A1
20040133640 Yeager et al. Jul 2004 A1
20040133665 Deboer et al. Jul 2004 A1
20040133703 Habelha Jul 2004 A1
20040135780 Nims Jul 2004 A1
20040139202 Talwar et al. Jul 2004 A1
20040139464 Ellis et al. Jul 2004 A1
20040141521 George Jul 2004 A1
20040143664 Usa et al. Jul 2004 A1
20040148326 Nadgir Jul 2004 A1
20040148390 Cleary et al. Jul 2004 A1
20040150664 Baudisch Aug 2004 A1
20040151181 Chu Aug 2004 A1
20040153563 Shay et al. Aug 2004 A1
20040158637 Lee Aug 2004 A1
20040162871 Pabla et al. Aug 2004 A1
20040165588 Pandya Aug 2004 A1
20040172464 Nag Sep 2004 A1
20040179528 Powers et al. Sep 2004 A1
20040181370 Froehlich et al. Sep 2004 A1
20040181476 Smith et al. Sep 2004 A1
20040189677 Amann et al. Sep 2004 A1
20040193674 Kurosawa et al. Sep 2004 A1
20040194061 Fujino Sep 2004 A1
20040194098 Chung et al. Sep 2004 A1
20040196308 Blomquist Oct 2004 A1
20040199566 Carlson Oct 2004 A1
20040199621 Lau Oct 2004 A1
20040199646 Susai et al. Oct 2004 A1
20040199918 Skovira Oct 2004 A1
20040203670 King et al. Oct 2004 A1
20040204978 Rayrole Oct 2004 A1
20040205101 Radhakrishnan Oct 2004 A1
20040205206 Naik et al. Oct 2004 A1
20040210624 Andrzejak et al. Oct 2004 A1
20040210632 Carlson Oct 2004 A1
20040210663 Phillips Oct 2004 A1
20040210693 Zeitler et al. Oct 2004 A1
20040213395 Ishii et al. Oct 2004 A1
20040215780 Kawato Oct 2004 A1
20040215858 Armstrong Oct 2004 A1
20040215864 Arimilli et al. Oct 2004 A1
20040215991 McAfee et al. Oct 2004 A1
20040216121 Jones et al. Oct 2004 A1
20040218615 Griffin et al. Nov 2004 A1
20040221038 Clarke et al. Nov 2004 A1
20040236852 Birkestrand et al. Nov 2004 A1
20040243378 Schnatterly et al. Dec 2004 A1
20040243466 Trybinski et al. Dec 2004 A1
20040244006 Kaufman et al. Dec 2004 A1
20040246900 Zhang et al. Dec 2004 A1
20040248576 Ghiglino Dec 2004 A1
20040260701 Lehikoinen Dec 2004 A1
20040260746 Brown et al. Dec 2004 A1
20040267486 Percer et al. Dec 2004 A1
20040267897 Hill et al. Dec 2004 A1
20040267901 Gomez Dec 2004 A1
20040268035 Ueno Dec 2004 A1
20040268315 Gouriou Dec 2004 A1
20050005200 Matena Jan 2005 A1
20050010465 Drew et al. Jan 2005 A1
20050010608 Horikawa Jan 2005 A1
20050015378 Gammel et al. Jan 2005 A1
20050015621 Ashley et al. Jan 2005 A1
20050018604 Dropps et al. Jan 2005 A1
20050018606 Dropps et al. Jan 2005 A1
20050018663 Dropps et al. Jan 2005 A1
20050021291 Retich Jan 2005 A1
20050021371 Basone et al. Jan 2005 A1
20050021606 Davies et al. Jan 2005 A1
20050021728 Sugimoto Jan 2005 A1
20050021759 Gupta et al. Jan 2005 A1
20050021862 Schroeder et al. Jan 2005 A1
20050022188 Tameshige et al. Jan 2005 A1
20050027863 Talwar et al. Feb 2005 A1
20050027864 Bozak et al. Feb 2005 A1
20050027865 Bozak et al. Feb 2005 A1
20050027870 Trebes et al. Feb 2005 A1
20050030954 Dropps et al. Feb 2005 A1
20050033742 Kamvar et al. Feb 2005 A1
20050033890 Lee Feb 2005 A1
20050034070 Meir et al. Feb 2005 A1
20050038808 Kutch Feb 2005 A1
20050038835 Chidambaran et al. Feb 2005 A1
20050039171 Avakian Feb 2005 A1
20050044167 Kobayashi Feb 2005 A1
20050044195 Westfall Feb 2005 A1
20050044205 Sankaranarayan et al. Feb 2005 A1
20050044226 McDermott et al. Feb 2005 A1
20050044228 Birkestrand et al. Feb 2005 A1
20050049884 Hunt et al. Mar 2005 A1
20050050057 Mital et al. Mar 2005 A1
20050050200 Mizoguchi Mar 2005 A1
20050050270 Horn et al. Mar 2005 A1
20050054354 Roman et al. Mar 2005 A1
20050055322 Masters et al. Mar 2005 A1
20050055442 Reeves, Jr. Mar 2005 A1
20050055694 Lee Mar 2005 A1
20050055697 Buco Mar 2005 A1
20050055698 Sasaki et al. Mar 2005 A1
20050060360 Doyle et al. Mar 2005 A1
20050060608 Marchand Mar 2005 A1
20050065826 Baker et al. Mar 2005 A1
20050066302 Kanade Mar 2005 A1
20050066358 Anderson et al. Mar 2005 A1
20050068922 Jalali Mar 2005 A1
20050071843 Guo et al. Mar 2005 A1
20050076145 Ben-Zvi et al. Apr 2005 A1
20050077921 Percer et al. Apr 2005 A1
20050080845 Gopinath Apr 2005 A1
20050080891 Cauthron Apr 2005 A1
20050080930 Joseph Apr 2005 A1
20050081210 Day Apr 2005 A1
20050086300 Yeager et al. Apr 2005 A1
20050086356 Shah Apr 2005 A1
20050091505 Riley et al. Apr 2005 A1
20050097560 Rolia et al. May 2005 A1
20050102396 Hipp May 2005 A1
20050102400 Nakahara May 2005 A1
20050102683 Branson May 2005 A1
20050105538 Perera et al. May 2005 A1
20050108407 Johnson et al. May 2005 A1
20050108703 Hellier May 2005 A1
20050113203 Mueller et al. May 2005 A1
20050114460 Chen May 2005 A1
20050114478 Popescu et al. May 2005 A1
20050114551 Basu et al. May 2005 A1
20050114862 Bisdikian et al. May 2005 A1
20050120160 Plouffe et al. Jun 2005 A1
20050125213 Chen et al. Jun 2005 A1
20050125537 Martins et al. Jun 2005 A1
20050125538 Tawil Jun 2005 A1
20050131898 Fatula, Jr. Jun 2005 A1
20050132378 Horvitz et al. Jun 2005 A1
20050132379 Sankaran et al. Jun 2005 A1
20050138618 Gebhart Jun 2005 A1
20050141424 Lim et al. Jun 2005 A1
20050144315 George et al. Jun 2005 A1
20050144619 Newman Jun 2005 A1
20050149940 Calinescu et al. Jul 2005 A1
20050154861 Arimilli et al. Jul 2005 A1
20050155033 Luoffo et al. Jul 2005 A1
20050156732 Matsumura Jul 2005 A1
20050160137 Ishikawa et al. Jul 2005 A1
20050160413 Broussard Jul 2005 A1
20050160424 Broussard Jul 2005 A1
20050163143 Kalantar et al. Jul 2005 A1
20050165925 Dan et al. Jul 2005 A1
20050169179 Antal Aug 2005 A1
20050172291 Das et al. Aug 2005 A1
20050177600 Eilam et al. Aug 2005 A1
20050187866 Lee Aug 2005 A1
20050188088 Fellenstein et al. Aug 2005 A1
20050188089 Lichtenstein et al. Aug 2005 A1
20050188091 Szabo et al. Aug 2005 A1
20050190236 Ishimoto Sep 2005 A1
20050192771 Fischer et al. Sep 2005 A1
20050193103 Drabik Sep 2005 A1
20050193225 Macbeth Sep 2005 A1
20050193231 Scheuren Sep 2005 A1
20050195075 McGraw Sep 2005 A1
20050197877 Kalinoski Sep 2005 A1
20050198200 Subramanian et al. Sep 2005 A1
20050198516 Marr Sep 2005 A1
20050202922 Thomas Sep 2005 A1
20050203761 Barr Sep 2005 A1
20050204040 Ferri et al. Sep 2005 A1
20050206917 Ferlitsch Sep 2005 A1
20050209892 Miller Sep 2005 A1
20050210470 Chung et al. Sep 2005 A1
20050213507 Banerjee et al. Sep 2005 A1
20050213560 Duvvury Sep 2005 A1
20050222885 Chen et al. Oct 2005 A1
20050228852 Santos et al. Oct 2005 A1
20050228856 Swildens Oct 2005 A1
20050228892 Riley et al. Oct 2005 A1
20050234846 Davidson et al. Oct 2005 A1
20050235137 Barr et al. Oct 2005 A1
20050235150 Kaler et al. Oct 2005 A1
20050240688 Moerman et al. Oct 2005 A1
20050243867 Petite Nov 2005 A1
20050246705 Etelson et al. Nov 2005 A1
20050249341 Mahone et al. Nov 2005 A1
20050256942 McCardle et al. Nov 2005 A1
20050256946 Childress et al. Nov 2005 A1
20050259397 Bash et al. Nov 2005 A1
20050259683 Bishop et al. Nov 2005 A1
20050262495 Fung et al. Nov 2005 A1
20050262508 Asano et al. Nov 2005 A1
20050267948 Mckinley et al. Dec 2005 A1
20050268063 Diao et al. Dec 2005 A1
20050278392 Hansen et al. Dec 2005 A1
20050278760 Dewar et al. Dec 2005 A1
20050283534 Bigagli et al. Dec 2005 A1
20050283782 Lu et al. Dec 2005 A1
20050283822 Appleby et al. Dec 2005 A1
20050288961 Tabrizi Dec 2005 A1
20050289540 Nguyen et al. Dec 2005 A1
20060002311 Iwanaga et al. Jan 2006 A1
20060008256 Khedouri et al. Jan 2006 A1
20060010445 Petersen et al. Jan 2006 A1
20060013132 Garnett et al. Jan 2006 A1
20060013218 Shore et al. Jan 2006 A1
20060015555 Douglass et al. Jan 2006 A1
20060015637 Chung Jan 2006 A1
20060015651 Freimuth Jan 2006 A1
20060015773 Singh et al. Jan 2006 A1
20060023245 Sato et al. Feb 2006 A1
20060028991 Tan et al. Feb 2006 A1
20060029053 Roberts et al. Feb 2006 A1
20060031379 Kasriel et al. Feb 2006 A1
20060031547 Tsui et al. Feb 2006 A1
20060031813 Bishop et al. Feb 2006 A1
20060036743 Deng et al. Feb 2006 A1
20060037016 Saha et al. Feb 2006 A1
20060039246 King et al. Feb 2006 A1
20060041444 Flores et al. Feb 2006 A1
20060047920 Moore et al. Mar 2006 A1
20060048157 Dawson et al. Mar 2006 A1
20060053215 Sharma Mar 2006 A1
20060053216 Deokar et al. Mar 2006 A1
20060056291 Baker et al. Mar 2006 A1
20060056373 Legg Mar 2006 A1
20060059253 Goodman et al. Mar 2006 A1
20060063690 Billiauw et al. Mar 2006 A1
20060069261 Bonneau Mar 2006 A1
20060069671 Conley et al. Mar 2006 A1
20060069774 Chen et al. Mar 2006 A1
20060069926 Ginter et al. Mar 2006 A1
20060074925 Bixby Apr 2006 A1
20060074940 Craft et al. Apr 2006 A1
20060088015 Kakivaya et al. Apr 2006 A1
20060089894 Balk et al. Apr 2006 A1
20060090003 Kakivaya et al. Apr 2006 A1
20060090025 Tufford et al. Apr 2006 A1
20060090136 Miller et al. Apr 2006 A1
20060092942 Newson May 2006 A1
20060095917 Black-Ziegelbein et al. May 2006 A1
20060097863 Horowitz et al. May 2006 A1
20060112184 Kuo May 2006 A1
20060112308 Crawford May 2006 A1
20060117064 Wilson Jun 2006 A1
20060117208 Davidson Jun 2006 A1
20060117317 Crawford et al. Jun 2006 A1
20060120322 Lindskog Jun 2006 A1
20060120411 Basu Jun 2006 A1
20060126619 Teisberg et al. Jun 2006 A1
20060126667 Smith et al. Jun 2006 A1
20060129667 Anderson Jun 2006 A1
20060129687 Goldszmidt et al. Jun 2006 A1
20060136235 Keohane et al. Jun 2006 A1
20060136570 Pandya Jun 2006 A1
20060136908 Gebhart et al. Jun 2006 A1
20060136928 Crawford et al. Jun 2006 A1
20060136929 Miller et al. Jun 2006 A1
20060140211 Huang et al. Jun 2006 A1
20060143350 Miloushev et al. Jun 2006 A1
20060149695 Bossman et al. Jul 2006 A1
20060153191 Rajsic et al. Jul 2006 A1
20060155740 Chen et al. Jul 2006 A1
20060155912 Singh et al. Jul 2006 A1
20060156273 Narayan et al. Jul 2006 A1
20060159088 Aghvami et al. Jul 2006 A1
20060161466 Trinon et al. Jul 2006 A1
20060161585 Clarke et al. Jul 2006 A1
20060165040 Rathod Jul 2006 A1
20060168107 Balan et al. Jul 2006 A1
20060168224 Midgley Jul 2006 A1
20060173730 Birkestrand Aug 2006 A1
20060174342 Zaheer et al. Aug 2006 A1
20060179241 Clark et al. Aug 2006 A1
20060182119 Li Aug 2006 A1
20060184939 Sahoo Aug 2006 A1
20060189349 Montulli et al. Aug 2006 A1
20060190775 Aggarwal et al. Aug 2006 A1
20060190975 Gonzalez Aug 2006 A1
20060200773 Nocera et al. Sep 2006 A1
20060206621 Toebes Sep 2006 A1
20060208870 Dousson Sep 2006 A1
20060212332 Jackson Sep 2006 A1
20060212333 Jackson Sep 2006 A1
20060212334 Jackson Sep 2006 A1
20060212740 Jackson Sep 2006 A1
20060218301 O'Toole et al. Sep 2006 A1
20060224725 Bali et al. Oct 2006 A1
20060224740 Sievers-Tostes Oct 2006 A1
20060224741 Jackson Oct 2006 A1
20060227810 Childress et al. Oct 2006 A1
20060229920 Favorel et al. Oct 2006 A1
20060230140 Aoyama et al. Oct 2006 A1
20060230149 Jackson Oct 2006 A1
20060236368 Raja et al. Oct 2006 A1
20060236371 Fish Oct 2006 A1
20060248141 Mukherjee Nov 2006 A1
20060248197 Evans et al. Nov 2006 A1
20060248359 Fung Nov 2006 A1
20060250971 Gammenthaler et al. Nov 2006 A1
20060251419 Zadikian et al. Nov 2006 A1
20060253570 Biswas et al. Nov 2006 A1
20060259734 Sheu et al. Nov 2006 A1
20060265508 Angel et al. Nov 2006 A1
20060265609 Fung Nov 2006 A1
20060268742 Chu Nov 2006 A1
20060271552 McChesney et al. Nov 2006 A1
20060271928 Gao et al. Nov 2006 A1
20060277278 Hegde Dec 2006 A1
20060282505 Hasha et al. Dec 2006 A1
20060282547 Hasha et al. Dec 2006 A1
20060294219 Ogawa Dec 2006 A1
20060294238 Naik et al. Dec 2006 A1
20070003051 Kiss et al. Jan 2007 A1
20070006001 Isobe et al. Jan 2007 A1
20070011224 Mena et al. Jan 2007 A1
20070011302 Groner et al. Jan 2007 A1
20070022425 Jackson Jan 2007 A1
20070028244 Landis et al. Feb 2007 A1
20070033292 Sull et al. Feb 2007 A1
20070033533 Sull et al. Feb 2007 A1
20070041335 Znamova et al. Feb 2007 A1
20070043591 Meretei Feb 2007 A1
20070044010 Sull et al. Feb 2007 A1
20070047195 Merkin et al. Mar 2007 A1
20070050777 Hutchinson et al. Mar 2007 A1
20070061441 Landis et al. Mar 2007 A1
20070067366 Landis Mar 2007 A1
20070067435 Landis et al. Mar 2007 A1
20070067766 Tal Mar 2007 A1
20070076653 Park et al. Apr 2007 A1
20070081315 Mondor et al. Apr 2007 A1
20070083899 Compton et al. Apr 2007 A1
20070088822 Coile et al. Apr 2007 A1
20070094002 Berstis Apr 2007 A1
20070094486 Moore et al. Apr 2007 A1
20070094665 Jackson Apr 2007 A1
20070094691 Gazdzinski Apr 2007 A1
20070109968 Hussain et al. May 2007 A1
20070118496 Bornhoevd May 2007 A1
20070124344 Rajakannimariyan et al. May 2007 A1
20070130397 Tsu Jun 2007 A1
20070143824 Shahbazi Jun 2007 A1
20070150426 Asher et al. Jun 2007 A1
20070150444 Chesnais et al. Jun 2007 A1
20070155406 Dowling et al. Jul 2007 A1
20070174390 Silvain et al. Jul 2007 A1
20070180310 Johnson et al. Aug 2007 A1
20070180380 Khavari et al. Aug 2007 A1
20070204036 Mohaban et al. Aug 2007 A1
20070209072 Chen Sep 2007 A1
20070220520 Tajima Sep 2007 A1
20070226313 Li et al. Sep 2007 A1
20070226795 Conti et al. Sep 2007 A1
20070233828 Gilbert et al. Oct 2007 A1
20070237115 Bae Oct 2007 A1
20070240162 Coleman et al. Oct 2007 A1
20070253017 Czyszczewski et al. Nov 2007 A1
20070260716 Gnanasambandam et al. Nov 2007 A1
20070264986 Warrillow et al. Nov 2007 A1
20070266136 Esfahany et al. Nov 2007 A1
20070268909 Chen Nov 2007 A1
20070271375 Hwang Nov 2007 A1
20070280230 Park Dec 2007 A1
20070286009 Norman Dec 2007 A1
20070288585 Sekiguchi et al. Dec 2007 A1
20070297350 Eilam et al. Dec 2007 A1
20070299946 El-Damhougy et al. Dec 2007 A1
20070299947 El-Damhougy et al. Dec 2007 A1
20070299950 Kulkarni et al. Dec 2007 A1
20080013453 Chiang et al. Jan 2008 A1
20080016198 Johnston-Watt et al. Jan 2008 A1
20080034082 McKinney Feb 2008 A1
20080040463 Brown et al. Feb 2008 A1
20080052437 Loffink et al. Feb 2008 A1
20080059782 Kruse et al. Mar 2008 A1
20080065835 Lacobovici Mar 2008 A1
20080075089 Evans et al. Mar 2008 A1
20080082663 Mouli et al. Apr 2008 A1
20080089358 Basso et al. Apr 2008 A1
20080104231 Dey et al. May 2008 A1
20080104264 Duerk et al. May 2008 A1
20080126523 Tantrum May 2008 A1
20080140771 Vass et al. Jun 2008 A1
20080140930 Hotchkiss Jun 2008 A1
20080155070 El-Damhougy et al. Jun 2008 A1
20080155100 Ahmed et al. Jun 2008 A1
20080159745 Segal Jul 2008 A1
20080162691 Zhang et al. Jul 2008 A1
20080168451 Challenger et al. Jul 2008 A1
20080183865 Appleby et al. Jul 2008 A1
20080183882 Flynn et al. Jul 2008 A1
20080184248 Barua et al. Jul 2008 A1
20080186965 Zheng et al. Aug 2008 A1
20080196043 Feinleib Aug 2008 A1
20080199133 Takizawa et al. Aug 2008 A1
20080212273 Bechtolsheim Sep 2008 A1
20080212276 Bottom et al. Sep 2008 A1
20080215730 Sundaram et al. Sep 2008 A1
20080216082 Eilam et al. Sep 2008 A1
20080217021 Lembcke et al. Sep 2008 A1
20080222434 Shimizu et al. Sep 2008 A1
20080232378 Moorthy Sep 2008 A1
20080235443 Chow et al. Sep 2008 A1
20080235702 Eilam et al. Sep 2008 A1
20080239649 Bradicich Oct 2008 A1
20080243634 Dworkin et al. Oct 2008 A1
20080250181 Li et al. Oct 2008 A1
20080255953 Chang et al. Oct 2008 A1
20080259555 Bechtolsheim et al. Oct 2008 A1
20080259788 Wang et al. Oct 2008 A1
20080263131 Hinni et al. Oct 2008 A1
20080263558 Lin et al. Oct 2008 A1
20080266793 Lee Oct 2008 A1
20080270599 Tamir et al. Oct 2008 A1
20080270731 Bryant et al. Oct 2008 A1
20080279167 Cardei et al. Nov 2008 A1
20080288646 Hasha et al. Nov 2008 A1
20080288659 Hasha et al. Nov 2008 A1
20080288660 Balasubramanian et al. Nov 2008 A1
20080288664 Pettey et al. Nov 2008 A1
20080288683 Ramey Nov 2008 A1
20080288873 McCardle et al. Nov 2008 A1
20080289029 Kim et al. Nov 2008 A1
20080301226 Cleary et al. Dec 2008 A1
20080301379 Pong Dec 2008 A1
20080301794 Lee Dec 2008 A1
20080304481 Gumey Dec 2008 A1
20080310848 Yasuda et al. Dec 2008 A1
20080313293 Jacobs Dec 2008 A1
20080313369 Verdoorn et al. Dec 2008 A1
20080313482 Karlapalem et al. Dec 2008 A1
20080320121 Altaf et al. Dec 2008 A1
20080320161 Maruccia et al. Dec 2008 A1
20080320482 Dawson Dec 2008 A1
20090010153 Filsfils et al. Jan 2009 A1
20090021907 Mann et al. Jan 2009 A1
20090043809 Fakhouri et al. Feb 2009 A1
20090043888 Jackson Feb 2009 A1
20090044036 Merkin Feb 2009 A1
20090049443 Powers et al. Feb 2009 A1
20090055542 Zhoa et al. Feb 2009 A1
20090055691 Ouksel et al. Feb 2009 A1
20090063443 Arimilli et al. Mar 2009 A1
20090063690 Verthein et al. Mar 2009 A1
20090064287 Bagepalli et al. Mar 2009 A1
20090070771 Yuyitung et al. Mar 2009 A1
20090080428 Witkowski et al. Mar 2009 A1
20090083390 Abu-Ghazaleh et al. Mar 2009 A1
20090089410 Vicente et al. Apr 2009 A1
20090094380 Qiu et al. Apr 2009 A1
20090097200 Sharma et al. Apr 2009 A1
20090100133 Giulio et al. Apr 2009 A1
20090103501 Farrag et al. Apr 2009 A1
20090105059 Dorry et al. Apr 2009 A1
20090113056 Tameshige et al. Apr 2009 A1
20090113130 He et al. Apr 2009 A1
20090133129 Jeong et al. May 2009 A1
20090135751 Hodges et al. May 2009 A1
20090135835 Gallatin et al. May 2009 A1
20090138594 Fellenstein et al. May 2009 A1
20090150566 Malkhi Jun 2009 A1
20090158070 Gruendler Jun 2009 A1
20090172423 Song et al. Jul 2009 A1
20090178132 Hudis et al. Jul 2009 A1
20090182836 Aviles Jul 2009 A1
20090187425 Thompson et al. Jul 2009 A1
20090198958 Arimilli et al. Aug 2009 A1
20090204834 Hendin et al. Aug 2009 A1
20090204837 Raval et al. Aug 2009 A1
20090210356 Abrams et al. Aug 2009 A1
20090210495 Wolfson et al. Aug 2009 A1
20090216881 Lovy et al. Aug 2009 A1
20090216910 Duchesneau Aug 2009 A1
20090216920 Lauterbach et al. Aug 2009 A1
20090217329 Riedl et al. Aug 2009 A1
20090219827 Chen et al. Sep 2009 A1
20090222884 Shaji et al. Sep 2009 A1
20090225360 Shirai Sep 2009 A1
20090225751 Koenck et al. Sep 2009 A1
20090234917 Despotovic et al. Sep 2009 A1
20090234962 Strong et al. Sep 2009 A1
20090234974 Arndt et al. Sep 2009 A1
20090235104 Fung Sep 2009 A1
20090238349 Pezzutti Sep 2009 A1
20090240547 Fellenstein et al. Sep 2009 A1
20090248943 Jiang et al. Oct 2009 A1
20090251867 Sharma Oct 2009 A1
20090257440 Yan Oct 2009 A1
20090259606 Seah et al. Oct 2009 A1
20090259863 Williams et al. Oct 2009 A1
20090259864 Li et al. Oct 2009 A1
20090265045 Coxe, III Oct 2009 A1
20090271656 Yokota et al. Oct 2009 A1
20090276666 Haley et al. Nov 2009 A1
20090279518 Falk et al. Nov 2009 A1
20090282274 Langgood et al. Nov 2009 A1
20090282419 Mejdrich et al. Nov 2009 A1
20090285136 Sun et al. Nov 2009 A1
20090287835 Jacobson et al. Nov 2009 A1
20090292824 Marashi et al. Nov 2009 A1
20090300608 Ferris et al. Dec 2009 A1
20090313390 Ahuja et al. Dec 2009 A1
20090316687 Kruppa et al. Dec 2009 A1
20090319684 Kakivaya et al. Dec 2009 A1
20090323691 Johnson Dec 2009 A1
20090327079 Parker et al. Dec 2009 A1
20090327489 Swildens et al. Dec 2009 A1
20100005331 Somasundaram et al. Jan 2010 A1
20100008038 Coglitore Jan 2010 A1
20100008365 Porat Jan 2010 A1
20100026408 Shau Feb 2010 A1
20100036945 Allibhoy et al. Feb 2010 A1
20100040053 Gottumukkula et al. Feb 2010 A1
20100049822 Davies et al. Feb 2010 A1
20100049931 Jacobson et al. Feb 2010 A1
20100051391 Jahkonen Mar 2010 A1
20100070675 Pong Mar 2010 A1
20100082788 Mundy Apr 2010 A1
20100088205 Robertson Apr 2010 A1
20100088490 Chakradhar Apr 2010 A1
20100091676 Moran et al. Apr 2010 A1
20100103837 Jungck et al. Apr 2010 A1
20100106987 Lambert et al. Apr 2010 A1
20100114531 Korn et al. May 2010 A1
20100118880 Kunz et al. May 2010 A1
20100121932 Joshi et al. May 2010 A1
20100121947 Pirzada et al. May 2010 A1
20100122251 Karc May 2010 A1
20100125742 Ohtani May 2010 A1
20100125915 Hall et al. May 2010 A1
20100131324 Ferris et al. May 2010 A1
20100131624 Ferris May 2010 A1
20100138481 Behrens Jun 2010 A1
20100153546 Clubb et al. Jun 2010 A1
20100158005 Mukhopadhyay et al. Jun 2010 A1
20100161909 Nation et al. Jun 2010 A1
20100165983 Aybay et al. Jul 2010 A1
20100169477 Stienhans et al. Jul 2010 A1
20100169479 Jeong et al. Jul 2010 A1
20100169888 Hare et al. Jul 2010 A1
20100174604 Mattingly et al. Jul 2010 A1
20100174813 Hildreth et al. Jul 2010 A1
20100198972 Umbehocker Aug 2010 A1
20100198985 Kanevsky Aug 2010 A1
20100217801 Leighton et al. Aug 2010 A1
20100218194 Dallman et al. Aug 2010 A1
20100220732 Hussain et al. Sep 2010 A1
20100223332 Maxemchuk et al. Sep 2010 A1
20100228848 Kis et al. Sep 2010 A1
20100235234 Shuster Sep 2010 A1
20100250914 Abdul et al. Sep 2010 A1
20100262650 Chauhan Oct 2010 A1
20100265650 Chen et al. Oct 2010 A1
20100281166 Buyya et al. Nov 2010 A1
20100281246 Bristow et al. Nov 2010 A1
20100299548 Chadirchi et al. Nov 2010 A1
20100302129 Kastrup et al. Dec 2010 A1
20100308897 Evoy et al. Dec 2010 A1
20100312910 Lin et al. Dec 2010 A1
20100312969 Yamazaki et al. Dec 2010 A1
20100318665 Demmer et al. Dec 2010 A1
20100318812 Auradkar et al. Dec 2010 A1
20100325371 Jagadish et al. Dec 2010 A1
20100332262 Horvitz et al. Dec 2010 A1
20100333116 Prahlad Dec 2010 A1
20110023104 Franklin Jan 2011 A1
20110026397 Saltsidis et al. Feb 2011 A1
20110029644 Gelvin et al. Feb 2011 A1
20110029652 Chhuor et al. Feb 2011 A1
20110035491 Gelvin et al. Feb 2011 A1
20110055627 Zawacki et al. Mar 2011 A1
20110058573 Balakavi et al. Mar 2011 A1
20110075369 Sun et al. Mar 2011 A1
20110082928 Hasha et al. Apr 2011 A1
20110090633 Rabinovitz Apr 2011 A1
20110103391 Davis May 2011 A1
20110113083 Shahar May 2011 A1
20110113115 Chang et al. May 2011 A1
20110119344 Eustis May 2011 A1
20110123014 Smith May 2011 A1
20110138046 Bonnier et al. Jun 2011 A1
20110145393 Ben-Zvi et al. Jun 2011 A1
20110153953 Khemani et al. Jun 2011 A1
20110154318 Oshins et al. Jun 2011 A1
20110154371 Beale Jun 2011 A1
20110167110 Hoffberg et al. Jul 2011 A1
20110173295 Bakke et al. Jul 2011 A1
20110173612 El Zur et al. Jul 2011 A1
20110179134 Mayo et al. Jul 2011 A1
20110185370 Tamir et al. Jul 2011 A1
20110188378 Collins Aug 2011 A1
20110191514 Wu et al. Aug 2011 A1
20110191610 Agarwal et al. Aug 2011 A1
20110197012 Liao et al. Aug 2011 A1
20110210975 Wong et al. Sep 2011 A1
20110213869 Korsunsky et al. Sep 2011 A1
20110231510 Korsunsky et al. Sep 2011 A1
20110231564 Korsunsky et al. Sep 2011 A1
20110238841 Kakivaya et al. Sep 2011 A1
20110238855 Korsunsky et al. Sep 2011 A1
20110239014 Karnowski Sep 2011 A1
20110271159 Ahn et al. Nov 2011 A1
20110273840 Chen Nov 2011 A1
20110274108 Fan Nov 2011 A1
20110295991 Aida Dec 2011 A1
20110296141 Daffron Dec 2011 A1
20110307887 Huang et al. Dec 2011 A1
20110314465 Smith et al. Dec 2011 A1
20110320540 Oostlander et al. Dec 2011 A1
20110320690 Petersen et al. Dec 2011 A1
20120011500 Faraboschi et al. Jan 2012 A1
20120020207 Corti et al. Jan 2012 A1
20120036237 Hasha et al. Feb 2012 A1
20120042196 Aron Feb 2012 A1
20120050981 Xu et al. Mar 2012 A1
20120054469 Ikeya et al. Mar 2012 A1
20120054511 Brinks et al. Mar 2012 A1
20120072997 Carlson et al. Mar 2012 A1
20120081850 Regimbal et al. Apr 2012 A1
20120096211 Davis et al. Apr 2012 A1
20120099265 Reber Apr 2012 A1
20120102457 Tal Apr 2012 A1
20120110055 Van Biljon et al. May 2012 A1
20120110180 Van Biljon et al. May 2012 A1
20120110188 Van Biljon et al. May 2012 A1
20120110651 Van Biljon et al. May 2012 A1
20120117229 Van Biljon et al. May 2012 A1
20120131201 Matthews et al. May 2012 A1
20120137004 Smith May 2012 A1
20120151476 Vincent Jun 2012 A1
20120155168 Kim et al. Jun 2012 A1
20120158925 Shen Jun 2012 A1
20120159116 Lim et al. Jun 2012 A1
20120167083 Suit Jun 2012 A1
20120167084 Suit Jun 2012 A1
20120167094 Suit Jun 2012 A1
20120185334 Sarkar et al. Jul 2012 A1
20120191860 Traversal et al. Jul 2012 A1
20120198075 Crowe Aug 2012 A1
20120198252 Kirschtein et al. Aug 2012 A1
20120207165 Davis Aug 2012 A1
20120209989 Stewart Aug 2012 A1
20120218901 Jungck et al. Aug 2012 A1
20120222033 Byrum Aug 2012 A1
20120226788 Jackson Sep 2012 A1
20120239479 Amaro et al. Sep 2012 A1
20120278378 Lehane et al. Nov 2012 A1
20120278430 Lehane et al. Nov 2012 A1
20120278464 Lehane et al. Nov 2012 A1
20120296974 Tabe et al. Nov 2012 A1
20120297042 Davis et al. Nov 2012 A1
20120324005 Nalawade Dec 2012 A1
20130010639 Armstrong et al. Jan 2013 A1
20130024645 Cheriton et al. Jan 2013 A1
20130031331 Cheriton et al. Jan 2013 A1
20130036236 Morales et al. Feb 2013 A1
20130058250 Casado et al. Mar 2013 A1
20130060839 Van Biljon et al. Mar 2013 A1
20130066940 Shao Mar 2013 A1
20130073602 Meadway et al. Mar 2013 A1
20130073724 Parashar et al. Mar 2013 A1
20130086298 Alanis Apr 2013 A1
20130094499 Davis et al. Apr 2013 A1
20130097351 Davis Apr 2013 A1
20130097448 Davis et al. Apr 2013 A1
20130107444 Schnell May 2013 A1
20130111107 Chang et al. May 2013 A1
20130124417 Spears et al. May 2013 A1
20130145375 Kang Jun 2013 A1
20130148667 Hama et al. Jun 2013 A1
20130163605 Chandra et al. Jun 2013 A1
20130191612 Li Jul 2013 A1
20130247064 Jackson Sep 2013 A1
20130268653 Deng et al. Oct 2013 A1
20130275703 Schenfeld et al. Oct 2013 A1
20130286840 Fan Oct 2013 A1
20130290643 Lim Oct 2013 A1
20130290650 Chang et al. Oct 2013 A1
20130298134 Jackson Nov 2013 A1
20130305093 Jayachandran et al. Nov 2013 A1
20130312006 Hardman Nov 2013 A1
20130318255 Karino Nov 2013 A1
20130318269 Dalal et al. Nov 2013 A1
20140052866 Jackson Feb 2014 A1
20140082614 Klein et al. Mar 2014 A1
20140104778 Schnell Apr 2014 A1
20140122833 Davis et al. May 2014 A1
20140135105 Quan et al. May 2014 A1
20140143773 Ciano et al. May 2014 A1
20140143781 Yao May 2014 A1
20140189039 Dalton Jul 2014 A1
20140201761 Dalal et al. Jul 2014 A1
20140317292 Odom Oct 2014 A1
20140348182 Chandra Nov 2014 A1
20140359044 Davis et al. Dec 2014 A1
20140359323 Fullerton et al. Dec 2014 A1
20140365596 Kanevsky Dec 2014 A1
20140379836 Zilberboim Dec 2014 A1
20150012679 Davis et al. Jan 2015 A1
20150039840 Chandra et al. Feb 2015 A1
20150103826 Davis Apr 2015 A1
20150229586 Jackson Aug 2015 A1
20150263913 De Temmerman Sep 2015 A1
20150293789 Jackson Oct 2015 A1
20150301880 Allu Oct 2015 A1
20150381521 Jackson Dec 2015 A1
20160154539 Buddhiraja Jun 2016 A1
20160161909 Wada Jun 2016 A1
20160306586 Dornemann Oct 2016 A1
20160378570 Ljubuncic Dec 2016 A1
20170111274 Bays Apr 2017 A1
20170115712 Davis Apr 2017 A1
20170127577 Rodriguez et al. May 2017 A1
20180018149 Cook Jan 2018 A1
20180054364 Jackson Feb 2018 A1
20190260689 Jackson Aug 2019 A1
20190286610 Dalton Sep 2019 A1
20200073722 Jackson Mar 2020 A1
20200159449 Davis et al. May 2020 A1
20200379819 Jackson Dec 2020 A1
20200382585 Abu-Ghazaleh et al. Dec 2020 A1
20210117130 Davis Apr 2021 A1
20210141671 Jackson May 2021 A1
20210250249 Jackson Aug 2021 A1
20210306284 Jackson Sep 2021 A1
20210311804 Jackson Oct 2021 A1
20220121545 Dalton Apr 2022 A1
20220206859 Jackson Jun 2022 A1
20220206861 Jackson Jun 2022 A1
20220214920 Jackson Jul 2022 A1
20220214921 Jackson Jul 2022 A1
20220214922 Jackson Jul 2022 A1
20220222119 Jackson Jul 2022 A1
20220222120 Jackson Jul 2022 A1
20220239606 Jackson Jul 2022 A1
20220239607 Jackson Jul 2022 A1
20220247694 Jackson Aug 2022 A1
20220300334 Jackson Sep 2022 A1
20220317692 Guim Bernat Oct 2022 A1
Foreign Referenced Citations (52)
Number Date Country
2496783 Mar 2004 CA
60216001 Jul 2007 DE
112008001875 Aug 2013 DE
0268435 May 1988 EP
0605106 Jul 1994 EP
0859314 Aug 1998 EP
1331564 Jul 2003 EP
1365545 Nov 2003 EP
1492309 Dec 2004 EP
1865684 Dec 2007 EP
2391744 Feb 2004 GB
2392265 Feb 2004 GB
8-212084 Aug 1996 JP
2002-207712 Jul 2002 JP
2005-165568 Jun 2005 JP
2005-223753 Aug 2005 JP
2005-536960 Dec 2005 JP
2006-309439 Nov 2006 JP
20040107934 Dec 2004 KR
M377621 Apr 2010 TW
201017430 May 2010 TW
WO1998011702 Mar 1998 WO
WO1998058518 Dec 1998 WO
WO1999015999 Apr 1999 WO
WO1999057660 Nov 1999 WO
WO2000014938 Mar 2000 WO
WO2000025485 May 2000 WO
WO2000060825 Oct 2000 WO
WO2001009791 Feb 2001 WO
WO2001014987 Mar 2001 WO
WO2001015397 Mar 2001 WO
WO2001039470 May 2001 WO
WO2001044271 Jun 2001 WO
WO2003046751 Jun 2003 WO
WO2003060798 Sep 2003 WO
WO2004021109 Mar 2004 WO
WO2004021641 Mar 2004 WO
WO2004046919 Jun 2004 WO
WO2004070547 Aug 2004 WO
WO2004092884 Oct 2004 WO
WO2005013143 Feb 2005 WO
WO2005017763 Feb 2005 WO
WO2005017783 Feb 2005 WO
WO2005089245 Sep 2005 WO
WO2005091136 Sep 2005 WO
WO2006036277 Apr 2006 WO
WO2006107531 Oct 2006 WO
WO2006108187 Oct 2006 WO
WO2006112981 Oct 2006 WO
WO2008000193 Jan 2008 WO
WO2011044271 Apr 2011 WO
WO2012037494 Mar 2012 WO
Non-Patent Literature Citations (600)
Entry
US 7,774,482 B1, 08/2010, Szeto et al. (withdrawn)
U.S. Appl. No. 11/279,007, filed Apr. 2006, Jackson.
U.S. Appl. No. 13/705,340, filed Apr. 2012, Davis et al.
U.S. Appl. No. 13/899,751, filed May 2013, Chandra.
U.S. Appl. No. 13/935,108, filed Jul. 2013, Davis.
U.S. Appl. No. 13/959,428, filed Aug. 2013, Chandra.
U.S. Appl. No. 60/662,240, filed Mar. 2005, Jackson.
U.S. Appl. No. 60/552,653, filed Apr. 2005, Jackson.
“Microsoft Computer Dictionary, 5th Ed.”; Microsoft Press; 3 pages; 2002.
“Random House Concise Dictionary of Science & Computers”; 3 pages; Helicon Publishing; 2004.
A Language Modeling Framework for Resource Selection and Results Merging Si et al. CIKM 2002, Proceedings of the eleventh international conference on Iformation and Knowledge Management.
Alhusaini et al. “A framework for mapping with resource co-allocation in heterogeneous computing systems,” Proceedings 9th Heterogeneous Computing Workshop (HCW 2000) (Cat. No. PR00556), Cancun, Mexico, 2000, pp. 273-286. (Year: 2000).
Ali et al., “Task Execution Time Modeling for Heterogeneous Computing System”, IEEE, 2000, pp. 1-15.
Amiri et al., “Dynamic Function Placement for Data-Intensive Cluster Computing,” Jun. 2000.
Bader et al.; “Applications”; The International Journal of High Performance Computing Applications, vol. 15, No. ; pp. 181-185; Summer 2001.
Banicescu et al., “Competitive Resource management in Distributed Computing Environments with Hectiling”, 1999, High Performance Computing Symposium, p. 1-7 (Year: 1999).
Banicescu et al., “Efficient Resource Management for Scientific Applications in Distributed Computing Environment” 1998, Mississippi State Univ. Dept. of Comp. Science, p. 45-54. (Year: 1998).
Buyya et al., “An Evaluation of Economy-based Resource Trading and Scheduling on Computational Power Grids for Parameter Sweep Applications,” Active Middleware Services, 2000, 10 pages.
Caesar et al., “Design and Implementation of a Routing Control Platform,” Usenix, NSDI '05 Paper, Technical Program , obtained from the Internet, on Apr. 13, 2021, at URL <https://www.usenix.org/legacy/event/nsdi05/tech/full_papers/caesar/ca-esar_html/>, 23 pages.
Chase et al., “Dynamic Virtual Clusters in a Grid Site Manager”, Proceedings of the 12.sup.th IEEE International Symposium on High Performance Distributed Computing (HPDC'03), 2003.
Chen et al., “A flexible service model for advance reservation”, Computer Networks, Elsevier science publishers, vol. 37, No. 3-4, pp. 251-262. Nov. 5, 2001.
Coomer et al.; “Introduction to the Cluster Grid—Part 1”; Sun Microsystems White Paper; 19 pages; Aug. 2002.
Exhibit 1002, Declaration of Dr. Andrew Wolfe, Ph.D., document filed on behalf of Unified Patents, LLC, in Case No. IPR2022-00136, 110 pages, Declaration dated Nov. 29, 2021.
Exhibit 1008, Declaration of Kevin Jakel, document filed on behalf of Unified Patents, LLC, in Case No. IPR2022-00136, 7 pages, Declaration dated Nov. 4, 2021.
Foster et al., “A Distributed Resource Management Architecture that Supports Advance Reservations and Co-Allocation,” Seventh International Workshop on Quality of Service (IWQoS '99), 1999, pp. 27-36.
Furmento et al. “An Integrated Grid Environment for Component Applications”, Proceedings of the Second International Workshop on Grid Computing table of contents, 2001, pp. 26-37.
He XiaoShan; QoS Guided Min-Min Heuristic for Grid Task Scheduling; Jul. 2003, vol. 18, No. 4, pp. 442-451 J. Comput. Sci. & Technol.
Huy Tuong Le, “The Data-AWare Resource Broker” Research Project Thesis, University of Adelaide, Nov. 2003, pp. 1-63.
IBM Tivoli “IBM Directory Integrator and Tivoli Identity Manager Integration” Apr. 2, 2003, pp. 1-13 online link “http:publib.boulder.ibm.com/tividd/td/ITIM/SC32-1683-00/en_US/HTML/idi_integration/index.html” (Year: 2003).
Intel, Architecture Guide: Intel® Active Management Technology, Intel.com, Oct. 10, 2008, pp. 1-23, (Year 2008).
Joseph et al.; “Evolution of grid computing architecture and grid adoption models”; IBM Systems Journal, vol. 43, No. 4; 22 pages; 2004.
Kafil et al., “Optimal Task Assignment in Herterogenous Computing Systems,” IEEE, 1997, pp. 135-146.
Kuan-Wei Cheng, Chao-Tung Yang, Chuan-Lin Lai and Shun-Chyi Change, “A parallel loop self-scheduling on grid computing environments,” 7th International Symposium on Parallel Architectures, Algorithms and Networks, 2004. Proceedings. 2004, pp. 409-414 (Year: 2004).
Luo Si et al. “A Language Modeling Framework for Resource Selection and Results Merging”, Conference on Information and Knowledge Management. 2002 ACM pp. 391-397.
Maheswaran et al., “Dynamic Matching and Scheduling of a Class of Independent Tasks onto Heterogeneous Computing Systems,” IEEE, 2000, pp. 1-15.
Mateescu et al., “Quality of service on the grid via metascheduling with resource co-scheduling and co-reservation,” The International Journal of High Performance Computing Applications, 2003, 10 pages.
Notice of Allowance on U.S. Appl. No. 11/207,438, dated Oct. 15, 2015.
Notice of Allowance on U.S. Appl. No. 11/207,438 dated Jan. 3, 2012.
Notice of Allowance on U.S. Appl. No. 11/276,852 dated Nov. 26, 2014.
Notice of Allowance on U.S. Appl. No. 11/276,853, dated Apr. 5, 2016.
Notice of Allowance on U.S. Appl. No. 11/276,854, dated Mar. 6, 2014.
Notice of Allowance on U.S. Appl. No. 11/276,855, dated Sep. 13, 2013.
Notice of Allowance on U.S. Appl. No. 11/616,156, dated Mar. 25, 2014.
Notice of Allowance on U.S. Appl. No. 11/718,867 dated May 25, 2012.
Notice of Allowance on U.S. Appl. No. 12/573,967, dated Jul. 20, 2015.
Notice of Allowance on U.S. Appl. No. 13/234,054, dated Sep. 19, 2017.
Notice of Allowance on U.S. Appl. No. 13/284,855, dated Jul. 14, 2014.
Notice of Allowance on U.S. Appl. No. 13/453,086, dated Jul. 18, 2013.
Notice of Allowance on U.S. Appl. No. 13/475,713, dated Feb. 5, 2015.
Notice of Allowance on U.S. Appl. No. 13/475,722, dated Feb. 27, 2015.
Notice of Allowance on U.S. Appl. No. 13/527,498, dated Feb. 23, 2015.
Notice of Allowance on U.S. Appl. No. 13/527,505, dated Mar. 6, 2015.
Notice of Allowance on U.S. Appl. No. 13/621,987 dated Jun. 4, 2015.
Notice of Allowance on U.S. Appl. No. 13/624,725, dated Mar. 30, 2016.
Notice of Allowance on U.S. Appl. No. 13/624,731, dated Mar. 5, 2015.
Notice of Allowance on U.S. Appl. No. 13/662,759 dated May 10, 2016.
Notice of Allowance on U.S. Appl. No. 13/692,741 dated Dec. 4, 2015.
Notice of Allowance on U.S. Appl. No. 13/705,286 dated Feb. 24, 2016.
Notice of Allowance on U.S. Appl. No. 13/705,340, dated Dec. 3, 2014.
Notice of Allowance on U.S. Appl. No. 13/705,340, dated Mar. 16, 2015.
Notice of Allowance on U.S. Appl. No. 13/705,386, dated Jan. 24, 2014.
Notice of Allowance on U.S. Appl. No. 13/705,414, dated Nov. 4, 2013.
Notice of Allowance on U.S. Appl. No. 13/728,308 dated Oct. 7, 2015.
Notice of Allowance on U.S. Appl. No. 13/728,428 dated Jul. 18, 2016.
Notice of Allowance on U.S. Appl. No. 13/758,164, dated Apr. 15, 2015.
Notice of Allowance on U.S. Appl. No. 13/760,600 dated Feb. 26, 2018.
Notice of Allowance on U.S. Appl. No. 13/760,600 dated Jan. 9, 2018.
Notice of Allowance on U.S. Appl. No. 13/855,241, dated Oct. 27, 2020.
Notice of Allowance on U.S. Appl. No. 13/855,241, dated Sep. 14, 2020.
Notice of Allowance on U.S. Appl. No. 14/052,723 dated Feb. 8, 2017.
Notice of Allowance on U.S. Appl. No. 14/106,254 dated May 25, 2017.
Notice of Allowance on U.S. Appl. No. 14/106,697 dated Oct. 24, 2016.
Notice of Allowance on U.S. Appl. No. 14/137,921 dated Aug. 12, 2021 and Jul. 16, 2021.
Notice of Allowance on U.S. Appl. No. 14/137,940 dated Jan. 30, 2019.
Notice of Allowance on U.S. Appl. No. 14/154,912 dated Apr. 25, 2019.
Notice of Allowance on U.S. Appl. No. 14/154,912, dated Apr. 3, 2019.
Notice of Allowance on U.S. Appl. No. 14/154,912, dated Feb. 7, 2019.
Notice of Allowance on U.S. Appl. No. 14/331,718 dated Jun. 7, 2017.
Notice of Allowance on U.S. Appl. No. 14/331,772, dated Jan. 10, 2018.
Notice of Allowance on U.S. Appl. No. 14/334,178 dated Aug. 19, 2016.
Notice of Allowance on U.S. Appl. No. 14/334,178 dated Jun. 8, 2016.
Notice of Allowance on U.S. Appl. No. 14/334,931 dated May 20, 2016.
Notice of Allowance on U.S. Appl. No. 14/454,049, dated Jan. 20, 2015.
Notice of Allowance on U.S. Appl. No. 14/590,102, dated Jan. 22, 2018.
Notice of Allowance on U.S. Appl. No. 14/704,231, dated Sep. 2, 2015.
Notice of Allowance on U.S. Appl. No. 14/709,642 dated Mar. 19, 2019.
Notice of Allowance on U.S. Appl. No. 14/709,642, dated May 9, 2019.
Notice of Allowance on U.S. Appl. No. 14/725,543 dated Jul. 21, 2016.
Notice of Allowance on U.S. Appl. No. 14/753,948 dated Jun. 14, 2017.
Notice of Allowance on U.S. Appl. No. 14/791,873 dated Dec. 20, 2018.
Notice of Allowance on U.S. Appl. No. 14/809,723 dated Jan. 11, 2018.
Notice of Allowance on U.S. Appl. No. 14/827,927 dated Jan. 21, 2022 and Dec. 9, 2021.
Notice of Allowance on U.S. Appl. No. 14/833,673, dated Dec. 2, 2016.
Notice of Allowance on U.S. Appl. No. 14/842,916 dated Oct. 2, 2017.
Notice of Allowance on U.S. Appl. No. 14/872,645 dated Oct. 13, 2016.
Notice of Allowance on U.S. Appl. No. 14/987,059, dated Feb. 14, 2020.
Notice of Allowance on U.S. Appl. No. 14/987,059, dated Jul. 8, 2019.
Notice of Allowance on U.S. Appl. No. 14/987,059, dated Nov. 7, 2019.
Notice of Allowance on U.S. Appl. No. 15/042,489 dated Jul. 16, 2018.
Notice of Allowance on U.S. Appl. No. 15/049,542 dated Feb. 28, 2018.
Notice of Allowance on U.S. Appl. No. 15/049,542 dated Jan. 4, 2018.
Notice of Allowance on U.S. Appl. No. 15/078,115 dated Jan. 8, 2018.
Notice of Allowance on U.S. Appl. No. 15/254,111 dated Nov. 13, 2017.
Notice of Allowance on U.S. Appl. No. 15/254,111 dated Sep. 1, 2017.
Notice of Allowance on U.S. Appl. No. 15/270,418 dated Nov. 2, 2017.
Notice of Allowance on U.S. Appl. No. 15/345,017 dated Feb. 2, 2021.
Notice of Allowance on U.S. Appl. No. 15/357,332 dated Jul. 12, 2018.
Notice of Allowance on U.S. Appl. No. 15/430,959, dated May 5, 2017.
Notice of Allowance on U.S. Appl. No. 15/430,959 dated Mar. 15, 2018.
Notice of Allowance on U.S. Appl. No. 15/478,467 dated May 30, 2019.
Notice of Allowance on U.S. Appl. No. 15/672,418 dated Apr. 4, 2018.
Notice of Allowance on U.S. Appl. No. 15/717,392 dated Mar. 22, 2019.
Notice of Allowance on U.S. Appl. No. 15/726,509, dated Sep. 25, 2019.
Office Action issued on U.S. Appl. No. 11/276,855, dated Jul. 22, 2010.
Office Action on U.S. Appl. No. 10/530,577, dated May 29, 2015.
Office Action on U.S. Appl. No. 11/207,438 dated Aug. 31, 2010.
Office Action on U.S. Appl. No. 11/207,438 dated Mar. 15, 2010.
Office Action on U.S. Appl. No. 11/276,852, dated Feb. 10, 2009.
Office Action on U.S. Appl. No. 11/276,852, dated Jan. 16, 2014.
Office Action on U.S. Appl. No. 11/276,852, dated Jun. 26, 2012.
Office Action on U.S. Appl. No. 11/276,852, dated Mar. 17, 2011.
Office Action on U.S. Appl. No. 11/276,852, dated Mar. 4, 2010.
Office Action on U.S. Appl. No. 11/276,852, dated Mar. 5, 2013.
Office Action on U.S. Appl. No. 11/276,852, dated Oct. 4, 2010.
Office Action on U.S. Appl. No. 11/276,852, dated Oct. 5, 2011.
Office Action on U.S. Appl. No. 11/276,852, dated Oct. 16, 2009.
Office Action on U.S. Appl. No. 11/276,853, dated Apr. 4, 2014.
Office Action on U.S. Appl. No. 11/276,853, dated Aug. 7, 2009.
Office Action on U.S. Appl. No. 11/276,853, dated Dec. 28, 2009.
Office Action on U.S. Appl. No. 11/276,853, dated Dec. 8, 2008.
Office Action on U.S. Appl. No. 11/276,853, dated Jul. 12, 2010.
Office Action on U.S. Appl. No. 11/276,853, dated May 26, 2011.
Office Action on U.S. Appl. No. 11/276,853, dated Nov. 23, 2010.
Office Action on U.S. Appl. No. 11/276,853, dated Oct. 16, 2009.
Office Action on U.S. Appl. No. 11/276,854, dated Apr. 18, 2011.
Office Action on U.S. Appl. No. 11/276,854, dated Aug. 1, 2012.
Office Action on U.S. Appl. No. 11/276,854, dated Jun. 10, 2009.
Office Action on U.S. Appl. No. 11/276,854, dated Jun. 5, 2013.
Office Action on U.S. Appl. No. 11/276,854, dated Jun. 8, 2010.
Office Action on U.S. Appl. No. 11/276,854, dated Nov. 26, 2008.
Office Action on U.S. Appl. No. 11/276,854, dated Oct. 27, 2010.
Office Action on U.S. Appl. No. 11/276,855, dated Aug. 13, 2009.
Office Action on U.S. Appl. No. 11/276,855, dated Dec. 30, 2008.
Office Action on U.S. Appl. No. 11/276,855, dated Dec. 31, 2009.
Office Action on U.S. Appl. No. 11/276,855, dated Dec. 7, 2010.
Office Action on U.S. Appl. No. 11/276,855, dated Jan. 26, 2012.
Office Action on U.S. Appl. No. 11/276,855, dated Jul. 22, 2010.
Office Action on U.S. Appl. No. 11/276,855, dated Jun. 27, 2011.
Office Action on U.S. Appl. No. 11/616,156, dated Jan. 18, 2011.
Office Action on U.S. Appl. No. 11/616,156, dated Oct. 13, 2011.
Office Action on U.S. Appl. No. 11/616,156, dated Sep. 17, 2013.
Office Action on U.S. Appl. No. 11/718,867 dated Dec. 29, 2009.
Office Action on U.S. Appl. No. 11/718,867 dated Jan. 8, 2009.
Office Action on U.S. Appl. No. 11/718,867 dated Jul. 11, 2008.
Office Action on U.S. Appl. No. 11/718,867 dated Jun. 15, 2009.
Office Action on U.S. Appl. No. 12/573,967, dated Apr. 1, 2014.
Office Action on U.S. Appl. No. 12/573,967, dated Aug. 13, 2012.
Office Action on U.S. Appl. No. 12/573,967, dated Mar. 1, 2012.
Office Action on U.S. Appl. No. 12/573,967, dated Nov. 21, 2014.
Office Action on U.S. Appl. No. 12/573,967, dated Oct. 10, 2013.
Office Action on U.S. Appl. No. 12/794,996, dated Jun. 19, 2013.
Office Action on U.S. Appl. No. 12/794,996, dated Sep. 17, 2012.
Office Action on U.S. Appl. No. 12/889,721 dated Aug. 2, 2016.
Office Action on U.S. Appl. No. 12/889,721, dated Apr. 17, 2014.
Office Action on U.S. Appl. No. 12/889,721, dated Feb. 24, 2016.
Office Action on U.S. Appl. No. 12/889,721, dated Jul. 2, 2013.
Office Action on U.S. Appl. No. 12/889,721, dated May 22, 2015.
Office Action on U.S. Appl. No. 12/889,721, dated Oct. 11, 2012.
Office Action on U.S. Appl. No. 12/889,721, dated Sep. 29, 2014.
Office Action on U.S. Appl. No. 13/234,054 dated May 31, 2017.
Office Action on U.S. Appl. No. 13/234,054 dated Oct. 20, 2016.
Office Action on U.S. Appl. No. 13/234,054, dated Apr. 16, 2015.
Office Action on U.S. Appl. No. 13/234,054, dated Aug. 6, 2015.
Office Action on U.S. Appl. No. 13/234,054, dated Jan. 26, 2016.
Office Action on U.S. Appl. No. 13/234,054, dated Oct. 23, 2014.
Office Action on U.S. Appl. No. 13/284,855, dated Dec. 19, 2013.
Office Action on U.S. Appl. No. 13/453,086, dated Mar. 12, 2013.
Office Action on U.S. Appl. No. 13/475,713, dated Apr. 1, 2014.
Office Action on U.S. Appl. No. 13/475,713, dated Oct. 17, 2014.
Office Action on U.S. Appl. No. 13/475,722, dated Jan. 17, 2014.
Office Action on U.S. Appl. No. 13/475,722, dated Oct. 20, 2014.
Office Action on U.S. Appl. No. 13/527,498, dated May 8, 2014.
Office Action on U.S. Appl. No. 13/527,498, dated Nov. 17, 2014.
Office Action on U.S. Appl. No. 13/527,505, dated Dec. 5, 2014.
Office Action on U.S. Appl. No. 13/527,505, dated May 8, 2014.
Office Action on U.S. Appl. No. 13/621,987 dated Feb. 27, 2015.
Office Action on U.S. Appl. No. 13/621,987 dated Oct. 8, 2014.
Office Action on U.S. Appl. No. 13/624,725 dated Mar. 10, 2016.
Office Action on U.S. Appl. No. 13/624,725, dated Apr. 23, 2015.
Office Action on U.S. Appl. No. 13/624,725, dated Jan. 10, 2013.
Office Action on U.S. Appl. No. 13/624,725, dated Nov. 4, 2015.
Office Action on U.S. Appl. No. 13/624,725, dated Nov. 13, 2013.
Office action on U.S. Appl. No. 13/624,731 dated Jan. 29, 2013.
Office Action on U.S. Appl. No. 13/624,731, dated Jul. 25, 2014.
Office Action on U.S. Appl. No. 13/662,759, dated Feb. 22, 2016.
Office Action on U.S. Appl. No. 13/662,759, dated Nov. 6, 2014.
Office Action on U.S. Appl. No. 13/692,741, dated Jul. 1, 2015.
Office Action on U.S. Appl. No. 13/692,741, dated Mar. 11, 2015.
Office Action on U.S. Appl. No. 13/692,741, dated Sep. 4, 2014.
Office Action on U.S. Appl. No. 13/705,286, dated May 13, 2013.
Office Action on U.S. Appl. No. 13/705,340, dated Aug. 2, 2013.
Office Action on U.S. Appl. No. 13/705,340, dated Mar. 12, 2014.
Office Action on U.S. Appl. No. 13/705,340, dated Mar. 29, 2013.
Office Action on U.S. Appl. No. 13/705,386, dated May 13, 2013.
Office Action on U.S. Appl. No. 13/705,414, dated Apr. 9, 2013.
Office Action on U.S. Appl. No. 13/705,414, dated Aug. 9, 2013.
Office Action on U.S. Appl. No. 13/705,428, dated Jul. 10, 2013.
Office Action on U.S. Appl. No. 13/728,308, dated May 14, 2015.
Office Action on U.S. Appl. No. 13/728,428 dated May 6, 2016.
Office Action on U.S. Appl. No. 13/728,428, dated Jun. 12, 2015.
Office Action on U.S. Appl. No. 13/760,600 dated Aug. 30, 2016.
Office Action on U.S. Appl. No. 13/760,600 dated Jan. 23, 2017.
Office Action on U.S. Appl. No. 13/760,600 dated Jun. 15, 2017.
Office Action on U.S. Appl. No. 13/760,600 dated Mar. 15, 2016.
Office Action on U.S. Appl. No. 13/760,600 dated Oct. 19, 2015.
Office Action on U.S. Appl. No. 13/760,600, dated Apr. 10, 2015.
Office Action on U.S. Appl. No. 13/855,241, dated Jan. 13, 2016.
Office Action on U.S. Appl. No. 13/855,241, dated Jul. 6, 2015.
Office Action on U.S. Appl. No. 13/855,241, dated Jun. 27, 2019.
Office Action on U.S. Appl. No. 13/855,241, dated Mar. 30, 2020.
Office Action on U.S. Appl. No. 13/855,241, dated Sep. 15, 2016.
Office Action on U.S. Appl. No. 14/052,723, dated Dec. 3, 2015.
Office Action on U.S. Appl. No. 14/052,723, dated May 1, 2015.
Office Action on U.S. Appl. No. 14/106,254 dated Aug. 12, 2016.
Office Action on U.S. Appl. No. 14/106,254 dated Feb. 15, 2017.
Office Action on U.S. Appl. No. 14/106,254, dated May 2, 2016.
Office Action on U.S. Appl. No. 14/106,697 dated Feb. 2, 2016.
Office Action on U.S. Appl. No. 14/106,697, dated Aug. 17, 2015.
Office Action on U.S. Appl. No. 14/106,698, dated Aug. 19, 2015.
Office Action on U.S. Appl. No. 14/106,698, dated Feb. 12, 2015.
Office Action on U.S. Appl. No. 14/137,921 dated Feb. 4, 2021.
Office Action on U.S. Appl. No. 14/137,921 dated Jun. 25, 2020.
Office Action on U.S. Appl. No. 14/137,921 dated May 31, 2017.
Office Action on U.S. Appl. No. 14/137,921 dated May 6, 2016.
Office Action on U.S. Appl. No. 14/137,921 dated Oct. 6, 2016.
Office Action on U.S. Appl. No. 14/137,921 dated Oct. 8, 2015.
Office Action on U.S. Appl. No. 14/137,940 dated Aug. 10, 2018.
Office Action on U.S. Appl. No. 14/137,940 dated Jan. 25, 2018.
Office Action on U.S. Appl. No. 14/137,940 dated Jun. 3, 2016.
Office Action on U.S. Appl. No. 14/137,940 dated Jun. 9, 2017.
Office Action on U.S. Appl. No. 14/137,940 dated Nov. 3, 2016.
Office Action on U.S. Appl. No. 14/154,912, dated Dec. 7, 2017.
Office Action on U.S. Appl. No. 14/154,912, dated Jul. 20, 2017.
Office Action on U.S. Appl. No. 14/154,912, dated May 8, 2018.
Office Action on U.S. Appl. No. 14/154,912, dated Oct. 11, 2018.
Office Action on U.S. Appl. No. 14/331,718 dated Feb. 28, 2017.
Office Action on U.S. Appl. No. 14/331,772, dated Aug. 11, 2017.
Office Action on U.S. Appl. No. 14/334,178 dated Dec. 18, 2015.
Office Action on U.S. Appl. No. 14/334,178, dated Nov. 4, 2015.
Office Action on U.S. Appl. No. 14/334,931 dated Dec. 11, 2015.
Office Action on U.S. Appl. No. 14/334,931, dated Jan. 5, 2015.
Office Action on U.S. Appl. No. 14/334,931, dated Jul. 9, 2015.
Office Action on U.S. Appl. No. 14/590,102, dated Aug. 15, 2017.
Office Action on U.S. Appl. No. 14/691,120 dated Mar. 10, 2022.
Office Action on U.S. Appl. No. 14/691,120 dated Mar. 30, 2020.
Office Action on U.S. Appl. No. 14/691,120 dated Oct. 3, 2019.
Office Action on U.S. Appl. No. 14/691,120 dated Oct. 20, 2020.
Office Action on U.S. Appl. No. 14/691,120 dated Sep. 29, 2021.
Office Action on U.S. Appl. No. 14/691,120, dated Aug. 27, 2018.
Office Action on U.S. Appl. No. 14/691,120, dated Feb. 12, 2018.
Office Action on U.S. Appl. No. 14/691,120, dated Mar. 2, 2017.
Office Action on U.S. Appl. No. 14/691,120, dated Mar. 22, 2019.
Office Action on U.S. Appl. No. 14/691,120, dated Sep. 13, 2017.
Office Action on U.S. Appl. No. 14/709,642 dated Feb. 7, 2018.
Office Action on U.S. Appl. No. 14/709,642 dated Feb. 17, 2016.
Office Action on U.S. Appl. No. 14/709,642 dated Jul. 12, 2017.
Office Action on U.S. Appl. No. 14/709,642 dated Sep. 12, 2016.
Office Action on U.S. Appl. No. 14/725,543 dated Apr. 7, 2016.
Office Action on U.S. Appl. No. 14/751,529 dated Aug. 9, 2017.
Office Action on U.S. Appl. No. 14/751,529 dated Oct. 3, 2018.
Office Action on U.S. Appl. No. 14/751,529, dated Jun. 6, 2016.
Office Action on U.S. Appl. No. 14/751,529, dated Nov. 14, 2016.
Office Action on U.S. Appl. No. 14/753,948 dated Nov. 4, 2016.
Office Action on U.S. Appl. No. 14/791,873 dated May 14, 2018.
Office Action on U.S. Appl. No. 14/809,723 dated Aug. 25, 2017.
Office Action on U.S. Appl. No. 14/809,723 dated Dec. 30, 2016.
Office Action on U.S. Appl. No. 14/827,927 dated Jan. 19, 2021.
Office Action on U.S. Appl. No. 14/827,927 dated Jan. 31, 2020.
Office Action on U.S. Appl. No. 14/827,927 dated May 16, 2018.
Office Action on U.S. Appl. No. 14/827,927 dated May 16, 2019.
Office Action on U.S. Appl. No. 14/827,927 dated Sep. 9, 2019.
Office Action on U.S. Appl. No. 14/827,927, dated Aug. 28, 2018.
Office Action on U.S. Appl. No. 14/827,927, dated Jan. 31, 2019.
Office Action on U.S. Appl. No. 14/833,673 dated Aug. 11, 2017.
Office Action on U.S. Appl. No. 14/833,673, dated Feb. 11, 2016.
Office Action on U.S. Appl. No. 14/833,673, dated Jun. 10, 2016.
Office Action on U.S. Appl. No. 14/833,673, dated Sep. 24, 2015.
Office Action on U.S. Appl. No. 14/842,916 dated May 5, 2017.
Office Action on U.S. Appl. No. 14/872,645 dated Feb. 16, 2016.
Office Action on U.S. Appl. No. 14/872,645 dated Jun. 29, 2016.
Office Action on U.S. Appl. No. 14/987,059, dated Jan. 31, 2019.
Office Action on U.S. Appl. No. 14/987,059, dated May 11, 2018.
Office Action on U.S. Appl. No. 14/987,059, dated Oct. 11, 2018.
Office Action on U.S. Appl. No. 15/042,489 dated Jan. 9, 2018.
Office Action on U.S. Appl. No. 15/078,115 dated Sep. 5, 2017.
Office Action on U.S. Appl. No. 15/254,111 dated Jun. 20, 2017.
Office Action on U.S. Appl. No. 15/281,462 dated Apr. 6, 2018.
Office Action on U.S. Appl. No. 15/281,462 dated Dec. 15, 2017.
Office Action on U.S. Appl. No. 15/281,462 dated Feb. 10, 2017.
Office Action on U.S. Appl. No. 15/281,462 dated Jun. 13, 2017.
Office Action on U.S. Appl. No. 15/345,017 dated Aug. 24, 2020.
Office Action on U.S. Appl. No. 15/345,017 dated Aug. 9, 2019.
Office Action on U.S. Appl. No. 15/345,017 dated Jan. 31, 2019.
Office Action on U.S. Appl. No. 15/345,017 dated Jul. 11, 2018.
Office Action on U.S. Appl. No. 15/345,017 dated Mar. 20, 2020.
Office Action on U.S. Appl. No. 15/345,017 dated Nov. 29, 2019.
Office Action on U.S. Appl. No. 15/357,332 dated May 9, 2018.
Office Action on U.S. Appl. No. 15/357,332 dated Nov. 9, 2017.
Office Action on U.S. Appl. No. 15/478,467, dated Jan. 11, 2019.
Office Action on U.S. Appl. No. 15/478,467, dated Jul. 13, 2018.
Office Action on U.S. Appl. No. 15/717,392 dated Dec. 3, 2018.
Office Action on U.S. Appl. No. 15/717,392 dated Jul. 5, 2018.
Office Action on U.S. Appl. No. 15/726,509, dated Jun. 3, 2019.
Office Action on U.S. Appl. No. 13/624,731, dated Nov. 12, 2013.
Office Action on U.S. Appl. No. 15/270,418 dated Apr. 21, 2017.
PCT/US2005/008296—International Search Report dated Aug. 3, 2005 for PCT Application No. PCT/US2005/008296, 1 page.
PCT/US2005/008297—International Search Report for Application No. PCT/US2005/008297, dated Sep. 29, 2005.
PCT/US2005/040669—International Preliminary Examination Report for PCT/US2005/040669, dated Apr. 29, 2008.
PCT/US2005/040669—Written Opinion for PCT/US2005/040669, dated Sep. 13, 2006.
PCT/US2009/044200—International Preliminary Report on Patentability for PCT/US2009/044200, dated Nov. 17, 2010.
PCT/US2009/044200—International Search Report and Written Opinion on PCT/US2009/044200, dated Jul. 1, 2009.
PCT/US2010/053227—International Preliminary Report on Patentability for PCT/US2010/053227, dated May 10, 2012.
PCT/US2010/053227—International Search Report and Written Opinion for PCT/US2010/053227, dated Dec 16, 2010.
PCT/US2011/051996—International Search Report and Written Opinion for PCT/US2011/051996, dated Jan. 19, 2012.
PCT/US2012/038986—International Preliminary Report on Patentability for PCT/US2012/038986 dated Nov. 26, 2013.
PCT/US2012/038986—International Search Report and Written Opinion on PCT/US2012/038986, dated Mar. 14, 2013.
PCT/US2012/038987—International Search Report and Written Opinion for PCT/US2012/038987, dated Aug. 16, 2012.
PCT/US2012/061747—International Preliminary Report on Patentability for PCT/US2012/061747, dated Apr. 29, 2014.
PCT/US2012/061747—International Search Report and Written Opinion for PCT/US2012/061747, dated Mar. 1, 2013.
PCT/US2012/062608—International Preliminary Report on Patentability issued on PCT/US2012/062608, dated May 6, 2014.
PCT/US2012/062608—International Search Report and Written Opinion for PCT/US2012/062608, dated Jan. 18, 2013.
Petition for Inter Partes Review of U.S. Pat. No. 8,271,980, Challenging Claims 1-5 and 14-15, document filed on behalf of Unified Patents, LLC, in Case No. IPR2022-00136, 92 pages, Petition document dated Nov. 29, 2021.
Roblitz et al., “Resource Reservations with Fuzzy Requests”, Con-currency and computation: Practice and Experience, 2005.
Smith et al.; “Grid computing”; MIT Sloan Management Review, vol. 46, Iss. 1.; 5 pages; Fall 2004.
Snell et al., “The Performance Impact of Advance Reservation Meta-Scheduling”, Springer-Verlag, Berlin, 2000, pp. 137-153.
Stankovic et al., “The Case for Feedback Control Real-Time Scheduling” 1999, IEEE pp. 1-13.
Takahashi et al. “A Programming Interface for Network Resource Management,” 1999 IEEE, pp. 34-44.
Tanaka et al. “Resource Manager for Globus-Based Wide-Area Cluster Computing,” 1999 IEEE, 8 pages.
U.S. Appl. No. 60/552,653, filed Apr. 19, 2005.
U.S. Appl. No. 60/662,240, filed Mar. 16, 2005, Jackson.
Brad Stone et al., UNIX Fault Management: A Guide for System Administration, Dec. 1, 1999, ISBN 0-13-026525-X, http://www.informit.com/content/images/013026525X/samplechapter/013026525-.pdf.
Chuang Liu et al. “Design and Evaluation of a Resource Selection Framework for Grid Applications” High Performance Distributed Computing, 2002. HPDC-11 2002. Proceedings S. 11.sup.th IEEE International Symposium on Jul. 23-26, 2002, Piscataway, NJ, USA IEEE, Jul. 23, 2002, pp. 63-72, XP010601162 ISBN: 978-0-7695-1686-8.
Design and Evaluation of a Resource Selection Framework for Grid Applications Liu et al. High Performance Distributed Computing, 2002.
IBM Tivoli Workload Scheduler job Scheduling Console User's Guide Feature Level 1.2 (Maintenance Release Oct. 2003). Oct. 2003, IBM Corporation, http://publib.boulder.ibm.com/tividd/td/TWS/SH19-4552-01/en.sub.--US/PDF/-jsc.sub.--user.pdf.
Lars C. Wolf et al. “Concepts for Resource Reservation in Advance” Multimedia Tools and Applications. [Online] 1997, pp. 255-278, XP009102070 The Netherlands Retreived from the Internet: URL: [retrieved on Jun. 23, 2008].
Leinberger, W. et al., “Gang Scheduling for Distributed Memory Systems”, University of Minnesota—Computer Science and Engineering—Technical Report, Feb. 16, 2000, vol. TR 00-014.
Roy, Alain, “Advance Reservation API”, University of Wisconsin-Madison, GFD-E.5, Scheduling Working Group, May 23, 2002.
Supercluster Research and Development Group, “Maui Administrator's Guide”, Internet citation, 2002.
Wolf et al. “Concepts for Resource Reservation in Advance” Multimedia Tools and Applications, 1997.
Abdelwahed, Sherif et al., “A Control-Based Framework for Self-Managing Distributed Computing Systems”, WOSS'04 Oct. 31-Nov. 1, 2004 Newport Beach, CA, USA. Copyright 2004 ACM 1-58113-989-6/04/0010.
Abdelzaher, Tarek, et al., “Performance Guarantees for Web Server End-Systems: A Control-Theoretical Approach”, IEEE Transactions on Parallel and Distributed Systems, vol. 13, No. 1, Jan. 2002.
Advanced Switching Technology Tech Brief, published 2005, 2 pages.
Amini, A. Shaikh, and H. Schulzrinne, “Effective Peering for Multi-provider Content Delivery Services”, In Proceedings of 23.sup.rd Annual IEEE Conference on Computer Communications (INFOCOM'04), pp. 850-861, 2004.
Amir and D. Shaw, “WALRUS—A Low Latency, High Throughput Web Service Using Internet-wide Replication”, In Proceedings of the 19.sup.th International Conference on Distributed Computing Systems Workshop, 1998.
Appleby, K., et al., “Oceano-SLA Based Management of a Computing Utility”, IBM T.J. Watson Research Center, P.O.Box 704, Yorktown Heights, New York 10598, USA. Proc. 7th IFIP/IEEE Int'l Symp. Integrated Network Management, IEEE Press 2001.
Aweya, James et al., “An adaptive load balancing scheme for web servers”, International Journal of Network Management 2002; 12: 3-39 (DOI: 10.1002/nem.421), Copyright 2002 John Wiley & Sons, Ltd.
Azuma, T. Okamoto, G. Hasegawa, and M. Murata, “Design, Implementation and Evaluation of Resource Management System for Internet Servers”, IOS Press, Journal of High Speed Networks, vol. 14 Issue 4, pp. 301-316, Oct. 2005.
Baentsch, Michael et al., “World Wide Web Caching: The Application-Level View of the Internet”, Communications Magazine, IEEE, vol. 35, Issue 6, pp. 170-178, Jun. 1997.
Banga, Gaurav et al., “Resource Containers: A New Facility for Resource Management in Server Systems”, Rice University, originally published in the Proceedings of the 3.sup.rd Symposium on Operating Systems Design and Implementation, New Orleans, Louisiana, Feb. 1999.
Belloum, A. et al., “A Scalable Web Server Architecture”, World Wide Web: Internet and Web Information Systems, 5, 5-23, 2002 Kluwer Academic Publishers. Manufactured in The Netherlands. 2000.
Benkner, Siegfried, et al., “VGE—A Service-Oriented Grid Environment for On-Demand Supercomputing”, Institute for Software Science, University of Vienna, Nordbergstrasse 15/C/3, A-1090 Vienna, Austria. Proceedings of the 5th IEEE/ACM International Workshop on Grid Computing. pp. 11-18. 2004.
Bent, Leeann et al., “Characterization of a Large Web Site Population with Implications for Content Delivery”, WWW2004, May 17-22, 2004, New York, New York, USA ACM 1-58113-844-X/04/0005, pp. 522-533.
Bian, Qiyong, et al., “Dynamic Flow Switching, A New Communication Service for ATM Networks”, 1997.
Bradford, S. Milliner, and M. Dumas, “Experience Using a Coordination-based Architecture for Adaptive Web Content Provision”, In Coordination, pp. 140-156. Springer, 2005.
Braumandl, R. et al., “ObjectGlobe: Ubiquitous query processing on the Internet”, Universitat Passau, Lehrstuhl fur Informatik, 94030 Passau, Germany. Technische Universitaat Muunchen, Institut fur Informatik, 81667 Munchen, Germany. Edited by F. Casati, M.-C. Shan, D. Georgakopoulos. Published online Jun. 7, 2001—.sub.—cSpringer-Verlag 2001.
Cardellini, Valeria et al., “Geographic Load Balancing for Scalable Distributed Web Systems”, Proceedings of the 8th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems, pp. 20-27. 2000.
Cardellini, Valeria et al., “The State of the Art in Locally Distributed Web-Server Systems”, ACM Computing Surveys, vol. 34, No. 2, Jun. 2002, pp. 263-311.
Casalicchio, Emiliano, et al., “Static and Dynamic Scheduling Algorithms for Scalable Web Server Farm”, University of Roma Tor Vergata, Roma, Italy, 00133.2001. In Proceedings of the IEEE 9.sup.th Euromicro Workshop on Parallel and Distributed Processing, pp. 369-376, 2001.
Chandra, Abhishek et al., “Dynamic Resource Allocation for Shared Data Centers Using Online Measurements” Proceedings of the 11th international conference on Quality of service, Berkeley, CA, USA pp. 381-398. 2003.
Chandra, Abhishek et al., “Quantifying the Benefits of Resource Multiplexing in On-Demand Data Centers”, Department of Computer Science, University of Massachusetts Amherst, 2003.
Chapter 1 Overview of the Origin Family Architecture from Origin and Onyx2 Theory of Operations Manual, published 1997, 18 pages.
Chawla, Hamesh et al., “HydraNet: Network Support for Scaling of Large-Scale Services”,Proceedings of 7th International Conference on Computer Communications and Networks, 1998. Oct. 1998.
Chellappa, Ramnath et al., “Managing Computing Resources in Active Intranets”, International Journal of Network Management, 2002, 12:117-128 (DOI:10.1002/nem.427).
Chen and G. Agrawal, “Resource Allocation in a Middleware for Streaming Data”, In Proceedings of the 2.sup.nd Workshop on Middleware for Grid Computing (MGC '04), pp. 5-10, Toronto, Canada, Oct. 2004.
Chen, et al., “Replicated Servers Allocation for Multiple Information Sources in a Distributed Environment”, Department of Computer Science, Hong Kong University of Science and Technology, Clear Water Bay, Kowloon, Hong Kong, Sep. 2000.
Chen, Thomas, “Increasing the Observability of Internet Behavior”, Communications of the ACM, vol. 44, No. 1, pp. 93-98, Jan. 2001.
Chen, Xiangping et al., “Performance Evaluation of Service Differentiating Internet Servers”, IEEE Transactions on Computers, vol. 51, No. 11, pp. 1368-1375, Nov. 2002.
Cisco MDS 9000 Family Multiprotocol Services Module, published 2006, 13 pages.
Clark, et al., “Providing Scalable Web Service Using Multicast Delivery”, College of Computing, Georgia Institute of Technology, Atlanta, GA 30332-0280, 1995.
Clarke and G. Coulson, “An Architecture for Dynamically Extensible Operating Systems”, In Proceedings of the 4th International Conference on Configurable Distributed Systems (ICCDS'98), Annapolis, MD, May 1998.
Colajanni, Michele et al., “Analysis of Task Assignment Policies in Scalable Distributed Web-server Systems”, IEEE Transactions on Parallel and Distributed Systes, vol. 9, No. 6, Jun. 1998.
Colajanni, P. Yu, V. Cardellini, M. Papazoglou, M. Takizawa, B. Cramer and S. Chanson, “Dynamic Load Balancing in Geographically Distributed Heterogeneous Web Servers”, In Proceedings of the 18.sup.th International Conference on Distributed Computing Systems, pp. 295-302, May 1998.
Comparing the I2C Bus to the SMBUS, Maxim Integrated, Dec. 1, 2000, p. 1.
Conti, Marco et al., “Quality of Service Issues in Internet Web Services”, IEEE Transactions on Computers, vol. 51, No. 6, pp. 593-594, Jun. 2002.
Conti, Marco, et al., “Client-side content delivery policies in replicated web services: parallel access versus single server approach”, Istituto di Informatica e Telematica (IIT), Italian National Research Council (CNR), Via G. Moruzzi, I. 56124 Pisa, Italy, Performance Evaluation 59 (2005) 137-157, Available online Sep. 11, 2004.
Das et al., “Unifying Packet and Circuit Switched Networks,” IEEE Globecom Workshops 2009, Nov. 30, 2009, pp. 1-6.
Deering, “IP Multicast Extensions for 4.3BSD Unix and related Systems,” Jun. 1999, 5 pages.
Devarakonda, V.K. Naik, N. Rajamanim, “Policy-based multi-datacenter resource management”, In 6.sup.th IEEE International Workshop on Policies for Distributed Systems and Networks, pp. 247-250, Jun. 2005.
Dilley, John, et al., “Globally Distributed Content Delivery”, IEEE Internet Computing, 1089-7801/02/$17.00 .Copyrgt. 2002 IEEE, pp. 50-58, Sep.-Oct. 2002.
Doyle, J. Chase, O. Asad, W. Jin, and A. Vahdat, “Model-Based Resource Provisioning in a Web Service Utility”, In Proceedings of the Fourth USENIX Symposium on Internet Technologies and Systems (USITS), Mar. 2003.
Elghany et al., “High Throughput High Performance NoC Switch,” NORCHIP 2008, Nov. 2008, pp. 237-240.
Ercetin, Ozgur et al., “Market-Based Resource Allocation for Content Delivery in the Internet”, IEEE Transactions on Computers, vol. 52, No. 12, pp. 1573-1585, Dec. 2003.
Fan, Li, et al., “Summary Cache: A Scalable Wide-Area Web Cache Sharing Protocol”, IEEE/ACM Transactions on networking, vol. 8, No. 3, Jun. 2000.
Feldmann, Anja, et al., “Efficient Policies for Carrying Web Traffic Over Flow-Switched Networks”, IEEE/ACM Transactions on Networking, vol. 6, No. 6, Dec. 1998.
Feldmann, Anja, et al., “Reducing Overhead in Flow-Switched Networks: An Empirical Study of Web Traffic”, AT&T Labs-Research, Florham Park, NJ, 1998.
Fong, L.L. et al., “Dynamic Resource Management in an eUtility”, IBM T. J. Watson Research Center, 0-7803-7382-0/02/$17.00 .Copyrgt. 2002 IEEE.
Foster, Ian et al., “The Anatomy of the Grid—Enabling Scalable Virtual Organizations”, To appear: Intl J. Supercomputer Applications, 2001.
Fox, Armando et al., “Cluster-Based Scalable Network Services”, University of California at Berkeley, SOSP—Oct. 16, 1997 Saint-Malo, France, ACM 1997.
fpga4fun.com, “What is JTAG?”, 2 pages, Jan. 31, 2010.
From AT to BTX: Motherboard Form Factor, Webopedia, Apr. 29, 2005, p. 1.
Furmento et al., “Building computational communities from federated resources.” European Conference on Parallel, Springer, Berlin, Heidelberg, pp. 855-863. (Year: 2001).
Garg, Rahul, et al., “A SLA Framework for QoS Provisioning and Dynamic Capacity Allocation”, 2002.
Gayek, P., et al., “A Web Content Serving Utility”, IBM Systems Journal, vol. 43, No. 1, pp. 43-63. 2004.
Genova, Zornitza et al., “Challenges in URL Switching for Implementing Globally Distributed Web Sites”, Department of Computer Science and Engineering, University of South Florida, Tampa, Florida 33620. 0-7695-077 I-9/00 $10.00—IEEE. 2000.
Grajcar, Martin, “Genetic List Scheduling Algorithm for Scheduling and Allocation on a Loosely Coupled Heterogeneous Multiprocessor System”, Proceedings of the 36.sup.th annual ACM/IEEE Design Automation Conference, New Orleans, Louisiana, pp. 280-285. 1999.
Grecu et al., “A Scalable Communication-Centric SoC Interconnect Architecture” Proceedings 5th International Symposium on Quality Electronic Design, 2005, pp. 343, 348 (full article included).
Grimm, Robert et al., “System Support for Pervasive Applications”, ACM Transactions on Computer Systems, vol. 22, No. 4, Nov. 2004, pp. 421-486.
Guo, L. Bhuyan, R. Kumar and S. Basu, “QoS Aware Job Scheduling in a Cluster-Based Web Server for Multimedia Applications”, In Proceedings of the 19.sup.th IEEE International Parallel and Distributed Processing Symposium (IPDPS'05), Apr. 2005.
Gupta, A., Kleinberg, J., Kumar, A., Rastogi, R. & Yener, B. “Provisioning a virtual private network: a network design problem for multicommodity flow,” Proceedings of the thirty-third annual ACM symposium on Theory of computing [online], Jul. 2001, pp. 389-398, abstract [retrieved on Jun. 14, 2007]. Retrieved from the Internet:<URL:http://portal.acm.org/citation.cfm?id=380830&dl=ACM&coll- - =GUIDE>.
Haddad and E. Paquin, “MOSIX: A Cluster Load-Balancing Solution for Linux”, In Linux Journal, vol. 2001 Issue 85es, Article No. 6, May 2001.
Hadjiefthymiades, Stathes et al., “Using Proxy Cache Relocation to Accelerate Web Browsing in Wireless/Mobile Communications”, University of Athens, Dept. of Informatics and Telecommunications, Panepistimioupolis, Ilisia, Athens, 15784, Greece. WWW10, May 1-5, 2001, Hong Kong.
He XiaoShan; QoS Guided Min-Min Heuristic for Grud Task Scheduling; Jul. 2003, vol. 18, No. 4, pp. 442-451 J. Comput. Sci. & Technol.
Hossain et al., “Extended Butterfly Fat Tree Interconnection (EFTI) Architecture for Network on CHIP,” 2005 IEEE Pacific Rim Conference on Communicatinos, Computers and Signal Processing, Aug. 2005, pp. 613-616.
HP “OpenView OS Manager using Radia software”, 5982-7478EN, Rev 1, Nov. 2005; (HP_Nov_2005.pdf; pp. 1-4).
HP ProLiant SL6500 Scalable System, Family data sheet, HP Technical sheet, Sep. 2010 4 pages.
HP Virtual Connect Traffic Flow—Technology brief, Jan. 2012, 22 pages.
Hu, E.C. et al., “Adaptive Fast Path Architecture”, Copyright 2001 by International Business Machines Corporation, pp. 191-206, IBM J. Res. & Dev. vol. 45 No. Mar. 2, 2001.
Huang, S. Sebastine and T. Abdelzaher, “An Architecture for Real-Time Active Content Distribution”, In Proceedings of the 16.sup.th Euromicro Conference on Real-Time Systems (ECRTS 04), pp. 271-280, 2004.
J. Chase, D. Irwin, L. Grit, J. Moore and S. Sprenkle, “Dynamic Virtual Clusters in a Grid Site Manager”, In Proceedings of the 12.sup.th IEEE International Symposium on High Performance Distributed Computing, pp. 90-100, 2003.
Jann, Joefon et al., “Web Applications and Dynamic Reconfiguration in UNIX Servers”, IBM, Thomos J. Watson Research Center, Yorktown Heights, New York 10598, 0-7803-7756-7/03/$17.00. 2003 IEEE. pp. 186-194.
Jansen et al., “SATA-IO to Develop Specification for Mini Interface Connector” Press Release Sep. 21, 2009, Serial ATA3 pages.
Jiang, Xuxian et al., “SODA: a Service-On-Demand Architecture for Application Service Hosting Utility Platforms”, Proceedings of the 12th IEEE International Symposium on High Performance Distributed Computing (HPDC'03) 1082-8907/03 $17.00 .COPYRGT. 2003 IEEE.
Kant, Krishna et al., “Server Capacity Planning for Web Traffic Workload”, IEEE Transactions on Knowledge and Data Engineering, vol. 11, No. 5, Sep./Oct. 1999, pp. 731-474.
Kapitza, F. J. Hauck, and H. P. Reiser, “Decentralized, Adaptive Services: The AspectIX Approach for a Flexible and Secure Grid Environment”, In Proceedings of the Grid Services Engineering and Management Conferences (GSEM, Erfurt, Germany, Nov. 2004), pp. 107-118, LNCS 3270, Springer, 2004.
Kavas et al., “Comparing Windows NT, Linux, and QNX as the Basis for Cluster Systems”, Concurrency and Computation Practice & Experience Wiley UK, vol. 13, No. 15, pp. 1303-1332, Dec. 25, 2001.
Koulopoulos, D. et al., “PLEIADES: An Internet-based parallel/distributed system”, Software-Practice and Experience 2002; 32:1035-1049 (DOI: 10.1002/spe.468).
Kuz, Ihor et al., “A Distributed-Object Infrastructure for Corporate Websites”, Delft University of Technology Vrije Universiteit Vrije Universiteit Delft, The Netherlands, 0-7695-0819-7/00 $10.00 0 2000 IEEE.
Liao, Raymond, et al., “Dynamic Core Provisioning for Quantitative Differentiated Services”, IEEE/ACM Transactions on Networking, vol. 12, No. 3, pp. 429-442, Jun. 2004.
Liu et al. “Design and Evaluation of a Resouce Selection Framework for Grid Applicaitons” High Performance Distributed Computing. 2002. HPDC-11 2002. Proceeding S. 11.sup.th IEEE International Symposium on Jul. 23-26, 2002, Piscataway, NJ, USA IEEE, Jul. 23, 2002, pp. 63-72, XP010601162 ISBN: 978-0-7695-1686-8.
Lowell, David et al., “Devirtualizable Virtual Machines Enabling General, Single-Node, Online Maintenance”, ASPLOS'04, Oct. 9-13, 2004, Boston, Massachusetts, USA. pp. 211-223, Copyright 2004 ACM.
Lu, Chenyang et al., “A Feedback Control Approach for Guaranteeing Relative Delays in Web Servers”, Department of Computer Science, University of Virginia, Charlottesville, VA 22903, 0-7695-1134-1/01 $10.00.2001 IEEE.
Mahon, Rob et al., “Cooperative Design in Grid Services”, The 8th International Conference on Computer Supported Cooperative Work in Design Proceedings. pp. 406-412. IEEE 2003.
McCann, Julie, et al., “Patia: Adaptive Distributed Webserver (A Position Paper)”, Department of Computing, Imperial College London, SW1 2BZ, UK. 2003.
Montez, Carlos et al., “Implementing Quality of Service in Web Servers”, LCMI—Depto de Automacao e Sistemas—Univ. Fed. de Santa Catarina, Caixa Postal 476-88040-900—Florianopolis—SC—Brasil, 1060-9857/02 $17.00. 2002 IEEE.
Naik, S. Sivasubramanian and S. Krishnan, “Adaptive Resource Sharing in a Web Services Environment”, In Proceedings of the 5.sup.th ACM/IFIP/USENIX International Conference on Middleware (Middleware '04), pp. 311-330, Springer-Verlag New York, Inc. New York, NY, USA, 2004.
Nakrani and C. Tovey, “On Honey Bees and Dynamic Server Allocation in Internet Hosting Centers”, Adaptive Behavior, vol. 12, No. 3-4, pp. 223-240, Dec. 2004.
Nawathe et al., “Implementation of an 8-Core, 64-Thread, Power Efficient SPARC Server on a Chip”, IEEE Journal of Solid-State Circuits, vol. 43, No. 1, Jan. 2008, pp. 6-20.
Notice of Allowance on U.S. Appl. No. 17/089,207, dated Jul. 7, 2022.
Notice of Allowance on U.S. Appl. No. 17/700,847, dated Jul. 7, 2022.
Office Action on U.S. Appl. No. 13/728,362, dated Feb. 21, 2014.
Office Action on U.S. Appl. No. 16/537,256 dated Jul. 7, 2022.
Pacifici, Giovanni et al., “Performance Management for Cluster Based Web Services”, IBM TJ Watson Research Center, May 13, 2003.
Pande et al., “Design of a Switch for Network on Chip Applications,” May 25-28, 2003 Proceedings of the 2003 International Symposium on Circuits and Systems, vol. 5, pp. V217-V220.
Ranjan, J. Rolia, H. Fu, and E. Knightly, “QoS-driven Server Migration for Internet Data Centers”, In Proceedings of the Tenth International Workshop on Quality of Service (IWQoS 2002), May 2002.
Rashid, Mohammad, et al., “An Analytical Approach to Providing Controllable Differentiated Quality of Service in Web Servers”, IEEE Transactions on Parallel and Distributed Systems, vol. 16, No. 11, pp. 1022-1033, Nov. 2005.
Raunak, Mohammad et al., “Implications of Proxy Caching for Provisioning Networks and Servers”, IEEE Journal on Selected Areas in Communications, vol. 20, No. 7, pp. 1276-1289, Sep. 2002.
Reed, Daniel et al., “The Next Frontier: Interactive and Closed Loop Performance Steering”, Department of Computer Science, University of Illinois, Urbana, Illinois 61801, International Conference on Parallel Processing Workshop, 1996.
Reumann, John et al., “Virtual Services: A New Abstraction for Server Consolidation”, Proceedings of 2000 USENIX Annual Technical Conference, San Diego, California, Jun. 18-23, 2000.
Rolia, S. Singhal, and R. Friedrich, “Adaptive Internet data centers”, In Proceedings of the International Conference on Advances in Infrastructure for Electronic Business, Science, and Education on the Internet (SSGRR '00), Jul. 2000.
Rolia, X. Zhu, and M. Arlitt, “Resource Access Management for a Utility Hosting Enterprise Applications”, In Proceedings of the 8th IFIP/IEEE International Symposium on Integrated Network Management (IM), pp. 549-562, Colorado Springs, Colorado, USA, Mar. 2003.
Ryu, Kyung Dong et al., “Resource Policing to Support Fine-Grain Cycle Stealing in Networks of Workstations”, IEEE Transactions on Parallel and Distributed Systems, vol. 15, No. 10, pp. 878-892, Oct. 2004.
Sacks, Lionel et al., “Active Robust Resource Management in Cluster Computing Using Policies”, Journal of Network and Systems Management, vol. 11, No. 3, pp. 329-350, Sep. 2003.
Shaikh, Anees et al., “Implementation of a Service Platform for Online Games”, Network Software and Services, IBM T.J. Watson Research Center, Hawthorne, NY 10532, SIGCOMM'04 Workshops, Aug. 30 & Sep. 3, 2004, Portland, Oregon, USA. Copyright 2004 ACM.
Shen, H. Tang, T. Yang, and L. Chu, “Integrated Resource Management for Cluster-based Internet Services”, In Proceedings of the 5.sup.th Symposium on Operating Systems Design and Implementation (OSDI '02), pp. 225-238, Dec. 2002.
Shen, L. Chu, and T. Yang, “Supporting Cluster-based Network Services on Functionally Symmetric Software Architecture”, In Proceedings of the ACM/IEEE SC2004 Conference, Nov. 2004.
Si et al., “Language Modeling Framework for Resource Selection and Results Merging”, SIKM 2002, Proceedings of the eleventh international conference on Information and Knowledge Management.
Sit, Yiu-Fai et al., “Cyclone: A High-Performance Cluster-Based Web Server with Socket Cloning”, Department of Computer Science and Information Systems, The University of Hong Kong, Cluster Computing vol. 7, issue 1, pp. 21-37, Jul. 2004, Kluwer Academic Publishers.
Sit, Yiu-Fai et al., “Socket Cloning for Cluster-BasedWeb Servers”, Department of Computer Science and Information Systems, The University of Hong Kong, Proceedings of the IEEE International Conference on Cluster Computing, IEEE 2002.
Snell, Quinn et al., “An Enterprise-Based Grid Resource Management System”, Brigham Young University, Provo, Utah 84602, Proceedings of the 11th IEEE International Symposium on High Performance Distributed Computing, 2002.
Soldatos, John, et al., “On the Building Blocks of Quality of Service in Heterogeneous IP Networks”, IEEE Communications Surveys, The Electronic Magazine of Original Peer-Reviewed Survey Articles, vol. 7, No. 1. First Quarter 2005.
Tang, Wenting et al., “Load Distribution via Static Scheduling and Client Redirection for Replicated Web Servers”, Department of Computer Science and Engineering, 3115 Engineering Building, Michigan State University, East Lansing, MI 48824-1226, Proceedings of the 2000 International Workshop on Parallel Processing, pp. 127-133, IEEE 2000.
Taylor, M. Surridge, and D. Marvin, “Grid Resources for Industrial Applications”, In Proceedings of the IEEE International Conference on Web Services (ICWS 04), pp. 402-409, San Diego, California, Jul. 2004.
Urgaonkar, Bhuvan, et al., “Sharc: Managing CPU and Network Bandwidth in Shared Clusters”, IEEE Transactions on Parallel and Distributed Systems, vol. 15, No. 1, pp. 2-17, Jan. 2004.
Venaas, “IPv4 Multicast Address Space Registry,” 2013, http://www.iana.org/assignments/multicast-addresses/multicast-addresses.x-html.
Vidyarthi, A. K. Tripathi, B. K. Sarker, A. Dhawan, and L. T. Yang, “Cluster-Based Multiple Task Allocation in Distributed Computing System”, In Proceedings of the 18.sup.th International Parallel and Distributed Processing Symposium (IPDPS'04), p. 239, Santa Fe, New Mexico, Apr. 2004.
Villela, P. Pradhan, and D. Rubenstein, “Provisioning Servers in the Application Tier for E-commerce Systems”, In Proceedings of the 12.sup.th IEEE International Workshop on Quality of Service (IWQoS '04), pp. 57-66, Jun. 2004.
Wang, Z., et al., “Resource Allocation for Elastic Traffic: Architecture and Mechanisms”, Bell Laboratories, Lucent Technologies, Network Operations and Management Symposium, 2000. 2000 IEEE/IFIP, pp. 157-170. Apr. 2000.
Wesley et al., “Taks Allocation and Precedence Relations for Distributed Real-Time Systems”, IEEE Transactions on Computers, vol. C-36, No. 6, pp. 667-679. Jun. 1987.
Workshop on Performance and Architecture of Web Servers (PAWS-2000) Jun. 17-18, 2000, Santa Clara, CA (Held in conjunction with SIGMETRICS-2000).
Xu, Jun, et al., “Sustaining Availability of Web Services under Distributed Denial of Service Attacks”, IEEE Transactions on Computers, vol. 52, No. 2, pp. 195-208, Feb. 2003.
Xu, Zhiwei et al., “Cluster and Grid Superservers: The Dawning Experiences in China”, Institute of Computing Technology, Chinese Academy of Sciences, P.O. Box 2704, Beijing 100080, China. Proceedings of the 2001 IEEE International Conference on Cluster Computing. IEEE 2002.
Yang, Chu-Sing, et al., “Building an Adaptable, Fault Tolerant, and Highly Manageable Web Server on Clusters of Non-dedicated Workstations”, Department of Computer Science and Engineering, National Sun Yat-Sen University, Kaohsiung, Taiwan, R.O.C.. 2000.
Zeng, Daniel et al., “Efficient Web Content Delivery Using Proxy Caching Techniques”, IEEE Transactions on Systems, Man, and Cybernetics—Part C: Applications and Reviews, vol. 34, No. 3, pp. 270-280, Aug. 2004.
Zhang, Qian et al., “Resource Allocation for Multimedia Streaming Over the Internet”, IEEE Transactions on Multimedia, vol. 3, No. 3, pp. 339-355, Sep. 2001.
Chen, Liang et al., “Resource Allocation in a Middleware for Streaming Data”, 2nd Workshop on Middleware for Grid Computing Toronto, Canada, pp. 5-10, Copyright 2004 ACM.
Extended European Search Report for EP 10827330.1, dated Jun. 5, 2013.
Jackson et al., “Grid Computing: Beyond Enablement”,; Cluster Resource, Inc., Jan. 21, 2005.
Office Action on Taiwan Application 101139729, dated May 25, 2015 (English translation not available).
Office Action on U.S. Appl. No. 17/711,214, dated Jul. 8, 2022.
Reexamination Report on Japanese Application 2012-536877, dated Jan. 22, 2015, including English Translation.
Search Report on EP Application 10827330.1, dated Feb. 12, 2015.
Office Action on U.S. Appl. No. 17/711,242, dated Jul. 28, 2022.
Office Action on U.S. Appl. No. 17/835,159 dated Aug. 31, 2022.
Office Action on U.S. Appl. No. 17/201,231 dated Oct. 5, 2022.
Notice of Allowance on U.S. Appl. No. 17/222,062 dated Oct. 7, 2022.
Liu, Simon: “Securing the Clouds: Methodologies and Practices.” Encyclopedia of Cloud Computing (2016): 220. (Year: 2016).
Notice of Allowance on U.S. Appl. No. 14/827,927 dated Apr. 25, 2022.
Notice of Allowance on U.S. Appl. No. 16/913,745, dated Jun. 9, 2022.
Notice of Allowance on U.S. Appl. No. 17/700,808, dated May 26, 2022 and Jun. 6, 2022.
Office Action on U.S. Appl. No. 16/913,745 dated Jan. 13, 2022.
Office Action on U.S. Appl. No. 17/089,207 dated Jan. 28, 2022.
Office Action on U.S. Appl. No. 17/201,245 dated Mar. 18, 2022.
Office Action on U.S. Appl. No. 17/697,235 dated May 25, 2022.
Office Action on U.S. Appl. No. 17/697,368 dated Jun. 7, 2022.
Office Action on U.S. Appl. No. 17/697,403 dated Jun. 7, 2022.
Office Acton on U.S. Appl. No. 16/537,256 dated Dec. 23, 2021.
Office Acton on U.S. Appl. No. 16/913,708 dated Jun. 7, 2022.
Office Acton on U.S. Appl. No. 17/722,037 dated Jun. 13, 2022.
Office Action on U.S. Appl. No. 14/691,120, dated Sep. 8, 2022.
Notice of Allowance on U.S. Appl. No. 17/700,767 dated Jun. 27, 2022.
Office Action on U.S. Appl. No. 17/722,076 dated Jun. 22, 2022.
Edited by William Gropp, Ewing Lusk and Thomas Sterling, “Beowulf Cluster Computing with Linux,” Massachusetts Institute of Technology, 2003.
Jarek Nabrzyski, Jennifer M. Schopf and Jan Weglarz, “Grid Resources Management, State of the Art and Future Trends,” Kluwer Academic Publishers, 2004.
Notice of Allowance on U.S. Appl. No. 17/722,037, dated Jul. 18, 2022.
Notice of Allowance on U.S. Appl. No. 17/722,062 dated Jun. 15, 2022.
IQSearchText-202206090108.txt, publication dated Apr. 6, 2005, 2 pages.
Notice of Allowance on U.S. Appl. No. 16/913,708 dated Aug. 24, 2022.
Office Action on U.S. Appl. No. 17/171,152 dated Aug. 16, 2022.
Notice of Allowance on U.S. Appl. No. 16/913,745, dated Sep. 27, 2022.
Notice of Allowance on U.S. Appl. No. 17/201,245, dated Sep. 22, 2022.
Notice of Allowance on U.S. Appl. No. 17/700,808, dated Sep. 26, 2022.
Office Action on U.S. Appl. No. 17/088,954, dated Mar. 15, 2023.
Office Action, Advisory Action, on U.S. Appl. No. 17/711,242, dated Mar. 3, 2023.
Notice of Allowance on U.S. Appl. No. 17/171,152 dated Feb. 27, 2023.
Office Action on U.S. Appl. No. 17/508,661 dated Feb. 27, 2023.
Office Action on U.S. Appl. No. 17/697,235 dated Feb. 28, 2023.
Office Action on U.S. Appl. No. 17/697,403 dated Feb. 28, 2023.
Office Action on U.S. Appl. No. 14/691,120, dated Feb. 9, 2023.
Notice of Allowance on U.S. Appl. No. 17/171,152 dated Feb. 6, 2023.
Notice of Allowance on U.S. Appl. No. 17/201,231 dated Feb. 6, 2023.
Office Action, Advisory Action, on U.S. Appl. No. 17/711,214, dated Feb. 14, 2023.
Office Action, Advisory Action, on U.S. Appl. No. 17/722,076, dated Feb. 17, 2023.
Office Action on U.S. Appl. No. 14/691,120, dated Nov. 18, 2022.
Office Action on U.S. Appl. No. 17/412,832, dated Oct. 14, 2022.
Notice of Allowance on U.S. Appl. No. 16/537,256 dated Jan. 12, 2023.
Office Action on U.S. Appl. No. 17/088,954, dated Sep. 13, 2022.
Notice of Allowance on U.S. Appl. No. 17/089,207, dated Oct. 31, 2022.
Office Action on U.S. Appl. No. 17/171,152 dated Dec. 21, 2022.
Notice of Allowance on U.S. Appl. No. 17/201,245 dated Sep. 14, 2022.
Office Action on U.S. Appl. No. 17/697,235 dated Sep. 20, 2022.
Advisory Action on U.S. Appl. No. 17/697,235 dated Dec. 5, 2022.
Office Action on U.S. Appl. No. 17/697,368 dated Oct. 18, 2022.
Advisory Action on U.S. Appl. No. 17/697,368 dated Jan. 13, 2023.
Office Action on U.S. Appl. No. 17/697,403 dated Oct. 18, 2022.
Advisory Action on U.S. Appl. No. 17/697,403 dated Jan. 13, 2023.
Notice of Allowance on U.S. Appl. No. 17/700,767 dated Jul, 11 2022.
Notice of Allowance on U.S. Appl. No. 17/700,767 dated Oct. 14 2022.
Notice of Allowance on U.S. Appl. No. 17/700,808, dated Sep. 14, 2022.
Notice of Allowance on U.S. Appl. No. 17/700,847, dated Oct. 26, 2022.
Office Action on U.S. Appl. No. 17/711,214, dated Nov. 16, 2022.
Office Action on U.S. Appl. No. 17/711,242, dated Dec. 12, 2022.
Notice of Allowance on U.S. Appl. No. 17/722,037, dated Oct. 27, 2022.
Office Action on U.S. Appl. No. 17/722,076, dated Nov. 28, 2022.
Office Action on U.S. Appl. No. 17/835,159 dated Jan. 13, 2023.
Notice of Allowance on U.S. Appl. No. 17/470,209, dated Mar. 21, 2023.
Office Action on U.S. Appl. No. 17/722,076, dated Mar. 21, 2023.
Office Action, Advisory Action, on U.S. Appl. No. 17/835,159 dated Mar. 22, 2023.
Office Action on U.S. Appl. No. 17/711,214, dated Apr. 25, 2023.
Notice of Allowance in U.S. Appl. No. 17/980,844, dated Jul. 5, 2023.
Notice of Allowance, Corrected NOA, in U.S. Appl. No. 17/532,667, dated May 9, 2023.
Office Action on U.S. Appl. No. 17/711,242, dated Jun. 7, 2023.
Notice of Allowance in U.S. Appl. No. 17/532,667, dated Apr. 26, 2023.
Office Action on U.S. Appl. No. 17/697,235 dated Jul. 14, 2023.
Office Action on U.S. Appl. No. 17/697,403 dated Jul. 14, 2023.
Notice of Allowance in U.S. Appl. No. 17/980,865, dated Jul. 18, 2023.
Notice of Allowance (Corrected NOA) in U.S. Appl. No. 17/411,616, dated Apr. 6, 2023.
Office Action on U.S. Appl. No. 17/412,832, dated Apr. 20, 2023.
Notice of Allowance in U.S. Appl. No. 17/411,616, dated Mar. 29, 2023.
Notice of Allowance in U.S. Appl. No. 17/985,241, dated Apr. 3, 2023.
Office Action in U.S. Appl. No. 17/508,661 dated Jul. 27, 2023.
Office Action in U.S. Appl. No. 17/960,251 dated Aug. 2, 2023.
Notice of Allowance in U.S. Appl. No. 17/985,252 dated Jul. 31, 2023.
Office Action in U.S. Appl. No. 18/194,783 dated Nov. 14, 2023.
Office Action on U.S. Appl. No. 17/412,832, dated Dec. 5, 2023.
Office Action on U.S. Appl. No. 17/711,214, dated Dec. 4, 2023.
Office Action in U.S. Appl. No. 17/960,251 dated Dec. 11, 2023.
Office Action in U.S. Appl. No. 17/697,235 dated Nov. 7, 2023.
Office Action in U.S. Appl. No. 14/691,120, dated Aug. 18, 2023.
Office Action in U.S. Appl. No. 17/088,954, dated Sep. 19, 2023.
Office Action, Advisory Action, in U.S. Appl. No. 17/697,235 dated Sep. 26, 2023.
Office Action on U.S. Appl. No. 17/697,368 dated Aug. 8, 2023.
Advisory Action on U.S. Appl. No. 17/697,368 dated Oct. 12, 2023.
Office Action, Advisory Action, in U.S. Appl. No. 17/697,403 dated Sep. 26, 2023.
Office Action on U.S. Appl. No. 17/711,242, dated Oct. 12, 2023.
Office Action in U.S. Appl. No. 17/835,159 dated Aug. 22, 2023.
Notice of Allowance in U.S. Appl. No. 17/960,228, dated Sep. 12, 2023.
Office Action in U.S. Appl. No. 17/960,244 dated Oct. 23, 2023.
Notice of Allowance in U.S. Appl. No. 17/985,267 dated Aug. 18, 2023.
Office Action in U.S. Appl. No. 18/120,123 dated Sep. 27, 2023.
Office Action in U.S. Appl. No. 18/295,344 dated Oct. 23, 2023.
Notice of Allowance in U.S. Appl. No. 18/194,783 dated Mar. 15, 2024.
Notice of Allowance in U.S. Appl. No. 18/232,512 dated Mar. 15, 2024.
Office Action in U.S. Appl. No. 14/691,120, dated Mar. 27, 2024.
Office Action in U.S. Appl. No. 17/088,954, dated Apr. 9, 2024.
Office Action in U.S. Appl. No. 17/697,235 dated Mar. 18, 2024.
Office Action in U.S. Appl. No. 17/711,242, dated Feb. 27, 2024.
Office Action in U.S. Appl. No. 17/902,525 dated Mar. 26, 2024.
Office Action in U.S. Appl. No. 17/960,244 dated May 20, 2024.
Office Action in U.S. Appl. No. 18/120,123 dated Apr. 9, 2024.
Office Action in U.S. Appl. No. 18/234,021 dated Apr. 19, 2024.
Office Action in U.S. Appl. No. 18/234,045 dated Apr. 19, 2024.
Office Action in U.S. Appl. No. 17/697,368 dated Dec. 19, 2023.
Notice of Allowance in U.S. Appl. No. 17/697,403 dated Dec. 18, 2023.
Office Action in U.S. Appl. No. 18/133,048 dated Dec. 18, 2023.
Office Action in U.S. Appl. No. 17/508,661 dated Jan. 26, 2024.
Office Action, Advisory Action, on U.S. Appl. No. 17/711,242, dated Dec. 20, 2023.
Office Action in U.S. Appl. No. 17/835,159 dated Jan. 12, 2024.
Notice of Allowance (Corrected) in U.S. Appl. No. 18/132,507, dated Feb. 27, 2024.
Notice of Allowance in U.S. Appl. No. 18/132,507, dated Feb. 12, 2024.
Office Action in U.S. Appl. No. 18/295,344 dated Feb. 12, 2024.
Related Publications (1)
Number Date Country
20220206860 A1 Jun 2022 US
Provisional Applications (1)
Number Date Country
60552653 Mar 2004 US
Continuations (3)
Number Date Country
Parent 17088954 Nov 2020 US
Child 17697368 US
Parent 13855241 Apr 2013 US
Child 17088954 US
Parent 10530581 US
Child 13855241 US