Method of delaying checkpoints by inspecting network packets

Information

  • Patent Grant
  • 9760442
  • Patent Number
    9,760,442
  • Date Filed
    Tuesday, December 16, 2014
    9 years ago
  • Date Issued
    Tuesday, September 12, 2017
    7 years ago
Abstract
A method of delaying checkpointing in a virtual machine system. In one embodiment, the method includes the steps of examining a network frame to determine if it is a deferrable frame and if the frame is a deferrable frame, delaying a checkpoint associated with the frame. In another embodiment, the deferrable frame is one of a group comprising: an IP packet tagged with the ‘more fragments’ attribute; TCP data segments that lack the PSH flag and carry no flags other than ‘ACK’; and TCP segments that contain no data and carry only the ‘ACK’ flag; and any frame originating from or destined to a designated network address or port number. In still another embodiment, the method includes the step of concatenating the delays due to deferrable frames. In still yet another embodiment, the method further includes setting an upper limit to the amount of delay that can be generated.
Description
FIELD OF THE INVENTION

The invention relates generally to the field of fault tolerant computing and more specifically to checkpointing, networking and virtualization systems.


BACKGROUND OF THE INVENTION

There are a variety of ways to achieve fault tolerant computing. Specifically, fault tolerant hardware and fault tolerant software may be used either alone or together. As an example, it is possible to connect two (or more) computers, such that one computer, the active computer or host, actively makes calculations while the other computer (or computers) is idle or on standby in case the active computer, or hardware or software component thereon, experiences some type of failure. In these systems, the information about the state of the active computer must be saved periodically to the standby computer so that the standby computer can substantially take over from the previously active computer at the point in the calculations where active computer experienced a failure. This function can be extended for use with the modern day practice of using a virtualized environment as part of a cloud or other computing system.


Virtualization is used in many fields to reduce the number of servers or other resources needed for a particular project or organization. Present day virtual machine computer systems utilize virtual machines (VM) operating as guests within a physical host computer. Each virtual machine includes its own virtual operating system and operates under the control of a managing operating system or hypervisor executing on the host physical machine. Each virtual machine executes one or more applications and accesses physical data storage and computer networks as required by the applications. In addition, each virtual machine may in turn act as the host computer system for another virtual machine.


Multiple virtual machines may be configured as a group to execute one or more of the same programs. Typically, one virtual machine in the group is the primary or active virtual machine, and the remaining virtual machines are the secondary or standby virtual machines. If something goes wrong with the primary virtual machine, one of the secondary virtual machines can take over and assume its role in the fault tolerant computing system. This redundancy allows the group of virtual machines to operate as a fault tolerant computing system. The primary virtual machine executes applications, receives and sends network data, and reads and writes to data storage while performing automated or user-initiated tasks or interactions. The secondary virtual machines have the same capabilities as the primary virtual machine, but do not take over the relevant tasks and activities until the primary virtual machine fails or is affected by an error.


For such a collection of virtual machines to function as a fault tolerant system, the operating state, which defines memory and data storage contents of a secondary virtual machine, should be equivalent to the operating state that is memory and data storage contents of the primary virtual machine. If this condition is met, the secondary virtual machine may take over for the primary virtual machine without a loss of any data. To assure that the state of the secondary machine and its memory is equivalent to the state of the primary machine and its memory, it is necessary for the primary virtual machine periodically to transfer its state and memory contents to the secondary virtual machine.


The periodic transfer of data to maintain synchrony between the states of the virtual machines is termed checkpointing. A checkpoint defines a point in time when the data is to be transferred. During a checkpoint, the processing on the primary virtual machine is paused, so that the final state of the virtual machine and associated memory is not changed during the checkpoint interval and once the relevant data is transferred, both the primary and secondary virtual machines are in the same state. The primary virtual machine is then resumed and continues to run the application until the next checkpoint, when the process repeats.


Checkpoints can be determined by either the passage of a fixed amount of elapsed time from the last checkpoint or by the occurrence of some event, such as: the number of memory accesses (termed dirty pages); the occurrence of a network event (such as network acknowledgement that is output from the primary virtual machine); or the occurrence of excessive buffering on the secondary virtual machine (as compared to available memory), during the execution of the application. Elapsed time checkpointing is considered fixed checkpointing, while event based checkpointing is considered dynamic or variable-rate checkpointing.


Excessive checkpointing can lead to performance degradation of the primary virtual machine. In turn, this performance degradation can result in delays and data loss, which can compromise the fault tolerant nature of the system. Triggering checkpoints in response to network traffic can be particularly taxing for a checkpointing system.


Therefore, a need exists for ways to reduce overhead in the system without reducing the reliability of the system.


Embodiments of the invention address this need and others.


SUMMARY OF THE INVENTION

In one aspect, the invention relates to a method for delaying checkpointing in a virtual machine system. In one embodiment, the method includes the steps of determining if a frame is deferrable; and if the frame is a deferrable frame delaying a checkpoint associated with the frame. In another embodiment the frame is examined to determine if it is a deferrable frame. In yet another embodiment the deferrable frame is a packet tagged with one or more of: not being the last data fragment of a multi-fragment message; a data segment that does not indicate a flush of the send buffer; and a segment that only serves as an acknowledge of some prior message. In still another embodiment the deferrable frame is an IP packet tagged with one or more of: a ‘more fragments’ attribute; a TCP data segment that lacks the PSH flag and that carries no flags other than ‘ACK’; and a TCP segment that contains no data and carries only the ‘ACK’ flag. In still yet another embodiment the deferrable frame is identified by the protocol in use.


In one embodiment, the deferrable frame is identified by the session-state of the protocol in use. In another embodiment, the deferrable frame is identified by port number. In yet another embodiment, the port number is a source port number. In still another embodiment, the port number is a destination port number. In still yet another embodiment, the deferrable frame is identified by network address. In one embodiment, the network address is a source network address. In another embodiment, the network address is a destination network address.


In one embodiment, the method includes the step of concatenating the checkpoint delays due to a plurality of deferrable frames. In another embodiment, the method includes the step of setting an upper limit to the amount of delay that can be generated. In yet another embodiment, the frame is determined to be deferrable based on the network interface in use.


In another aspect, the invention relates to a fault tolerant computing system. In one embodiment, the fault tolerant computer system includes a first computing device comprising a primary virtual machine; and a second computing device comprising a secondary virtual machine, wherein the first computing device and the second computing device are networked, wherein the primary virtual machine comprises a first checkpointing engine and a first network interface, wherein the secondary virtual machine comprises a second checkpointing engine and a second network interface, and wherein the first checkpointing engine delays a checkpoint for one or more deferrable frames of data. In another embodiment, the first checkpoint engine pauses the primary virtual machine when a checkpoint is declared. In yet another embodiment, the first checkpoint engine adds an additional delay before declaring a subsequent checkpoint. In still yet another embodiment, the first checkpoint engine captures the state of the primary virtual machine while the primary virtual machine is paused.





BRIEF DESCRIPTION OF THE DRAWINGS

The structure and function of the invention can be best understood from the description herein in conjunction with the accompanying figures. The figures are not necessarily to scale, emphasis instead generally being placed upon illustrative principles. The figures are to be considered illustrative in all aspects and are not intended to limit the invention, the scope of which is defined only by the claims.



FIG. 1 is a block diagram of an embodiment of a checkpointing engine in a virtual machine constructed in accordance with the invention.



FIGS. 2 (a) and (b) are timing diagrams known to the prior art for an embodiment of a virtual machine and a checkpointing engine, respectively.



FIGS. 3 (a) and (b) are timing diagrams for an embodiment of a virtual machine and a checkpointing engine, respectively, constructed in accordance with the invention.





DESCRIPTION OF A PREFERRED EMBODIMENT

Detailed embodiments of the invention are disclosed herein, however, it is to be understood that the disclosed embodiments are merely exemplary of the invention, which may be embodied in various forms. Therefore, specific functional details disclosed herein are not to be interpreted as limiting, but merely as a basis for the claims and as a representative basis for teaching one skilled in the art to variously employ the invention in virtually any appropriately detailed embodiment.


In a variable-rate checkpoint system, an excessive checkpoint rate can occur when checkpoint cycles are initiated on the primary virtual machine in response to the occurrence of an event on the primary virtual machine such as outbound network data transmission or traffic. Normally, such outbound network traffic causes an immediate checkpoint cycle to ensure the occurrence of lower-latency exchanges between the primary virtual machine and the computer on the network receiving the transmission from the virtual machine. This is indeed the case for file-level operations such as folder enumeration, file deletion, attribute manipulation, and even single-threaded transaction exchanges.


Under these types of latency-sensitive exchanges, it is desirable to have a rapid and responsive checkpoint rate. Latency-sensitive exchanges such as certain client requests and server responses benefit from a very responsive checkpoint mechanism. However, other exchanges, such as certain communication protocol flags and parameters, between a client and server are “housekeeping” in nature. Communications of this housekeeping or procedural-type activity produce outbound traffic from the primary virtual machine that results in excessive checkpoint activity. This excessive checkpoint activity is detrimental to application efficiency and requires substantial processing and inter-host communication overhead to handle.


An example of a network exchange of data that is housekeeping in nature is a simple TCP acknowledge packet or ‘ACK’. When a client is transferring a file to the guest virtual machine using TCP protocol, the guest virtual machine's TCP stack typically responds to each message/segment with an ‘ACK’. Hundreds of such rapid ‘ACKs’ can be generated as this incoming stream of segments is underway. The computer receiving the ACK is not dependent on the ACK in order to continue to operate or handle other necessary processing tasks. Further, with few exceptions, it is not necessary that the client computer receive these ‘ACKs’ in real time. Performing an immediate checkpoint cycle for each such ‘ACK’ would interfere greatly with the incoming message/segment flow of client frames while providing little or no benefit.


Another example of such a housekeeping network exchange of data is an IP packet tagged with the IP_MF flag, indicating that more data packet fragments are coming. These types of outbound frames increase the checkpoint rate and associated system overhead, thereby resulting in lower overall system efficiency. As a result, the checkpoints that arise from these types of network data exchanges are not critical and can be deferred for processing at a later time.


According to one embodiment, outbound network traffic and operational parameters can be monitored to detect data such as frames that can be categorized as deferrable. By deferrable, it is meant that the computer receiving the data, such as an ACK, is not typically waiting for the specific frame in order to continue processing or otherwise operating. Deferrable data examples include: IP packets tagged with the ‘more fragments’ attribute; TCP data segments that lack the PSH flag (push) and carry no flags other than ‘ACK’; and TCP segments that contain no data and only the ‘ACK’ flag. Other candidate frames and data types may exist and are handled similarly.


The above embodiments describe TCP/IP specific packets and field definitions that can be generalized to almost any network protocol. Housekeeping only packets such as IP specific “ACK” are candidates for deferrable packets. Fragmented packets such as those identified in IP with IP_MF are candidates for deferrable packets. Likewise, packets that contain no data and are not part of a send buffer flush (as in the TCP PSH flag), are also candidates for deferrable packets. Many other packet types and conditions can be identified depending on the specific protocol being used. There are hundreds of registered network protocols, each of which has its own behavior characteristics. Every packet in a given protocol may be deferrable.


Referring to FIG. 1, components of a checkpointing engine 122 with a virtual machine 100 are shown. The virtual machine 100 would typically be connected to a second virtual machine (not shown) to achieve a fault tolerant computing system. The virtual machine 100 includes a checkpoint module 101. The checkpoint module 101 buffers network data 109 in an input/output module (I/O) 110 and a notification event 112 alerts the checkpointing engine 122. Storage data (referred to generically as block—IO) is allowed to move 138 more freely between input/output module (I/O) 110 and the block interface 114. The checkpointing engine 122 is in communication with a memory tracking module 126 and a buffer cache 130. When the checkpointing engine 122 is idle, network data (if any) remains in a buffered state by input/output module (I/O) 110, and will only be released by signal 113 to input/output module (I/O) 110 after the next checkpoint reaches its commit/release stage. At that time, network data (if any) is transferred 134 from the input/output module (I/O) 110 to the network interface 118. When a checkpoint is declared, the virtual machine 100 is paused and the dirty page data is processed by the checkpointing engine 122. In the case of the present invention, when a checkpoint is declared, for example by an ACK frame being received by the I/O module 110, the frame is made available to the checkpoint engine 122 for inspection.


In more detail and referring to FIG. 2(a), a timing diagram of events and methods occurring in sequence for a primary virtual machine is shown for a checkpoint system known to the prior art. The primary virtual machine is either running (executing the application) 210, 210′ or paused 214. The checkpointing engine includes additional timing conditions as shown in FIG. 2(b). The checkpointing engine may be idle 220 until a checkpoint (P) occurs 224 at which time the checkpointing engine causes 226 the primary virtual machine to pause 214. While the primary virtual machine is paused 214, the state of the primary virtual machine is captured 228 by the checkpoint engine. Capture is a page-copying stage that allows the virtual machine to be resumed and executed in parallel with checkpoint ‘transfer’ and ‘ACK-wait’.


In another embodiment, page-copying is avoided all together and pages are just transferred as they are. In this case, the guest remains paused during the transfer and ACK-wait stages. Combinations of these two approaches may also be used. Once the state data is captured 228, the checkpointing engine resumes 232 the primary virtual machine which again executes 210′ the application.


The captured state data is then transferred 236 to the secondary virtual machine. Once the data has been transferred, the checkpointing engine waits for an acknowledgement (ACK) to be received 240 from the secondary virtual machine, indicating that the data has been received. At this point, the checkpoint cycle is committed and the buffered network data released 244 to the network. At this point the checkpointing engine enters an idle state 220′ until the next checkpoint 224′ occurs.


When certain triggering events occur, which are not deferrable, a checkpoint is declared. Some non-limiting examples of checkpoint triggering events are as follows: a timer initiating a checkpoint period, a network event such as an ACK reply being required from the primary virtual machine to another computer on the network, and a write operation to a memory page occurring. Even though a checkpoint is declared, until the virtual machine is paused by the checkpointing engine, additional checkpoint generating events may occur. Thus, many checkpoints may be queued with each requiring action by the checkpointing engine. For example, each subsequent checkpoint will, at a minimum, require a pausing and restarting of the virtual machine, contributing to the time overhead of the system in which the primary virtual machine is not doing useful work.


When a deferrable frame or other deferrable data such as an ACK is detected, the system causes a checkpoint to be time-scheduled at some near point in the future, rather than being generated immediately. In some embodiments, subsequent deferrable frames can modify the timer, as appropriate, to extend, shorten, or cancel the delay to the scheduled checkpoint cycle. The delay timer is automatically cancelled when non-deferrable data is detected, or another triggering event for a checkpoint occurs (which overrides the delay of a checkpoint), thereby reverting to normal checkpoint initiation policy at that instant.


The amount of deferrable frame delay can be fixed or calculated based on other inputs, but should be sufficiently large to allow the system to aggregate other deferrable data frames likely to appear within the delay. However, the delay should also be sufficiently short to avoid problems in the event that a frame deemed as deferrable is, for some reason, time-sensitive to the receiving computer. In one embodiment, a delay of from about 5 to about 10 milliseconds is suitable.


Referring to FIG. 3(a), the timing of the virtual machine is shown for a checkpoint system according to the present invention. In this embodiment, the primary virtual machine is either running (executing one or more applications) 310, 310′ or paused 314. The checkpointing engine operates under the timing conditions as shown in FIG. 3(b). A number of checkpointing engine timing transitions are the same as shown in FIG. 2(b). For example, the checkpointing engine may be idle 320 until a checkpoint (P) occurs 324, at which time the checkpointing engine causes 326 the primary virtual machine to pause 314. While the primary virtual machine is paused 314, the state of the primary virtual machine is captured 328 by the checkpoint engine. Once the state data is captured 328, the checkpointing engine resumes 332 the primary virtual machine which again executes 310′ one or more applications.


The captured state data is then transferred 336 to the secondary virtual machine. Once the data has been transferred, the checkpointing engine waits for an acknowledgement (ACK) 340 to be received from the secondary virtual machine, indicating that the data has been received. At this point, the checkpoint data is committed 344, for example by releasing one or more frames of data to the network.


The checkpointing engine then enters an idle state 348 but in this embodiment, the idle state includes an additional delay period if the events generating a subsequent checkpoint are deferrable. If the events are deferrable, a delay is added to the idle period before the next checkpoint 324′ is declared. This increased delay permits additional checkpoint events to be queued prior to a checkpoint being declared, thus allowing multiple checkpoint events to be covered by one set of checkpoint operations. This reduces the overhead per event for handling a checkpoint.


The method for determining the existence of a deferrable frame should be applied easily and quickly, without the need to track the history of a connection/session. In this sense, the test for a deferrable frame is stateless with regard to the ongoing dialogue between the virtual machine and the other computer on the network. Some exemplary characteristics of implementing a checkpoint delay and deferrable data-based method are: acting on the most commonly used transport protocols (but the method can also be adapted to custom protocols); not causing frames to be dropped or re-ordered as a means of improving checkpoint performance; and controlling selectivity by applying different rules to various deferrable frame types.


Additionally, the checkpoint delay can be applied not only according to whether the frame itself is deferrable, but also whether the frame is directed to or originating from specific network (MAC, IP, etc.) addresses and/or port numbers, some of which are typically used by convention for certain applications. Thus, the system can be set to defer any packet (or prevent the deferral and checkpoint delay of any packet) originating from/destined to a specific network address and/or port number (e.g. port 89) based on convention or rules provided to the checkpointing engine. Various special handling of data which is important to a networked computer can be used to override deferrable checkpoints and checkpoint delays or to more aggressively cause delays for selected network addresses and/or ports.


In another embodiment, the virtual machine has multiple network interfaces. The network traffic on a specific network interface may be of a nature such that all of its packets are deferrable. It is also possible that a specific network interface only carries non-deferrable packets.


An example of special case handling in the checkpoint context is TCP ACKs known as ‘window updates’. These ACKs provide specific information about the sender's TCP send window, and delaying these may occasionally cause the sender (the client) to pause its transmission. Accurately detecting when ‘window update’ information is important to the sender is difficult and adds little improvement to the general approach of deferrable frame management, especially when the send window size reaches a full 64 KB (typical) or larger (window scaling) size.


However, such send-window stalls may still occur, which is an important reason for keeping the deferrable delay from getting too large. Thus, it is important to limit the amount of delay that can be generated so as not to allow the delay to affect other functions of the virtual machine. It is also acceptable for partial or full session-state tracking to be used to selectively avoid delaying a “window update” packet


Unless specifically stated otherwise as apparent from the following discussion, it is appreciated that throughout the description, discussions utilizing terms such as “processing” or “computing” or “calculating” or “delaying” or “comparing”, “generating” or “determining” or “deferring” or “committing” or “checkpointing” or “interrupting” or “handling” or “receiving” or “buffering” or “allocating” or “displaying” or “flagging” or Boolean logic or other set related operations or the like, refer to the action and processes of a computer system, or electronic device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system's or electronic devices' registers and memories into other data similarly represented as physical quantities within electronic memories or registers or other such information storage, transmission or display devices.


The algorithms presented herein are not inherently related to any particular computer or other apparatus. Various general purpose systems may be used with programs in accordance with the teachings herein, or it may prove convenient to construct a more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will be understood from the description above. In addition, the present invention is not described with reference to any particular programming language, and various embodiments may thus be implemented using a variety of programming languages.


The aspects, embodiments, features, and examples of the invention are to be considered illustrative in all respects and are not intended to limit the invention, the scope of which is defined only by the claims. Other embodiments, modifications, and usages will be apparent to those skilled in the art without departing from the spirit and scope of the claimed invention.


In the application, where an element or component is said to be included in and/or selected from a list of recited elements or components, it should be understood that the element or component can be any one of the recited elements or components and can be selected from a group consisting of two or more of the recited elements or components. Further, it should be understood that elements and/or features of a composition, an apparatus, or a method described herein can be combined in a variety of ways without departing from the spirit and scope of the present teachings, whether explicit or implicit herein.


The use of the terms “include,” “includes,” “including,” “have,” “has,” or “having” should be generally understood as open-ended and non-limiting unless specifically stated otherwise.


It should be understood that the order of steps or order for performing certain actions is immaterial so long as the present teachings remain operable. Moreover, two or more steps or actions may be conducted simultaneously.


It is to be understood that the figures and descriptions of the invention have been simplified to illustrate elements that are relevant for a clear understanding of the invention, while eliminating, for purposes of clarity, other elements. Those of ordinary skill in the art will recognize, however, that these and other elements may be desirable. However, because such elements are well known in the art, and because they do not facilitate a better understanding of the invention, a discussion of such elements is not provided herein. It should be appreciated that the figures are presented for illustrative purposes and not as construction drawings. Omitted details and modifications or alternative embodiments are within the purview of persons of ordinary skill in the art.


The invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. The foregoing embodiments are therefore to be considered in all respects illustrative rather than limiting on the invention described herein. Scope of the invention is thus indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are intended to be embraced therein.

Claims
  • 1. A method of delaying an initiation of a checkpoint in a virtual machine system comprising the steps of: establishing an input/output module;buffering a deferrable or non-deferrable frame in the input/output module;determining whether the frame is a deferrable frame or a non-deferrable frame;if the frame is a deferrable frame, delaying an initiation of a checkpoint; andnot transmitting the frame until the checkpoint occurs.
  • 2. The method of claim 1 wherein the frame is examined to determine if it is a deferrable frame.
  • 3. The method of claim 2 wherein the deferrable frame is a packet tagged with one or more of: not being a last data fragment of a multi-fragment message;a data segment that does not indicate a flush of a send buffer; anda segment that only serves as an acknowledgment of some prior message.
  • 4. The method of claim 2 wherein the deferrable frame is an internet protocol (IP) packet tagged with one or more of: a ‘more fragments’ attribute;a transmission control protocol (TCP) data segment that lacks a push (PSH) flag and that carries no flags other than an acknowledgement (ACK); anda transmission control protocol (TCP) segment that contains no data and carries only the acknowledgement (ACK) flag.
  • 5. The method of claim 2 wherein the deferrable frame is identified by a protocol in use.
  • 6. The method of claim 5 wherein the deferrable frame is identified by a session-state of the protocol in use.
  • 7. The method of claim 2 wherein the deferrable frame is identified by port number.
  • 8. The method of claim 7 wherein the port number is a source port number.
  • 9. The method of claim 7 wherein the port number is a destination port number.
  • 10. The method of claim 2 wherein the deferrable frame is identified by network address.
  • 11. The method of claim 10 wherein the network address is a source network address.
  • 12. The method of claim 10 wherein the network address is a destination network address.
  • 13. The method of claim 1 further comprising the step of concatenating the checkpoint delays due to a plurality of deferrable frames.
  • 14. The method of claim 13 further comprising the step of setting an upper limit to the amount of delay that can be generated.
  • 15. The method of claim 1 wherein the frame is determined to be deferrable based on a network interface in use.
  • 16. A fault tolerant computing system comprising: a first computing device comprising a primary virtual machine; anda second computing device comprising a secondary virtual machine,wherein the first computing device and the second computing device are networked,wherein the primary virtual machine comprises a first checkpointing engine and a first network interface having a first input/output module for buffering frames,wherein the secondary virtual machine comprises a second checkpointing engine and a second network interface,wherein the primary virtual machine buffers a frame in the first input/output module,wherein the primary virtual machine determines whether the frame is a deferrable frame or a non-deferrable frame,wherein the first checkpointing engine delays an initiation of a checkpoint for each of the one or more deferrable frames of data in the first input/output module, andwherein the deferrable and non-deferrable frames in the first input/output module are not transmitted until the checkpoint occurs.
  • 17. The fault tolerant computing system of claim 16 wherein the first checkpoint engine pauses the primary virtual machine when a checkpoint is declared.
  • 18. The fault tolerant computing system of claim 17 wherein the first checkpoint engine adds an additional delay before declaring a subsequent checkpoint.
  • 19. The fault tolerant computing system 16 wherein the first checkpoint engine captures the state of the primary virtual machine while the primary virtual machine is paused.
RELATED APPLICATIONS

This application claims priority to U.S. provisional patent application 61/921,720 filed on Dec. 30, 2013 and owned by the assignee of the current application, the contents of which are herein incorporated by reference in their entirety.

US Referenced Citations (146)
Number Name Date Kind
3736566 Anderson et al. May 1973 A
4590554 Glazer et al. May 1986 A
4751702 Beier et al. Jun 1988 A
4758951 Sznyter, III Jul 1988 A
4831581 Rubinfeld May 1989 A
5099485 Bruckert et al. Mar 1992 A
5155809 Baker et al. Oct 1992 A
5157663 Major et al. Oct 1992 A
5193162 Bordsen et al. Mar 1993 A
5235700 Alaiwan et al. Aug 1993 A
5333265 Orimo et al. Jul 1994 A
5357612 Alaiwan Oct 1994 A
5404361 Casorso et al. Apr 1995 A
5465328 Dievendorff et al. Nov 1995 A
5568380 Brodnax et al. Oct 1996 A
5615403 Bissett et al. Mar 1997 A
5621885 Del Vigna, Jr. Apr 1997 A
5694541 Service et al. Dec 1997 A
5721918 Nilsson et al. Feb 1998 A
5724581 Kozakura Mar 1998 A
5745905 Larsson et al. Apr 1998 A
5787485 Fitzgerald et al. Jul 1998 A
5790397 Bissett et al. Aug 1998 A
5802265 Bressoud et al. Sep 1998 A
5892928 Wallach et al. Apr 1999 A
5896523 Bissett et al. Apr 1999 A
5913021 Masubuchi Jun 1999 A
5918229 Davis et al. Jun 1999 A
5923832 Shirakihara et al. Jul 1999 A
5933838 Lomet Aug 1999 A
5948112 Shimada et al. Sep 1999 A
5958070 Stiffler Sep 1999 A
5968185 Bressoud et al. Oct 1999 A
6023772 Fleming Feb 2000 A
6035415 Fleming Mar 2000 A
6067550 Lomet May 2000 A
6088773 Kano et al. Jul 2000 A
6098137 Goodrum et al. Aug 2000 A
6105148 Chung et al. Aug 2000 A
6141769 Petivan et al. Oct 2000 A
6158019 Squibb Dec 2000 A
6289474 Beckerle Sep 2001 B1
6301677 Squibb Oct 2001 B1
6397242 Devine et al. May 2002 B1
6401216 Meth et al. Jun 2002 B1
6438705 Chao et al. Aug 2002 B1
6453343 Housel, III et al. Sep 2002 B1
6526447 Giammaria Feb 2003 B1
6622263 Stiffler et al. Sep 2003 B1
6678369 DeMent Jan 2004 B2
6687849 Cherf Feb 2004 B1
6718538 Mathiske Apr 2004 B1
6823474 Kampe et al. Nov 2004 B2
6907477 Altman et al. Jun 2005 B2
6954877 Earl et al. Oct 2005 B2
6978400 Hickson et al. Dec 2005 B2
7039663 Federwisch et al. May 2006 B1
7055063 Leymann et al. May 2006 B2
7058846 Kelkar et al. Jun 2006 B1
7076555 Orman et al. Jul 2006 B1
7080375 Martin Jul 2006 B2
7162698 Huntington et al. Jan 2007 B2
7206964 Moser et al. Apr 2007 B2
7240172 Chong, Jr. Jul 2007 B2
7249118 Sandler et al. Jul 2007 B2
7289433 Chmara et al. Oct 2007 B1
7299337 Traut et al. Nov 2007 B2
7310721 Cohen Dec 2007 B2
7363538 Kundu et al. Apr 2008 B1
7380039 Miloushev et al. May 2008 B2
7392433 Sahoo et al. Jun 2008 B2
7545752 Krautkremer et al. Jun 2009 B2
7577934 Anonsen et al. Aug 2009 B2
7603391 Federwisch et al. Oct 2009 B1
7743381 Tran Jun 2010 B1
7934035 Miloushev et al. Apr 2011 B2
8117496 Bashir et al. Feb 2012 B2
8145945 Lee Mar 2012 B2
8533382 Scales Sep 2013 B2
8769533 North Jul 2014 B2
8806266 Qu et al. Aug 2014 B1
8812907 Bissett et al. Aug 2014 B1
9323553 Larson Apr 2016 B2
9459895 Venkitachalam et al. Oct 2016 B2
20020073249 French et al. Jun 2002 A1
20020073276 Howard et al. Jun 2002 A1
20020078308 Altman et al. Jun 2002 A1
20020147890 Saulsbury et al. Oct 2002 A1
20030005102 Russell Jan 2003 A1
20030005356 Franckowiak et al. Jan 2003 A1
20030084376 Nash et al. May 2003 A1
20030163763 DeLano Aug 2003 A1
20040001435 Wong Jan 2004 A1
20040010663 Prabhu Jan 2004 A1
20040143776 Cox Jul 2004 A1
20040158549 Matena et al. Aug 2004 A1
20040193658 Kawamura et al. Sep 2004 A1
20040193945 Eguchi et al. Sep 2004 A1
20040199812 Earl et al. Oct 2004 A1
20040267897 Hill et al. Dec 2004 A1
20050015702 Shier et al. Jan 2005 A1
20050201373 Shimazu et al. Sep 2005 A1
20050229039 Anderson et al. Oct 2005 A1
20050251785 Vertes et al. Nov 2005 A1
20050256826 Hambrick et al. Nov 2005 A1
20060047925 Perry Mar 2006 A1
20060062142 Appanna et al. Mar 2006 A1
20060112297 Davidson May 2006 A1
20060179147 Tran et al. Aug 2006 A1
20070027985 Ramany et al. Feb 2007 A1
20070174484 Lussier et al. Jul 2007 A1
20090113109 Nelson et al. Apr 2009 A1
20090193298 Mukherjee Jul 2009 A1
20100077164 Stiffler et al. Mar 2010 A1
20100095074 Ganesh et al. Apr 2010 A1
20100107158 Chen et al. Apr 2010 A1
20100235591 Akutsu Sep 2010 A1
20100325500 Bashir et al. Dec 2010 A1
20110161639 Knauth et al. Jun 2011 A1
20110167194 Scales et al. Jul 2011 A1
20110167195 Scales et al. Jul 2011 A1
20110167196 Scales et al. Jul 2011 A1
20110167298 Lee Jul 2011 A1
20110258625 Waldspurger Oct 2011 A1
20110289345 Agesen et al. Nov 2011 A1
20110289417 Schaefer et al. Nov 2011 A1
20120011401 Ranganathan et al. Jan 2012 A1
20120011508 Ahmad Jan 2012 A1
20120084520 Chou et al. Apr 2012 A1
20120266018 Tanaka Oct 2012 A1
20120284721 Antani et al. Nov 2012 A1
20130024855 North Jan 2013 A1
20130198139 Skidanov Aug 2013 A1
20130212205 Flockhart et al. Aug 2013 A1
20130311992 Fuente et al. Nov 2013 A1
20140040206 Ramakrishnan et al. Feb 2014 A1
20140201574 Manchek et al. Jul 2014 A1
20140215461 Laor Jul 2014 A1
20150007172 Hudzia et al. Jan 2015 A1
20150009804 Koponen et al. Jan 2015 A1
20150082087 Jacobs et al. Mar 2015 A1
20150095907 Jacobs et al. Apr 2015 A1
20150127809 Akella et al. May 2015 A1
20150149999 Ramanathan et al. May 2015 A1
20150180761 Watanabe Jun 2015 A1
20150205672 Bissett et al. Jul 2015 A1
Foreign Referenced Citations (7)
Number Date Country
0 750 261 Dec 1996 EP
0 974 912 Jan 2000 EP
2 268 817 Jan 1994 GB
10-320326 Dec 1998 JP
2004-206212 Jul 2004 JP
9217842 Oct 1992 WO
9926133 May 1999 WO
Non-Patent Literature Citations (16)
Entry
PCT International Search Report and the Written Opinion of the International Searching Authority for International Application No. PCT/US2014/070453, mailed Aug. 21, 2015, 24 pgs.
International Search Report and the Written Opinion of the International Searching Authority for International Application No. PCT/US2014/070456, mailed Jun. 22, 2015, 16 pages.
International Preliminary Report on Patentability from PCT/US2014/070461 dated Jul. 5, 2016.
International Preliminary Report on Patentability from PCT/US2014/070453 dated Jul. 5, 2016.
International Preliminary Report on Patentability from PCT/US2014/070456 dated Jul. 5, 2016.
Clark et al., “Live Migration of Virtual Machines,” NSDI '05 2nd Symposium on Networked Systems Design & Implementation, USENIX Association, May 4, 2005, pp. 273-286.
Cully et al., “Remus: High Availability via Asynchronous Virtual Machine Replication,” The University of British Columbia, Department of Computer Science, 14 pages.
Kim et al., “Process Allocation for Load Distribution in Fault-Tolerant Multicomputers,” IEEE Publication, pp. 174-183, 1995.
Nakamikawa et al., “High Performance Fault Tolerant Computer and its Fault Recovery,” 1997 IEEE, pp. 2-6.
Tamura, “Modernization of Kemari using HVM with PV Drivers,” Nippon Telegraph and Telephone Corporation, NTT Cyber Space Labs, Nov. 20, 2008, 23 pages.
Tamura et al., “Kemari: Virtual Machine Synchronization for Fault Tolerance,” NTT Cyber Space Labs, 2 pages.
Tamura, “Kemari: Virtual Machine Synchronization for Fault Tolerance using DomT,” Nippon Telegraph and Telephone Corporation, NTT Cyber Space Labs, Jun. 24, 2008, 19 pages.
Tamura et al., “Virtual Machine Synchronization for High Availability Clusters,” Nippon Telegraph and Telephone Corporation, NTT Cyber Space Labs, Apr. 17, 2007, 17 pages.
PCT International Search Report for International Application No. PCT/US06/19754, dated Oct. 17, 2007, 3 pages.
Supplementary European Search Report for Application No. EP 06770856, completed Nov. 10, 2010, 6 pages.
PCT International Search Report and PCT Written Opinion of the International Searching Authority for International Application No. PCT/US2014/070461, mailed Apr. 2, 2015, 9 pages.
Related Publications (1)
Number Date Country
20150205672 A1 Jul 2015 US
Provisional Applications (1)
Number Date Country
61921720 Dec 2013 US