The present invention relates generally to data storage within a data center. More specifically, the present invention relates to reserving a virtual disk within a data center using multiple virtual machines.
In the field of data storage, enterprises have used a variety of techniques in order to store the data that their software applications use. At one point in time, each individual computer server within an enterprise running a particular software application (such as a database or e-mail application) would store data from that application in any number of attached local disks. Although this technique was relatively straightforward, it led to storage manageability problems in that the data was stored in many different places throughout the enterprise.
These problems led to the introduction of the storage area network in which each computer server within an enterprise communicated with a central storage computer node that included all of the storage disks. The application data that used to be stored locally at each computer server was now stored centrally on the central storage node via a fiber channel switch, for example. Although such a storage area network was easier to manage, changes in computer server architecture created new problems.
With the advent of virtualization, each computer server can now host dozens of software applications through the use of a hypervisor on each computer server and the use of virtual machines. Thus, computer servers which had been underutilized could now host many different server applications, each application needing to store its data within the storage area network. Weaknesses in the storage area network were revealed by the sheer number of server applications needing to access disks within the central storage node.
For example, the same application executing on different hypervisors (whether on different computers or not) is required to use the same target virtual machine (under the iSCSI protocol) when accessing a particular virtual disk. This causes latency issues and may cause a computer or virtual machine to crash. Improvements are needed in order to provide greater efficiency and to ensure continual access to storage data in the cloud by remote applications.
To achieve the foregoing, and in accordance with the purpose of the present invention, methods and a system are disclosed that provide the advantages discussed below. For one, any number of virtual machines operating within different hypervisors (and perhaps on different computers) are allowed to access the same virtual disk within a storage platform using a virtual machine within their own hypervisor. It is not necessary for various virtual machines on different hypervisors to use the same virtual machine within a single hypervisor to access the virtual disk. Each virtual machine acts as an iSCSI Initiator and it treats a controller virtual machine within its own hypervisor as an iSCSI Target. Moreover, the registration list and current reservation holder information for a particular virtual disk are stored within the storage platform itself rather than within a virtual machine of the remote computer that hosts an application. And, a coordinator module within the storage platform implements a lock mechanism in order to coordinate the requests from the various virtual machines to reserve a virtual disk, release a reservation for a virtual disk, or preempt a reservation for virtual disk.
In a first embodiment, a method reserves a virtual disk in a storage platform having a number of storage nodes. The method begins by receiving a request at a first process on one of the storage nodes to reserve a specific virtual disk in said storage platform. The request originates at an iSCSI Initiator virtual machine of a remote computer and comes via an iSCSI Target of said remote computer. The first process requests of a coordinator process on one of said storage nodes a lock for the specific virtual disk. The coordinator process gives the lock to the first process. Thus, an application of the virtual machine is able to write to, or read from, the virtual disk.
In a second embodiment, a method releases a reservation for a virtual disk in a storage platform having a number of storage nodes. The method begins by receiving a request at a first process on one of the storage nodes to release a reservation for a specific virtual disk in the storage platform. The request originates at an iSCSI Initiator virtual machine of a remote computer and comes via an iSCSI Target of the remote computer. The first process requests of a coordinator process on one of the storage nodes a lock for the specific virtual disk. The coordinator process gives it the lock to the first process, and then the first process resets a current reservation holder field in persistent storage on one of the storage nodes that pertains to the specific virtual disk.
In a third embodiment, a method of preempting a reservation for a virtual disk in a storage platform having a number of storage nodes begins by receiving a request at a first process on one of the storage nodes to preempt a reservation for a specific virtual disk in the storage platform. The request originates at an iSCSI Initiator virtual machine of a remote computer and comes via an iSCSI Target of the remote computer. The first process requests of a coordinator process on one of said storage nodes a lock for the specific virtual disk. The coordinator process giving the lock to the first process, and then the first process sets a current reservation holder field in persistent storage on one of the storage nodes that pertains to the specific virtual disk. The current reservation holder field is set to identify the iSCSI Initiator virtual machine.
The invention, together with further advantages thereof, may best be understood by reference to the following description taken in conjunction with the accompanying drawings in which:
Computers nodes 30-40 are shown logically being grouped together, although they may be spread across data centers and may be in different geographic locations. A management console 40 used for provisioning virtual disks within the storage platform communicates with the platform over a link 44. Any number of remotely located computer servers 50-52 each typically executes a hypervisor in order to host any number of virtual machines. Server computers 50-52 form what is typically referred to as a compute farm.
As shown, these virtual machines may be implementing any of a variety of applications such as a database server, an e-mail server, etc., including applications from companies such as Oracle, Microsoft, etc. These applications write to and read data from the storage platform using a suitable storage protocol such as iSCSI or NFS, although each application will not be aware that data is being transferred over link 54 using a generic protocol. In some situations, the same application may be running on different server computers and this application may need to access the same virtual disk within storage platform 20.
Management console 40 is any suitable computer able to communicate over an Internet connection or link 44 with storage platform 20. When an administrator wishes to manage the storage platform (e.g., provisioning a virtual disk, snapshots, revert, clone, analyze metrics, determine health of cluster, etc.) he or she uses the management console to access the storage platform and is put in communication with a management console routine executing as part of metadata module 130 (shown in
In order to provision a new virtual disk within storage platform 20 for a particular application running on a virtual machine, the virtual disk is first created and then attached to a particular virtual machine. In order to create a virtual disk, a user uses the management console to first select the size of the virtual disk (e.g., 100 GB), and then selects the individual policies that will apply to that virtual disk. For example, the user selects a replication factor, a data center aware policy and other policies concerning whether or not to compress the data, the type of disk storage, etc. Once the virtual disk has been created, it is then attached to a particular virtual machine within one of the computer servers 50-52 and the provisioning process is complete.
Advantageously, storage platform 20 is able to simulate prior art central storage nodes (such as the VMAX and CLARION products from EMC, VMWARE products, etc.) and the virtual machines and application servers will be unaware that they are communicating with storage platform 20 instead of with a prior art central storage node. In addition, the provisioning process can be completed on the order of minutes or less, rather than in four to eight weeks as was typical with prior art techniques. The advantage is that one only need to add metadata concerning a new virtual disk in order to provision the disk and have the disk ready to perform writes and reads. No allocation of actual physical storage is needed.
Although shown as three modules, each of the modules runs independently on each of the computer nodes within the platform 20. Also, associated with each module on each node is a memory cache 122, 132 and 142 that stores information used by that module; each module on each computer node may also use persistent storage on that node. A file (for example) that is stored on nodes 32, 34 and 36, is referred to as being stored on its “data nodes 32, 34 and 36.” The metadata for that file may be stored on three different nodes, and those nodes for that file are referred to as the “metadata nodes 30, 36 and 38.” The data nodes and metadata nodes for a particular stored file may be the same or may be different. The modules communicate with each other via a modified version of Gossip over TCP, and work in concert to manage the storage platform.
Controller Virtual Machine
As shown, server 51 includes a hypervisor and virtual machines 182 and 186 that execute respectively a database application and an e-mail application, and that desire to perform I/O handling using the iSCSI protocol 187 and 183. Server 51 also includes a specialized controller virtual machine (CVM) 180 that is adapted to handle communications with the virtual machines using the iSCSI protocol (and other protocols), yet communicates with the storage platform using a single proprietary protocol 189. Protocol 189 may be any suitable protocol for passing data between storage platform 20 and a remote computer server 51 such as TCP. In addition, the CVM may also communicate with public cloud storage 190 using the same or different protocol 191. Advantageously, the CVM need not communicate any “liveness” information between itself and the computer nodes of the platform. There is no need for any CVM to track the status of nodes in the cluster. The CVM need only talk to a node in the platform, which is then able to route requests to other nodes and public storage nodes.
CVM 180 handles different protocols by simulating an entity that the protocol would expect. For example, when communicating under the iSCSI protocol, CVM responds to an iSCSI Initiator (such as the database or e-mail application) by behaving as an iSCSI Target. In other words, when the e-mail application of virtual machine 186 performs I/O handling, it is the iSCSI Initiator and the controller virtual machine is the iSCSI Target. In other words, when an application is using the block protocol, the CVM masquerades as the iSCSI Target, traps the iSCSI CDBs, translates this information into its own protocol, and then communicates this information to the storage platform. When the CVM presents itself as an iSCSI Target, the application simply talks to a block device as it would do normally. The application is unaware that the CVM is trapping and intercepting its calls under the iSCSI protocol or that the CVM even exists. One advantage is that an application need not be changed in order to write to and read from the storage platform. Use of the CVM allows an application executing upon a virtual machine to continue using the protocol it expects, yet allows these applications on the various computer servers to write data to and read data from the same storage platform 20.
Each CVM also uses a memory cache of the computer server. Preferably, all information concerning a particular virtual disk attached to a CVM will be organized into a virtual disk object and then stored into the memory cache. A hash table is used to store these virtual disk objects and the key to find each object is the name of the virtual disk. Stored within this cache is virtual disk information 862 and the identifiers of the metadata nodes on which the metadata for this virtual disk is stored. Also in communication with each computer and CVM are also any number of local solid-state disks (or other similar persistent storage). These disks may be used as a data cache to also store data blocks that are written into storage platform 20. This cache may be used to rapidly retrieve data blocks instead of retrieving them from the remote storage platform.
Persistent Reservations
Typically, once a virtual disk is created within the storage platform, the “Add LUN” command is used to add the virtual disk as a LUN within the target, i.e., the controller virtual machine. Next, the “Add Access” command is used to provide the IP address or host name of the particular application, such as the e-mail application of virtual machine 186. Now the controller virtual machine 180 can see the virtual disk and can allow its clients (i.e., applications) to access that disk.
As known in the art, iSCSI reservations are used to control access by a single target to a shared iSCSI device such as a disk or tape drive, including the virtual disks discussed herein. In general, an Initiator makes a reservation of a storage asset such as a disk and no other Initiator may use that asset until the Initiator gives it up. A variety of commands are available under the iSCSI-3 RFC, and it will be useful to review these commands in the context of the present invention. A single target (e.g., controller virtual machine 180) is used to control access to a particular virtual disk by multiple Initiators, and this CVM maintains an in memory data structure that keeps track of a registration list of virtual machines and the current reservation holder.
In order to access a particular virtual disk by an application running within a virtual machine, each Initiator must first issue the “register” command to CVM 180; when done, a unique identifier for the virtual machine (such as its IQN, “iSCSI Qualified Name”) is added to the registration list of the CVM. Next, in order to actually reserve the particular disk for access (i.e., a particular LUN), the virtual machine issues a “reserve” command to the CVM. The CVM check to see if the virtual machine has registered, and if there is not currently an existing reservation then the reservation is given to the Initiator and its IQN is assigned as the current reservation holder. At that point no other Initiator may make changes to that LUN. When done, the current reservation holder issues a “release” command to the CVM in which case the current reservation holder is reset. A “preempt” command may be issued by a virtual machine if it is apparent that something has happened to the current reservation holder (i.e., computer crash, hang up, etc.). If so, this command resets both the registration list and the current reservation holder. The “clear” command may be issued in order to reset both the registration list and the current reservation holder.
It is important to note that a single target is responsible for controlling access to a particular virtual disk by any number of virtual machines. As shown in
Requiring this e-mail application on different hypervisors (and potentially on different computers) to use a single virtual machine in order to access a particular virtual disk can create problems. For one, there may be multiple (or even dozens of) e-mail applications all attempting to access a virtual disk using controller virtual machine 180; this overburdens the CVM and can create greater latency and can even cause the CVM (or its hypervisor or its host computer) to crash, thus preventing access to data. Use of a single CVM to access one particular virtual disk can also create problems for other applications. For example, even though there is only a single database application on virtual machine 182, it must also use CVM 180 because they are within the same hypervisor. If dozens of e-mail applications overburden CVM 180, the performance of the database application will be affected even though it is only a single application.
Another disadvantage of the prior art approach is that should CVM 180 crash, all the data it has held regarding registration lists and current reservation holders for any number virtual disks within the storage platform will be lost. Each Initiator would then be required to again issue the register and reserve commands.
Persistent Reservations Using Multiple Targets
By way of the present invention, each e-mail application may now communicate with its own CVM via links 363, 187 and 373, respectively, in order to write to or to read data from the same virtual disk. Thus, any number of virtual machines may now use multiple targets to access a single virtual disk. This allows any number of different Initiators within different hypervisors to access a single virtual disk without using a single target, such as a single controller virtual machine. Because multiple targets may be used, latency will be improved, a crash is less likely, and recovery is faster because persistent reservation data will be stored within the storage platform and not locally at a CVM.
Such information may be distributed and stored within the nodes of the storage platform in a variety of manners. In one particular embodiment, the registration and reservations information for a particular virtual disk is stored upon three different nodes within the platform (or upon fewer or greater nodes if the replication factor is different). In order to select the first node for storage, a hash function is taken over the virtual disk name to produce a numeral which is mapped to a unique one of the storage nodes. I.e., given the six nodes as shown in
Once the three storage nodes are determined for the requested virtual disk, the metadata module of the first node contacted then stores this registration information onto these three storage nodes using the metadata modules of those nodes as shown in greater detail in
Using this embodiment of the present invention, it is now possible for other virtual machines within system 10 to also make a reservation and perhaps access the same virtual disk “Avinash” by using their own local virtual machine as a Target. Accordingly, in step 516 a second virtual machine 172 registers with its CVM 170 in order to access the same virtual disk “Avinash.” In step 520 CVM 170 passes the name of this disk and the IQN of virtual machine 172 to any node of the storage platform. Next, in step 524, the node that receives this information stores the received registration information on the three particular storage nodes that pertain to the virtual disk named “Avinash” using the hash function and procedure discussed above in step 512. Accordingly, the registration list for this particular virtual disk may appear as shown in 864 where the first and third virtual machines have registered.
Once registered, a virtual machine may issue the reserve, release, clear and preempt commands described below. In order to issue these commands, it is a prerequisite that the virtual machine first have registered for that particular virtual disk. Typically, the CVM corresponding to the virtual machine making a request will first check with the metadata module on the storage node which holds the virtual disk information to make sure that the requesting virtual machine is actually registered for that virtual disk. If so, then the command is passed to the metadata module, if not, then a “Fail” status is returned to the virtual machine.
In a first step 604, an application within a first virtual machine (such as virtual machine 162) desires to access a particular virtual disk within the storage platform and thus issues a reserve command via its CVM 160 and link 363, providing not only the disk name “Avinash,” but also the IQN of the virtual machine. The CVM passes the reserve command and this information to any node within the storage platform. The metadata module 130 on this node receives this registration information. At more or less the same time, or perhaps simultaneously, in step 608 another virtual machine (such as virtual machine 172) within system 10 also desires to reserve the same disk and also issues a reserve command via its CVM 170, providing not only the disk name “Avinash,” but also the IQN of the virtual machine. CVM 170 also passes this reserve command and information to any node within the storage platform. The metadata module 130 on this other node receives this registration information. It may be that both reserve commands are handled by the same storage node within the storage platform or they may be handled by different storage nodes.
At this time, each storage node that has received the reserve command may check field 864 of the storage node that holds information for the particular virtual disk in order to make sure that the requesting virtual machine has already registered. If not, then in error is generated and any virtual disk that has not registered is provided with a “Fail” status via its local CVM that had sent the reserve command. This check may also be done in conjunction with step 620 below, or performed by the CVM as mentioned above.
In step 612 each of the storage nodes that has received the reserve command from the different virtual machines will make a request of the corresponding pod module for the Lock on that virtual disk. In the same way that a particular storage node (or nodes, if replication is used) is responsible for storing information concerned with a particular virtual disk (such as shown in
Accordingly, in step 616 the pod module will give the Lock for a particular virtual disk to the first request that it receives from the metadata module of one of the storage nodes, i.e., first come first served. The second, latter request (or any other later request), will fail and the requesting virtual machine will receive a “Fail” status via its local CVM that had sent the reserve command. Of course, if only a single virtual machine is attempting to reserve the particular virtual disk, then the Lock will always be given to that virtual machine.
Step 620 is a check to determine whether or not the current reservation holder field 866 has already been set. For example, field 866 of
Finally, in step 632 the successful metadata module of the storage node that had locked the virtual disk returns that Lock by again communicating with the pod module responsible for handling locks for that particular virtual disk. Thus, the Lock is not in effect for the entire reservation, but only long enough to ensure that two or more virtual machines do not attempt to reserve the disk at the same time. The pod module will only give the Lock to one virtual machine at a time (or, rather, to a metadata module representing a virtual machine), and, each metadata module may not set a reservation unless it holds the lock.
In a first step 704, an application within a virtual machine (such as virtual machine 162) desires to access a particular virtual disk within the storage platform and reserves the virtual disk as described above with reference to
The pod module checks to make sure that the requesting virtual machine is actually the same virtual machine that is the current reservation holder (by checking field 866), and then gives the Lock to the metadata module that is representing the virtual machine. If the virtual machine requesting the Lock is not the current reservation holder then a “Fail” status is returned as in step 624. Once the metadata module has the Lock, then in step 716 it resets the current reservation holder field 866 for the virtual disk in question (on the node or nodes that hold metadata information for that particular virtual disk) and then returns the Lock to the pod module in step 720. While the metadata module holds the Lock, no other virtual machine is able to change the current reservation holder field by setting it to a particular virtual machine or by resetting it. Once the Lock has been returned, then the metadata module returns a “Success” status to the requesting virtual machine via its CVM.
In step 804 a first virtual machine (such as virtual machine 186) reserves a particular virtual disk within storage platform 20 on behalf of a software application as has been described above in
Next, in step 812 the second virtual machine issues a preempt command providing the name of the virtual disk and its own IQN. The preempt command from the virtual machine is passed to CVM 170, which in turn communicates with the metadata module of any node of the storage platform in order to find the actual storage node which holds the information for that particular virtual disk.
In step 816, this metadata module then requests the Lock for that particular virtual disk from the pod module responsible for that virtual disk (as has been described above) and receives the Lock from that pod module. Of course, if the Lock for that particular virtual disk is currently held by another metadata module (on behalf of another virtual machine) then the request will fail. Once the Lock is received, the metadata module in step 820 resets the current reservation holder field 866 to null, sets the current reservation holder field to be the IQN of the second virtual machine which is performing the preempt command, and then clears the IQN of the first virtual machine from the registration list 864.
Finally, the metadata module then returns the Lock in step 824 to the pod module responsible for that virtual disk. In this fashion, a preempt of the first virtual machine is performed and allows the second virtual machine to hold the reservation for the particular virtual disk.
The clear command is another command that utilizes the Lock mechanism. Similar to use of the reserve, release and preempt commands, a metadata module (representing a particular virtual machine), must first obtain the Lock from the corresponding pod module before being able to execute the clear command for a particular virtual disk. When executed, the clear command clears the registration list 864 and the current reservation holder 866 for the particular virtual disk. When done, the metadata module returns the Lock to the pod module. No other metadata module is permitted to perform commands such as reserve, release or preempt on a particular virtual disk when it does not hold the Lock for that disk.
Storage of Metadata
In particular, information associated with persistent reservations for a particular virtual disk is also stored using the metadata module of a particular storage node. Shown at 864 is the registration list of those virtual machines within system 10 which have registered to access the particular virtual disk. Also, shown at 866 is a particular virtual machine that is the current reservation holder for the particular virtual disk. This information is stored within persistent storage on a storage node A and may also be present in the memory of that storage node.
Computer System Embodiment
CPU 922 is also coupled to a variety of input/output devices such as display 904, keyboard 910, mouse 912 and speakers 930. In general, an input/output device may be any of: video displays, track balls, mice, keyboards, microphones, touch-sensitive displays, transducer card readers, magnetic or paper tape readers, tablets, styluses, voice or handwriting recognizers, biometrics readers, or other computers. CPU 922 optionally may be coupled to another computer or telecommunications network using network interface 940. With such a network interface, it is contemplated that the CPU might receive information from the network, or might output information to the network in the course of performing the above-described method steps. Furthermore, method embodiments of the present invention may execute solely upon CPU 922 or may execute over a network such as the Internet in conjunction with a remote CPU that shares a portion of the processing.
In addition, embodiments of the present invention further relate to computer storage products with a computer-readable medium that have computer code thereon for performing various computer-implemented operations. The media and computer code may be those specially designed and constructed for the purposes of the present invention, or they may be of the kind well known and available to those having skill in the computer software arts. Examples of computer-readable media include, but are not limited to: magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROMs and holographic devices; magneto-optical media such as floptical disks; and hardware devices that are specially configured to store and execute program code, such as application-specific integrated circuits (ASICs), programmable logic devices (PLDs) and ROM and RAM devices. Examples of computer code include machine code, such as produced by a compiler, and files containing higher-level code that are executed by a computer using an interpreter.
Although the foregoing invention has been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications may be practiced within the scope of the appended claims. Therefore, the described embodiments should be taken as illustrative and not restrictive, and the invention should not be limited to the details given herein but should be defined by the following claims and their full scope of equivalents.
This application is a continuation of U.S. patent application Ser. No. 16/274,014, filed Feb. 12, 2019, which is a continuation of U.S. patent application Ser. No. 15/163,446, filed May 24, 2016, which are hereby incorporated by reference. This application is related to U.S. patent application Ser. Nos. 14/322,813, 14/322,832, 14/684,086, 14/322,850, 14/322,855, 14/322,867, 14/322,868, 14/322871, and 14/723,380, which are all hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
4084231 | Capozzi et al. | Apr 1978 | A |
4267568 | Dechant et al. | May 1981 | A |
4283787 | Chambers | Aug 1981 | A |
4417321 | Chang et al. | Nov 1983 | A |
4641274 | Swank | Feb 1987 | A |
4654819 | Stiffler et al. | Mar 1987 | A |
4686620 | Ng | Aug 1987 | A |
4912637 | Sheedy et al. | Mar 1990 | A |
4995035 | Cole et al. | Feb 1991 | A |
5005122 | Griffin et al. | Apr 1991 | A |
5093912 | Dong et al. | Mar 1992 | A |
5133065 | Cheffetz et al. | Jul 1992 | A |
5193154 | Kitajima et al. | Mar 1993 | A |
5212772 | Masters | May 1993 | A |
5226157 | Nakano et al. | Jul 1993 | A |
5239647 | Anglin et al. | Aug 1993 | A |
5241668 | Eastridge et al. | Aug 1993 | A |
5241670 | Eastridge et al. | Aug 1993 | A |
5276860 | Fortier et al. | Jan 1994 | A |
5276867 | Kenley et al. | Jan 1994 | A |
5287500 | Stoppani, Jr. | Feb 1994 | A |
5301286 | Rajani | Apr 1994 | A |
5321816 | Rogan et al. | Jun 1994 | A |
5347653 | Flynn et al. | Sep 1994 | A |
5410700 | Fecteau et al. | Apr 1995 | A |
5420996 | Aoyagi | May 1995 | A |
5454099 | Myers et al. | Sep 1995 | A |
5559991 | Kanfi | Sep 1996 | A |
5642496 | Kanfi | Jun 1997 | A |
6418478 | Ignatius et al. | Jul 2002 | B1 |
6473774 | Cellis et al. | Oct 2002 | B1 |
6542972 | Ignatius et al. | Apr 2003 | B2 |
6658436 | Insky et al. | Dec 2003 | B2 |
6721767 | De Meno et al. | Apr 2004 | B2 |
6760723 | Insky et al. | Jul 2004 | B2 |
7003641 | Prahlad et al. | Feb 2006 | B2 |
7035880 | Crescenti et al. | Apr 2006 | B1 |
7076509 | Chen et al. | Jul 2006 | B1 |
7107298 | Prahlad et al. | Sep 2006 | B2 |
7130970 | Devassy et al. | Oct 2006 | B2 |
7162496 | Amarendran et al. | Jan 2007 | B2 |
7174433 | Kottomtharayil et al. | Feb 2007 | B2 |
7246207 | Kottomtharayil et al. | Jul 2007 | B2 |
7315923 | Retnamma et al. | Jan 2008 | B2 |
7343453 | Prahlad et al. | Mar 2008 | B2 |
7373451 | Lam et al. | May 2008 | B2 |
7389311 | Crescenti et al. | Jun 2008 | B1 |
7395282 | Crescenti et al. | Jul 2008 | B1 |
7440982 | Lu et al. | Oct 2008 | B2 |
7454569 | Kavuri et al. | Nov 2008 | B2 |
7490207 | Amarendran et al. | Feb 2009 | B2 |
7500053 | Kavuri et al. | Mar 2009 | B1 |
7529782 | Prahlad et al. | May 2009 | B2 |
7536291 | Vijayan Retnamma et al. | May 2009 | B1 |
7543125 | Gokhale | Jun 2009 | B2 |
7546324 | Prahlad et al. | Jun 2009 | B2 |
7603386 | Amarendran et al. | Oct 2009 | B2 |
7606844 | Kottomtharayil | Oct 2009 | B2 |
7613752 | Prahlad et al. | Nov 2009 | B2 |
7617253 | Prahlad et al. | Nov 2009 | B2 |
7617262 | Prahlad et al. | Nov 2009 | B2 |
7620710 | Kottomtharayil et al. | Nov 2009 | B2 |
7636743 | Erofeev | Dec 2009 | B2 |
7651593 | Prahlad et al. | Jan 2010 | B2 |
7653669 | Kapoor et al. | Jan 2010 | B2 |
7657550 | Prahlad et al. | Feb 2010 | B2 |
7660807 | Prahlad et al. | Feb 2010 | B2 |
7661028 | Erofeev | Feb 2010 | B2 |
7734669 | Kottomtharayil et al. | Jun 2010 | B2 |
7747579 | Prahlad et al. | Jun 2010 | B2 |
7801864 | Prahlad et al. | Sep 2010 | B2 |
7809914 | Kottomtharayil et al. | Oct 2010 | B2 |
7925624 | Vosshall et al. | Apr 2011 | B2 |
8001322 | Lecomte et al. | Aug 2011 | B2 |
8156086 | Lu et al. | Apr 2012 | B2 |
8170995 | Prahlad et al. | May 2012 | B2 |
8229954 | Kottomtharayil et al. | Jul 2012 | B2 |
8260751 | Johnson et al. | Sep 2012 | B2 |
8285681 | Prahlad et al. | Oct 2012 | B2 |
8307177 | Prahlad et al. | Nov 2012 | B2 |
8312046 | Eisler et al. | Nov 2012 | B1 |
8364639 | Koryakina et al. | Jan 2013 | B1 |
8364652 | Vijayan et al. | Jan 2013 | B2 |
8370542 | Lu et al. | Feb 2013 | B2 |
8386838 | Byan | Feb 2013 | B1 |
8413142 | Sharp et al. | Apr 2013 | B2 |
8453145 | Naik | May 2013 | B1 |
8458422 | Holdman et al. | Jun 2013 | B1 |
8578120 | Attarde et al. | Nov 2013 | B2 |
8601473 | Aron et al. | Dec 2013 | B1 |
8706867 | Vijayan | Apr 2014 | B2 |
8707070 | Muller | Apr 2014 | B2 |
8762430 | Hall | Jun 2014 | B1 |
8850130 | Aron et al. | Sep 2014 | B1 |
8861527 | Bao et al. | Oct 2014 | B1 |
8892549 | Thakur | Nov 2014 | B1 |
8950009 | Vijayan et al. | Feb 2015 | B2 |
8954446 | Vijayan Retnamma et al. | Feb 2015 | B2 |
9009106 | Aron et al. | Apr 2015 | B1 |
9020900 | Vijayan Retnamma et al. | Apr 2015 | B2 |
9053339 | Pate et al. | Jun 2015 | B2 |
9098495 | Gokhale | Aug 2015 | B2 |
9116633 | Sancheti et al. | Aug 2015 | B2 |
9189170 | Kripalani et al. | Nov 2015 | B2 |
9239687 | Vijayan et al. | Jan 2016 | B2 |
9262496 | Kumarasamy et al. | Feb 2016 | B2 |
9286110 | Mitkar et al. | Mar 2016 | B2 |
9311121 | Deshpande et al. | Apr 2016 | B2 |
9378035 | Kripalani | Jun 2016 | B2 |
9411534 | Lakshman et al. | Aug 2016 | B2 |
9424151 | Lakshman et al. | Aug 2016 | B2 |
9461881 | Kumarasamy et al. | Oct 2016 | B2 |
9483205 | Lakshman | Nov 2016 | B2 |
9495404 | Kumarasamy et al. | Nov 2016 | B2 |
9558085 | Lakshman | Jan 2017 | B2 |
9575673 | Mitkar et al. | Feb 2017 | B2 |
9633033 | Vijayan et al. | Apr 2017 | B2 |
9639274 | Maranna et al. | May 2017 | B2 |
9641388 | Kripalani et al. | May 2017 | B2 |
9652265 | Narayanasamy et al. | May 2017 | B1 |
9710465 | Dornemann et al. | Jul 2017 | B2 |
9798489 | Lakshman et al. | Oct 2017 | B2 |
9875063 | Lakshman | Jan 2018 | B2 |
9977687 | Kottomtharayil et al. | May 2018 | B2 |
9983936 | Dornemann et al. | May 2018 | B2 |
10067722 | Lakshman | Sep 2018 | B2 |
10084873 | Dornemann | Sep 2018 | B2 |
10162528 | Sancheti et al. | Dec 2018 | B2 |
10210048 | Sancheti | Feb 2019 | B2 |
10228962 | Dornemann et al. | Mar 2019 | B2 |
10248174 | Lakshman et al. | Apr 2019 | B2 |
10296368 | Dornemann et al. | May 2019 | B2 |
10346259 | Gokhale et al. | Jul 2019 | B2 |
10691187 | Lakshman et al. | Jun 2020 | B2 |
20020019908 | Reuter et al. | Feb 2002 | A1 |
20020133669 | Devireddy et al. | Sep 2002 | A1 |
20030145167 | Tomita | Jul 2003 | A1 |
20030182494 | Rodrigues et al. | Sep 2003 | A1 |
20040098373 | Bayliss et al. | May 2004 | A1 |
20040250033 | Prahlad et al. | Dec 2004 | A1 |
20050108292 | Burton et al. | May 2005 | A1 |
20050246393 | Coates et al. | Nov 2005 | A1 |
20050246397 | Edwards et al. | Nov 2005 | A1 |
20050256972 | Cochran et al. | Nov 2005 | A1 |
20060168402 | Ahmad et al. | Jul 2006 | A1 |
20060224846 | Amarendran et al. | Oct 2006 | A1 |
20070038888 | Kariv | Feb 2007 | A1 |
20070106706 | Ahrens et al. | May 2007 | A1 |
20080183973 | Aguilera et al. | Jul 2008 | A1 |
20080201535 | Hara | Aug 2008 | A1 |
20080244028 | Le et al. | Oct 2008 | A1 |
20090210875 | Boles et al. | Aug 2009 | A1 |
20090217265 | Ishikawa | Aug 2009 | A1 |
20090282404 | Khandekar et al. | Nov 2009 | A1 |
20090319534 | Gokhale | Dec 2009 | A1 |
20100058013 | Gelson et al. | Mar 2010 | A1 |
20100199126 | Noguchi et al. | Aug 2010 | A1 |
20100223368 | Runcie et al. | Sep 2010 | A1 |
20100257403 | Virk et al. | Oct 2010 | A1 |
20110185355 | Chawla et al. | Jul 2011 | A1 |
20110246984 | Sharp et al. | Oct 2011 | A1 |
20110265083 | Davis | Oct 2011 | A1 |
20120084598 | Alibakhsh et al. | Apr 2012 | A1 |
20120110275 | Ganti et al. | May 2012 | A1 |
20120110328 | Pate et al. | May 2012 | A1 |
20120150818 | Vijayan Retnamma et al. | Jun 2012 | A1 |
20120150826 | Vijayan Retnamma et al. | Jun 2012 | A1 |
20120323853 | Fries et al. | Dec 2012 | A1 |
20130007183 | Sorenson, III et al. | Jan 2013 | A1 |
20130007436 | Bookman et al. | Jan 2013 | A1 |
20130055371 | Kumamo et al. | Feb 2013 | A1 |
20130086018 | Horii | Apr 2013 | A1 |
20130111471 | Chandrasekaran | May 2013 | A1 |
20130117337 | Dunham | May 2013 | A1 |
20130166849 | Beukema et al. | Jun 2013 | A1 |
20130191577 | Thomas et al. | Jul 2013 | A1 |
20130227552 | Reddin et al. | Aug 2013 | A1 |
20130239106 | Srinivasan et al. | Sep 2013 | A1 |
20130262387 | Varadharajan et al. | Oct 2013 | A1 |
20130263114 | Watkins et al. | Oct 2013 | A1 |
20130305002 | Hallak et al. | Nov 2013 | A1 |
20130326053 | Bauer et al. | Dec 2013 | A1 |
20130339407 | Sharpe et al. | Dec 2013 | A1 |
20130346709 | Wang | Dec 2013 | A1 |
20140040206 | Ramakrishnan et al. | Feb 2014 | A1 |
20140059292 | Phelan et al. | Feb 2014 | A1 |
20140059375 | Mcelhoe et al. | Feb 2014 | A1 |
20140156925 | Baron et al. | Jun 2014 | A1 |
20140173181 | Beveridge | Jun 2014 | A1 |
20140181397 | Bonzini | Jun 2014 | A1 |
20140188825 | Muthukkaruppan et al. | Jul 2014 | A1 |
20140196038 | Kottomtharayil et al. | Jul 2014 | A1 |
20140201140 | Vibhor et al. | Jul 2014 | A1 |
20140201157 | Pawar et al. | Jul 2014 | A1 |
20140201170 | Vijayan et al. | Jul 2014 | A1 |
20140282509 | Zheng et al. | Sep 2014 | A1 |
20140317618 | Behera et al. | Oct 2014 | A1 |
20140324793 | Glazemakers | Oct 2014 | A1 |
20140344809 | Jin et al. | Nov 2014 | A1 |
20140365549 | Jenkins | Dec 2014 | A1 |
20140379983 | Sasaki | Dec 2014 | A1 |
20150074536 | Varadharajan et al. | Mar 2015 | A1 |
20150088586 | Pavlas et al. | Mar 2015 | A1 |
20150127658 | Ding et al. | May 2015 | A1 |
20150134616 | Zheng et al. | May 2015 | A1 |
20150227535 | Avati et al. | Aug 2015 | A1 |
20150370495 | Georgiev | Dec 2015 | A1 |
20160132400 | Pawar et al. | May 2016 | A1 |
20160142485 | Mitkar et al. | May 2016 | A1 |
20160202916 | Cui et al. | Jul 2016 | A1 |
20160350302 | Lakshman | Dec 2016 | A1 |
20160350391 | Vijayan et al. | Dec 2016 | A1 |
20170168903 | Dornemann et al. | Jun 2017 | A1 |
20170185488 | Kumarasamy et al. | Jun 2017 | A1 |
20170192866 | Vijayan et al. | Jul 2017 | A1 |
20170193003 | Vijayan et al. | Jul 2017 | A1 |
20170235647 | Kilaru et al. | Aug 2017 | A1 |
20170242871 | Kilaru et al. | Aug 2017 | A1 |
20170262347 | Dornemann et al. | Sep 2017 | A1 |
20180095846 | Sanakkayala et al. | Apr 2018 | A1 |
20180137139 | Bangalore et al. | May 2018 | A1 |
20180143879 | Dornemann | May 2018 | A1 |
20180253361 | Dhatrak | Sep 2018 | A1 |
20180276085 | Mitkar et al. | Sep 2018 | A1 |
20180285202 | Bhagi et al. | Oct 2018 | A1 |
20180285205 | Mehta et al. | Oct 2018 | A1 |
20180373597 | Rana | Dec 2018 | A1 |
20190278662 | Nagrale et al. | Sep 2019 | A1 |
20200241613 | Lakshman et al. | Jul 2020 | A1 |
Number | Date | Country |
---|---|---|
0259912 | Mar 1988 | EP |
0405926 | Jan 1991 | EP |
0467546 | Jan 1992 | EP |
0541281 | May 1993 | EP |
0774715 | May 1997 | EP |
0809184 | Nov 1997 | EP |
0899662 | Mar 1999 | EP |
0981090 | Feb 2000 | EP |
WO 9513580 | May 1995 | WO |
WO 9912098 | Mar 1999 | WO |
WO2001061507 | Feb 2002 | WO |
WO 2006052872 | May 2006 | WO |
Entry |
---|
Office Action dated Feb. 8, 2018, for U.S. Appl. No. 15/155,838. |
Office Action dated Feb. 8, 2018, for U.S. Appl. No. 15/156,015. |
Office Action dated Jul. 3, 2017, for U.S. Appl. No. 15/155,838. |
Office Action dated Jul. 6, 2017, for U.S. Appl. No. 15/156,015. |
Office Action dated Nov. 17, 2018, for U.S. Appl. No. 15/156,015. |
Office Action dated Nov. 19, 2018, for U.S. Appl. No. 15/155,838. |
Office Action dated Oct. 12, 2018, for U.S. Appl. No. 14/723,380. |
Office Action dated Sep. 1, 2017, for U.S. Appl. No. 14/723,380. |
U.S. Appl. No. 16/121,883, filed Sep. 5, 2018, title: Data Migration Using Stubs, Applicant: Commvault Systems, Inc. |
U.S. Appl. No. 16/378,328, filed Apr. 8, 2019, title: Power Management of Components Within a Storage Management System, Applicant: Commvault Systems, Inc. |
U.S. Appl. No. 16/262,721, filed Jan. 30, 2019, title: Cross-Hypervisor Live Mount of Backed Up Virtual Machine Data, Applicant: Commvault Systems, Inc. |
U.S. Appl. No. 16/253,643, filed Jan. 22, 2019, title: File Indexing for Virtual Machine Backups Based on Using Live Browse Features, Applicant: Commvault Systems, Inc. |
U.S. Appl. No. 16/568,075, filed Sep. 11, 2019, title: Data Protection Component Scaling in a Cloud-Based Data Storage System, Applicant: Commvault Systems, Inc. |
Final Office Action dated Aug. 7, 2019 for U.S. Appl. No. 15/155,838. |
Office Action dated Mar. 19, 2020 for U.S. Appl. No. 15/155,838. |
Final Office Action dated Aug. 8, 2019 for U.S. Appl. No. 15/156,015. |
Office Action dated Jan. 9, 2020 for U.S. Appl. No. 15/156,015. |
Arneson, “Mass Storage Archiving in Network Environments” IEEE, Oct. 31-Nov. 1998, pp. 45-50. |
Cabrera, et al. “ADSM: A Multi-Platform, Scalable, Back-up and Archive Mass Storage System,” Digest of Papers, Compcon '95, Proceedings of the 40th IEEE Computer Society International Conference, Mar. 5, 1995-Mar. 9, 1995, pp. 420-427, San Francisco, CA. |
Eitel, “Backup and Storage Management in Distributed Heterogeneous Environments,” IEEE, 1994, pp. 124-126. |
Huff, KL, “Data Set Usage Sequence Number,” IBM Technical Disclosure Bulletin, vol. 24, No. 5, Oct. 1981 New York, US, pp. 2404-2406. |
Rosenblum et al., “The Design and Implementation of a Log-Structure File System,” Operating Systems Review SIGOPS, vol. 25, No. 5, May 1991, New York, US, pp. 1-15. |
Swiftstack, Inc., The OpenStack Object Storage System, Feb. 2012, pp. 1-29. |
International Search Report dated Jan. 13, 2016 from International Application No. PCT/US2015/038687. |
Written Opinion dated Jan. 13, 2016 from International Application No. PCT/US2015/038687. |
Lakshman et al., “Cassandra—A Decentralized Structured Storage System”, https://doi.org/10.1145/1773912.1773922, ACM SIGOPS Operating Systems Review, vol. 44, Issue 2, Apr. 2010, pp. 35-40. |
PTAB-IPR2021-00673—('723) POPR Final, filed Jun. 30, 2021, in 70 pages. |
PTAB-IPR2021-00673—('723) Sur-Reply Final, filed Aug. 16, 2021, in 7 pages. |
PTAB-IPR2021-00673—723 patent IPR—Reply to POPR, filed Aug. 9, 2021, in 6 pages. |
PTAB-IPR2021-00673—Mar. 17, 2021_Petition_723, filed Mar. 17, 2021, in 98 pages. |
PTAB-IPR2021-00673—Exhibit 1001—U.S. Pat. No. 9,740,723, Issue Date Aug. 22, 2017, in 51 pages. |
PTAB-IPR2021-00673—Exhibit 1002—Declaration_Jagadish_EXSRanger, filed Mar. 16, 2021, in 191 pages. |
PTAB-IPR2021-00673—Exhibit 1003—FH 9740723, Issue Date Aug. 22, 2017, in 594 pages. |
PTAB-IPR2021-00673—Exhibit 1004—esxRangerProfessionalUserManual v.3.1, 2006 in 102 pages. |
PTAB-IPR2021-00673—Exhibit 1005—VC_Users_Manual_11_NoRestriction, Copyright date 1998-2004, in 466 pages. |
PTAB-IPR2021-00673—Exhibit 1006—U.S. Pat. No. 8,635,429—Naftel, Issue Date Jan. 21, 2014, in 12 pages. |
PTAB-IPR2021-00673—Exhibit 1007—US20070288536A1—Sen, Issue Date Dec. 13, 2007, in 12 pages. |
PTAB-IPR2021-00673—Exhibit 1008—US20060224846A1—Amarendran, Oct. 5, 2006, in 15 pages. |
PTAB-IPR2021-00673—Exhibit 1009—U.S. Pat. No. 8,209,680—Le, Issue Date Jun. 26, 2012, in 55 pages. |
PTAB-IPR2021-00673—Exhibit 1010—Virtual Machine Monitors Current Technology and Future Trends, May 2005 in 9 pages. |
PTAB-IPR2021-00673—Exhibit 1011—Virtualization Overview, Copyright 2005, VMware, Inc., 11 pages. |
PTAB-IPR2021-00673—Exhibit 1012—Let's Get Virtual A Look at Today's Virtual Server, May 14, 2007 in 42 pages. |
PTAB-IPR2021-00673—Exhibit 1013—U.S. Pat. No. 8,135,930—Mattox, Issue Date Mar. 13, 2012, in 19 pages. |
PTAB-IPR2021-00673—Exhibit 1014—U.S. Pat. No. 8,060,476—Afonso, Issue Date Nov. 15, 2011, in 46 pages. |
PTAB-IPR2021-00673—Exhibit 1015—U.S. Pat. No. 7,823,145—Le 145, Issue Date Oct. 26, 2010, in 24 pages. |
PTAB-IPR2021-00673—Exhibit 1016—US20080091655A1—Gokhale, Publication Date Apr. 17, 2008, in 14 pages. |
PTAB-IPR2021-00673—Exhibit 1017—US20060259908A1—Bayer, Publication Date Nov. 16, 2006, in 8 pages. |
PTAB-IPR2021-00673—Exhibit 1018—U.S. Pat. No. 8,037,016—Odulinski, Issue Date Oct. 11, 2011, in 20 pages. |
PTAB-IPR2021-00673—Exhibit 1019—U.S. Pat. No. 7,925,850—Waldspurger, Issue Date Apr. 12, 2011, in 23 pages. |
PTAB-IPR2021-00673—Exhibit 1020—U.S. Pat. No. 8,191,063—Shingai, May 29, 2012, in 18 pages. |
PTAB-IPR2021-00673—Exhibit 1021—U.S. Pat. No. 8,959,509B1—Sobel, Issue Date Feb. 17, 2015, in 9 pages. |
PTAB-IPR2021-00673—Exhibit 1022—U.S. Pat. No. 8,458,419—Basler, Issue Date Jun. 4, 2013, in 14 pages. |
PTAB-IPR2021-00673—Exhibit 1023—D. Hall_lnternet Archive Affidavit & Ex. A, dated Jan. 20, 2021, in 106 pages. |
PTAB-IPR2021-00673—Exhibit 1024—esxRangerProfessionalUserManual, 2006, in 103 pages. |
PTAB-IPR2021-00673—Exhibit 1025—D.Hall_lnternet Archive Affidavit & Ex. A (source html view), dated Jan. 27, 2021, in 94 pages. |
PTAB-IPR2021-00673—Exhibit 1026—Scripting VMware (excerpted) (GMU), 2006, in 19 pages. |
PTAB-IPR2021-00673—Exhibit 1027—How to cheat at configuring VMware ESX server (excerpted), 2007, in 16 pages. |
PTAB-IPR2021-00673—Exhibit 1028—Robs Guide to Using VMware (excerpted), Sep. 2005 in 28 pages. |
PTAB-IPR2021-00673—Exhibit 1029—Hall-Ellis Declaration, dated Feb. 15, 2021, in 55 pages. |
PTAB-IPR2021-00673—Exhibit 1030—B. Dowell declaration, dated Oct. 15, 2020, in 3 pages. |
PTAB-IPR2021-00673—Exhibit 1031—Vizioncore esxEssentials Review ZDNet, Aug. 21, 2007, in 12 pages. |
PTAB-IPR2021-00673—Exhibit 1032—ZDNet Search on_howorth—p. 6 _, printed on Jan. 15, 2021, ZDNet 3 pages. |
PTAB-IPR2021-00673—Exhibit 1033—ZDNet _ Reviews _ ZDNet, printed on Jan. 15,02021, in 33 pages. |
PTAB-IPR2021-00673—Exhibit 1034—Understanding VMware Consolidated Backup, 2007, 11 pages. |
PTAB-IPR2021-00673—Exhibit 1035—techtarget.com news links—May 2007, May 20, 2007, in 39 pages. |
PTAB-IPR2021-00673—Exhibit 1036—ITPro 2007 Issue 5 (excerpted), Sep.-Oct. 2007 in 11 pages. |
PTAB-IPR2021-00673—Exhibit 1037—InfoWorld—Feb. 13, 2006, Feb. 13, 2006, in 17 pages. |
PTAB-IPR2021-00673—Exhibit 1038—InfoWorld—Mar. 6, 2006, Mar. 6, 2006, in 18 pages. |
PTAB-IPR2021-00673—Exhibit 1039—InfoWorld—Apr. 10, 2006, Apr. 10, 2006, in 18 pages. |
PTAB-IPR2021-00673—Exhibit 1040—InfoWorld—Apr. 17, 2006, Apr. 17, 2006, in 4 pages. |
PTAB-IPR2021-00673—Exhibit 1041—InfoWorld—May 1, 2006, May 1, 2006, in 15 pages. |
PTAB-IPR2021-00673—Exhibit 1042—InfoWorld—Sep. 25, 2006, Sep. 25, 2006, in 19 pages. |
PTAB-IPR2021-00673—Exhibit 1043—InfoWorld—Feb. 5, 2007, Feb. 5, 2007, in 22 pages. |
PTAB-IPR2021-00673—Exhibit 1044—InfoWorld—Feb. 12, 2007, Feb. 12, 2007, in 20 pages. |
PTAB-IPR2021-00673—Exhibit 1045—Informationweek—Aug. 14, 2006, Aug. 14, 2006, in 17 pages. |
PTAB-IPR2021-00673—Exhibit 1046—esxRanger Ably Backs Up VMs, May 2, 2007 in 6 pages. |
PTAB-IPR2021-00673—Exhibit 1047—Businesswire—Vizioncore Inc. Releases First Enterprise-Class Hot Backup and Recovery Solution for VMware Infrastructure, Aug. 31, 2006 in 2 pages. |
PTAB-IPR2021-00673—Exhibit 1048—Vizioncore Offers Advice to Help Users Understand VCB for VMwar, Jan. 23, 2007 in 3 pages. |
PTAB-IPR2021-00673—Exhibit 1049—Dell Power Solutions—Aug. 2007 (excerpted), Aug. 2007 in 21 pages. |
PTAB-IPR2021-00673—Exhibit 1050—communities-vmware-t5-VI-VMware-ESX-3-5-Discussions, Jun. 28, 2007, in 2 pages. |
PTAB-IPR2021-00673—Exhibit 1051—Distributed_File_System_Virtualization, Jan. 2006, pp. 45-56, in 12 pages. |
PTAB-IPR2021-00673—Exhibit 1052—Distributed File System Virtualization article abstract, 2006, in 12 pages. |
PTAB-IPR2021-00673—Exhibit 1053—Cluster Computing _ vol. 9, issue 1, Jan. 2006 in 5 pages. |
PTAB-IPR2021-00673—Exhibit 1054—redp3939—Server Consolidation with VMware ESX Server, Jan. 12, 2005 in 159 pages. |
PTAB-IPR2021-00673—Exhibit 1055—Server Consolidation with VMware ESX Server _ Index Page, Jan. 12, 2005 in 2 pages. |
PTAB-IPR2021-00673—Exhibit 1056—Apr. 21, 2020 [1] Complaint, filed Apr. 21, 2020, in 300 pages. |
PTAB-IPR2021-00673—Exhibit 1057—Feb. 17, 2021 (0046) Scheduling Order, filed Feb. 17, 2021, in 15 pages. |
PTAB-IPR2021-00673—Exhibit 1058—Novell Netware 5.0-5.1 Network Administration (Doering), Copyright 2001 in 40 pages. |
PTAB-IPR2021-00673—Exhibit 1059—US20060064555A1 (Prahlad 555), Publication Date Mar. 23, 006, in 33 pages. |
PTAB-IPR2021-00673—Exhibit 1060—Carrier Book, 2005 in 94 pages. |
PTAB-IPR2021-00673—Exhibit 2001 Jones Declaration, filed Jun. 30, 2021, in 35 pages. |
PTAB-IPR2021-00673—Exhibit 2002 VM Backup Guide 3.0.1, updated Nov. 21, 2007, 74 pages. |
PTAB-IPR2021-00673—Exhibit 2003 VM Backup Guide 3.5, updated Feb. 21, 2008, 78 pages. |
PTAB-IPR2021-00673—Exhibit 3001 RE_IPR2021-00535, 2021-00589, 2021-00590, 2021-00609, 2021-00673, 2021-00674, 2021-00675, Aug. 30, 2021, in 2 pages. |
PTAB-IPR2021-00673—Joint Motion to Terminate, filed Aug. 31, 2021, in 7 pages. |
PTAB-IPR2021-00673—Joint Request to Seal Settlement Agreement, filed Aug. 31, 2021, in 4 pages. |
PTAB-IPR2021-00673—673 674 Termination Order, Sep. 1, 2021, in 4 pages. |
PTAB-IPR2021-00673—Patent Owner Mandatory Notices, filed Apr. 7, 2021, 6 pages. |
Case No. 1:20-cv-00525-MN, Amended Complaint DDE-1-20-CV-00525-15, filed Jul. 27, 2020, in 30 pages. |
Case No. 1:20-cv-00525-MN, Complaint DDE-1-20-cv-00525-1, Apr. 21, 2020, in 28 pages. |
Case No. 1:20-cv-00525-MN, First Amended Answer DDE-1-20-cv-00525-95, filed Jul. 23, 2021, in 38 pages. |
Case No. 1:20-cv-00525-MN, Joint Claim Construction Brief Exhibits DDE-1-20-cv-00525-107-1, filed Oct. 1, 2021, in 488 pages (in 7 parts). |
Case No. 1:20-cv-00525-MN, Oral Order DDE-1-20-cv-00524-78_DDE-1-20-cv-00525-77, dated May 24, 2021, in 1 page. |
Case No. 1:20-cv-00525-MN, Oral Order DDE-1-20-cv-00524-86_DDE-1-20-cv-00525-87, dated Jun. 29, 2021, in 1 page. |
Case No. 1:20-cv-00525-MN, Order DDE-1-20-cv-00525-38_DDE-1-20-cv-00524-42, filed Feb. 10, 2021, in 4 pages. |
Case No. 20-525-MN-CJB, Joint Claim Construction Brief DDE-1-20-cv-00525-107, filed Oct. 1, 2021, in 79 pages. |
Case No. 20-525-MN-CJB, Joint Claim Construction Statement DDE-1-20-cv-00525-119, filed Oct. 29, 2021, in 12 pages. |
Case No. 20-525-MN-CJB, Farnan Letter DDE-1-20-cv-00525-111, filed Oct. 6, 2021, in 2 pages. |
Case No. 20-525-MN-CJB, Farnan Letter Exhibit A DDE-1-20-cv-00525-111-1, filed Oct. 6, 2021, in 7 pages. |
Case No. 1:20-cv-00524-MN, Order Dismissing with Prejudice DDE-1-20-cv-00524-101, filed Aug. 31, 2021, in 1 page. |
Case No. 1:20-cv-00524-MN, Amended_Complaint_DDE-1-20-cv-00524-13, filed on Jul. 27, 2020, in 30 pages. |
Case No. 1:20-cv-00524-MN, First_Amended_Answer_DDE-1-20-cv-00524-96, filed Jul. 23, 2021, in 41 pages. |
Case No. 1:20-cv-00524-MN, Stipulation_DDE-1-20-cv-00524-93, filed Jul. 14, 2021, in 3 pages. |
Case No. 1:20-cv-00524-MN, Oral Order DDE-1-20-cv-00524-86_DDE-1-20-cv-00525-87, filed Jun. 29, 2021, in 1 page. |
Case No. 1:20-cv-00524-MN, Complaint_DDE-1-20-cv-00524-1, filed on Apr. 21, 2020 in 28 pages. |
Case No. 1:20-cv-00524-MN, Answer_DDE-1-20-cv-00524-45, filed Feb. 16, 2021, in 25 pages. |
Case No. 1:20-cv-00524-MN, Order_DDE-1-20-cv-00525-38_DDE-1-20-cv-00524-42, filed Feb. 10, 2021, in 4 pages. |
PTAB-IPR2021-00674—('723) POPR Final, filed Jul. 8, 2021, in 70 pages. |
PTAB-IPR2021-00674—2021-03-31 723 Petition, filed Mar. 31, 2021, in 87 pages. |
PTAB-IPR2021-00674—Mar. 31, 2021 Explanation for Two Petitions, filed Mar. 31, 2021, in 9 pages. |
PTAB-IPR2021-00674—Exhibit 1001—U.S. Pat. No. 9,740,723, Issue Date Aug. 22, 2017, in 51 pages. |
PTAB-IPR2021-00674—Exhibit 1002—Jagadish Declaration, dated Mar. 31, 2021, in 200 pages. |
PTAB-IPR2021-00674—Exhibit 1003—U.S. Pat. No. 9,740,723 file history, Issue Date Aug. 22, 2017, in 594 pages. |
PTAB-IPR2021-00674—Exhibit 1004—Virtual Machine Monitors Current Technology and Future Trends, May 2005, in 9 pages. |
PTAB-IPR2021-00674—Exhibit 1005—Virtualization Overview, 2005, 11 pages. |
PTAB-IPR2021-00674—Exhibit 1006—Let's Get Virtual_Final Stamped, May 14, 2007, in 42 pages. |
PTAB-IPR2021-00674—Exhibit 1007—U.S. Pat. No. 8,458,419—Basler, Issue Date Jun. 4, 2013, in 14 pages. |
PTAB-IPR2021-00674—Exhibit 1008—US20080244028A1 (Le), Publication Date Oct. 2, 2008, in 22 pages. |
PTAB-IPR2021-00674—Exhibit 1009—60920847 (Le Provisional), Filed Mar. 29, 2007, in 70 pages. |
PTAB-IPR2021-00674—Exhibit 1010—Discovery Systems in Ubiquitous Computing (Edwards), 2006, in 8 pages. |
PTAB-IPR2021-00674—Exhibit 1011—HTTP The Definitive Guide excerpts (Gourley), 2002, in 77 pages. |
PTAB-IPR2021-00674—Exhibit 1012—VCB White Paper (Wayback Mar. 21, 2007), retrieved Mar. 21, 2007, Coypyright Date 1998-2006, in 6 pages. |
PTAB-IPR2021-00674—Exhibit 1013—Scripting VMware excerpts (Muller), 2006, in 66 pages. |
PTAB-IPR2021-00674—Exhibit 1014—Rob's Guide to Using VMWare excerpts (Bastiaansen), Sep. 2005, in 178 pages. |
PTAB-IPR2021-00674—Exhibit 1015—Carrier, 2005 in 94 pages. |
PTAB-IPR2021-00674—Exhibit 1016—U.S. Pat. No. 7,716,171 (Kryger), Issue Date May 11, 2010, in 18 pages. |
PTAB-IPR2021-00674—Exhibit 1017—RFC2609, Jun. 1999, in 33 pages. |
PTAB-IPR2021-00674—Exhibit 1018—MS Dictionary excerpt, 2002, in 3 pages. |
PTAB-IPR2021-00674—Exhibit 1019—Commvault v. Rubrik Complaint, Filed Apr. 21, 2020, in 29 pages. |
PTAB-IPR2021-00674—Exhibit 1020—Commvault v. Rubrik Scheduling Order, Filed Feb. 17, 2021, in 15 pages. |
PTAB-IPR2021-00674—Exhibit 1021—Duncan Affidavit, Dated Mar. 3, 2021, in 16 pages. |
PTAB-IPR2021-00674—Exhibit 1022—Hall-Ellis Declaration, dated Mar. 30, 2021, in 291 pages. |
PTAB-IPR2021-00674—Exhibit 1023—Digital_Data_lntegrity_2007_Appendix_A_UMCP, 2007, in 24 pages. |
PTAB-IPR2021-00674—Exhibit 1024—Rob's Guide—Amazon review (Jan. 4, 2007), retrieved Jan. 4, 2007, in 5 pages. |
PTAB-IPR2021-00674—Exhibit 2001—esxRanger, 2006, in 102 pages. |
PTAB-IPR2021-00674—Exhibit 2002—Want, 1995, in 31 pages. |
PTAB-IPR2021-00674—Exhibit 2003—Shea, retrieved Jun. 10, 2021, in 5 pages. |
PTAB-IPR2021-00674—Exhibit 2004—Jones Declaration, Dated Jul. 8, 2021, in 36 pages. |
PTAB-IPR2021-00674—Exhibit 3001, dated Aug. 30, 2021, in 2 pages. |
PTAB-IPR2021-00674—Exhibit IPR2021-00674 Joint Request to Seal Settlement Agreement, dated Aug. 31, 2021, in 4 pages. |
PTAB-IPR2021-00674—Joint Motion to Terminate, filed Aug. 31, 2021, in 7 pages. |
PTAB-IPR2021-00674—Response to Notice Ranking Petitions Final, filed Jul. 8, 2021, in 7 pages. |
PTAB-IPR2021-00674—Termination Order, filed Sep. 1, 2021, in 4 pages. |
PTAB-IPR2021-00609—('048) POPR Final, filed Jun. 16, 2021, in 28 pages. |
PTAB-IPR2021-00609—Mar. 10, 2021 IPR Petition—pty, Mar. 10, 2021, in 89 pages. |
PTAB-IPR2021-00609—Exhibit 1001—U.S. Pat. No. 10,210,048, Issue Date Feb. 19, 2019, in 49 pages. |
PTAB-IPR2021-00609—Exhibit 1002—Sandeep Expert Declaration,dated Mar. 10, 2021, in 176 pages. |
PTAB-IPR2021-00609—Exhibit 1003—U.S. Pat. No. 9,354,927 (Hiltgen), Issue Date May 31, 2016, in 35 pages. |
PTAB-IPR2021-00609—Exhibit 1004—U.S. Pat. No. 8,677,085 (Vaghani), Issue Date Mar. 18, 2014, in 44 pages. |
PTAB-IPR2021-00609—Exhibit 1005—U.S. Pat. No. 9,639,428 (Boda), Issue Date May 2, 2017, in 12 pages. |
PTAB-IPR2021-00609—Exhibit 1006—US20150212895A1 (Pawar), Publication Date Jul. 30, 2015, in 60 pages. |
PTAB-IPR2021-00609—Exhibit 1007—U.S. Pat. No. 9,665,386 (Bayapuneni), Issue Date May 30, 2017, in 18 pages. |
PTAB-IPR2021-00609—Exhibit 1008—Popek and Golberg, Jul. 1974, in 10 pages. |
PTAB-IPR2021-00609—Exhibit 1009—Virtualization Essentials—First Edition (2012)—Excerpted, 2012, in 106 pages. |
PTAB-IPR2021-00609—Exhibit 1010—Virtual Machine Monitors Current Technology and Future Trends, May 2005, in 9 pages. |
PTAB-IPR2021-00609—Exhibit 1011—Virtualization Overview, 2005, in 11 pages. |
PTAB-IPR2021-00609—Exhibit 1012—Let's Get Virtual A Look at Today's Virtual Server, May 14, 2007, in 42 pages. |
PTAB-IPR2021-00609—Exhibit 1013—Virtual volumes, Jul. 22, 2016, in 2 pages. |
PTAB-IPR2021-00609—Exhibit 1014—Virtual volumes and the SDDC—Virtual Blocks, Internet Archives on Sep. 29, 2015, in 4 pages. |
PTAB-IPR2021-00609—Exhibit 1015—NEC White Paper-VMWare vSphere Virtual Volumes (2015), Internet Archives Dec. 4, 2015 in 13 pages. |
PTAB-IPR2021-00609—Exhibit 1016—EMC Storage and Virtual volumes, Sep. 16, 2015 in 5 pages. |
PTAB-IPR2021-00609—Exhibit 1017—U.S. Pat. No. 8,621,460 (Evans), Issue Date Dec. 31, 2013, in 39 pages. |
PTAB-IPR2021-00609—Exhibit 1018—U.S. Pat. No. 7,725,671 (Prahlad), Issue Date May 25, 2010, in 48 pages. |
PTAB-IPR2021-00609—Exhibit 1019—Assignment—Vaghani to VMWare, Feb. 8, 2012, in 8 pages. |
PTAB-IPR2021-00609—Exhibit 1020—Assignment Docket—Vaghani, Nov. 11, 2011, in 1 page. |
PTAB-IPR2021-00609—Exhibit 1021—Dive into the VMware ESX Server hypervisor-IBM Developer, Sep. 23, 2011, in 8 pages. |
PTAB-IPR2021-00609—Exhibit 1022—MS Computer Dictionary Backup labeled, 2002 in 3 pages. |
PTAB-IPR2021-00609—Exhibit 1023—Jul. 7, 2014_VMware vSphere Blog, Jun. 30, 2014, 4 pages. |
PTAB-IPR2021-00609—Exhibit 1024—CommVault v. Rubrik Complaint, filed on Apr. 21, 2020, in 29 pages. |
PTAB-IPR2021-00609—Exhibit 1025—CommVault v. Cohesity Complaint, filed on Apr. 21, 2020, in 28 pages. |
PTAB-IPR2021-00609—Exhibit 1026—Feb. 17, 2021 (0046) Scheduling Order, filed on Feb. 17, 2021, in 15 pages. |
PTAB-IPR2021-00609—Exhibit 2001—Prosecution History_Part1, Issue Date Feb. 19, 2019, in 300 pages, Part 1 of 2. |
PTAB-IPR2021-00609—Exhibit 2001—Prosecution History_Part2, Issue Date Feb. 19, 2019, in 265 pages, Part 2 of 2. |
PTAB-IPR2021-00609—Exhibit 2002—Jones Declaration, dated Jun. 16, 2021, in 38 pages. |
PTAB-IPR2021-00609—Exhibit 3001—RE_IPR2021-00535, 2021-00589, 2021-00590, 2021-00609, 2021-00673, 2021-00674, 2021-00675, dated Aug. 30, 2021, in 2 pages. |
PTAB-IPR2021-00609—Joint Motion to Terminate. Filed Aug. 31, 2021, in 7 pages. |
PTAB-IPR2021-00609—Joint Request to Seal Settlement Agreement, filed Aug. 31, 2021, in 4 pages. |
PTAB-IPR2021-00609—Termination Order, Sep. 1, 2021, in 4 pages. |
Case No. No. 6:21-CV-00634-ADA, Answer WDTX-6-21-CV-00634-19, filed Aug. 27, 2021, in 23 pages. |
Case No. 1:21-cv-00537, Complaint WDTX-1-21-cv-00537-1_WDTX-6-21-cv-00634-1, filed Jun. 18, 2021, in 44 pages. |
Case No. 6:21-cv-00634-ADA, Order Dismissing with Prejudice WDTX-6-21-CV-00634-22, filed Sep. 1, 2021, in 1 page. |
Number | Date | Country | |
---|---|---|---|
20200241613 A1 | Jul 2020 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16274014 | Feb 2019 | US |
Child | 16848799 | US | |
Parent | 15163446 | May 2016 | US |
Child | 16274014 | US |