System and method for performing an image level snapshot and for restoring partial volume data

Abstract
The present invention relates to a method for performing an image level copy of an information store. The present invention comprises performing a snapshot of an information store that indexes the contents of the information store, retrieving data associated with the contents of the information store from a file allocation table, copying the contents of the information store to a storage device based on the snapshot, and associating the retrieved data with the copied contents to provide file system information for the copied contents.
Description
COPYRIGHT NOTICE

A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosures, as it appears in the Patent and Trademark Office patent files or records, but otherwise reserves all copyright rights whatsoever.


RELATED APPLICATIONS

This application is related to the following patents, each of which is hereby incorporated herein by reference in its entirety:

    • U.S. Pat. No. 6,418,478, entitled “PIPELINED HIGH SPEED DATA TRANSFER MECHANISM,” issued Jul. 9, 2002;
    • application Ser. No. 09/610,738, entitled “MODULAR BACKUP AND RETRIEVAL SYSTEM USED IN CONJUNCTION WITH A STORAGE AREA NETWORK,” filed Jul. 6, 2000;
    • application Ser. No. 09/744,268, entitled “LOGICAL VIEW AND ACCESS TO PHYSICAL STORAGE IN MODULAR DATA AND STORAGE MANAGEMENT SYSTEM,” filed Jan. 30, 2001;
    • Application Ser. No. 60/409,183, entitled “DYNAMIC STORAGE DEVICE POOLING IN A COMPUTER SYSTEM,” filed Sep. 9, 2002;
    • application Ser. No. 10/681,386 entitled “SYSTEM AND METHOD FOR MANAGING STORED DATA,” filed Oct. 7, 2003; and
    • Application Ser. No. 60/460,234, entitled “SYSTEM AND METHOD FOR PERFORMING STORAGE OPERATIONS IN A COMPUTER NETWORK,” filed Apr. 3, 2003.


BACKGROUND OF THE INVENTION

The invention disclosed herein relates generally to a system and method for performing a snapshot and for restoring data. More particularly, the present invention relates to a system and method for performing snapshots of an information store, which are stored across multiple storage devices, and for restoring partial or full snapshots.


To obtain a more thorough understanding of the present invention, the following discussion provides additional understanding regarding the manner is which magnetic media is used to store information. Using traditional techniques, copies of an information store are performed using the operating system's file system. Copying is done by accessing the operating system's (OS) file system for the information store to be backed-up, such as the Windows NTFS file system. The file allocation system of the operating system typically uses a file allocation table to keep track of the physical or logical clusters across which each file in the information store is stored. Also called an allocation unit, a cluster is a given number of disk sectors that are treated as a unit, each disk sector storing a number of bytes of data. This unit, the cluster, is the smallest unit of storage the operating system can manage. For example, on a computer running Microsoft's Windows 95 operating system, the OS uses the Windows F AT32 32-bit file allocation table having a cluster size to 4K. The number of sectors is determined when the disk is formatted by a formatting program, generally, but not necessarily, when the OS is installed.


The operating system allocates disk space for a file only when needed. That is, the data space is not preallocated but allocated dynamically. The space is allocated one cluster at a time, where a cluster is a given number of consecutive disk sectors. The clusters for a file are chained together, and kept track of, by entries in a file allocation table (FAT).


The clusters are arranged on the disk to minimize the disk head movement. For example, all of the space on a track is allocated before moving on to the next track. This is accomplished by using the sequential sectors on the lowest-numbered cylinder of the lowest numbered platter, then all sectors in the cylinder on the next platter, and so on, until all sectors on all platters of the cylinder are used. This is performed sequentially across the entire disk, for example, the next sector to be used will be sector 1 on platter 0 of the next cylinder.


For a hard (fixed) disk, FAT, sector, cluster, etc. size is determined when a disk formatting program formats the disk, and are based on the size of the partition. To locate all of the data that is associated with a particular file stored on a hard disk, the starting cluster of the file is obtained from the directory entry, then the FAT is referenced to locate the next cluster associated with the file. Essentially, the FAT is a linked list of pointers to clusters on the disk, e.g., each 16-bit FAT entry for a file points to the next sequential cluster used for that file. The last entry for a file in the FAT has a number indicating that no more clusters follow. This number can be from FFF8 to FFFF (base 16) inclusive.



FIG. 1 shows an example directory entry 2 of a Windows-formatted hard disk and accompanying FAT 20. The exemplary directory entry 2 consists of 32 bytes of data. The name of the file and its extension are stored in the first eleven bytes 4 of the directory entry 2 and a file attribute byte 6 is provided. By definition, ten bytes 8 are reserved for future use and four bytes are provided to store time 10 and date 12 information (two bytes each). Two cluster bytes 14 point to the first cluster of sectors used to store the file information. The last four bytes 18 of the directory entry 2 are used to store the size of the file.


A sixteen-byte section of a FAT 20 is depicted. The first four bytes 21 store system information. A two-byte pair, bytes four and five (16), are the beginning bytes of the FAT 20 used to track file information. The first cluster for data space on all disks is cluster “02.” Therefore, bytes four and five (16) are associated with the first cluster of disk sectors “02” used to store file information. Bytes six and seven (22) are associated with cluster “03” . . . and bytes fourteen and fifteen (24) are associated with cluster “07.”


This example illustrates how sectors associated with a file referenced in a directory are located. The cluster information bytes 14 in the directory 2 point to cluster number “02.” The sectors in cluster “02” (not shown), contain the initial sector of data for the referenced file. Next, the FAT is referenced to see if additional clusters are used to store the file information. FAT bytes four and five (16) were pointed to by the cluster information bytes 14, and the information stored in bytes four and five (16) in the FAT 20 point to the next cluster used for the file. Here, the next cluster is “OS”. Accordingly, cluster “OS” contains the next sector of data for the referenced file. FAT bytes ten and eleven (26) contain an end-of-file flag, “FFFF,” indicating there are no more clusters associated with the referenced file. All of the information comprising the referenced file, therefore, is contained in clusters “02” and “05” on the disk.


As with other applications running on the computer, a typical backup application provides a read request to the operating system, which handles interpretation of the information contained in the FAT and reading of each file for the backup application. A file system is provided on the storage device that is used by the backup application to write files that are copied to the device. Similarly, the recovery portion of the backup application, or a separate recovery application, may read files from the storage device for recovery of the information.


Inherent problems and disadvantages have been discovered with currently available systems and methods for archiving data contained in an information store. One technique is to perform a full copy of the data contained in the information store. Utilizing this technique results in two separate copies of the information store, and the length of time it takes to make this kind of copy is related to the amount of data copied and the speed of the disk subsystem. For example, assuming a transfer rate of 25 MB/sec, the approach will take one hour to copy 90 GB of data. These techniques, however, in addition to other disadvantages, require the applications on the information store to be quiesced during the copy routine. This places a significant burden on system administrators to complete copying and get critical systems back into the production environment as, absent a high-speed data bus, the copying may consume a significant amount of time to complete.


Administrators typically keep multiple copies of a given information store. Unfortunately, this has the drawback of requiring n times the amount of space of the information store to maintain n copies, which can be quite expensive to store, in addition to requiring complex and time consuming techniques for restoration of the copied data.


One currently available alternative is to perform snapshots of an information store. With current snapshot systems and methods, administrators create an incremental copy that is an exact point-in-time replica of the source volume each time a snapshot is taken. A series of snapshot are stored locally on the information store from which it was taken and track incremental changes to the data in the information store. Furthermore, changed data is written to a new location in the information store as tracked by the snapshot. With knowledge regarding the change, as well as the changed data, the snapshot can be used to “roll back” changes to an information store to the point in time when the snapshot was taken. If there should be any logical corruption in the information store's data that went undetected for a period of time, however, these incremental updates faithfully replicate that logical corruption to the data when copying. Additionally, other drawbacks are associated with currently know snapshot techniques, including the significant drawback of preventing restoration from the snapshot in the event that the information store fails, as both the snapshot and the information store become unavailable.


Systems and methods are needed, therefore, that overcome problems associated with currently known techniques for taking, maintaining and restoring snapshots.


SUMMARY OF THE INVENTION

The present invention addresses, among other things, the problems discussed above with copying up data using systems and methods known to those of skill in the art. The invention provides systems and methods for performing n snapshots of an information store, without requiring n times the space of the information store, and storing those snapshots in multiple destinations across a network.


One embodiment of the system of the present invention creates the snapshots by taking a snapshot that indexes only clusters for files that were created or changed since the last snapshot. A snapshots, tn, is restored by restoring the clusters from the snapshot tn. The clusters that were not restored from snapshot tn are restored from snapshot tn−}, etc., until the remaining clusters are restored from the first snapshot, snapshot t0.


In accordance with some aspects of the present invention, multiple snapshots are kept on a storage device, without requiring n times the space of the total volume of the information store. The system creates snapshots at various points in time that index only clusters for files that were created or changed since the last snapshot, and creates a copy of the data that has been changed or created. This allows users to keep several snapshots without requiring n times the space of the total volume of the information store.


In some embodiments, the system stores a map, which may be part of a snapshot, to track specific files and folders with their corresponding copied clusters. The map created by reading data from the file allocation table of the information store and associates files and folders with the clusters stored in the snapshots. In this way, even though the snapshot was performed at the cluster level, individual or groups of files and/or folders may be restored without unnecessarily restoring the entire information store.





BRIEF DESCRIPTION OF THE DRAWINGS

The invention is illustrated in the figures of the accompanying drawings which are meant to be exemplary and not limiting, in which like references are intended to refer to like or corresponding parts, and in which:



FIG. 1 is an example directory entry for a file in a prior art FAT of a Windows-formatted hard disk;



FIG. 2 is a block diagram illustrating a network architecture for performing snapshot operations according to one embodiment of the present invention;



FIG. 3 is a flow diagram illustrating a method for creating a snapshot according to one embodiment of the present invention;



FIG. 4 is a block diagram illustrating the relationship between a map and a snapshot according to one embodiment of the present invention;



FIG. 5 is a flow diagram illustrating a method for restoring a snapshot according to one embodiment of the present invention; and



FIG. 6 is a flow diagram illustrating a method for restoring specific files or folders from a snapshot according to one embodiment of the present invention.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT

With reference to FIGS. 2 through 6, embodiments of the present invention are shown. FIG. 2 presents a block diagram illustrating the components of a system for performing storage and restoration operations on electronic data in a computer network according to one embodiment of the invention. It should be understood that the invention is not limited to networked environments, and may also be implemented on a stand-alone computer or electronic device.


As shown, the system of FIG. 2 includes a storage manager 100, including a volume replication table 102 and a storage manager index cache 120, and one or more of the following: a client 85, an information store 90, a data agent 95, a media agent 105, a media agent index cache 110, and a storage device 115. One exemplary embodiment of the present system is the CommVault QuNetix three-tier system available from CommVault Systems, Inc. of Oceanport, N.J., further described in U.S. patent application Ser. No. 09/610,738 and hereby incorporated by reference in its entirety.


A data agent 95 is a software module that is generally responsible for retrieving data from an information store 90 for copies, snapshots, archiving, migration, and recovery of data stored in an information store 90 or other memory location, e.g., hard disc drive. Each client computer 85 preferably has at least one data agent 95 and the system can support many client computers 85. The data agent 95 provides an interface to an information store 90 to execute copies, snapshots, archiving, migration, recovery and other storage operations on data in conjunction with one or more media agents 105. According to one embodiment, each client 85 runs a number of data agents 95, wherein each data agent is configured to interface with data generated by or from a specific application, e.g., a first data agent to interface with Microsoft Exchange data and a second data agent to interface with Oracle database data. As is explained in greater detail herein, a data agent 95 is III communication with one or more media agents 105 to effect the distributed storage of snapshots on one or more storage devices 115 that are remote from the information store that is the source of the snapshot 90.


The storage manager 100 is a software module or application that coordinates and controls other components comprising the system, e.g., data and media agents, 95 and 105, respectively. The storage manager 100 communicates with data 95 and media 105 agents to control and manage snapshot creation, migration, recovery and other storage operations. According to one embodiment, the storage manger 100 maintains data in a storage manager index cache 120 that instructs a given data agent 95 to work in conjunction with a specific media agent 105 to store snapshots on one or more storage devices 115.


The storage manager 100 maintains a storage manager index cache 120. Data in the storage manager index cache 120, which the storage manager 100 collects from data agents 95, media agents 105, user and other applications, is used to indicate, track and associate: logical relationships and associations between components of the system, user preferences, management tasks, and other data that is useful to the system. For example, the storage manager index cache 120 may contain data that tracks logical associations between media agents 105 and storage devices 115. The storage manager index cache 120 may also contain data that tracks the status of storage operations to be performed, storage patterns such as media use, storage space growth, network bandwidth, service level agreement (“SLA”) compliance levels, data protection levels, storage policy information, storage criteria associated with user preferences, data retention criteria, storage operation preferences, and other storage-related information.


A media agent 105 is a software module that transfers data in conjunction with one or more data agents 95, as directed by the storage manager 100, between an information store 90 and one or more storage devices 115, such as a tape library, a magnetic media storage device, an optical media storage device, or other storage device. The media agent 105 communicates with and controls the one or more storage devices 115. According to one embodiment, the media agent 105 may communicate with the storage device 115 via a local bus, such as a SCSI adaptor. Alternatively, the storage device 115 may communicate with the data agent 105 via a Storage Area Network (“SAN”). Other types of communication techniques, protocols and media are contemplated as falling within the scope of the invention.


The media agent 105 receives snapshots, preferably with the changed data that is tracked by the snapshot, from one or more data agents 95 and determines one or more storage devices 115 to which it should write the snapshot. According to one embodiment, the media agent 105 applies load-balancing algorithms to select a storage device 115 to which it writes the snapshot. Alternatively, the storage manager 100 instructs the media agent 105 as to which storage device 115 the snapshot should be written. In this manner, snapshots from a given information store 90 may be written to one or more storage devices 115, ensuring data is available for restoration purposes in the event that the information store fails. Either the media agent or the storage manager 100 records the storage device on which the snapshot is written in a replication volume table 102, thereby allowing the snapshot to be located when required for restoring the information store 90.


A media agent 105 maintains a media agent index cache 110 that stores index data the system generates during snapshot, migration, and restore operations. For example, storage operations for Microsoft Exchange data generate application specific index data regarding the substantive Exchange data. Similarly, other applications may be capable of generating application specific data during a copy or snapshot. This data is generally described as metadata, and may be stored in the media agent index cache 110. The media agent index cache 110 may track data that includes, for example, information regarding the location of stored data on a given volume. The media agent index cache 110 may also track data that includes, but is not limited to, file names, sizes, creation dates, formats, application types, and other file-related information, information regarding one or more clients associated stored data, information regarding one or more storage policies, storage criteria, storage preferences, compression information, retention related information, encryption related information, and stream related information. Index data provides the system with an efficient mechanism for locating user files during storage operations such as copying, performing snapshots and recovery.


This index data is preferably stored with the snapshot that is backed up to the storage device 115, although it is not required, and the media agent 105 that controls the storage operation may also write an additional copy of the index data to its media agent index cache 110. The data in the media agent index cache 110 is thus readily available to the system for use in storage operations and other activities without having to be first retrieved from the storage device 115.


In order to track the location of snapshots, the system uses a database table or similar data structure, referred to herein as a replication volume table 102. The replication volume table 102, among other advantages, facilitates the tracking of multiple snapshots across multiple storage devices 115. For example, the system might, as directed by a policy or a user, store a first snapshot to on first storage device A, such as a tape drive or library, and then store subsequent snapshots containing only the changed cluster(s), tn, on a second storage device B, such as an optical drive or library. Alternatively, instructions may be stored within system components, e.g., a storage manger 100 or media agent 105, directing the storage devices 115 used to store snapshots. Information regarding the storage device 115 to which the snapshot is written, as well as other information regarding the snapshot generally, is written to the replication volume table 102. An exemplary structure according to one embodiment is as follows:


















{







id
serial,
//
PRIMARY KEY FOR THIS






TABLE



PointInTime
integer,
//




CreationTime
integer,
//
Timestamp of RV creation



ModifyTime
integer,
//
Timestamp of last






RV update



Current State
integer,
//
Current state of R V



CurrentRole
integer,
//
Current role of RV



PrimaryVolumeId
integer,
//
FOREIGN KEY FOR



PhysicalVolumeID
integer,
//
SNR Volume TABLE



ReplicationPolicyId
integer,
//
FOREIGN KEY FOR



RVScratch
integer,
//
SNR Volume TABLE



Flags
integer,

FOREIGN KEY FOR



JobId
longlong,

Replication Policy TABLE



Snap VolumeId
integer,
//
FOREIGN KEY FOR


}









In the exemplary replication volume table, id is a unique identification number assigned by the system to the snapshot; PointInTime represents the date and time that the snapshot was created; CreationTime represents the date and time that the snapshot was completed; ModifyTime is the recorded date and time of the snapshot taken prior to the current snapshot; Current state is an identifier used to indicate a current status of the snapshot (e.g. pending, completed, unfinished, etc.); PrimaryVolumeId is the identifier for the information store 90 from which the snapshot is being made; PhysicalVolumeId is a hardware identifier for the information store 90; RVScratchVolumeId is an identifier for a scratch volume, which in some embodiments may be used to buffer additional memory as known to those of skill in the art; Flags contains a 32 bit word for various settings such as whether a snapshot has been taken previously, etc.; JobId stores the identifier for the job as assigned by a storage management module; and the Snap VolumeId points to the physical destination storage device 115 to which the snapshot is written.


As each snapshot indexes an information store at a given point in time, a mechanism must be provided that allows the snapshots taken of an information store to be chronologically related so that they are properly used for restoring an information store 90. According to the replication volume table 102, the CurrentRole integer may store a value for the relative position of a given snapshot in hierarchy of snapshots taken from a given information store 90 (e.g. first (to), second (tl), t2, t3, etc.)


In some embodiments, components of the system may reside on and be executed by a single computer. According to this embodiment, a data agent 95, media agent 105 and storage manager 100 are located at the client computer 85 to coordinate and direct local copying, archiving, migration, and retrieval application functions among one or more storage devices 115 that are remote or distinct from the information store 90. This embodiment is further described in U.S. patent application Ser. No. 09/610,738.


One embodiment of a method for using the system of the present invention to perform snapshots is illustrated in the flow diagram of FIG. 3. When the system is initialized, or at other times as directed by a user or rules, e.g., policies or other instructions, the storage manager directs the data agent to perform an initial full snapshot of the data stored in the information store, e.g., indexing the location of all data in the information store, in conjunction with one or more media agents. The system copies all of the data on the information store with the initial snapshot to a storage device, step 300.


Advantageously, the snapshot and data copied from the information store may be written to a storage device that is remote or different from the information store, step 302, e.g., local data from a given information store is written to a storage device attached to a network. The selection of a destination storage device for the snapshot may be accomplished using one or more techniques known to those of skill in the art. For example, a fixed mapping may be provided indicating a storage device for which all snapshots and copied or changed data should be written. Alternatively, an algorithm may be implemented to dynamically select a storage device from among a number of storage devices available on a network. For example, a storage manager may select a media agent to handle the transfer of the snapshot and copied data to a specific storage device based on criteria such as available bandwidth, other scheduled storage operations, media availability, storage policies, storage preferences, or other consider considerations. The snapshot, preferably along with the data from the information store, is written to the selected destination storage device, step 304. According to certain embodiments, the snapshot contains information regarding the files and folders that are tracked by the snapshot. Alternatively, the information regarding the files and folders that are indexed by the snapshot, e.g., file system information, are stored on the storage device.


One embodiment of a snapshot used to track clusters read from the information store to clusters in a snapshot, as well as to map file and folder names corresponding to the snapshot clusters, is illustrated in FIG. 4. It should be noted that clusters are but one level of granularity that may be indexed by a snapshot, e.g., blocks, extents, etc. During the scan, the data agent creates a snapshot 350 and writes data, e.g., new or changed data, to a storage device 115. According to the present embodiment, the snapshot is illustrated as a flat file data structure, although those of skill in the art will recognize that the snapshot may be embodied in a number of disparate types of data structures.


The snapshot 350 is used to associate the original cluster numbers from an information store with clusters on a storage device, which in the present embodiment is a magnetic tape. It should be appreciated by those of skill in the art that the present invention is not limited to magnetic tape, and that the systems and methods described herein may be applicable to using snapshots with other storage technologies, e.g., storing disk geometry data to identify the location of a cluster on a storage device, such as a hard disk drive.


The tape offsets 356 for the clusters 372 in the snapshot 370 are mapped to original disk cluster information 352. File and folder names 354 may be scanned from the information store's FAT and also mapped to the tape offsets 356. A file part column 358 in the snapshot tracks the clusters 372 for each file and folder where each file and folder contains an entry for the first cluster 372. For files or folders that are stored in more than one cluster, sometimes not in contiguous clusters, the offset table entry for each further cluster is numbered consecutively 358.


In order to identify the files and folders represented by the stored clusters 372, e.g., changed data, in the snapshot 370, the map may exclude data from columns relating to the original disc clusters 352 and last snapshot 360. In order to keep track of changed verses unchanged clusters, however, the original disk cluster information 352 is stored in the map 350. Other information may also be stored in the map 350, such as timestamps for last edit and creation dates of the files.


For each snapshot, even though only clusters that have been changed or created since a previous snapshot are tracked in a given snapshot after the initial snapshot to, the snapshot may be provided with the data from all previous snapshots to provide the latest snapshot with folder and file information such that an index of the entire information store is maintained concurrently each snapshot. Alternatively, this may be bypassed in favor of creating a snapshot that indexes all data at a given point in time in the information store and copying only changed data.


Entries from each snapshot 350 may also contain a last-snapshot field 360 that holds an identifier for the last snapshot containing the cluster indexed by the entry at the time the current snapshot was created. According to an alternative embodiment, e.g., for snapshots that do not store the information from the information store's FAT, the snapshot only tracks clusters stored in the information store with the clusters indexed by the snapshot. For those embodiments, the snapshot 350 contains neither file and folder information 345 nor file part information 358.


Returning to FIG. 3, once the first full snapshot to has been taken, step 300, the storage manager may implement a rule, policy, or similar set of instructions that require snapshots to be taken at certain time intervals. Accordingly, at each time interval where a subsequent snapshot tn is taken, the data agent works in conjunction with one or more of the media agents to perform and store snapshot and accompanying data that changed since the subsequent snapshot, tn-1, loop 306.


For each snapshot, tn, that is taken of the information store, a comparison is performed such that only the clusters which have changed or been created since the last snapshot, tn.)' was taken of that volume are stored, step 310. For example, in some embodiments the data agent employs a block filter or similar construct known to those of skill in the art to compare snapshot tn with tn-i and thereby detect changed clusters on an information store. Alternatively, the data agent may use other techniques know in the art, such as Copy on Write (“COW”), to identify changed data on an information store. If a given cluster in the information store has changed since the last snapshot in which the cluster appears, or if the cluster from the information store was created subsequent to the last snapshot, then the cluster is read from information store and stored with the new snapshot being written to the storage device, step 314.


A determination is made regarding the given storage device to which the snapshot and changed data (which may also include newly created data) is to be written, step 316. Techniques such as those described in conjunction with storage of the initial snapshot, steps 302 and 304, may also be employed regarding storage of subsequent snapshots. Advantageously, the initial snapshot and any subsequent snapshot may written to any storage device available in the network. Furthermore, there is no limitation to the combination of devices used to store the snapshots for a given information store. For example, an initial snapshot may be written to storage device A, a second and third snapshots may be written to storage device B, and a fourth snapshot may be written to storage device C. Regardless of the storage device that is selected, step 316, the replication volume table is updated to reflect the location, step 318, allowing snapshots to be located when a user requests to restore the information store from which the snapshots were taken.


System administrators use stored snapshots, in conjunction with the changed data that the snapshot indexes or tracks, to recover lost or corrupted information. FIG. 5 presents a flow diagram illustrating one embodiment of a method for restoring an information store from one or more snapshots. If the user or a system process wants to restore an information store from one or more snapshots, an interface is presented to restore the snapshot, step 400. The interface may be, for example, a graphical user interface (“GUI”), and Application Programming Interface (“API”) or other interface known to those of skill in the art. The storage manager scans the replication volume table to identify available snapshots for presentation in a menu that allows selection of an available snapshot, step 402.


When the user selects a snapshot, the storage manager performs a query of the replication volume table to identify all previous snapshots for an information store from which the selected snapshot was taken, step 404. This may be accomplished by performing a search on the replication volume table for all snapshots with the same PrimaryVolumeId or PhysicalVolumeId. Starting with the selected snapshot, for each snapshot in the query result, loop 406, the storage manager directs a given media agent, in conjunction with a given data agent, to read and restore all clusters of changed data not already restored from clusters indexed by a prior snapshot, e.g., the latest version of each cluster, step 408. According to one embodiment, this is accomplished by restoring the clusters indexed by each of the snapshots in the query result, starting with the original snapshot, and overwriting clusters indexed by the original snapshot with changed clusters indexed by subsequent snapshots up to the snapshot representing the point in time selected by the user or system process. As an alternative, the last snapshot field of the selected snapshot may be utilized to determine the snapshots that should be utilized in the restore operation. The latest version of each cluster, starting with those indexed by the selected snapshot, is then restored, step 408.


As discussed above, embodiments of the invention are contemplated wherein FAT information of the information store is stored in conjunction with a given snapshot, e.g. the file and folder information corresponding to the clusters of changed data indexed by a given snapshot. Accordingly, the storage manager may allow the user to select individual files and/or folders to be selected for restoration from a snapshot. With reference to FIG. 6, a flow diagram is presented illustrating one embodiment of a method for restoring individual files and/or folders indexed by a snapshot.


When the user desires to restore the information store to a given point in time, the user interface allows the user to view the files and folders indexed by a snapshot representing the point in time as if the user were viewing a folder structure on a storage device, step 500. The storage manager retrieves the file and folder information for changed data that is indexed by one or more snapshots for display. Once one or more files and/or folders are selected, step 502, the storage manager selects those snapshots that index the given version of the files and/or folders using the replication volume table, step 502. Each snapshot indexing data for the one or more files to be restored are opened serially, loop 506. The changed data for the selected files and folders that are indexed by the snapshots are restored from clusters indexed by each snapshot, step 508, but not overwriting clusters indexed by prior snapshots.


While the invention has been described and illustrated in connection with preferred embodiments, many variations and modifications as will be evident to those skilled in this art may be made without departing from the spirit and scope of the. invention, and the invention is thus not to be limited to the precise details of methodology or construction set forth above as such variations and modification are intended to be included within the scope of the invention.

Claims
  • 1. A method in a computing system of creating snapshots for an information store, the method comprising: performing a first snapshot of one or more files in an information store at a first time, the one or more files being associated with clusters of data stored in the information store;storing the first snapshot in computer storage that is separate from the information store;performing a second snapshot of the one or more files in the information store at a second time, the second time occurring after the first time;comparing the second snapshot to the first snapshot to identify clusters of data associated with the one or more files that changed after the first time and before the second time;storing the second snapshot and copies of the changed clusters of data in computer storage that is separate from the information store; andautomatically indexing the second snapshot, wherein said indexing includes storing data indicating that the copies of the changed clusters are stored in association with the second snapshot.
  • 2. The method of claim 1, wherein the second snapshot is automatically indexed in a system catalog containing file system information.
  • 3. The method of claim 1, wherein said automatically indexing includes storing data indicating where in the second storage device the copies of the changed clusters are stored.
  • 4. The method of claim 1, wherein said automatically indexing the second snapshot includes identifying one or more previously saved snapshots where previous copies of the changed clusters were stored, if any.
  • 5. The method of claim 1, wherein said automatically indexing the second snapshot includes storing file data indicating what files or folders the changed clusters are associated with.
  • 6. The method of claim 5, wherein the file data includes a file name associated with one or more of the changed clusters.
  • 7. The method of claim 1, further comprising storing copies of each of the one or more files in association with the first snapshot.
  • 8. The method of claim 1, further comprising identifying a plurality of snapshots stored in association with a copy of a first cluster or file, and restoring the first cluster or file at least partially by reading each of the copies of the first cluster or file.
  • 9. An electronic file storage system configured to create snapshots for an information store, the system comprising: computer hardware comprising one or more computer processors;a storage manager module configured to execute in the one or more computer processors and configured to: direct performance of a first snapshot of one or more files in an information store at a first time, the one or more files being associated with clusters of data stored in the information store;store the first snapshot in a first storage device that is different than the information store;direct performance of a second snapshot of the one or more files in the information store at a second time, the second time occurring after the first time;compare the second snapshot to the first snapshot to identify clusters of data associated with the one or more files that changed after the first time and before the second time;store the second snapshot and copies of the changed clusters of data in a second storage device that is different than the information store; andautomatically index the second snapshot, wherein said indexing includes storing data indicating that the copies of the changed clusters are stored in association with the second snapshot.
  • 10. The system of claim 9, wherein the storage manager is configured to automatically index the second snapshot in a system catalog containing file system information.
  • 11. The system of claim 9, wherein the storage manager is configured to automatically index the second snapshot at least partially by storing data indicating where in the second storage device the copies of the changed clusters are stored.
  • 12. The system of claim 9, wherein the storage manager is configured to automatically index the second snapshot at least partially by identifying one or more previously-saved snapshots where previous copies of the changed clusters were stored, in any.
  • 13. The system of claim 9, wherein the storage manager is configured to automatically index the second snapshot at least partially by storing file data indicating what files or folders the changed clusters are associated with.
  • 14. The system of claim 13, wherein the file data includes a file name associated with one or more of the changed clusters.
  • 15. The system of claim 9, wherein the storage manager is further configured to store copies of each of the one or more files in association with the first snapshot.
  • 16. The method of claim 9, wherein the storage manager is further configured to identify a plurality of snapshots stored in association with a copy of a first cluster or file, and restore the first cluster or file at least partially by reading each of the copies of the first cluster or file.
  • 17. A computing system comprising: an information store having clusters of data associated with a plurality of files stored thereon;computer storage separate from the information store, the computer storage having stored thereon: a first snapshot of the plurality of files, the first snapshot having been stored in the computer storage at a first time;a second snapshot of the plurality of files, the second snapshot having been stored in the computer storage at a second time;copies of clusters of data associated with the plurality of files that changed between the first time and the second time; andan index of the second snapshot including data indicating that the copies of the changed clusters are stored in association with the second snapshot.
Parent Case Info

The present application is a continuation of U.S. patent application Ser. No. 12/951,773, filed on Nov. 22, 2010, which is a continuation of U.S. patent application Ser. No. 12/433,238, filed Apr. 30, 2009, now U.S. Pat. No. 7,840,533, issued Nov. 23, 2010, which is a continuation of U.S. patent application Ser. No. 10/990,353, filed Nov. 15, 2004, U.S. Pat. No. 7,539,707, issued May 26, 2009, which is based on and claims the benefit of U.S. Provisional Patent Application Nos. 60/519,876 and 60/519,576, each filed on Nov. 13, 2003. These applications are incorporated by reference herein in their entirety.

US Referenced Citations (493)
Number Name Date Kind
4296465 Lemak Oct 1981 A
4686620 Ng Aug 1987 A
4995035 Cole et al. Feb 1991 A
5005122 Griffin et al. Apr 1991 A
5093912 Dong et al. Mar 1992 A
5133065 Cheffetz et al. Jul 1992 A
5193154 Kitajima et al. Mar 1993 A
5212772 Masters May 1993 A
5226157 Nakano et al. Jul 1993 A
5239647 Anglin et al. Aug 1993 A
5241668 Eastridge et al. Aug 1993 A
5241670 Eastridge et al. Aug 1993 A
5263154 Eastridge et al. Nov 1993 A
5265159 Kung Nov 1993 A
5276860 Fortier et al. Jan 1994 A
5276867 Kenley et al. Jan 1994 A
5287500 Stoppani, Jr. Feb 1994 A
5301351 Jippo Apr 1994 A
5311509 Heddes et al. May 1994 A
5317731 Dias et al. May 1994 A
5321816 Rogan et al. Jun 1994 A
5333315 Saether et al. Jul 1994 A
5347653 Flynn et al. Sep 1994 A
5369757 Spiro et al. Nov 1994 A
5403639 Belsan et al. Apr 1995 A
5410700 Fecteau et al. Apr 1995 A
5448724 Hayashi et al. Sep 1995 A
5455926 Keele et al. Oct 1995 A
5487072 Kant Jan 1996 A
5491810 Allen Feb 1996 A
5495607 Pisello et al. Feb 1996 A
5504873 Martin et al. Apr 1996 A
5544345 Carpenter et al. Aug 1996 A
5544347 Yanai et al. Aug 1996 A
5546536 Davis et al. Aug 1996 A
5555404 Torbjornsen et al. Sep 1996 A
5559957 Balk Sep 1996 A
5559991 Kanfi Sep 1996 A
5598546 Blomgren Jan 1997 A
5604862 Midgely et al. Feb 1997 A
5615392 Harrison et al. Mar 1997 A
5619644 Crockett et al. Apr 1997 A
5638509 Dunphy et al. Jun 1997 A
5642496 Kanfi Jun 1997 A
5673381 Huai et al. Sep 1997 A
5675511 Prasad et al. Oct 1997 A
5677900 Nishida et al. Oct 1997 A
5682513 Candelaria et al. Oct 1997 A
5687343 Fecteau et al. Nov 1997 A
5689706 Rao et al. Nov 1997 A
5699361 Ding et al. Dec 1997 A
5719786 Nelson et al. Feb 1998 A
5720026 Uemura et al. Feb 1998 A
5729743 Squibb Mar 1998 A
5737747 Vishlitzky et al. Apr 1998 A
5742792 Yanai et al. Apr 1998 A
5751997 Kullick et al. May 1998 A
5758359 Saxon May 1998 A
5761677 Senator et al. Jun 1998 A
5761734 Pfeffer et al. Jun 1998 A
5764972 Crouse et al. Jun 1998 A
5765173 Cane et al. Jun 1998 A
5778395 Whiting et al. Jul 1998 A
5790114 Geaghan et al. Aug 1998 A
5790828 Jost Aug 1998 A
5805920 Sprenkle et al. Sep 1998 A
5812398 Nielsen Sep 1998 A
5813009 Johnson et al. Sep 1998 A
5813017 Morris Sep 1998 A
5829046 Tzelnic et al. Oct 1998 A
5860104 Witt et al. Jan 1999 A
5875478 Blumenau Feb 1999 A
5875481 Ashton et al. Feb 1999 A
5878408 Van Huben et al. Mar 1999 A
5887134 Ebrahim Mar 1999 A
5901327 Ofek May 1999 A
5907621 Bachman et al. May 1999 A
5907672 Matze et al. May 1999 A
5924102 Perks Jul 1999 A
5926836 Blumenau Jul 1999 A
5933104 Kimura Aug 1999 A
5933601 Fanshier et al. Aug 1999 A
5938135 Sasaki et al. Aug 1999 A
5950205 Aviani, Jr. Sep 1999 A
5956519 Wise et al. Sep 1999 A
5958005 Thorne et al. Sep 1999 A
5970233 Liu et al. Oct 1999 A
5970255 Tran et al. Oct 1999 A
5974563 Beeler, Jr. Oct 1999 A
5987478 See et al. Nov 1999 A
5991779 Bejar Nov 1999 A
5995091 Near et al. Nov 1999 A
6003089 Shaffer et al. Dec 1999 A
6009274 Fletcher et al. Dec 1999 A
6012090 Chung et al. Jan 2000 A
6021415 Cannon et al. Feb 2000 A
6021475 Nguyen et al. Feb 2000 A
6023710 Steiner et al. Feb 2000 A
6026414 Anglin Feb 2000 A
6049889 Steely, Jr. et al. Apr 2000 A
6052735 Ulrich et al. Apr 2000 A
6061692 Thomas et al. May 2000 A
6072490 Bates et al. Jun 2000 A
6076148 Kedem et al. Jun 2000 A
6094416 Ying Jul 2000 A
6105129 Meier et al. Aug 2000 A
6112239 Kenner et al. Aug 2000 A
6122668 Teng et al. Sep 2000 A
6131095 Low et al. Oct 2000 A
6131148 West et al. Oct 2000 A
6131190 Sidwell Oct 2000 A
6137864 Yaker Oct 2000 A
6148377 Carter et al. Nov 2000 A
6148412 Cannon et al. Nov 2000 A
6154787 Urevig et al. Nov 2000 A
6154852 Amundson et al. Nov 2000 A
6158044 Tibbetts Dec 2000 A
6161111 Mutalik et al. Dec 2000 A
6163856 Dion et al. Dec 2000 A
6167402 Yeager Dec 2000 A
6175829 Li et al. Jan 2001 B1
6195695 Cheston et al. Feb 2001 B1
6205450 Kanome et al. Mar 2001 B1
6212512 Barney et al. Apr 2001 B1
6212521 Minami et al. Apr 2001 B1
6230164 Rikieta et al. May 2001 B1
6260069 Anglin Jul 2001 B1
6269431 Dunham Jul 2001 B1
6275953 Vahalia et al. Aug 2001 B1
6279078 Sicola et al. Aug 2001 B1
6292783 Rohler Sep 2001 B1
6301592 Aoyama et al. Oct 2001 B1
6304880 Kishi Oct 2001 B1
6311193 Sekido et al. Oct 2001 B1
6324581 Xu et al. Nov 2001 B1
6328766 Long Dec 2001 B1
6330570 Crighton Dec 2001 B1
6330642 Carteau Dec 2001 B1
6343324 Hubis et al. Jan 2002 B1
6350199 Williams et al. Feb 2002 B1
RE37601 Eastridge et al. Mar 2002 E
6353878 Dunham Mar 2002 B1
6356801 Goodman et al. Mar 2002 B1
6363464 Mangione Mar 2002 B1
6366986 St. Pierre et al. Apr 2002 B1
6366988 Skiba et al. Apr 2002 B1
6374336 Peters et al. Apr 2002 B1
6374363 Wu et al. Apr 2002 B1
6389432 Pothapragada et al. May 2002 B1
6397308 Ofek et al. May 2002 B1
6418478 Ignatius et al. Jul 2002 B1
6421711 Blumenau et al. Jul 2002 B1
6434681 Amangau Aug 2002 B1
6473775 Kusters et al. Oct 2002 B1
6487561 Ofek et al. Nov 2002 B1
6487644 Huebsch et al. Nov 2002 B1
6487645 Clark et al. Nov 2002 B1
6502205 Yanai et al. Dec 2002 B1
6519679 Devireddy et al. Feb 2003 B2
6538669 Lagueux, Jr. et al. Mar 2003 B1
6539462 Mikkelsen et al. Mar 2003 B1
6542909 Tamer et al. Apr 2003 B1
6542972 Ignatius et al. Apr 2003 B2
6564228 O'Connor May 2003 B1
6581143 Gagne et al. Jun 2003 B2
6604118 Kleinman et al. Aug 2003 B2
6604149 Deo et al. Aug 2003 B1
6615223 Shih et al. Sep 2003 B1
6631477 LeCrone et al. Oct 2003 B1
6631493 Ottesen et al. Oct 2003 B2
6643671 Milillo et al. Nov 2003 B2
6647396 Parnell et al. Nov 2003 B2
6647473 Golds et al. Nov 2003 B1
6651075 Kusters et al. Nov 2003 B1
6654825 Clapp et al. Nov 2003 B2
6658436 Oshinsky et al. Dec 2003 B2
6658526 Nguyen et al. Dec 2003 B2
6662198 Satyanarayanan et al. Dec 2003 B2
6665815 Goldstein et al. Dec 2003 B1
6681230 Blott et al. Jan 2004 B1
6721767 De Meno et al. Apr 2004 B2
6728733 Tokui Apr 2004 B2
6728736 Hostetter et al. Apr 2004 B2
6732124 Koseki et al. May 2004 B1
6732125 Autrey et al. May 2004 B1
6742092 Huebsch et al. May 2004 B1
6748504 Sawdon et al. Jun 2004 B2
6751635 Chen et al. Jun 2004 B1
6757794 Cabrera et al. Jun 2004 B2
6760723 Oshinsky et al. Jul 2004 B2
6763351 Subramaniam et al. Jul 2004 B1
6789161 Blendermann et al. Sep 2004 B1
6792518 Armangau et al. Sep 2004 B2
6799258 Linde Sep 2004 B1
6836779 Poulin Dec 2004 B2
6871163 Hiller et al. Mar 2005 B2
6871271 Ohran et al. Mar 2005 B2
6880051 Timpanaro-Perrotta Apr 2005 B2
6886020 Zahavi et al. Apr 2005 B1
6892211 Hitz et al. May 2005 B2
6898688 Martin et al. May 2005 B2
6912482 Kaiser Jun 2005 B2
6925512 Louzoun et al. Aug 2005 B2
6938135 Kekre et al. Aug 2005 B1
6938180 Dysert et al. Aug 2005 B1
6941393 Secatch Sep 2005 B2
6944796 Joshi et al. Sep 2005 B2
6948038 Berkowitz et al. Sep 2005 B2
6948089 Fujibayashi Sep 2005 B2
6952705 Knoblock et al. Oct 2005 B2
6952758 Chron et al. Oct 2005 B2
6954834 Slater et al. Oct 2005 B2
6968351 Butterworth Nov 2005 B2
6973553 Archibald, Jr. et al. Dec 2005 B1
6978265 Schumacher Dec 2005 B2
6981177 Beattie Dec 2005 B2
6983351 Gibble et al. Jan 2006 B2
6993539 Federwisch et al. Jan 2006 B2
7003519 Biettron et al. Feb 2006 B1
7003641 Prahlad et al. Feb 2006 B2
7032131 Lubbers et al. Apr 2006 B2
7035880 Crescenti et al. Apr 2006 B1
7039661 Ranade May 2006 B1
7051050 Chen et al. May 2006 B2
7062761 Slavin et al. Jun 2006 B2
7065538 Aronoff et al. Jun 2006 B2
7082441 Zahavi et al. Jul 2006 B1
7085787 Beier et al. Aug 2006 B2
7085904 Mizuno et al. Aug 2006 B2
7096315 Takeda et al. Aug 2006 B2
7103731 Gibble et al. Sep 2006 B2
7103740 Colgrove et al. Sep 2006 B1
7107298 Prahlad et al. Sep 2006 B2
7107395 Ofek et al. Sep 2006 B1
7120757 Tsuge Oct 2006 B2
7130860 Pachet Oct 2006 B2
7130970 Devassy et al. Oct 2006 B2
7139887 Colgrove et al. Nov 2006 B2
7139932 Watanabe Nov 2006 B2
7155465 Lee et al. Dec 2006 B2
7155633 Tuma et al. Dec 2006 B2
7165079 Chen et al. Jan 2007 B1
7174352 Kleinman et al. Feb 2007 B2
7181477 Saika et al. Feb 2007 B2
7188292 Cordina et al. Mar 2007 B2
7191198 Asano et al. Mar 2007 B2
7194454 Hansen et al. Mar 2007 B2
7203807 Urabe et al. Apr 2007 B2
7209972 Ignatius et al. Apr 2007 B1
7225204 Manley et al. May 2007 B2
7225208 Midgley et al. May 2007 B2
7225210 Guthrie, II May 2007 B2
7228456 Lecrone et al. Jun 2007 B2
7231544 Tan et al. Jun 2007 B2
7234115 Sprauve et al. Jun 2007 B1
7237075 Welsh et al. Jun 2007 B2
7246140 Therrien et al. Jul 2007 B2
7246207 Kottomtharayil et al. Jul 2007 B2
7257689 Baird Aug 2007 B1
7269612 Devarakonda et al. Sep 2007 B2
7269641 Powers et al. Sep 2007 B2
7272606 Borthakur et al. Sep 2007 B2
7275138 Saika Sep 2007 B2
7275177 Armangau et al. Sep 2007 B2
7275277 Moskovich et al. Oct 2007 B2
7278142 Bandhole et al. Oct 2007 B2
7287047 Kavuri Oct 2007 B2
7293133 Colgrove et al. Nov 2007 B1
7296125 Ohran Nov 2007 B2
7315923 Retnamma et al. Jan 2008 B2
7343356 Prahlad et al. Mar 2008 B2
7343365 Farnham et al. Mar 2008 B2
7343453 Prahlad et al. Mar 2008 B2
7343459 Prahlad et al. Mar 2008 B2
7346623 Prahlad et al. Mar 2008 B2
7346751 Prahlad et al. Mar 2008 B2
7356657 Mikami Apr 2008 B2
7359917 Winter et al. Apr 2008 B2
7370232 Safford May 2008 B2
7373364 Chapman May 2008 B1
7380072 Kottomtharayil et al. May 2008 B2
7383538 Bates et al. Jun 2008 B2
7389311 Crescenti et al. Jun 2008 B1
7392360 Aharoni et al. Jun 2008 B1
7395282 Crescenti et al. Jul 2008 B1
7409509 Devassy et al. Aug 2008 B2
7412583 Burton et al. Aug 2008 B2
7415488 Muth et al. Aug 2008 B1
7421554 Colgrove et al. Sep 2008 B2
7430587 Malone et al. Sep 2008 B2
7433301 Akahane et al. Oct 2008 B2
7440982 Lu et al. Oct 2008 B2
7454569 Kavuri et al. Nov 2008 B2
7467167 Patterson Dec 2008 B2
7467267 Mayock Dec 2008 B1
7472238 Gokhale Dec 2008 B1
7475284 Koike Jan 2009 B2
7484054 Kottomtharayil et al. Jan 2009 B2
7490207 Amarendran Feb 2009 B2
7496589 Jain et al. Feb 2009 B1
7496690 Beverly et al. Feb 2009 B2
7500053 Kavuri et al. Mar 2009 B1
7500150 Sharma et al. Mar 2009 B2
7509316 Greenblatt et al. Mar 2009 B2
7512601 Cucerzan et al. Mar 2009 B2
7516088 Johnson et al. Apr 2009 B2
7519726 Palliyll et al. Apr 2009 B2
7523483 Dogan Apr 2009 B2
7529748 Wen et al. May 2009 B2
7529782 Prahlad et al. May 2009 B2
7532340 Koppich et al. May 2009 B2
7536291 Retnamma et al. May 2009 B1
7539707 Prahlad et al. May 2009 B2
7539835 Kaiser May 2009 B2
7543125 Gokhale Jun 2009 B2
7546324 Prahlad et al. Jun 2009 B2
7565572 Yamasaki Jul 2009 B2
7567991 Armangau et al. Jul 2009 B2
7581077 Ignatius et al. Aug 2009 B2
7596586 Gokhale et al. Sep 2009 B2
7606841 Ranade Oct 2009 B1
7606844 Kottomtharayil Oct 2009 B2
7613748 Brockway et al. Nov 2009 B2
7613750 Valiyaparambil et al. Nov 2009 B2
7617253 Prahlad et al. Nov 2009 B2
7617262 Prahlad et al. Nov 2009 B2
7617541 Plotkin et al. Nov 2009 B2
7620666 Root et al. Nov 2009 B1
7627598 Burke Dec 2009 B1
7627617 Kavuri et al. Dec 2009 B2
7636743 Erofeev Dec 2009 B2
7651593 Prahlad et al. Jan 2010 B2
7661028 Erofeev Feb 2010 B2
7668798 Scanlon et al. Feb 2010 B2
7669029 Mishra et al. Feb 2010 B1
7685126 Patel et al. Mar 2010 B2
7689467 Belanger et al. Mar 2010 B1
7702533 Barnard et al. Apr 2010 B2
7716171 Kryger May 2010 B2
7734578 Prahlad et al. Jun 2010 B2
7734715 Hyakutake et al. Jun 2010 B2
7739235 Rousseau et al. Jun 2010 B2
7810067 Kaelicke et al. Oct 2010 B2
7831553 Prahlad et al. Nov 2010 B2
7831622 Prahlad et al. Nov 2010 B2
7840533 Prahlad et al. Nov 2010 B2
7840537 Gokhale et al. Nov 2010 B2
7844577 Becker et al. Nov 2010 B2
7870355 Erofeev Jan 2011 B2
7930476 Castelli et al. Apr 2011 B1
7962455 Erofeev Jun 2011 B2
7962709 Agrawal Jun 2011 B2
8024294 Kottomtharayil Sep 2011 B2
8121983 Prahlad et al. Feb 2012 B2
8166263 Prahlad Apr 2012 B2
8190565 Prahlad et al. May 2012 B2
8204859 Ngo Jun 2012 B2
8219524 Gokhale Jul 2012 B2
8271830 Erofeev Sep 2012 B2
8352422 Prahlad et al. Jan 2013 B2
8463751 Kottomtharayil Jun 2013 B2
8489656 Erofeev Jul 2013 B2
8504517 Agrawal Aug 2013 B2
20010029512 Oshinsky et al. Oct 2001 A1
20010029517 De Meno et al. Oct 2001 A1
20010032172 Moulinet et al. Oct 2001 A1
20010042222 Kedem et al. Nov 2001 A1
20010044807 Kleiman et al. Nov 2001 A1
20020002557 Straube et al. Jan 2002 A1
20020004883 Nguyen et al. Jan 2002 A1
20020023051 Kunzle et al. Feb 2002 A1
20020040376 Yamanaka et al. Apr 2002 A1
20020042869 Tate et al. Apr 2002 A1
20020049626 Mathias et al. Apr 2002 A1
20020049718 Kleiman et al. Apr 2002 A1
20020049738 Epstein Apr 2002 A1
20020049778 Bell et al. Apr 2002 A1
20020062230 Morag et al. May 2002 A1
20020069324 Gerasimov et al. Jun 2002 A1
20020083055 Pachet et al. Jun 2002 A1
20020103848 Giacomini et al. Aug 2002 A1
20020107877 Whiting et al. Aug 2002 A1
20020112134 Ohran et al. Aug 2002 A1
20020133511 Hostetter et al. Sep 2002 A1
20020133512 Milillo et al. Sep 2002 A1
20020161753 Inaba et al. Oct 2002 A1
20020174107 Poulin Nov 2002 A1
20020174416 Bates et al. Nov 2002 A1
20030018657 Monday Jan 2003 A1
20030028736 Berkowitz et al. Feb 2003 A1
20030061491 Jaskiewicz et al. Mar 2003 A1
20030097296 Putt May 2003 A1
20030131278 Fujibayashi Jul 2003 A1
20030135783 Martin et al. Jul 2003 A1
20030167380 Green et al. Sep 2003 A1
20030177149 Coombs Sep 2003 A1
20030177321 Watanabe Sep 2003 A1
20030187847 Lubbers et al. Oct 2003 A1
20030225800 Kavuri Dec 2003 A1
20040015468 Beier et al. Jan 2004 A1
20040039679 Norton et al. Feb 2004 A1
20040098425 Wiss et al. May 2004 A1
20040107199 Dairymple et al. Jun 2004 A1
20040117572 Welsh et al. Jun 2004 A1
20040139128 Becker et al. Jul 2004 A1
20040193953 Callahan et al. Sep 2004 A1
20040205206 Naik et al. Oct 2004 A1
20040230829 Dogan et al. Nov 2004 A1
20040236958 Teicher et al. Nov 2004 A1
20040249883 Srinivasan et al. Dec 2004 A1
20040250033 Prahlad et al. Dec 2004 A1
20040254919 Giuseppini Dec 2004 A1
20040260678 Verbowski et al. Dec 2004 A1
20040267835 Zwilling et al. Dec 2004 A1
20040267836 Armangau et al. Dec 2004 A1
20050027892 McCabe et al. Feb 2005 A1
20050033800 Kavuri et al. Feb 2005 A1
20050080928 Beverly et al. Apr 2005 A1
20050108292 Burton et al. May 2005 A1
20050138306 Panchbudhe et al. Jun 2005 A1
20050144202 Chen Jun 2005 A1
20050187982 Sato Aug 2005 A1
20050187992 Prahlad et al. Aug 2005 A1
20050188254 Urabe et al. Aug 2005 A1
20050193026 Prahlad et al. Sep 2005 A1
20050198083 Saika et al. Sep 2005 A1
20050246376 Lu et al. Nov 2005 A1
20050246510 Retnamma et al. Nov 2005 A1
20060005048 Osaki et al. Jan 2006 A1
20060010154 Prahlad et al. Jan 2006 A1
20060010227 Atluri Jan 2006 A1
20060020616 Hardy et al. Jan 2006 A1
20060047805 Byrd et al. Mar 2006 A1
20060120401 Harada et al. Jun 2006 A1
20060129537 Torii et al. Jun 2006 A1
20060136685 Griv et al. Jun 2006 A1
20060242371 Shono et al. Oct 2006 A1
20060242489 Brockway et al. Oct 2006 A1
20070043956 El Far et al. Feb 2007 A1
20070094467 Yamasaki Apr 2007 A1
20070100867 Celik et al. May 2007 A1
20070112897 Asano et al. May 2007 A1
20070113006 Elliott et al. May 2007 A1
20070124347 Vivian et al. May 2007 A1
20070124348 Claborn et al. May 2007 A1
20070143371 Kottomtharayil Jun 2007 A1
20070143756 Gokhale Jun 2007 A1
20070179990 Zimran et al. Aug 2007 A1
20070183224 Erofeev Aug 2007 A1
20070185937 Prahlad et al. Aug 2007 A1
20070185938 Prahlad et al. Aug 2007 A1
20070185939 Prahlad et al. Aug 2007 A1
20070185940 Prahlad et al. Aug 2007 A1
20070186068 Agrawal Aug 2007 A1
20070198602 Ngo et al. Aug 2007 A1
20070226438 Erofeev Sep 2007 A1
20070244571 Wilson et al. Oct 2007 A1
20070288536 Sen et al. Dec 2007 A1
20080016293 Saika Jan 2008 A1
20080028009 Ngo Jan 2008 A1
20080059515 Fulton Mar 2008 A1
20080103916 Camarador et al. May 2008 A1
20080183775 Prahlad et al. Jul 2008 A1
20080229037 Bunte et al. Sep 2008 A1
20080243914 Prahlad et al. Oct 2008 A1
20080243957 Prahlad et al. Oct 2008 A1
20080243958 Prahlad et al. Oct 2008 A1
20080244177 Crescenti et al. Oct 2008 A1
20080244205 Amano et al. Oct 2008 A1
20080306954 Hornqvist Dec 2008 A1
20090044046 Yamasaki Feb 2009 A1
20090150462 McClanahan et al. Jun 2009 A1
20090182963 Prahlad et al. Jul 2009 A1
20090187944 White et al. Jul 2009 A1
20090319534 Gokhale Dec 2009 A1
20090319585 Gokhale Dec 2009 A1
20100005259 Prahlad Jan 2010 A1
20100049753 Prahlad et al. Feb 2010 A1
20100094808 Erofeev Apr 2010 A1
20100100529 Erofeev Apr 2010 A1
20100122053 Prahlad et al. May 2010 A1
20100131461 Prahlad et al. May 2010 A1
20100145909 Ngo Jun 2010 A1
20100179941 Agrawal et al. Jul 2010 A1
20100205150 Prahlad et al. Aug 2010 A1
20110066599 Prahlad et al. Mar 2011 A1
20110246416 Prahlad et al. Oct 2011 A1
20110246429 Prahlad et al. Oct 2011 A1
20110295804 Erofeev Dec 2011 A1
20110295806 Erofeev Dec 2011 A1
20120317074 Ngo Dec 2012 A1
20130006926 Erofeev Jan 2013 A1
20130006938 Prahlad et al. Jan 2013 A1
Foreign Referenced Citations (33)
Number Date Country
2006331932 Dec 2006 AU
2632935 Dec 2006 CA
0259912 Mar 1988 EP
0405926 Jan 1991 EP
0467546 Jan 1992 EP
0774715 May 1997 EP
0809184 Nov 1997 EP
0862304 Sep 1998 EP
0899662 Mar 1999 EP
0981090 Feb 2000 EP
1174795 Jan 2002 EP
1349089 Jan 2003 EP
1349088 Oct 2003 EP
1579331 Sep 2005 EP
2256952 Dec 1992 GB
2411030 Aug 2005 GB
05189281 Jul 1993 JP
06274605 Sep 1994 JP
09016463 Jan 1997 JP
11259348 Sep 1999 JP
200347811 Dec 2000 JP
WO 9303549 Feb 1993 WO
WO 9513580 May 1995 WO
WO 9839707 Sep 1998 WO
WO 9912098 Mar 1999 WO
WO 9914692 Mar 1999 WO
WO 02095632 Nov 2002 WO
WO 03028183 Apr 2003 WO
WO 2004034197 Apr 2004 WO
WO 2005055093 Jun 2005 WO
WO 2005086032 Sep 2005 WO
WO 2007053314 May 2007 WO
WO 2007075587 Jul 2007 WO
Non-Patent Literature Citations (44)
Entry
U.S. Appl. No. 12/433,238, filed Apr. 30, 2009, Prahlad et al.
Armstead et al., “Implementation of a Campus-Wide Distributed Mass Storage Service: The Dream vs. Reality,” IEEE, 1995, pp. 190-199.
Arneson, “Development of Omniserver; Mass Storage Systems,” Control Data Corporation, 1990, pp. 88-93.
Arneson, “Mass Storage Archiving in Network Environments” IEEE, 1998, pp. 45-50.
Ashton, et al., “Two Decades of policy-based storage management for the IBM mainframe computer”, www.research.ibm.com, 19 pages, published Apr. 10, 2003, printed Jan. 3, 2009, www.research.ibm.com, Apr. 10, 2003, pp. 19.
Cabrera, et al. “ADSM: A Multi-Platform, Scalable, Back-up and Archive Mass Storage System,” Digest of Papers, Compcon '95, Proceedings of the 40th IEEE Computer Society International Conference, Mar. 5-9, 1995, pp. 420-427, San Francisco, CA.
Canadian Office Action dated Dec. 29, 2010, Application No. CA2546304.
Canadian Office Action dated Dec. 10, 2010, Application No. CA2544063.
Eitel, “Backup and Storage Management in Distributed Heterogeneous Environments,” IEEE, 1994, pp. 124-126.
European Examination Report; Application No. 06848901.2, Apr. 1, 2009, pp. 7.
Examiner's First Report; Application No. 2006331932 May 11, 2011 in 2 pages.
Examiner's Report for Australian Application No. 2003279847, Dated Dec. 9, 2008, Examiner Pahwa, 4 pages.
Final Office Action for Japanese Application No. 2003531581, Examiner Kudoh, Mail Date Mar. 24, 2009, 6 pages.
First Office Action for Japanese Application No. 2003531581, Mail Date Jul. 8, 2008, Examiner Kudoh, 8 pages.
Gait, “The Optical File Cabinet: A Random-Access File system for Write-Once Optical Disks,” IEEE Computer, vol. 21, No. 6, pp. 11-22 (1988).
Gray, et al. “Transaction processing: concepts and techniques” 1994, Morgan Kaufmann Publishers, USA, pp. 604-609, 646-655.
Harrington, Lisa H., “The RFP Process: How To Hire a Third Party”, Transportation & Distribution, Sep. 1988, vol. 39, Issue 9, in 5 pages.
http://en.wikipedia.org/wiki/Naive—Bayes—classifier, printed on Jun. 1, 2010, in 7 pages.
International Preliminary Report on Patentability, PCT Application No. PCT/US2009/066880, mailed Jun. 23, 2011, in 9 pages.
International Search Report and Written Opinion dated Jan. 11, 2006 , PCT/US2004/038455.
International Search Report and Written Opinion dated Mar. 25, 2010, PCT/US2009/066880.
International Search Report and Written Opinion dated Nov. 13, 2009, PCT/US2007/081681.
International Search Report and Written Opinion issued in PCT Application No. PCT/US2011/030396, mailed Jul. 18, 2011, in 20 pages.
International Search Report and Written Opinion issued in PCT Application No. PCT/US2011/38436, mailed Sep. 21, 2011, in 18 pages.
International Search Report dated Dec. 28, 2009, PCT/US2004/038324.
International Search Report from International Application No. PCT/US2006/048273, dated May 15, 2007.
Jander, “Launching Storage-Area Net,” Data Communications, US, McGraw Hill, NY, vol. 27, No. 4(Mar. 21, 1998), pp. 64-72.
Kashyap, et al., “Professional Services Automation: A knowledge Management approach using LSI and Domain specific Ontologies”, FLAIRS-01 Proceedings, 2001, pp. 300-302.
Lyon J., Design considerations in replicated database systems for disaster protection, COMPCON 1988, Feb. 29, 1988, pp. 428-430.
Microsoft Corporation, “Microsoft Exchange Server: Best Practices for Exchange Database Management,” 1998.
Non-Final Office Action for U.S. Appl. No. 10/681,386, Mailed Oct. 28, 2008, 16 pages.
Non-Final Office Action for U.S. Appl. No. 10/990,353, Mailed Sep. 15, 2008, 8 pages.
Non-Final Office Action for U.S. Appl. No. 11/672,926, Mailed Nov. 25, 2008, 16 pages.
Notice of Allowance for U.S. Appl. No. 10/681,386 Mailed Apr. 21, 2009, 10 pages.
Notice of Allowance for U.S. Appl. No. 10/990,353 Mailed Apr. 7, 2009, 11 pages.
Rosenblum et al., “The Design and Implementation of a Log-Structure File System,” Operating Systems Review SIGOPS, vol. 25, No. 5, New York, US, pp. 1-15 (May 1991).
Second Examination Report in EU Appl. No. 06 848 901.2-2201 dated Dec. 3, 2010.
The Oracle8 Replication Manual, Part No. A58245-01; Chapters 1-2; Dec. 1, 1997; obtained from website: http://download-west.oracle.com/docs/cd/A64702—01/doc/server.805/a58245/toc.htm on May 20, 2009.
Veritas Software Corporation, “Veritas Volume Manager 3.2, Administrator's Guide,” Aug. 2001, 360 pages.
Wiesmann M., Database replication techniques: a three parameter classification, Oct. 16, 2000, pp. 206-215.
First Office Action in Indian Application No. 3359/DELNP/2006 dated Feb. 11, 2013.
U.S. Appl. No. 13/887,103, filed May 3, 2013, Kottomtharayil.
U.S. Appl. No. 13/955,445, filed Jul. 31, 2013, Prahlad et al.
U.S. Appl. No. 13/957,824, filed Aug. 2, 2013, Agrawal.
Related Publications (1)
Number Date Country
20130006942 A1 Jan 2013 US
Provisional Applications (2)
Number Date Country
60519876 Nov 2003 US
60519576 Nov 2003 US
Continuations (3)
Number Date Country
Parent 12951773 Nov 2010 US
Child 13480321 US
Parent 12433238 Apr 2009 US
Child 12951773 US
Parent 10990353 Nov 2004 US
Child 12433238 US