File system virtualization

Information

  • Patent Grant
  • 9785647
  • Patent Number
    9,785,647
  • Date Filed
    Tuesday, October 2, 2012
    11 years ago
  • Date Issued
    Tuesday, October 10, 2017
    6 years ago
Abstract
Systems, methods, and media file system virtualization. Exemplary methods may include rendering a sparse file system and creating a block map for the sparse file system. The block map may include a map record for each file record included in the sparse file system. A map record may include extents for blocks associated with a file record, the map record providing a mapping between allocated block sectors for a file record in the sparse file system and corresponding blocks associated with one or more data sources.
Description
FIELD OF THE INVENTION

The present technology may be generally described as providing systems and methods for file system virtualization. More specifically, but not by way of limitation, the present technology may allow for the creation of virtual disks that are utilizable by virtual machines. These virtual disks may comprise a sparse file system that is mountable by a virtual machine. That is, the file system includes allocated block sectors that correspond to blocks of file records. These allocated block sectors are not populated with actual block data but remain sparse. The present technology may provide access to actual block data of files using a block map that provides a mapping between the allocated block sectors of the sparse file system and actual blocks of files in a data source(s). Thus, a sparse file system may be used by a virtual machine to obtain actual data blocks (e.g., files) without requiring the virtual machine to create a fully populated file system.


BACKGROUND

Customer data for physical systems may be backed up to a backing store, as is common with remote backup systems. Mirrors and incremental files may be captured periodically. These backup files may be used to recreate files of the physical system at a later point in time. Backup files captured over time may be used to recreate the file system of the physical system at an arbitrary point in time in the past (e.g., prior to the failure event).


According to known methods, to access blocks of files from a backing store, a physical or virtual disk is partitioned and formatted. Files from a backing store are then populated onto the disk. The disk is then mounted for use. Therefore, when an end user desires to obtain only a part of a file, or even a selected file or group of files, the end user recreates the entire physical disk, which is an onerous task given the desired outcome.


SUMMARY OF THE PRESENT TECHNOLOGY

According to some embodiments, the present technology may be directed to methods that comprise: (a) rendering a sparse file system; and (b) creating a block map for the sparse file system, the block map comprising mappings between allocated block sectors of file records of the sparse file system and corresponding disperse blocks stored in a data source.


In some embodiments, the present technology may be directed to systems that comprise: (a) at least one server comprising: a memory that includes executable instructions and a processor executing the instructions, the instructions comprising: (i) a rendering module that renders a sparse file system; and (ii) an extent mapper that creates a block map for the sparse file system, the block map comprising a map record for each file record included in the sparse file system, a map record comprising extents for blocks associated with a file record, the map record providing a mapping between allocated block sectors for a file record and corresponding blocks associated with a data source.


According to some embodiments, the present technology may be directed to methods that comprise: (a) requesting blocks from a virtual disk, the blocks corresponding to a file that is to be assembled; and (b) assembling the file by: (i) evaluating a block map to obtain the requested blocks from any of a data source and a backup data source, the block map providing a mapping between allocated block sectors of the sparse file system and corresponding block data in the data source; (ii) for blocks not included in the block map, obtaining blocks not included in the block map from a sparse file system directly; and (iii) assembling the obtained blocks to assemble the file (iii) assembling the obtained blocks to assemble the file.


According to some embodiments, the present technology may be directed to a non-transitory machine-readable storage medium having embodied thereon a program. In some embodiments the program may be executed by a machine to perform a method. The method may comprise: (a) requesting blocks from a virtual disk, the blocks corresponding to a file that is to be assembled; and (b) assembling the file by: (i) evaluating a block map to obtain the requested blocks from a data source, the block map providing a mapping between allocated block sectors of the sparse file system and corresponding block data in the data source; (ii) for blocks not included in the block map, obtaining blocks not included in the block map from a sparse file system directly; and (iii) assembling the obtained blocks to assemble the file.





BRIEF DESCRIPTION OF THE DRAWINGS

Certain embodiments of the present technology are illustrated by the accompanying figures. It will be understood that the figures are not necessarily to scale and that details not necessary for an understanding of the technology or that render other details difficult to perceive may be omitted. It will be understood that the technology is not necessarily limited to the particular embodiments illustrated herein.



FIG. 1 is a block diagram of an exemplary architecture in which embodiments of the present technology may be practiced;



FIG. 2 is a flowchart of an exemplary method of file system virtualization.



FIG. 3 is a flowchart of an exemplary method for assembling a file using a spares file system and block map.



FIG. 4 illustrates an exemplary computing system that may be used to implement embodiments according to the present technology.





DESCRIPTION OF EXEMPLARY EMBODIMENTS

While this technology is susceptible of embodiment in many different forms, there is shown in the drawings and will herein be described in detail several specific embodiments with the understanding that the present disclosure is to be considered as an exemplification of the principles of the technology and is not intended to limit the technology to the embodiments illustrated.


It will be understood that like or analogous elements and/or components, referred to herein, may be identified throughout the drawings with like reference characters. It will be further understood that several of the figures are merely schematic representations of the present technology. As such, some of the components may have been distorted from their actual scale for pictorial clarity.


Generally speaking, the present technology allows for efficient retrieval of blocks from data sources such as files and/or backing stores without the need to recreate an entire file system that includes the blocks. More specifically, but not by way of limitation, the present technology allows virtual or physical machines to request blocks from various data sources.


Rather than recreating the entire file system, a sparse file system may be created from known metadata that includes extents for blocks. The extents for a block may comprise a file, offset, and/or length (e.g., size). The blocks may be associated with file records for files. In some embodiments, the metadata of the backing store contains two types of information, full mirror files and incremental file fragments. The metadata maps to both of these types of files.


The sparse file system may be created from a spare file that has been formatted with a particular file system such as new technology file system (NTFS), although other file system types that would be known to one of ordinary skill in the art may also likewise be utilized in accordance with the present technology.


Using the metadata that comprises extents for the blocks, spaces for file records may be allocated to the sparse file system in such a way that appropriate spaces for file records are established in the file system without actually writing the data blocks for the file records into the file system. Stated otherwise, the sparse file system may comprise allocated block sectors that remain sparse (e.g., do not include actual data).


The present technology also creates a block map using the metadata. The block map provides a mapping between allocated block sectors of the sparse file system and the actual (e.g., corresponding) blocks stored in one or more data sources. In some instances the data sources may comprise backup data sources. Therefore, corresponding blocks may be stored in mirrors (also referred to as core files or snapshots) or incremental files that comprise the backup data source. It is noteworthy that in some embodiments, the backup data source may comprise distributed hash tables that are provisioned within a cloud.


When data blocks are requested by a physical or virtual machine, in lieu of writing all data blocks directly into the file system, the block map may be consulted to obtain the desired blocks from various data sources such as files and/or backup trees. Advantageously, blocks representing a file or a portion of a file, as they existed at an arbitrary point in time, may be obtained without needing to recreate the entire file system to which the blocks and files belong. These and other advantages of the present technology will be discussed in greater detail herein.


Referring now to the drawings, and more particularly, to FIG. 1, which includes a schematic diagram of an exemplary architecture 100 for practicing the present invention. Architecture 100 may include one or more data sources such as data source 105. The data source(s) 105 may comprise a file system of a computing device that includes files that comprise data blocks.


Generally speaking, the backup system 110 provides remote backup services for files that reside on a physical system such as an end user computing system (not shown). In some instances, the backup system 110 may be implemented within a cloud-based computing environment. In general, a cloud-based computing environment is a resource that typically combines the computational power of a large model of processors and/or that combines the storage capacity of a large model of computer memories or storage devices. For example, systems that provide a cloud resource may be utilized exclusively by their owners, such as Google™ or Yahoo!™; or such systems may be accessible to outside users who deploy applications within the computing infrastructure to obtain the benefit of large computational or storage resources.


The cloud may be formed, for example, by a network of servers, with each server (or at least a plurality thereof) providing processor and/or storage resources. These servers may manage workloads provided by multiple users (e.g., cloud resource consumers or other users). Typically, each user places workload demands upon the cloud that vary in real-time, sometimes dramatically. The nature and extent of these variations typically depend on the type of business associated with the user.


In greater detail, the backup system 110 may include one or more VPN devices adapted to receive information from the gateway router of the end user computing system. In some embodiments, the repository may include an Ethernet switch operatively coupling the VPN devices to a remote backup application server adapted execute at least a portion of methods for restoring files. The backup system 110 may also include disaster recovery servers, physical media input servers, and one or more virus scanning servers. It will be understood that the primary storage devices for the backup system 110 may include, for example, RAID redundant storage servers, although other types of servers that would be known to one of ordinary skill in the art with the present disclosure before them are likewise contemplated for use in accordance with the present invention.


According to some embodiments, the backup system 110 may comprise various data sources such as backup data source (or data source) 120. The backup data source 120 may comprise various files such as mirrors (also known as core files and/or snapshots), incremental files (both forward and reverse differentials), or other data structures such as distributed hash tables. The present technology may utilize any backup data source that would be known to one of ordinary skill in the art with the present disclosure before them. The backup data sources may exist on a single backup device, such as a failover appliance (which can be local or remotely positioned relative to an end user computing system) or they may be distributed over a plurality of systems which are arranged in a cloud configuration.


Additionally, the architecture 100 may comprise a virtual environment 125 that comprises a virtual machine 130. The virtual machine 130 may comprise any virtual instantiation of a computing device that utilizes any desired operating system. One of ordinary skill in the art would be well versed in various functional and operational details regarding the use of a virtual machine. Therefore, a detailed discussion regarding these features will be omitted so as not to obscure the advantages of the present technology described herein.


According to some embodiments, the architecture 100 may comprise a file system virtualization device 135. The file system virtualization device 135 may comprise a rendering module 140, an extent mapper module 145, a block retrieval module 150, and a copy on write (COW) source 155, which are each stored in memory and/or a physical storage media (e.g., disk, RAID, distributed hash table, etc.) and executable by a processor of the file system virtualization device 135. It is noteworthy that the file system virtualization device 135 may include additional or fewer modules, engines, or components, and still fall within the scope of the present technology. As used herein, the term “module” may also refer to any of an application-specific integrated circuit (“ASIC”), an electronic circuit, a processor (shared, dedicated, or group) that executes one or more software or firmware programs, a combinational logic circuit, and/or other suitable components that provide the described functionality. In other embodiments, individual modules of the file system virtualization device 135 may include separately configured servers. The file system virtualization device 135 may also be integrated within the virtual environment 125.


Prior to the virtual machine 130 requesting blocks from a virtual disk, such as virtual disk 160, the file system virtualization device 135 may be utilized to create a sparse file system and a block map for the virtual disk 160. Traditionally, data sources may contain files that are comprised of actual blocks of data. Additionally, blocks of data may be obtained from mirrors and/or incremental files, such as those stored on the backup data source 120. Files may be recreated from the mirrors and/or incremental files as they existed at an arbitrary point in time in the past. Using the obtained mirrors and/or increments, a file system may be created by formatting a target disk image with a file system, such as NTFS. Again, the target disk image may be provided on either a physical or virtual disk, which may be provisioned with a cloud.


After formatting the target disk image, the files may be walked to rebuild the file system of the physical device. Additional details regarding systems and methods for restoring a file are included in U.S. patent application Ser. No. 12/895,275, filed Sep. 30, 2010, entitled “Systems and Methods for Restoring a File,” now U.S. Pat. No. 8,924,360, issued Dec. 30, 2014, which is hereby incorporated by reference herein in its entirety including all reference cited therein. Advantageously, rather than using traditional methods for obtaining blocks or files, the present technology may create sparse file systems and corresponding block maps, also referred to as extent maps.


The rendering module 140 may be executed to create a new file system in a sparse file. Again, this file system could include NTFS or any other file system desired. Generally, the sparse file may be created from an analysis of metadata obtained from about the data source 105 and/or the backup data source 120. The metadata may be utilized to lay out a file system and extents for blocks (e.g., allocated block sectors) within the file system. An extent for a block may be understood to include an identification of the file to which the block belongs, an offset, and a length (e.g., size) for a particular block of data. Again, the block of data may be part of a file record. Moreover, a file record may comprise a plurality of blocks. It will be understood that most file systems utilize specific sector sizes and block sizes, such that blocks rarely cross boundaries within the file system.


In contrast with traditional methods, the rendering module 140 may not write actual data into the sparse file system. The rendering module 140 may allocate block sectors for blocks of file records which are to be included in the file system. The allocation of block sectors for blocks of each file record that is to be included in the file system, using metadata obtained from the backup data source 120, produces a sparse file system. The sparse file system resides on the virtual disk 160. While the target disk image has been described as residing in a virtual disk, the target disk image may also reside on a physical disk.


Additionally, the extent mapper module 145 may be executed to create a block map for the sparse file system. Generally, the block map is used to map allocated block sectors included in the sparse file system to their actual blocks stored in the data source 105, for example in files. Thus, the allocated block sectors of the sparse file system are “sparse” inasmuch as the blocks are merely referenced by their extents in the sparse file system. That is, the sparse file system knows locations of blocks based upon their extents, but the actual data included in those blocks remains sparse or unknown to the file system. The allocated block sectors of the sparse file system are functionally placeholders that represent the location and size of actual block data.


The block map links the actual blocks of data on the data source 105 to the sparse file system using the same extent metadata used to create the sparse file system. More specifically, the block map may comprise a plurality of map records that correspond to file records. The file records correspond to files that are to be included in the sparse file system. Thus, a map record contains a mapping of blocks for a specific file record to corresponding blocks stored in the data source 105.


Once the sparse file system and the block map have been created, the virtual disk 160, which includes the sparse virtual file system, may be mounted by the virtual machine 130. Even though the sparse system file does not include actual block data, the virtual machine 130 may utilize the virtual disk 160 in the same manner as a traditional disk that includes all corresponding block data.


When the virtual machine 130 desires to obtain blocks from virtual disk 160, for example when restoring a file, a COW source 155 may first be used to process the request. Initially, partially written requests may be fulfilled by the COW source 155. If the COW source 155 is unable to fulfill the block request, the block request may be subsequently passed to the block retrieval module 150 for fulfillment.


The block retrieval module 150 may receive the block request and read the block map to determine mappings for the requested blocks. Once the mappings have been determined, the block retrieval module 150 may obtain the blocks from the appropriately mapped backup data source(s), such as the data source 105.


For blocks that do not fall within the block map, the block retrieval module 150 may obtain the block data from the sparse file system and may in some instances include file system metadata. According to some embodiments, the metadata associated with the backup data source 120 may be evaluated to determine the location of blocks included in the request.


Blocks obtained from the data source 105 may be assembled together to recreate a file, or at least a portion of a file. Additionally, these blocks obtained from the data source 105 may be combined with blocks obtained from the backup data source 120. The assembled blocks may be served by the virtual machine 130 without the requirement to move the data into the sparse file system, which may result in the mutation of the block data.


Although not shown, the virtual machine 130 may utilize a virtual device driver to open files, cache data, and serve block requests to provide nearly instant failover for the data source 105 using a relatively small virtual disk creation operation as described above.



FIG. 2 is a flowchart of an exemplary method 200 of file system virtualization. The method 200 may comprise a step 205 of rendering a sparse file system according to the processes previously described. The method may also comprise a step 210 of creating a block map for the sparse file system. Again, the block map may comprise mappings between allocated block sectors of file records of the sparse file system and corresponding disperse blocks stored in a data source.


Next, the method may comprise a step 215 of receiving a request for blocks, which may comprise a request for a file from a virtual disk. Again, a file is comprised of one or more blocks of data. In some instances, the request may explicitly request data blocks from a particular location on the virtual disk.


In response to the request, the method may comprise a step 220 of determining if the requested blocks have been at least partially written. If so, the method may comprise a step 225 of fulfilling the block request using a COW source. If the requested blocks have not been partially written, the method may comprise a step 230 of determining if the requested blocks are included in the block map.


If the requested blocks are included in the block map, method may comprise a step 235 of reading the block map for the sparse file system to obtain the requested blocks from a data source(s). If the requested blocks are not included in the block map, the method may comprise a step 240 of obtaining the requested blocks from the sparse file system. The method may comprise a step 245 of assembling the requested blocks to fulfill the request.



FIG. 3 is a flowchart of an exemplary method 300 for assembling a file using a sparse file system and a block map. The method may comprise a step 305 of requesting blocks from a sparse file system. As mentioned previously, the blocks may correspond to a file that is to be assembled.


Next, the method may comprise a step 310 of assembling the file by executing a sub-step 315 of evaluating a block map to obtain the requested blocks from a data source. In some instances, the block map may provide a mapping between allocated block sectors of the sparse file system and corresponding block data in the data source.


Advantageously, for blocks not included in the block map, the method may comprise a sub-step 320 of obtaining blocks not included in the block map from the sparse file system. Thus, mapped blocks may be gathered from a data source such as a current file(s), as well as a backup data source, while unmapped blocks may be gathered from the sparse file system. In some instances, these non-mapped blocks are fulfilled by the sparse file system which contains only file system metadata blocks. In contrast, blocks that have been written at any time through the virtual disk system are stored in the COW and provided by the COW on any future request for the block. This applies to both blocks in the sparse file system and mapped blocks.


The method may also comprise a sub-step 325 of assembling the obtained blocks to assemble the file. Again, in some embodiments the assembled blocks may comprise blocks obtained from only a data source such as current files. Alternatively, the assembled blocks may comprise blocks obtained from only the backup data source. In other embodiments, the assembled blocks may comprise blocks obtained from a data source(s) and/or backup data source (s).



FIG. 4 illustrates an exemplary computing system 400 that may be used to implement an embodiment of the present technology. The computing system 400 of FIG. 4 includes one or more processors 410 and memory 420. Main memory 420 stores, in part, instructions and data for execution by processor 410. Main memory 420 can store the executable code when the system 400 is in operation. The system 400 of FIG. 4 may further include a mass storage device 430, portable storage medium drive(s) 440, output devices 450, user input devices 460, a graphics display 470, and other peripheral devices 480. The system 400 may also comprise network storage 445.


The components shown in FIG. 4 are depicted as being connected via a single bus 490. The components may be connected through one or more data transport means. Processor unit 410 and main memory 420 may be connected via a local microprocessor bus, and the mass storage device 430, peripheral device(s) 480, portable storage device 440, and graphics display 470 may be connected via one or more input/output (I/O) buses.


Mass storage device 430, which may be implemented with a magnetic disk drive or an optical disk drive, is a non-volatile storage device for storing data and instructions for use by processor unit 410. Mass storage device 430 can store the system software for implementing embodiments of the present technology for purposes of loading that software into main memory 420.


Portable storage device 440 operates in conjunction with a portable non-volatile storage medium, such as a floppy disk, compact disk or digital video disc, to input and output data and code to and from the computing system 400 of FIG. 4. The system software for implementing embodiments of the present technology may be stored on such a portable medium and input to the computing system 400 via the portable storage device 440.


Input devices 460 provide a portion of a user interface. Input devices 460 may include an alphanumeric keypad, such as a keyboard, for inputting alphanumeric and other information, or a pointing device, such as a mouse, a trackball, stylus, or cursor direction keys. Additionally, the system 400 as shown in FIG. 4 includes output devices 450. Suitable output devices include speakers, printers, network interfaces, and monitors.


Graphics display 470 may include a liquid crystal display (LCD) or other suitable display device. Graphics display 470 receives textual and graphical information, and processes the information for output to the display device.


Peripherals 480 may include any type of computer support device to add additional functionality to the computing system. Peripheral device(s) 480 may include a modem or a router.


The components contained in the computing system 400 of FIG. 4 are those typically found in computing systems that may be suitable for use with embodiments of the present technology and are intended to represent a broad category of such computer components that are well known in the art. Thus, the computing system 400 of FIG. 4 can be a personal computer, hand held computing system, telephone, mobile computing system, workstation, server, minicomputer, mainframe computer, or any other computing system. The computer can also include different bus configurations, networked platforms, multi-processor platforms, etc. Various operating systems can be used including UNIX, Linux, Windows, Macintosh OS, Palm OS, and other suitable operating systems.


Some of the above-described functions may be composed of instructions that are stored on storage media (e.g., computer-readable medium). The instructions may be retrieved and executed by the processor. Some examples of storage media are memory devices, tapes, disks, and the like. The instructions are operational when executed by the processor to direct the processor to operate in accord with the technology. Those skilled in the art are familiar with instructions, processor(s), and storage media.


It is noteworthy that any hardware platform suitable for performing the processing described herein is suitable for use with the technology. The terms “computer-readable storage medium” and “computer-readable storage media” as used herein refer to any medium or media that participate in providing instructions to a CPU for execution. Such media can take many forms, including, but not limited to, non-volatile media, volatile media and transmission media. Non-volatile media include, for example, optical or magnetic disks, such as a fixed disk. Volatile media include dynamic memory, such as system RAM. Transmission media include coaxial cables, copper wire and fiber optics, among others, including the wires that comprise one embodiment of a bus. Transmission media can also take the form of acoustic or light waves, such as those generated during radio frequency (RF) and infrared (IR) data communications. Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, a hard disk, magnetic tape, any other magnetic medium, a CD-ROM disk, digital video disk (DVD), any other optical medium, any other physical medium with patterns of marks or holes, a RAM, a PROM, an EPROM, an EEPROM, a FLASHEPROM, any other memory chip or data exchange adapter, a carrier wave, or any other medium from which a computer can read.


Various forms of computer-readable media may be involved in carrying one or more sequences of one or more instructions to a CPU for execution. A bus carries the data to system RAM, from which a CPU retrieves and executes the instructions. The instructions received by system RAM can optionally be stored on a fixed disk either before or after execution by a CPU.


It is noteworthy that various modules and engines may be located in different places in various embodiments. Modules and engines mentioned herein can be stored as software, firmware, hardware, as a combination, or in various other ways. It is contemplated that various modules and engines can be removed or included in other suitable locations besides those locations specifically disclosed herein. In various embodiments, additional modules and engines can be included in the exemplary embodiments described herein.


While various embodiments have been described above, it should be understood that they have been presented by way of example only, and not limitation. The descriptions are not intended to limit the scope of the technology to the particular forms set forth herein. Thus, the breadth and scope of a preferred embodiment should not be limited by any of the above-described exemplary embodiments. It should be understood that the above description is illustrative and not restrictive. To the contrary, the present descriptions are intended to cover such alternatives, modifications, and equivalents as may be included within the spirit and scope of the technology as defined by the appended claims and otherwise appreciated by one of ordinary skill in the art. The scope of the technology should, therefore, be determined not with reference to the above description, but instead should be determined with reference to the appended claims along with their full scope of equivalents.

Claims
  • 1. A method, comprising: storing a backup of a file system in a backup data source;receiving a request to render a portion of file records of the file system stored in the backup data source;rendering a sparse file system by allocating block sectors for blocks of the portion of the file records in a formatted sparse file, without having to render the file system entirely; andcreating a block map for the sparse file system, the block map comprising mappings between allocated block sectors of file records of the sparse file system and corresponding disperse blocks stored in all the backup data source, wherein all block sectors for blocks in the formatted sparse file are allocated without writing data associated with the file records into the allocated block sectors.
  • 2. The method according to claim 1, wherein the allocated block sectors are placeholders that represent a location and size of actual block data.
  • 3. The method according to claim 2, wherein block sectors are allocated using metadata that comprises extents for blocks included in the portion of file records.
  • 4. The method according to claim 1, wherein the block map comprises a map record for each file record included in the sparse file system, the map record comprising extents for blocks associated with a file record.
  • 5. The method according to claim 1, further comprising: receiving a request for blocks within a sector of the sparse file system; andobtaining the requested blocks by: reading the block map for the sparse file system to locate the requested blocks in any of a data source or the backup data source;obtaining the requested blocks from the data source or the backup data source;for blocks not included in the block map, obtaining blocks not included in the block map from the sparse file system; andfulfilling the request with the obtained blocks.
  • 6. The method according to claim 1, further comprising fulfilling a request for blocks using a copy on write source if the requested blocks have been partially written.
  • 7. The method according to claim 5, wherein the backup data source comprises a distributed hash table provisioned within a cloud.
  • 8. The method according to claim 5, wherein the backup data source comprises one or more of a mirror, an incremental file, and combinations thereof.
  • 9. The method according to claim 1, further comprising mounting the sparse file system as a virtual disk.
  • 10. The method according to claim 1, wherein rendering a sparse file system comprises formatting a portion of a virtual disk into a sparse file.
  • 11. The method according to claim 1, wherein the sparse file system comprises a virtual sparse file system.
  • 12. A system, the system comprising: at least one server having a memory that includes executable instructions and a processor executing the instructions, the instructions comprising: a rendering module that:receives a request to render a portion of file records of a file system stored in a backup data source; andrenders a sparse file system by allocating block sectors for blocks of the portion of the file records in a formatted sparse file, without having to render the file system entirely;an extent mapper that creates a block map for the sparse file system, the block map comprising a map record for each file record included in the sparse file system, the map record comprising extents for blocks associated with a file record, the map record providing a mapping between allocated block sectors for a file record and corresponding blocks associated with a data source, wherein all block sectors for blocks in the formatted sparse file are allocated without writing data associated with the file records into the allocated block sectors; anda block retriever module that: receives a request for blocks; andobtains the requested blocks by:reading the block map for the sparse file system to obtain the requested blocks from any of a data source and the backup data source;for blocks not included in the block map, obtaining blocks not included in the block map from the sparse file system; andfulfilling the request with the obtained blocks, wherein the block retriever module fulfills a request for blocks using copy on write source if the requested blocks have been partially written.
  • 13. The system according to claim 12, wherein the rendering module allocates block sectors for blocks of the portion of the file records in a formatted sparse file.
  • 14. The system according to claim 13, wherein the rendering module allocates block sectors using metadata that comprises extents for blocks included in the portion of the file records.
  • 15. The system according to claim 12, wherein the rendering module renders a sparse file system by formatting a portion of a virtual disk into a sparse file, wherein the sparse file system comprises a virtual sparse file system.
  • 16. The system according to claim 12, wherein the rendering module creates a sparse file system using metadata that includes extents for blocks included in the portion of the file records.
  • 17. A method, comprising: storing a backup of a file system in a backup data source;receiving a request to render a file of the file system stored in the backup data source;requesting blocks from a sparse file system by allocating block sectors for blocks of the file in a formatted sparse file, without having to render the file system entirely, the blocks corresponding to a file; andassembling the file by: evaluating a block map to obtain the requested blocks from a data source, the block map providing a mapping between allocated block sectors of the sparse file system and corresponding block data in a data source, wherein all block sectors for blocks in the formatted sparse file are allocated without writing data associated with the file into the allocated block sectors;for blocks not included in the block map, obtaining blocks not included in the block map from the sparse file system; andassembling the obtained blocks to assemble the file.
US Referenced Citations (236)
Number Name Date Kind
5379412 Eastridge et al. Jan 1995 A
5574905 deCarmo Nov 1996 A
5860107 Patel Jan 1999 A
6122629 Walker et al. Sep 2000 A
6205527 Goshey et al. Mar 2001 B1
6233589 Balcha et al. May 2001 B1
6272492 Kay Aug 2001 B1
6411985 Fujita et al. Jun 2002 B1
6604236 Draper et al. Aug 2003 B1
6629110 Cane et al. Sep 2003 B2
6651075 Kusters et al. Nov 2003 B1
6971018 Witt et al. Nov 2005 B1
7024581 Wang et al. Apr 2006 B1
7085904 Mizuno et al. Aug 2006 B2
7266655 Escabi, II et al. Sep 2007 B1
7401192 Stakutis et al. Jul 2008 B2
7406488 Stager et al. Jul 2008 B2
7546323 Timmins et al. Jun 2009 B1
7620765 Ohr et al. Nov 2009 B1
7631155 Bono Dec 2009 B1
7647338 Lazier et al. Jan 2010 B2
7676763 Rummel Mar 2010 B2
7730425 de los Reyes et al. Jun 2010 B2
7743038 Goldick Jun 2010 B1
7752487 Feeser et al. Jul 2010 B1
7769731 O'Brien Aug 2010 B2
7797582 Stager et al. Sep 2010 B1
7809688 Cisler et al. Oct 2010 B2
7832008 Kraemer Nov 2010 B1
7844850 Yasuzato Nov 2010 B2
7873601 Kushwah Jan 2011 B1
7930275 Chen et al. Apr 2011 B2
7966293 Owara et al. Jun 2011 B1
8037345 Iyer et al. Oct 2011 B1
8046632 Miwa et al. Oct 2011 B2
8060476 Afonso et al. Nov 2011 B1
8099391 Monckton Jan 2012 B1
8099572 Arora et al. Jan 2012 B1
8117163 Brown et al. Feb 2012 B2
8200926 Stringham Jun 2012 B1
8224935 Bandopadhyay et al. Jul 2012 B1
8244678 Hartland et al. Aug 2012 B1
8244914 Nagarkar Aug 2012 B1
8245156 Mouilleseaux et al. Aug 2012 B2
8260742 Cognigni et al. Sep 2012 B2
8279174 Jee et al. Oct 2012 B2
8296410 Myhill et al. Oct 2012 B1
8321688 Auradkar et al. Nov 2012 B2
8332442 Greene Dec 2012 B1
8352717 Campbell et al. Jan 2013 B2
8381133 Iwema et al. Feb 2013 B2
8402087 O'Shea et al. Mar 2013 B2
8407190 Prahlad et al. Mar 2013 B2
8412680 Gokhale et al. Apr 2013 B1
8417674 Provenzano Apr 2013 B2
8504785 Clifford et al. Aug 2013 B1
8533608 Tantiprasut Sep 2013 B1
8549432 Warner Oct 2013 B2
8554918 Douglis Oct 2013 B1
8572337 Gokhale et al. Oct 2013 B1
8589350 Lalonde et al. Nov 2013 B1
8589913 Jelvis et al. Nov 2013 B2
8600947 Freiheit et al. Dec 2013 B1
8601389 Schulz et al. Dec 2013 B2
8606752 Beatty et al. Dec 2013 B1
8639917 Ben-Shaul et al. Jan 2014 B1
8676273 Fujisaki Mar 2014 B1
8745003 Patterson Jun 2014 B1
8767593 Allen et al. Jul 2014 B1
8886611 Caputo Nov 2014 B2
8924360 Caputo Dec 2014 B1
8954544 Edwards Feb 2015 B2
9104621 Caputo Aug 2015 B1
9110964 Shilane et al. Aug 2015 B1
9213607 Lalonde et al. Dec 2015 B2
9235474 Petri et al. Jan 2016 B1
9292153 Edwards et al. Mar 2016 B1
9397907 Edwards et al. Jul 2016 B1
9559903 Edwards Jan 2017 B2
9705730 Petri et al. Jul 2017 B1
20010034737 Cane et al. Oct 2001 A1
20010056503 Hibbard Dec 2001 A1
20020124124 Matsumoto et al. Sep 2002 A1
20020169740 Korn Nov 2002 A1
20030011638 Chung Jan 2003 A1
20030158873 Sawdon et al. Aug 2003 A1
20030177293 Bilak et al. Sep 2003 A1
20030208492 Winiger et al. Nov 2003 A1
20040030852 Coombs et al. Feb 2004 A1
20040044707 Richard Mar 2004 A1
20040073560 Edwards Apr 2004 A1
20040073677 Honma et al. Apr 2004 A1
20040093474 Lin et al. May 2004 A1
20040098423 Chigusa et al. May 2004 A1
20040233924 Bilak et al. Nov 2004 A1
20040260973 Michelman Dec 2004 A1
20050010835 Childs et al. Jan 2005 A1
20050027748 Kisley Feb 2005 A1
20050114290 Borthakur et al. May 2005 A1
20050154937 Achiwa Jul 2005 A1
20050171979 Stager et al. Aug 2005 A1
20050223043 Randal et al. Oct 2005 A1
20050273654 Chen et al. Dec 2005 A1
20050278583 Lennert et al. Dec 2005 A1
20050278647 Leavitt et al. Dec 2005 A1
20060013462 Sadikali Jan 2006 A1
20060047720 Kulkarni et al. Mar 2006 A1
20060064416 Sim-Tang Mar 2006 A1
20060224636 Kathuria et al. Oct 2006 A1
20070033301 Aloni et al. Feb 2007 A1
20070038884 Campbell et al. Feb 2007 A1
20070061385 Clark et al. Mar 2007 A1
20070112895 Ahrens et al. May 2007 A1
20070113032 Kameyama et al. May 2007 A1
20070143497 Kottomtharayil et al. Jun 2007 A1
20070168413 Barletta et al. Jul 2007 A1
20070176898 Suh Aug 2007 A1
20070179995 Prahlad et al. Aug 2007 A1
20070180207 Garfinkle Aug 2007 A1
20070198789 Clark et al. Aug 2007 A1
20070204166 Tome et al. Aug 2007 A1
20070208918 Harbin et al. Sep 2007 A1
20070220029 Jones et al. Sep 2007 A1
20070226400 Tsukazaki Sep 2007 A1
20070233699 Taniguchi et al. Oct 2007 A1
20070250302 Xu et al. Oct 2007 A1
20070260842 Faibish et al. Nov 2007 A1
20070276916 McLoughlin et al. Nov 2007 A1
20070283017 Anand et al. Dec 2007 A1
20070283343 Aridor et al. Dec 2007 A1
20070288525 Stakutis et al. Dec 2007 A1
20070288533 Srivastava et al. Dec 2007 A1
20070294321 Midgley et al. Dec 2007 A1
20080005468 Faibish et al. Jan 2008 A1
20080010422 Suzuki et al. Jan 2008 A1
20080027998 Hara Jan 2008 A1
20080036743 Westerman et al. Feb 2008 A1
20080082310 Sandorfi et al. Apr 2008 A1
20080133622 Brown et al. Jun 2008 A1
20080141018 Tanaka et al. Jun 2008 A1
20080154979 Saitoh et al. Jun 2008 A1
20080162590 Kundu et al. Jul 2008 A1
20080162607 Torii et al. Jul 2008 A1
20080201315 Lazier et al. Aug 2008 A1
20080229050 Tillgren Sep 2008 A1
20080307345 Hart et al. Dec 2008 A1
20080307527 Kaczmarski et al. Dec 2008 A1
20090089253 Huang et al. Apr 2009 A1
20090094427 Sano Apr 2009 A1
20090164527 Spektor et al. Jun 2009 A1
20090185500 Mower et al. Jul 2009 A1
20090216973 Nakajima et al. Aug 2009 A1
20090248755 Watanabe et al. Oct 2009 A1
20090276771 Nickolov et al. Nov 2009 A1
20090309849 Iwema et al. Dec 2009 A1
20090319653 Lorenz et al. Dec 2009 A1
20090327964 Mouilleseaux et al. Dec 2009 A1
20100042449 Thomas Feb 2010 A1
20100077165 Lu et al. Mar 2010 A1
20100095077 Lockwood Apr 2010 A1
20100104105 Schmidt et al. Apr 2010 A1
20100107155 Banerjee et al. Apr 2010 A1
20100114832 Lillibridge et al. May 2010 A1
20100165947 Taniuchi et al. Jul 2010 A1
20100179941 Agrawal et al. Jul 2010 A1
20100179973 Carruzzo Jul 2010 A1
20100192103 Cragun et al. Jul 2010 A1
20100205152 Ansari et al. Aug 2010 A1
20100223368 Runcie et al. Sep 2010 A1
20100228999 Maheshwari et al. Sep 2010 A1
20100235831 Dittmer Sep 2010 A1
20100262637 Akagawa et al. Oct 2010 A1
20100268689 Gates et al. Oct 2010 A1
20100318748 Ko et al. Dec 2010 A1
20100325377 Lango Dec 2010 A1
20100332454 Prahlad et al. Dec 2010 A1
20110041004 Miwa et al. Feb 2011 A1
20110047405 Marowsky-Bree et al. Feb 2011 A1
20110055399 Tung et al. Mar 2011 A1
20110055471 Thatcher et al. Mar 2011 A1
20110055500 Sasson et al. Mar 2011 A1
20110082998 Boldy et al. Apr 2011 A1
20110106768 Khanzode et al. May 2011 A1
20110154268 Trent, Jr. et al. Jun 2011 A1
20110218966 Barnes et al. Sep 2011 A1
20110238937 Murotani et al. Sep 2011 A1
20110264785 Newman et al. Oct 2011 A1
20110265143 Grube et al. Oct 2011 A1
20110295811 Cherkasova et al. Dec 2011 A1
20110307681 Piry et al. Dec 2011 A1
20120013540 Hogan Jan 2012 A1
20120065802 Seeber et al. Mar 2012 A1
20120084501 Watanabe et al. Apr 2012 A1
20120124307 Ashutosh et al. May 2012 A1
20120130956 Caputo May 2012 A1
20120131235 Nageshappa et al. May 2012 A1
20120179655 Beatty et al. Jul 2012 A1
20120204060 Swift et al. Aug 2012 A1
20120210398 Triantafillos et al. Aug 2012 A1
20120215743 Triantafillos et al. Aug 2012 A1
20130018946 Brown et al. Jan 2013 A1
20130024426 Flowers et al. Jan 2013 A1
20130036095 Titchener et al. Feb 2013 A1
20130091183 Edwards et al. Apr 2013 A1
20130091471 Gutt et al. Apr 2013 A1
20130125097 Ebcioglu et al. May 2013 A1
20130166511 Ghatty et al. Jun 2013 A1
20130234402 Fietz Sep 2013 A1
20130238562 Kumarasamy et al. Sep 2013 A1
20130238752 Park et al. Sep 2013 A1
20130318046 Clifford et al. Nov 2013 A1
20140006858 Helfman et al. Jan 2014 A1
20140032498 Lalonde et al. Jan 2014 A1
20140040580 Kripalani Feb 2014 A1
20140047081 Edwards Feb 2014 A1
20140052694 Dasari et al. Feb 2014 A1
20140053022 Forgette et al. Feb 2014 A1
20140089619 Khanna et al. Mar 2014 A1
20140101113 Zhang et al. Apr 2014 A1
20140149358 Aphale et al. May 2014 A1
20140189680 Kripalani Jul 2014 A1
20140201137 Vibhor et al. Jul 2014 A1
20140201144 Vibhor et al. Jul 2014 A1
20140244599 Zhang et al. Aug 2014 A1
20140303961 Leydon et al. Oct 2014 A1
20150046404 Caputo Feb 2015 A1
20150095691 Edwards Apr 2015 A1
20160054911 Edwards et al. Feb 2016 A1
20160055062 Petri et al. Feb 2016 A1
20160105312 Edwards Apr 2016 A1
20160110261 Parab et al. Apr 2016 A1
20160162349 Edwards et al. Jun 2016 A1
20160182170 Daoura et al. Jun 2016 A1
20170075719 Scallan et al. Mar 2017 A1
20170090786 Parab et al. Mar 2017 A1
20170177452 Parab et al. Jun 2017 A1
Non-Patent Literature Citations (55)
Entry
Final, Jan. 14, 2014, U.S. Appl. No. 13/030,073, filed Feb. 17, 2011.
Notice of Allowance, Sep. 12, 2013, U.S. Appl. No. 13/437,738, filed Apr. 2, 2012.
Final, Dec. 24, 2013, U.S. Appl. No. 13/363,234, filed Jan. 31, 2012.
Non-Final, Jul. 28, 2014, U.S. Appl. No. 13/671,498, filed Nov. 7, 2012.
Notice of Allowance, Sep. 26, 2014, U.S. Appl. No. 12/895,275, filed Sep. 30, 2010.
Non-Final Office Action, Oct. 1, 2014, U.S. Appl. No. 13/030,073, filed Feb. 17, 2011.
Notice of Allowance, Sep. 15, 2014, U.S. Appl. No. 13/363,234, filed Jan. 31, 2012.
Notice of Allowance, Oct. 20, 2014, U.S. Appl. No. 13/570,161, filed Aug. 8, 2012.
Corrected Notice of Allowability, Nov. 3, 2014, U.S. Appl. No. 13/570,161, filed Aug. 8, 2012.
Non-Final Office Action, Nov. 5, 2014, U.S. Appl. No. 13/789,578, filed Mar. 7, 2013.
Li et al., “Efficient File Replication,” U.S. Appl. No. 13/671,498, filed Nov. 7, 2012.
Non-Final Office Action, Jun. 17, 2015, U.S. Appl. No. 13/889,164, filed May 7, 2013.
Advisory Action, May 12, 2015, U.S. Appl. No. 13/671,498, filed Nov. 7, 2012.
Advisory Action, Jun. 10, 2015, U.S. Appl. No. 13/671,498, filed Nov. 7, 2012.
Final Office Action, May 27, 2015, U.S. Appl. No. 13/789,565, Mar. 7, 2013.
Final Office Action, Jun. 3, 2015, U.S. Appl. No. 13/789,578, filed Mar. 7, 2013.
Advisory Action, Jun. 10, 2015, U.S. Appl. No. 14/037,231, filed Sep. 25, 2013.
Non-Final Office Action, Jul. 9, 2015, U.S. Appl. No. 14/037,231, filed Sep. 25, 2013.
Notice of Allowance, Jun. 5, 2015, U.S. Appl. No. 14/534,017, filed Nov. 5, 2014.
Non-Final Office Action, Jul. 27, 2015, U.S. Appl. No. 14/564,082, filed Dec. 8, 2014.
Caputo, “Systems and Methods for Restoring a File”, U.S. Appl. No. 12/895,275, filed Sep. 30, 2010.
Corrected Notice of Allowability, Dec. 30, 2014, U.S. Appl. No. 13/570,161, filed Aug. 8, 2012.
Non-Final Office Action, Nov. 12, 2014, U.S. Appl. No. 14/037,231, filed Sep. 25, 2013.
Notice of Allowance, Sep. 8, 2015, U.S. Appl. No. 14/037,231, filed Sep. 25, 2013.
Notice of Allowance, Oct. 22, 2015, U.S. Appl. No. 13/030,073, filed Feb. 17, 2011.
Non-Final Office Action, Oct. 30, 2015, U.S. Appl. No. 13/789,565, filed Mar. 7, 2013.
Notice of Allowance, Oct. 21, 2015, U.S. Appl. No. 13/789,578, filed Mar. 7, 2013.
Non-Final Office Action, Jun. 1, 2016, U.S. Appl. No. 15/019,951, filed Feb. 9, 2016.
Advisory Action, Jun. 8, 2016, U.S. Appl. No. 14/564,082, filed Dec. 8, 2014.
Non-Final Office Action, Jun. 28, 2016, U.S. Appl. No. 13/671,498, filed Nov. 7, 2012.
Final Office Action, Jul. 26, 2016, U.S. Appl. No. 13/889,164, filed May 7, 2013.
Non-Final Office Action, Aug. 10, 2016, U.S. Appl. No. 14/522,527, filed Oct. 23, 2014.
Non-Final Office Action, Oct. 5, 2016, U.S. Appl. No. 14/970,480, filed Dec. 15, 2015.
Notice of Allowance, Oct. 18, 2016, U.S. Appl. No. 14/564,082, filed Dec. 8, 2014.
Advisory Action, Oct. 26, 2016, U.S. Appl. No. 13/889,164, filed May 7, 2013.
Final Office Action, Oct. 27, 2016, U.S. Appl. No. 14/977,581, filed Dec. 21, 2015.
Corrected Notice of Allowability, Dec. 5, 2016, U.S. Appl. No. 14/564,082, filed Dec. 8, 2014.
Non-Final Office Action, Dec. 9, 2016, U.S. Appl. No. 15/019,951, filed Feb. 9, 2016.
Final Office Action, Dec. 27, 2016, U.S. Appl. No. 13/671,498, filed Nov. 7, 2012.
Advisory Action, Dec. 30, 2016, U.S. Appl. No. 13/889,164, filed May 7, 2013.
Advisory Action, Feb. 7, 2017, U.S. Appl. No. 14/977,581, filed Dec. 21, 2015.
Final Office Action, Feb. 9, 2017, U.S. Appl. No. 14/522,527, filed Oct. 23, 2014.
Advisory Action, Feb. 13, 2017, U.S. Appl. No. 15/019,951, filed Feb. 9, 2016.
Non-Final Office Action, Mar. 22, 2017, U.S. Appl. No. 14/864,850, filed Sep. 24, 2015.
Advisory Action, Mar. 30, 2017, U.S. Appl. No. 14/977,581, filed Dec. 21, 2015.
Notice of Allowance, Apr. 6, 2017, U.S. Appl. No. 13/889,164, filed May 7, 2013.
Advisory Action, Apr. 20, 2017, U.S. Appl. No. 13/671,498, filed Nov. 7, 2012.
Advisory Action, Apr. 25, 2017, U.S. Appl. No. 14/522,527, filed Oct. 23, 2014.
Non-Final, Feb. 19, 2013, U.S. Appl. No. 13/437,738, filed Apr. 2, 2012.
Non-Final, Apr. 30, 2013, U.S. Appl. No. 13/363,234, filed Jan. 31, 2012.
Final, Dec. 27, 2012, U.S. Appl. No. 12/895,275, filed Sep. 30, 2010.
Non-Final Office Action, dated Jun. 14, 2017, U.S. Appl. No. 15/360,836, filed Nov. 23, 2016.
Non-Final Office Action, dated Jun. 29, 2017, U.S. Appl. No. 15/019,951, filed Feb. 9, 2016.
Final Office Action, dated Jul. 24, 2017, U.S. Appl. No. 14/864,850, filed Sep. 24, 2015.
Non-Final Office Action, dated Jul. 26, 2017, U.S. Appl. No. 14/977,581, filed Dec. 21, 2015.