This invention relates generally to storage networks with two or more tiers of storage servers, and more specifically, relates to a more efficient way of storing files that have identical contents in a storage network.
In enterprises today, employees tend to keep copies of all of the necessary documents and data that they access often. This is so that they can find the documents and data easily (central locations tend to change at least every so often). Furthermore, employees also tend to forget where certain things were found (in the central location), or never even knew where the document originated (they are sent a copy of the document via email). Finally, multiple employees may each keep a copy of the latest mp3 file, or video file, even if it is against company policy.
This can lead to duplicate copies of the same document or data residing in individually owned locations, so that the individual's themselves can easily find the document. However, this also means a lot of wasted space to store all of these copies of the document or data. And these copies are often stored on more expensive (and higher performance) tiers of storage, since the employees tend not to focus on costs, but rather on performance (they will store data on the location that they can most easily remember that gives them the best performance in retrieving the data).
Deduplication is a technique where files with identical contents are first identified and then only one copy of the identical contents, the single-instance copy, is kept in the physical storage while the storage space for the remaining identical contents is reclaimed and reused. Files whose contents have been deduped because of identical contents are hereafter referred to as deduplicated files. Thus, deduplication achieves what is called “Single-Instance Storage” where only the single-instance copy is stored in the physical storage, resulting in more efficient use of the physical storage space. File deduplication thus creates a domino effect of efficiency, reducing capital, administrative, and facility costs and is considered one of the most important and valuable technologies in storage.
U.S. Pat. Nos. 6,389,433 and 6,477,544 are examples of how a file system provides the single-instance-storage.
While single-instance-storage is conceptually simple, implementing it without sacrificing read/write performance is difficult. Files are deduped without the owners being aware of it. The owners of deduplicated files therefore have the same performance expectation as other files that have no duplicated copies. Since many deduplicated files are sharing one single-instance copy of the contents, it is important to prevent the single-instance copy from being modified. Typically, a file system uses the copy-on-write technique to protect the single-instance copy. When an update is pending on a deduplicated file, the file system creates a partial or full copy of the single-instance copy, and the update is allowed to proceed only after the (partial) copied data has been created and only on the copied data. The delay to wait for the creation of a (partial) copy of the single-instance data before an update can proceed introduces significant performance degradation. In addition, the process to identify and dedupe replicated files also puts a strain on file system resources. Because of the performance degradation, deduplication or single-instance copy is deemed not acceptable for normal use. In reality, deduplication is of no (obvious) benefit to the end-user. Thus, while the feature of deduplication or single-instance storage has been available in a few file systems, it is not commonly used and many file systems do not even offer this feature due to its adverse performance impact.
File system level deduplication offers many advantages for the IT administrators. However, it generally offers no direct benefits to the users of the file system other than performance degradation for those files that have been deduped. Therefore, the success of deduplication in the market place depends on reducing performance degradation to an acceptable level.
Another aspect of the file system level deduplication is that deduplication is usually done on a per file system basis. It is more desirable if deduplication is done together on one or more file systems. For example, the more file systems that are deduped together, the more chances that files with identical contents will be found and more storage space will be reclaimed. For example, if there is only one copy of file A in a file system, file A will not be deduped. On the other hand, if there is a copy of file A in another file system, then together, file A in the two file systems can be deduped. Furthermore, since there is only one single-instance copy for all of the deduplicated files from one or more file systems, the more file systems that are deduped together, the more efficient the deduplication process becomes.
Thus, it is desirable to achieve deduplication with acceptable performance. It is even more desirable to be able to dedupe across more file systems to achieve more deduplication efficiency.
In accordance with one aspect of the invention there are provided a method and an apparatus for deduplicating files in a file storage system having a primary storage tier and a secondary storage tier. In such embodiments, file deduplication involves identifying a plurality of files stored in the primary storage tier having identical file contents; copying the plurality of files to the secondary storage tier; storing in the primary storage tier a single copy of the file contents; and storing metadata for each of the plurality of files, the metadata associating each of the file copies in the secondary storage tier with the single copy of the file contents stored in the primary storage tier.
In various alternative embodiments, identifying the plurality of files stored in the primary storage tier having identical file contents may involve computing, for each of the plurality of files, a hash value based on the contents of the file; and identifying the files having identical file contents based on the hash values. Storing the single copy of the file contents in the primary storage tier may involve copying the file contents to a designated mirror server; and deleting the remaining file contents from each of the plurality of files in the primary storage tier. Upon a read access to one of the plurality of files, the read access may be directed to the single copy of the file contents maintained in the primary storage tier. Upon a write access to one of the plurality of files, the association between the file copy in the secondary storage tier and the single copy of the file contents stored in the primary storage tier may be broken the file copy stored in the secondary storage tier may be modified. The modified file copy subsequently may be migrated from the secondary storage tier to the primary storage tier based on a migration policy.
In other embodiments, deduplicating a selected file in the primary storage tier may involve determining whether the file contents of the selected file match the file contents of a previously deduplicated file having a single copy of file contents stored in the primary storage tier; when the file contents of the selected file match the file contents of a previously deduplicated file, deduplicating the selected file; otherwise determining whether the file contents of the selected file match the file contents of a non-duplicate file in the first storage tier; and when the file contents of the selected file match the file contents of a non-duplicate file, deduplicating both the selected file and the non-duplicate file. Determining whether the file contents of the selected file match the file contents of a previously deduplicated file may involve comparing a hash value associated with the selected file to a distinct hash value associated with each single copy of file contents stored in the primary storage tier. Deduplicating the selected file may involve copying the selected file to the secondary storage tier; deleting the file contents from the selected file; and storing metadata for the selected file, the metadata associating the file copy in the secondary storage tier with the single copy of the file contents for the previously deduplicated file stored in the primary storage tier. Deduplicating both the selected file and the non-duplicate file may involve copying the selected file and the non-duplicate file to the secondary storage tier; storing in the primary storage tier a single copy of the file contents; and storing metadata for each of the first and second selected files, the metadata associating each of the file copies in the secondary storage tier with the single copy of the file contents stored in the primary storage tier. Storing the single copy of the file contents for deduplicating both the selected file and the non-duplicate file may involve copying the file contents to the designated mirror server; and deleting the remaining file contents from the selected file and the non-duplicate file. Determining whether the file contents of the selected file match the file contents of a non-duplicate file in the primary storage tier may involve maintaining a list of non-duplicate files in the primary storage tier, the list including a distinct hash value for each non-duplicate file; and comparing a hash value associated with the selected file to the hash values associated with the non-duplicate files in the list, and when the file contents of the selected file do not match the file contents of any non-duplicate file, may involve adding the selected file to the list of non-duplicate files (e.g., by storing a pathname and a hash value associated with the selected file). Deduplicating both the selected file and the non-duplicate file may further involve removing the non-duplicate file from the list of non-duplicate files.
Deduplication may be implemented in a file switch or other device that manages file storage.
The foregoing features of the invention will be more readily understood by reference to the following detailed description, taken with reference to the accompanying drawings, in which:
This patent application relates generally to a method for performing deduplication on a global namespace using file virtualization when the global namespace is constructed from one or more storage servers, and to enable deduplication as a storage placement policy in a tiered storage environment.
A traditional file system manages the storage space by providing a hierarchical namespace. The hierarchical namespace starts from the root directory, which contains files and subdirectories. Each directory may also contain files and subdirectories identifying other files or subdirectories. Data is stored in files. Every file and directory is identified by a name. The full name of a file or directory is constructed by concatenating the name of the root directory and the names of each subdirectory that finally leads to the subdirectory containing the identified file or directory, together with the name of the file or the directory.
The full name of a file thus carries with it two pieces of information: (1) the identification of the file and (2) the physical storage location where the file is stored. If the physical storage location of a file is changed (for example, moved from one partition mounted on a system to another), the identification of the file changes as well.
For ease of management, as well as for a variety of other reasons, the administrator would like to control the physical storage location of a file. For example, important files might be stored on expensive, high-performance file servers, while less important files could be stored on less expensive and less capable file servers.
Unfortunately, moving files from one server to another usually changes the full name of the files and thus, their identification, as well. This is usually a very disruptive process, since after the move users may not be able to remember the new location of their files. Thus, it is desirable to separate the physical storage location of a file from its identification. With this separation, IT and system administrators will be able to control the physical storage location of a file while preserving what the user perceives as the location of the file (and thus its identity).
File virtualization is a technology that separates the full name of a file from its physical storage location. File virtualization is usually implemented as a hardware appliance that is located in the data path between users and the file servers. For users, a file virtualization appliance appears as a file server that exports the namespace of a file system. From the file servers' perspective, the file virtualization appliance appears as just a normal user. Attune System's Maestro File Manager (MFM) is an example of a file virtualization appliance.
As a result of separating the full name of a file from the file's physical storage location, file virtualization provides the following capabilities:
1) Creation of a Synthetic Namespace
2) Allows Having Many Full Filenames to Refer to a Single File
3) Allows Having One Full Name to Refer to Many Files
Deduplication is of no obvious benefit to the end users of a file system. Instead of using deduplication as a management policy to reduce storage space and subsequently cause inconvenience to the end users of the deduplicated files, this invention uses deduplication as a storage placement policy to intelligently managed the storage assets of an enterprise, with relatively little inconvenience to the end users.
In embodiments of the present invention, a set of file servers is designated as tier 1 where data stored in these file servers is considered more important to the enterprise. Another (typically non-overlapping) set of file servers is designated as tier 2 storage where data stored in these file servers is considered less important to the business. By using these two storage tiers to identify data important to the business, the system administrators can spend more time and resources to provide faster access and more frequent backup on the data stored on the tier 1 file servers.
Deduplication typically is treated as one of the storage placement policies that decides where data should be stored, e.g., on a tier 1 or tier 2 file server.
In embodiments of the present invention, duplicated data is automatically moved from tier 1 to tier 2. The total storage space used by the deduplicated data on tier 1 and tier 2 remains the same (or perhaps even increases slightly). However, there is more storage space available on tier 1 file servers as a result of deduplication, since all the duplicated data is now stored on tier 2.
There may be performance differences between tier 1 and tier 2 file servers. However, these differences tend to be small since the relatively inexpensive file servers are still very capable. To maintain the same level of performance when accessing the deduplicated files, as each set of duplicated files is moved from the tier 1 file servers, a single instance copy of the file is left behind as a mirror copy. One of the tier 1 file servers is designated as a mirror server where all of the mirror copies are stored. Read access to a deduplicated file is redirected to the deduplicated file's mirror copy. When the first write to a deduplicated file is received, the association from the deduplicated file stored in a tier 2 server to its mirror copy that is stored in a tier 1 server is discarded. Accesses to the “modified” duplicated file will then resume normally from the tier 2 file server. At a certain time, the “modified” deduplicated file is then migrated back to tier 1 storage.
Extending file virtualization to support deduplication is relatively straight forward. First, a set of tier-1 servers is identified as a target for deduplication, and a set of tier 2 servers is identified for receiving deduplicated data. One of the tier 1 file servers is chosen as the mirror server. The mirror server is used to store the mirror copy of each set of deduplicated files with identical contents.
A background deduplication process typically is run periodically within the file virtualization appliance to perform the deduplication. Exemplary embodiments use a sha1 digest computed from the contents of a file to identify files that have identical contents. A sha1 digest value is a 160-bit globally unique value for any given set of data (contents) of a file. Therefore, if two files are identical in contents (but not necessarily name or location), they should always have the same sha1 digest values. And conversely, if two files are different in contents, they should always have different sha1 digest values.
An exemplary deduplication process for the namespace is as follows:
It is possible that the sha1 digest value for a file marked for deduplication may have changed before it is actually deduped. This case should occur relatively infrequently. If it does occur, essentially the worst that can happen is that a file that really has no duplicate files in tier 1 gets deduplicated and migrated to tier 2. However, the deduplicated file eventually should be migrated back to the tier 1 storage tier.
An exemplary process to dedupe a single file (called from the deduplication process for the namespace) is as follows:
When a non-deduplicated file that has a sha1 digest is opened for update, its sha1 digest is immediately cleared.
When a deduplicated file is opened for update, its sha1 digest is immediately cleared. The mirror association between the deduplicated copy and the mirror copy is immediately broken. The file is no longer a deduplicated file (its deduplicated flag is cleared), and an entry is added to a to-do list to migrate this file back to tier 1 storage in the future.
When a deduplicated file is open for read, a check is made to see if there is a mirror copy stored in the mirror server. If there is, subsequent read requests on the deduplicated file will be switched to the mirror server for processing. Otherwise, the read request is switched to the tier 2 file server containing the actual data of the deduplicated file.
It should be noted that file deduplication as discussed herein may be implemented using a file switches of the types described above and in the provisional patent application Ser. No. 60/923,765. It should also be noted that embodiments of the present invention may incorporate, utilize, supplement, or be combined with various features described in one or more of the other referenced patent applications.
It should be noted that terms such as “client,” “server,” “switch,” and “node” may be used herein to describe devices that may be used in certain embodiments of the present invention and should not be construed to limit the present invention to any particular device type unless the context otherwise requires. Thus, a device may include, without limitation, a bridge, router, bridge-router (brouter), switch, node, server, computer, appliance, or other type of device. Such devices typically include one or more network interfaces for communicating over a communication network and a processor (e.g., a microprocessor with memory and other peripherals and/or application-specific hardware) configured accordingly to perform device functions. Communication networks generally may include public and/or private networks; may include local-area, wide-area, metropolitan-area, storage, and/or other types of networks; and may employ communication technologies including, but in no way limited to, analog technologies, digital technologies, optical technologies, wireless technologies (e.g., Bluetooth), networking technologies, and internetworking technologies.
It should also be noted that devices may use communication protocols and messages (e.g., messages created, transmitted, received, stored, and/or processed by the device), and such messages may be conveyed by a communication network or medium. Unless the context otherwise requires, the present invention should not be construed as being limited to any particular communication message type, communication message format, or communication protocol. Thus, a communication message generally may include, without limitation, a frame, packet, datagram, user datagram, cell, or other type of communication message.
It should also be noted that logic flows may be described herein to demonstrate various aspects of the invention, and should not be construed to limit the present invention to any particular logic flow or logic implementation. The described logic may be partitioned into different logic blocks (e.g., programs, modules, functions, or subroutines) without changing the overall results or otherwise departing from the true scope of the invention. Often times, logic elements may be added, modified, omitted, performed in a different order, or implemented using different logic constructs (e.g., logic gates, looping primitives, conditional logic, and other logic constructs) without changing the overall results or otherwise departing from the true scope of the invention.
The present invention may be embodied in many different forms, including, but in no way limited to, computer program logic for use with a processor (e.g., a microprocessor, microcontroller, digital signal processor, or general purpose computer), programmable logic for use with a programmable logic device (e.g., a Field Programmable Gate Array (FPGA) or other PLD), discrete components, integrated circuitry (e.g., an Application Specific Integrated Circuit (ASIC)), or any other means including any combination thereof. In a typical embodiment of the present invention, predominantly all of the described logic is implemented as a set of computer program instructions that is converted into a computer executable form, stored as such in a computer readable medium, and executed by a microprocessor under the control of an operating system.
Computer program logic implementing all or part of the functionality previously described herein may be embodied in various forms, including, but in no way limited to, a source code form, a computer executable form, and various intermediate forms (e.g., forms generated by an assembler, compiler, linker, or locator). Source code may include a series of computer program instructions implemented in any of various programming languages (e.g., an object code, an assembly language, or a high-level language such as Fortran, C, C++, JAVA, or HTML) for use with various operating systems or operating environments. The source code may define and use various data structures and communication messages. The source code may be in a computer executable form (e.g., via an interpreter), or the source code may be converted (e.g., via a translator, assembler, or compiler) into a computer executable form.
The computer program may be fixed in any form (e.g., source code form, computer executable form, or an intermediate form) either permanently or transitorily in a tangible storage medium, such as a semiconductor memory device (e.g., a RAM, ROM, PROM, EEPROM, or Flash-Programmable RAM), a magnetic memory device (e.g., a diskette or fixed disk), an optical memory device (e.g., a CD-ROM), a PC card (e.g., PCMCIA card), or other memory device. The computer program may be fixed in any form in a signal that is transmittable to a computer using any of various communication technologies, including, but in no way limited to, analog technologies, digital technologies, optical technologies, wireless technologies (e.g., Bluetooth), networking technologies, and internetworking technologies. The computer program may be distributed in any form as a removable storage medium with accompanying printed or electronic documentation (e.g., shrink wrapped software), preloaded with a computer system (e.g., on system ROM or fixed disk), or distributed from a server or electronic bulletin board over the communication system (e.g., the Internet or World Wide Web).
Hardware logic (including programmable logic for use with a programmable logic device) implementing all or part of the functionality previously described herein may be designed using traditional manual methods, or may be designed, captured, simulated, or documented electronically using various tools, such as Computer Aided Design (CAD), a hardware description language (e.g., VHDL or AHDL), or a PLD programming language (e.g., PALASM, ABEL, or CUPL).
Programmable logic may be fixed either permanently or transitorily in a tangible storage medium, such as a semiconductor memory device (e.g., a RAM, ROM, PROM, EEPROM, or Flash-Programmable RAM), a magnetic memory device (e.g., a diskette or fixed disk), an optical memory device (e.g., a CD-ROM), or other memory device. The programmable logic may be fixed in a signal that is transmittable to a computer using any of various communication technologies, including, but in no way limited to, analog technologies, digital technologies, optical technologies, wireless technologies (e.g., Bluetooth), networking technologies, and internetworking technologies. The programmable logic may be distributed as a removable storage medium with accompanying printed or electronic documentation (e.g., shrink wrapped software), preloaded with a computer system (e.g., on system ROM or fixed disk), or distributed from a server or electronic bulletin board over the communication system (e.g., the Internet or World Wide Web).
The present invention may be embodied in other specific forms without departing from the true scope of the invention. Any references to the “invention” are intended to refer to exemplary embodiments of the invention and should not be construed to refer to all embodiments of the invention unless the context otherwise requires. The described embodiments are to be considered in all respects only as illustrative and not restrictive.
This patent application claims priority from U.S. Provisional Patent Application No. 60/987,181 entitled FILE DEDUPLICATION USING STORAGE TIERS filed on Nov. 12, 2007. This patent application also may be related to one or more of the following patent applications: U.S. Provisional Patent Application No. 60/923,765 entitled NETWORK FILE MANAGEMENT SYSTEMS, APPARATUS, AND METHODS filed on Apr. 16, 2007. U.S. Provisional Patent Application No. 60/940,104 entitled REMOTE FILE VIRTUALIZATION filed on May 25, 2007. U.S. Provisional Patent Application No. 60/987,161 entitled REMOTE FILE VIRTUALIZATION METADATA MIRRORING filed Nov. 12, 2007. U.S. Provisional Patent Application No. 60/987,165 entitled REMOTE FILE VIRTUALIZATION DATA MIRRORING filed Nov. 12, 2007. U.S. Provisional Patent Application No. 60/987,170 entitled REMOTE FILE VIRTUALIZATION WITH NO EDGE SERVERS filed Nov. 12, 2007. U.S. Provisional Patent Application No. 60/987,174 entitled LOAD SHARING CLUSTER FILE SYSTEM filed Nov. 12, 2007. U.S. Provisional Patent Application No. 60/987,206 entitled NON-DISRUPTIVE FILE MIGRATION filed Nov. 12, 2007. U.S. Provisional Patent Application No. 60/987,197 entitled HOTSPOT MITIGATION IN LOAD SHARING CLUSTER FILE SYSTEMS filed Nov. 12, 2007. U.S. Provisional Patent Application No. 60/987,194 entitled ON DEMAND FILE VIRTUALIZATION FOR SERVER CONFIGURATION MANAGEMENT WITH LIMITED INTERRUPTION filed Nov. 12, 2007. U.S. patent application Ser. No. 12/104,197 entitled FILE AGGREGATION IN A SWITCHED FILE SYSTEM filed Apr. 16, 2008. U.S. patent application Ser. No. 12/103,989 entitled FILE AGGREGATION IN A SWITCHED FILE SYSTEM filed Apr. 16, 2008. U.S. patent application Ser. No. 12/126,129 entitled REMOTE FILE VIRTUALIZATION IN A SWITCHED FILE SYSTEM filed May 23, 2008. All of the above-referenced patent applications are hereby incorporated herein by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
4993030 | Krakauer et al. | Feb 1991 | A |
5218695 | Noveck et al. | Jun 1993 | A |
5303368 | Kotaki | Apr 1994 | A |
5473362 | Fitzgerald et al. | Dec 1995 | A |
5511177 | Kagimasa et al. | Apr 1996 | A |
5537585 | Blickenstaff et al. | Jul 1996 | A |
5548724 | Akizawa et al. | Aug 1996 | A |
5550965 | Gabbe et al. | Aug 1996 | A |
5583995 | Gardner et al. | Dec 1996 | A |
5586260 | Hu | Dec 1996 | A |
5590320 | Maxey | Dec 1996 | A |
5623490 | Richter et al. | Apr 1997 | A |
5649194 | Miller et al. | Jul 1997 | A |
5649200 | Leblang et al. | Jul 1997 | A |
5668943 | Attanasio et al. | Sep 1997 | A |
5692180 | Lee | Nov 1997 | A |
5721779 | Funk | Feb 1998 | A |
5724512 | Winterbottom | Mar 1998 | A |
5806061 | Chaudhuri et al. | Sep 1998 | A |
5832496 | Anand et al. | Nov 1998 | A |
5832522 | Blickenstaff et al. | Nov 1998 | A |
5838970 | Thomas | Nov 1998 | A |
5862325 | Reed et al. | Jan 1999 | A |
5884303 | Brown | Mar 1999 | A |
5893086 | Schmuck et al. | Apr 1999 | A |
5897638 | Lasser et al. | Apr 1999 | A |
5905990 | Inglett | May 1999 | A |
5917998 | Cabrera et al. | Jun 1999 | A |
5920873 | Van Huben et al. | Jul 1999 | A |
5926816 | Bauer et al. | Jul 1999 | A |
5937406 | Balabine et al. | Aug 1999 | A |
5991302 | Berl et al. | Nov 1999 | A |
5995491 | Richter et al. | Nov 1999 | A |
5999664 | Mahoney et al. | Dec 1999 | A |
6012083 | Savitzky et al. | Jan 2000 | A |
6029168 | Frey | Feb 2000 | A |
6044367 | Wolff | Mar 2000 | A |
6047129 | Frye | Apr 2000 | A |
6072942 | Stockwell et al. | Jun 2000 | A |
6078929 | Rao | Jun 2000 | A |
6085234 | Pitts et al. | Jul 2000 | A |
6088694 | Burns et al. | Jul 2000 | A |
6104706 | Richter et al. | Aug 2000 | A |
6128627 | Mattis et al. | Oct 2000 | A |
6128717 | Harrison et al. | Oct 2000 | A |
6161145 | Bainbridge et al. | Dec 2000 | A |
6161185 | Guthrie et al. | Dec 2000 | A |
6181336 | Chiu et al. | Jan 2001 | B1 |
6202156 | Kalajan | Mar 2001 | B1 |
6223206 | Dan et al. | Apr 2001 | B1 |
6233648 | Tomita | May 2001 | B1 |
6237008 | Beal et al. | May 2001 | B1 |
6256031 | Meijer et al. | Jul 2001 | B1 |
6282610 | Bergsten | Aug 2001 | B1 |
6289345 | Yasue | Sep 2001 | B1 |
6308162 | Ouimet et al. | Oct 2001 | B1 |
6324581 | Xu et al. | Nov 2001 | B1 |
6329985 | Tamer et al. | Dec 2001 | B1 |
6339785 | Feigenbaum | Jan 2002 | B1 |
6349343 | Foody et al. | Feb 2002 | B1 |
6374263 | Bunger et al. | Apr 2002 | B1 |
6389433 | Bolosky et al. | May 2002 | B1 |
6393581 | Friedman et al. | May 2002 | B1 |
6397246 | Wolfe | May 2002 | B1 |
6412004 | Chen et al. | Jun 2002 | B1 |
6438595 | Blumenau et al. | Aug 2002 | B1 |
6466580 | Leung | Oct 2002 | B1 |
6469983 | Narayana et al. | Oct 2002 | B2 |
6477544 | Bolosky et al. | Nov 2002 | B1 |
6487561 | Ofek et al. | Nov 2002 | B1 |
6493804 | Soltis et al. | Dec 2002 | B1 |
6516350 | Lumelsky et al. | Feb 2003 | B1 |
6516351 | Borr | Feb 2003 | B2 |
6542909 | Tamer et al. | Apr 2003 | B1 |
6549916 | Sedlar | Apr 2003 | B1 |
6553352 | Delurgio et al. | Apr 2003 | B2 |
6556997 | Levy | Apr 2003 | B1 |
6556998 | Mukherjee et al. | Apr 2003 | B1 |
6560230 | Li et al. | May 2003 | B1 |
6601101 | Lee et al. | Jul 2003 | B1 |
6606663 | Liao et al. | Aug 2003 | B1 |
6612490 | Herrendoerfer et al. | Sep 2003 | B1 |
6654346 | Mahalingaiah et al. | Nov 2003 | B1 |
6721794 | Taylor et al. | Apr 2004 | B2 |
6728265 | Yavatkar et al. | Apr 2004 | B1 |
6738357 | Richter et al. | May 2004 | B1 |
6738790 | Klein et al. | May 2004 | B1 |
6742035 | Zayas et al. | May 2004 | B1 |
6744776 | Kalkunte et al. | Jun 2004 | B1 |
6748420 | Quatrano et al. | Jun 2004 | B1 |
6754215 | Arikawa et al. | Jun 2004 | B1 |
6757706 | Dong et al. | Jun 2004 | B1 |
6775672 | Mahalingam et al. | Aug 2004 | B2 |
6775673 | Mahalingam et al. | Aug 2004 | B2 |
6775679 | Gupta | Aug 2004 | B2 |
6782450 | Arnott et al. | Aug 2004 | B2 |
6801960 | Ericson et al. | Oct 2004 | B1 |
6826613 | Wang et al. | Nov 2004 | B1 |
6839761 | Kadyk et al. | Jan 2005 | B2 |
6847959 | Arrouye et al. | Jan 2005 | B1 |
6847970 | Kar et al. | Jan 2005 | B2 |
6850997 | Rooney et al. | Feb 2005 | B1 |
6871245 | Bradley | Mar 2005 | B2 |
6880017 | Marce et al. | Apr 2005 | B1 |
6889249 | Miloushev et al. | May 2005 | B2 |
6914881 | Mansfield et al. | Jul 2005 | B1 |
6922688 | Frey, Jr. | Jul 2005 | B1 |
6934706 | Mancuso et al. | Aug 2005 | B1 |
6938039 | Bober et al. | Aug 2005 | B1 |
6938059 | Tamer et al. | Aug 2005 | B2 |
6959373 | Testardi | Oct 2005 | B2 |
6961815 | Kistler et al. | Nov 2005 | B2 |
6973455 | Vahalia et al. | Dec 2005 | B1 |
6973549 | Testardi | Dec 2005 | B1 |
6975592 | Seddigh et al. | Dec 2005 | B1 |
6985936 | Agarwalla et al. | Jan 2006 | B2 |
6985956 | Luke et al. | Jan 2006 | B2 |
6986015 | Testardi | Jan 2006 | B2 |
6990114 | Erimli et al. | Jan 2006 | B1 |
6990547 | Ulrich et al. | Jan 2006 | B2 |
6990667 | Ulrich et al. | Jan 2006 | B2 |
6996841 | Kadyk et al. | Feb 2006 | B2 |
7010553 | Chen et al. | Mar 2006 | B2 |
7013379 | Testardi | Mar 2006 | B1 |
7039061 | Connor et al. | May 2006 | B2 |
7051112 | Dawson | May 2006 | B2 |
7072917 | Wong et al. | Jul 2006 | B2 |
7075924 | Richter et al. | Jul 2006 | B2 |
7089286 | Malik | Aug 2006 | B1 |
7111115 | Peters et al. | Sep 2006 | B2 |
7113962 | Kee et al. | Sep 2006 | B1 |
7120746 | Campbell et al. | Oct 2006 | B2 |
7127556 | Blumenau et al. | Oct 2006 | B2 |
7133967 | Fujie et al. | Nov 2006 | B2 |
7146524 | Patel et al. | Dec 2006 | B2 |
7152184 | Maeda et al. | Dec 2006 | B2 |
7155466 | Rodriguez et al. | Dec 2006 | B2 |
7165095 | Sim | Jan 2007 | B2 |
7167821 | Hardwick et al. | Jan 2007 | B2 |
7173929 | Testardi | Feb 2007 | B1 |
7194579 | Robinson et al. | Mar 2007 | B2 |
7234074 | Cohn et al. | Jun 2007 | B2 |
7236491 | Tsao et al. | Jun 2007 | B2 |
7280536 | Testardi | Oct 2007 | B2 |
7284150 | Ma et al. | Oct 2007 | B2 |
7293097 | Borr | Nov 2007 | B2 |
7293099 | Kalajan | Nov 2007 | B1 |
7293133 | Colgrove et al. | Nov 2007 | B1 |
7343398 | Lownsbrough | Mar 2008 | B1 |
7346664 | Wong et al. | Mar 2008 | B2 |
7383288 | Miloushev et al. | Jun 2008 | B2 |
7401220 | Bolosky et al. | Jul 2008 | B2 |
7406484 | Srinivasan et al. | Jul 2008 | B1 |
7415488 | Muth et al. | Aug 2008 | B1 |
7415608 | Bolosky et al. | Aug 2008 | B2 |
7440982 | Lu et al. | Oct 2008 | B2 |
7475241 | Patel et al. | Jan 2009 | B2 |
7477796 | Sasaki et al. | Jan 2009 | B2 |
7509322 | Miloushev et al. | Mar 2009 | B2 |
7512673 | Miloushev et al. | Mar 2009 | B2 |
7519813 | Cox et al. | Apr 2009 | B1 |
7562110 | Miloushev et al. | Jul 2009 | B2 |
7571168 | Bahar et al. | Aug 2009 | B2 |
7574433 | Engel | Aug 2009 | B2 |
7599941 | Bahar et al. | Oct 2009 | B2 |
7610307 | Havewala et al. | Oct 2009 | B2 |
7624109 | Testardi | Nov 2009 | B2 |
7639883 | Gill | Dec 2009 | B2 |
7653699 | Colgrove et al. | Jan 2010 | B1 |
7685177 | Hagerstrom et al. | Mar 2010 | B1 |
7734603 | McManis | Jun 2010 | B1 |
7788335 | Miloushev et al. | Aug 2010 | B2 |
7809691 | Karmarkar et al. | Oct 2010 | B1 |
7818299 | Federwisch et al. | Oct 2010 | B1 |
7822939 | Veprinsky et al. | Oct 2010 | B1 |
7831639 | Panchbudhe et al. | Nov 2010 | B1 |
7870154 | Shitomi et al. | Jan 2011 | B2 |
7877511 | Berger et al. | Jan 2011 | B1 |
7885970 | Lacapra | Feb 2011 | B2 |
7904466 | Valencia et al. | Mar 2011 | B1 |
7913053 | Newland | Mar 2011 | B1 |
7953701 | Okitsu et al. | May 2011 | B2 |
7958347 | Ferguson | Jun 2011 | B1 |
8005953 | Miloushev et al. | Aug 2011 | B2 |
8046547 | Chatterjee et al. | Oct 2011 | B1 |
8103622 | Karinta | Jan 2012 | B1 |
8112392 | Bunnell et al. | Feb 2012 | B1 |
8271751 | Hinrichs, Jr. | Sep 2012 | B2 |
8326798 | Driscoll et al. | Dec 2012 | B1 |
8351600 | Resch | Jan 2013 | B2 |
20010007560 | Masuda et al. | Jul 2001 | A1 |
20010014891 | Hoffert et al. | Aug 2001 | A1 |
20010047293 | Waller et al. | Nov 2001 | A1 |
20010051955 | Wong | Dec 2001 | A1 |
20020035537 | Waller et al. | Mar 2002 | A1 |
20020059263 | Shima et al. | May 2002 | A1 |
20020065810 | Bradley | May 2002 | A1 |
20020073105 | Noguchi et al. | Jun 2002 | A1 |
20020083118 | Sim | Jun 2002 | A1 |
20020087887 | Busam et al. | Jul 2002 | A1 |
20020106263 | Winker | Aug 2002 | A1 |
20020120763 | Miloushev et al. | Aug 2002 | A1 |
20020133330 | Loisey et al. | Sep 2002 | A1 |
20020133491 | Sim et al. | Sep 2002 | A1 |
20020138502 | Gupta | Sep 2002 | A1 |
20020143909 | Botz et al. | Oct 2002 | A1 |
20020147630 | Rose et al. | Oct 2002 | A1 |
20020150253 | Brezak et al. | Oct 2002 | A1 |
20020156905 | Weissman | Oct 2002 | A1 |
20020161911 | Pinckney, III et al. | Oct 2002 | A1 |
20020188667 | Kirnos | Dec 2002 | A1 |
20020194342 | Lu et al. | Dec 2002 | A1 |
20030009429 | Jameson | Jan 2003 | A1 |
20030012382 | Ferchichi et al. | Jan 2003 | A1 |
20030028514 | Lord et al. | Feb 2003 | A1 |
20030033308 | Patel et al. | Feb 2003 | A1 |
20030033535 | Fisher et al. | Feb 2003 | A1 |
20030061240 | McCann et al. | Mar 2003 | A1 |
20030065956 | Belapurkar et al. | Apr 2003 | A1 |
20030115218 | Bobbitt et al. | Jun 2003 | A1 |
20030115439 | Mahalingam et al. | Jun 2003 | A1 |
20030128708 | Inoue et al. | Jul 2003 | A1 |
20030135514 | Patel et al. | Jul 2003 | A1 |
20030149781 | Yared et al. | Aug 2003 | A1 |
20030156586 | Lee et al. | Aug 2003 | A1 |
20030159072 | Bellinger et al. | Aug 2003 | A1 |
20030171978 | Jenkins et al. | Sep 2003 | A1 |
20030177364 | Walsh et al. | Sep 2003 | A1 |
20030177388 | Botz et al. | Sep 2003 | A1 |
20030179755 | Fraser | Sep 2003 | A1 |
20030200207 | Dickinson | Oct 2003 | A1 |
20030204635 | Ko et al. | Oct 2003 | A1 |
20040003266 | Moshir et al. | Jan 2004 | A1 |
20040006575 | Visharam et al. | Jan 2004 | A1 |
20040010654 | Yasuda et al. | Jan 2004 | A1 |
20040017825 | Stanwood et al. | Jan 2004 | A1 |
20040025013 | Parker et al. | Feb 2004 | A1 |
20040028043 | Maveli et al. | Feb 2004 | A1 |
20040028063 | Roy et al. | Feb 2004 | A1 |
20040030857 | Krakirian et al. | Feb 2004 | A1 |
20040044705 | Stager et al. | Mar 2004 | A1 |
20040054748 | Ackaouy et al. | Mar 2004 | A1 |
20040054777 | Ackaouy et al. | Mar 2004 | A1 |
20040093474 | Lin et al. | May 2004 | A1 |
20040098383 | Tabellion et al. | May 2004 | A1 |
20040098595 | Aupperle et al. | May 2004 | A1 |
20040133573 | Miloushev et al. | Jul 2004 | A1 |
20040133577 | Miloushev et al. | Jul 2004 | A1 |
20040133606 | Miloushev et al. | Jul 2004 | A1 |
20040133607 | Miloushev et al. | Jul 2004 | A1 |
20040133650 | Miloushev et al. | Jul 2004 | A1 |
20040139355 | Axel et al. | Jul 2004 | A1 |
20040148380 | Meyer et al. | Jul 2004 | A1 |
20040153479 | Mikesell et al. | Aug 2004 | A1 |
20040181605 | Nakatani et al. | Sep 2004 | A1 |
20040199547 | Winter et al. | Oct 2004 | A1 |
20040213156 | Smallwood et al. | Oct 2004 | A1 |
20040236798 | Srinivasan et al. | Nov 2004 | A1 |
20040267830 | Wong et al. | Dec 2004 | A1 |
20050021615 | Arnott et al. | Jan 2005 | A1 |
20050050107 | Mane et al. | Mar 2005 | A1 |
20050091214 | Probert et al. | Apr 2005 | A1 |
20050108575 | Yung | May 2005 | A1 |
20050114291 | Becker-Szendy et al. | May 2005 | A1 |
20050114701 | Atkins et al. | May 2005 | A1 |
20050117589 | Douady et al. | Jun 2005 | A1 |
20050160161 | Barrett et al. | Jul 2005 | A1 |
20050175013 | Le Pennec et al. | Aug 2005 | A1 |
20050187866 | Lee | Aug 2005 | A1 |
20050198501 | Andreev et al. | Sep 2005 | A1 |
20050213587 | Cho et al. | Sep 2005 | A1 |
20050246393 | Coates et al. | Nov 2005 | A1 |
20050289109 | Arrouye et al. | Dec 2005 | A1 |
20050289111 | Tribble et al. | Dec 2005 | A1 |
20060010502 | Mimatsu et al. | Jan 2006 | A1 |
20060045096 | Farmer et al. | Mar 2006 | A1 |
20060074922 | Nishimura | Apr 2006 | A1 |
20060075475 | Boulos et al. | Apr 2006 | A1 |
20060080353 | Miloushev et al. | Apr 2006 | A1 |
20060106882 | Douceur et al. | May 2006 | A1 |
20060112151 | Manley et al. | May 2006 | A1 |
20060123062 | Bobbitt et al. | Jun 2006 | A1 |
20060140193 | Kakani et al. | Jun 2006 | A1 |
20060153201 | Hepper et al. | Jul 2006 | A1 |
20060161518 | Lacapra | Jul 2006 | A1 |
20060167838 | Lacapra | Jul 2006 | A1 |
20060179261 | Rajan | Aug 2006 | A1 |
20060184589 | Lees et al. | Aug 2006 | A1 |
20060190496 | Tsunoda | Aug 2006 | A1 |
20060200470 | Lacapra et al. | Sep 2006 | A1 |
20060206547 | Kulkarni et al. | Sep 2006 | A1 |
20060212746 | Amegadzie et al. | Sep 2006 | A1 |
20060218135 | Bisson et al. | Sep 2006 | A1 |
20060224636 | Kathuria et al. | Oct 2006 | A1 |
20060224687 | Popkin et al. | Oct 2006 | A1 |
20060230265 | Krishna | Oct 2006 | A1 |
20060242179 | Chen et al. | Oct 2006 | A1 |
20060259949 | Schaefer et al. | Nov 2006 | A1 |
20060268692 | Wright et al. | Nov 2006 | A1 |
20060271598 | Wong et al. | Nov 2006 | A1 |
20060277225 | Mark et al. | Dec 2006 | A1 |
20060282461 | Marinescu | Dec 2006 | A1 |
20060282471 | Mark et al. | Dec 2006 | A1 |
20070022121 | Bahar et al. | Jan 2007 | A1 |
20070024919 | Wong et al. | Feb 2007 | A1 |
20070027929 | Whelan | Feb 2007 | A1 |
20070027935 | Haselton et al. | Feb 2007 | A1 |
20070028068 | Golding et al. | Feb 2007 | A1 |
20070088702 | Fridella et al. | Apr 2007 | A1 |
20070098284 | Sasaki et al. | May 2007 | A1 |
20070136308 | Tsirigotis et al. | Jun 2007 | A1 |
20070139227 | Speirs, II et al. | Jun 2007 | A1 |
20070180314 | Kawashima et al. | Aug 2007 | A1 |
20070208748 | Li | Sep 2007 | A1 |
20070209075 | Coffman | Sep 2007 | A1 |
20070226331 | Srinivasan et al. | Sep 2007 | A1 |
20080046432 | Anderson et al. | Feb 2008 | A1 |
20080070575 | Claussen et al. | Mar 2008 | A1 |
20080104443 | Akutsu et al. | May 2008 | A1 |
20080114718 | Anderson et al. | May 2008 | A1 |
20080189468 | Schmidt et al. | Aug 2008 | A1 |
20080200207 | Donahue et al. | Aug 2008 | A1 |
20080209073 | Tang | Aug 2008 | A1 |
20080215836 | Sutoh et al. | Sep 2008 | A1 |
20080222223 | Srinivasan et al. | Sep 2008 | A1 |
20080243769 | Arbour et al. | Oct 2008 | A1 |
20080282047 | Arakawa et al. | Nov 2008 | A1 |
20080294446 | Guo et al. | Nov 2008 | A1 |
20090007162 | Sheehan | Jan 2009 | A1 |
20090013138 | Sudhakar | Jan 2009 | A1 |
20090037975 | Ishikawa et al. | Feb 2009 | A1 |
20090041230 | Williams | Feb 2009 | A1 |
20090055507 | Oeda | Feb 2009 | A1 |
20090055607 | Schack et al. | Feb 2009 | A1 |
20090077097 | Lacapra et al. | Mar 2009 | A1 |
20090089344 | Brown et al. | Apr 2009 | A1 |
20090094252 | Wong et al. | Apr 2009 | A1 |
20090106255 | Lacapra et al. | Apr 2009 | A1 |
20090106263 | Khalid et al. | Apr 2009 | A1 |
20090132616 | Winter et al. | May 2009 | A1 |
20090204649 | Wong et al. | Aug 2009 | A1 |
20090204650 | Wong et al. | Aug 2009 | A1 |
20090204705 | Marinov et al. | Aug 2009 | A1 |
20090210431 | Marinkovic et al. | Aug 2009 | A1 |
20090210875 | Bolles et al. | Aug 2009 | A1 |
20090240705 | Miloushev et al. | Sep 2009 | A1 |
20090240899 | Akagawa et al. | Sep 2009 | A1 |
20090254592 | Marinov et al. | Oct 2009 | A1 |
20090265396 | Ram et al. | Oct 2009 | A1 |
20100017643 | Baba et al. | Jan 2010 | A1 |
20100077294 | Watson | Mar 2010 | A1 |
20100082542 | Feng et al. | Apr 2010 | A1 |
20100205206 | Rabines et al. | Aug 2010 | A1 |
20100211547 | Kamei et al. | Aug 2010 | A1 |
20100325634 | Ichikawa et al. | Dec 2010 | A1 |
20110083185 | Sheleheda et al. | Apr 2011 | A1 |
20110087696 | Lacapra | Apr 2011 | A1 |
20110093471 | Brockway et al. | Apr 2011 | A1 |
20110107112 | Resch | May 2011 | A1 |
20110119234 | Schack et al. | May 2011 | A1 |
20110320882 | Beaty et al. | Dec 2011 | A1 |
20120144229 | Nadolski | Jun 2012 | A1 |
20120150699 | Trapp et al. | Jun 2012 | A1 |
Number | Date | Country |
---|---|---|
2003300350 | Jul 2004 | AU |
2080530 | Apr 1994 | CA |
2512312 | Jul 2004 | CA |
0605088 | Feb 1996 | EP |
0 738 970 | Oct 1996 | EP |
63010250 | Jan 1988 | JP |
6205006 | Jul 1994 | JP |
06-332782 | Dec 1994 | JP |
8021924 | Mar 1996 | JP |
08-328760 | Dec 1996 | JP |
08-339355 | Dec 1996 | JP |
9016510 | Jan 1997 | JP |
11282741 | Oct 1999 | JP |
2000-183935 | Jun 2000 | JP |
566291 | Dec 2008 | NZ |
0239696 | May 2002 | WO |
WO 02056181 | Jul 2002 | WO |
WO 2004061605 | Jul 2004 | WO |
2006091040 | Aug 2006 | WO |
WO 2008130983 | Oct 2008 | WO |
WO 2008147973 | Dec 2008 | WO |
Entry |
---|
“The AFS File System in Distributed Computing Environment”, www.transarc.ibm.com/Library/whitepapers/AFS/afsoverview.html, last accessed on Dec. 20, 2002. |
Aguilera, Marcos K. et al., “Improving recoverability in multi-tier storage systems”, International Conference on Dependable Systems and Networks (DSN-2007), Edinburgh, Scotland, Jun. 2007, 10 pages. |
Anderson, Darrell C. et al., “Interposed Request Routing for Scalable Network Storage”, ACM Transactions on Computer Systems 20(1): (Feb. 2002), pp. 1-24. |
Anderson et al., “Serverless Network File System”, in the 15th Symposium on Operating Systems Principles, Dec. 1995, Association for Computing Machinery, Inc. |
Anonymous, “How DFS Works: Remote File Systems”, Distributed File System (DFS) Technical Reference, retrieved from the Internet on Feb. 13, 2009: URL<:http://technetmicrosoft.com/en-us/library/cc782417W5.10,printer).aspx> (Mar. 2003). |
Apple, Inc., “Mac OS X Tiger Keynote Intro. Part 2”, Jun. 2004, www.youtube.com <http://www.youtube.com/watch?v=zSBJwEmR.JbY>, p. 1. |
Apple, Inc., “Tiger Developer Overview Series: Working with Spotlight”, Nov. 23, 2004, www.apple.com using www.archive.org <http ://web.archive.org/web/20041123005335/developer.apple.com/macosx/tiger/sp otlight.html>, pp. 1-6. |
“Auspex Storage Architecture Guide”, Second Edition, 2001, Auspex Systems, Inc., www.auspex.com, last accessed on Dec. 30, 2002. |
Cabrera et al., “Swift: Storage Architecture for Large Objects”, in Proceedings of the-Eleventh IEEE Symposium on Mass Storage Systems, pp. 123-128, Oct. 1991. |
Cabrera et al., “Swift: Using Distributed Disk Striping to Provide High I/O Data Rates”, Computing Systems 4, 4 (Fall 1991), pp. 405-436. |
Cabrera et al., “Using Data Striping in a Local Area Network”, 1992, technical report No. UCSC-CRL-92-09 of the Computer & Information Sciences Department of University of California at Santa Cruz. |
Callaghan et al., “NFS Version 3 Protocol Specifications” (RFC 1813), Jun. 1995, The Internet Engineering Task Force (IETN), www.ietf.org, last accessed on Dec. 30, 2002. |
Carns et al., “PVFS: A Parallel File System for Linux Clusters”, in Proceedings of the Extreme Linux Track: 4th Annual Linux Showcase and Conference, pp. 317-327, Atlanta, Georgia, Oct. 2000, USENIX Association. |
Cavale, M. R., “Introducing Microsoft Cluster Service (MSCS) in the Windows Server 2003”, Microsoft Corporation, Nov. 2002. |
“CSA Persistent File System Technology”, Colorado Software Architecture, Inc.: A White Paper, Jan. 1, 1999, p. 1-3, <http://www.cosoa.com/white—papers/pfs.php>. |
“Distributed File System: Logical View of Physical Storage: White Paper”, 1999, Microsoft Corp., www.microsoft.com, <http://www.eu.microsoft.com/TechNet/prodtechnol/windows2000serv/maintain/DFSnt95>, pp. 1-26, last accessed on Dec. 20, 2002. |
English Language Abstract of JP 08-328760 from Patent Abstracts of Japan. |
English Language Abstract of JP 08-339355 from Patent Abstracts of Japan. |
English Translation of paragraphs 17, 32, and 40-52 of JP 08-328760. |
English Translation of Notification of Reason(s) for Refusal for JP 2002-556371 (Dispatch Date: Jan. 22, 2007). |
Fan et al., “Summary Cache: A Scalable Wide-Area Protocol”, Computer Communications Review, Association Machinery, New York, USA, Oct. 1998, vol. 28, Web Cache Sharing for Computing No. 4, pp. 254-265. |
Farley, M., “Building Storage Networks”, Jan. 2000, McGraw Hill, ISBN 0072120509. |
Gibson et al., “File Server Scaling with Network-Attached Secure Disks”, in Proceedings of the ACM International Conference on Measurement and Modeling of Computer Systems (Sigmetrics '97), Jun. 15-18, 1997, Association for Computing Machinery, Inc. |
Gibson et al., “NASD Scalable Storage Systems”, Jun. 1999, USENIX99, Extreme Linux Workshop, Monterey, California. |
Harrison, C., May 19, 2008 response to Communication pursuant to Article 96(2) EPC dated Nov. 9, 2007 in corresponding European patent application No. 02718824.2. |
Hartman, J., “The Zebra Striped Network File System”, 1994, Ph.D. dissertation submitted in the Graduate Division of the University of California at Berkeley. |
Haskin et al., “The Tiger Shark File System”, 1996, in proceedings of IEEE, Spring COMPCON, Santa Clara, CA, www.research.ibm.com, last accessed on Dec. 30, 2002. |
Hu, J., Final Office action dated Sep. 21, 2007 for related U.S. Appl. No. 10/336,784. |
Hu, J., Office action dated Feb. 6, 2007 for related U.S. Appl. No. 10/336,784. |
Hwang et al., “Designing SSI Clusters with Hierarchical Checkpointing and Single 1/0 Space”, IEEE Concurrency, pp. 60-69, Jan.-Mar. 1999. |
International Search Report for International Patent Application No. PCT/US2008/083117 (Jun. 23, 2009). |
International Search Report for International Patent Application No. PCT/US2008/060449 (Apr. 9, 2008). |
International Search Report for International Patent Application No. PCT/US2008/064677 (Sep. 6, 2009). |
International Search Report for International Patent Application No. PCT/US02/00720, Jul. 8, 2004. |
International Search Report from International Application No. PCT/US03/41202, mailed Sep. 15, 2005. |
Karamanolis, C. et al., “An Architecture for Scalable and Manageable File Services”, HPL-2001-173, Jul. 26, 2001. p. 1-114. |
Katsurashima, W. et al., “NAS Switch: A Novel CIFS Server Virtualization, Proceedings”, 20th IEEE/11th NASA Goddard Conference on Mass Storage Systems and Technologies, 2003 (MSST 2003), Apr. 2003. |
Kimball, C.E. et al., “Automated Client-Side Integration of Distributed Application Servers”, 13Th LISA Conf., 1999, pp. 275-282 of the Proceedings. |
Klayman, J., Nov. 13, 2008 e-mail to Japanese associate including instructions for response to office action dated May 26, 2008 in corresponding Japanese patent application No. 2002-556371. |
Klayman, J., Response filed by Japanese associate to office action dated Jan. 22, 2007 in corresponding Japanese patent application No. 2002-556371. |
Klayman, J., Jul. 18, 2007 e-mail to Japanese associate including instructions for response to office action dated Jan. 22, 2007 in corresponding Japanese patent application No. 2002-556371. |
Kohl et al., “The Kerberos Network Authentication Service (V5)”, RFC 1510, Sep. 1993. (http://www.ietf.org/ rfc/rfc1510.txt?number=1510). |
Korkuzas, V., Communication pursuant to Article 96(2) EPC dated Sep. 11, 2007 in corresponding European patent application No. 02718824.2-2201. |
Lelil, S., “Storage Technology News: AutoVirt adds tool to help data migration projects”, Feb. 25, 2011, last accessed Mar. 17, 2011, <http://searchstorage.techtarget.com/news/article/0,289142, sid5—gci1527986,00. html>. |
Long et al., “Swift/RAID: A distributed RAID System”, Computing Systems, vol. 7, pp. 333-359, Summer 1994. |
“NERSC Tutorials: I/O on the Cray T3E, 'Chapter 8, Disk Striping'”, National Energy Research Scientific Computing Center (NERSC), http://hpcfnersc.gov, last accessed on Dec. 27, 2002. |
Noghani et al., “A Novel Approach to Reduce Latency on the Internet: 'Component-Based Download”', Proceedings of the Computing, Las Vegas, NV, Jun. 2000, pp. 1-6 on the Internet: Intl Conf on Internet. |
Norton et al., “CIFS Protocol Version CIFS-Spec 0.9”, 2001, Storage Networking Industry Association (SNIA), www.snia.org, last accessed on Mar. 26, 2001. |
Patterson et al., “A case for redundant arrays of inexpensive disks (RAID)”, Chicago, Illinois, Jun. 1-3, 1998, in Proceedings of ACM SIGMOND conference on the Management of Data, pp. 109-116, Association for Computing Machinery, Inc., www.acm.org, last accessed on Dec. 20, 2002. |
Pearson, P.K., “Fast Hashing of Variable-Length Text Strings”, Comm. of the ACM, vol. 33, No. 6, Jun. 1990. |
Peterson, M., “Introducing Storage Area Networks”, Feb 1998, InfoStor, www.infostor.com, last accessed on Dec. 20, 2002. |
Preslan et al., “Scalability and Failure Recovery in a Linux Cluster File System”, in Proceedings of the 4th Annual Linux Showcase & Conference, Atlanta, Georgia, Oct. 10-14,2000, pp. 169-180 of the Proceedings, www.usenix.org, last accessed on Dec. 20, 2002. |
Response filed Jul. 6, 2007 to Office action dated Feb. 6, 2007 for related U.S. Appl. No. 10/336,784. |
Response filed Mar. 20, 2008 to Final Office action dated Sep. 21,2007 for related U.S. Appl. No. 10/336,784. |
Rodriguez et al., “Parallel-access for mirror sites in the Internet”, InfoCom 2000. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings. IEEE Tel Aviv, Israel Mar. 26-30, 2000, Piscataway, NJ, USA, IEEE, US, pp. 864-873, XP010376176 ISBN: 0/7803-5880—5 p. 867, col. 2, last paragraph—p. 868, col. 1, paragraph 1. |
Rsync, “Welcome to the RSYNC Web Pages”, Retrieved from the Internet URL: http://samba.anu.edu.ut.rsync/. (Retrieved on Dec. 18, 2009). |
Savage, et al., “AFRAID—A Frequently Redundant Array of Independent Disks”, 1996 USENIX Technical Conf., San Diego, California, Jan. 22-26, 1996. |
“Scaling Next Generation Web Infrastructure with Content-Intelligent Switching: White Paper”, Apr. 2000, p. 1-9 Alteon Web Systems, Inc. |
Soltis et al., “The Design and Performance of a Shared Disk File System for IRIX”, in Sixth NASA Goddard Space Flight Center Conference on Mass Storage and Technologies in cooperation with the Fifteenth IEEE Symposium on Mass Storage Systems, Mar. 23-26, 1998. |
Soltis, et al., “The Global File System”, in Proceedings of the Fifth NASA Goddard Space Flight Center Conference on Mass Storage Systems and Technologies, Sep. 17-19, 1996, College Park, Maryland. |
Sorenson, K.M., “Installation and Administration: Kimberlite Cluster Version 1.1.0, Rev. Dec. 2000”, Mission Critical Linux, http://oss.missioncriticallinux.corn/kimberlite/kimberlite.pdf. |
Stakutis, C., “Benefits of SAN-based file system sharing”, Jul. 2000, InfoStor, www.infostor.com, last accessed on Dec. 30, 2002. |
Thekkath et al., “Frangipani: A Scalable Distributed File System”, in Proceedings of the 16th ACM Symposium on Operating Systems Principles, Oct. 1997, Association for Computing Machinery, Inc. |
Uesugi, H., Nov. 26, 2008 amendment filed by Japanese associate in response to office action dated May 26, 2008 in corresponding Japanese patent application No. 2002556371. |
Uesugi, H., English translation of office action dated May 26, 2008 in corresponding Japanese patent application No. 2002-556371. |
Uesugi, H., Jul. 15, 2008 letter from Japanese associate reporting office action dated May 26, 2008 in corresponding Japanese patent application No. 2002-556371. |
“VERITAS SANPoint Foundation Suite(tm) and SANPoint Foundation Suite(tm) HA: New VERITAS vol. Management and File System Technology for Cluster Environments”, Sep. 2001, VERITAS Software Corp. |
Wilkes, J., et al., “The HP AutoRAID Hierarchical Storage System”, ACM Transactions on Computer Systems, vol. 14, No. 1, Feb. 1996. |
“Windows Clustering Technologies—An Overview”, Nov. 2001, Microsoft Corp., www.microsoft.com, last accessed on Dec. 30, 2002. |
Zayas, E., “AFS-3 Programmer's Reference: Architectural Overview”, Transarc Corp., version 1.0 of Sep. 2, 1991, doc. No. FS-00-D160. |
Basney, Jim et al., “Credential Wallets: A Classification of Credential Repositories Highlighting MyProxy,” TPRC 2003, Sep. 19-21, 2003. |
Botzum, Keys, “Single Sign On—A Contrarian View,” Open Group Website, <http://www.opengroup.org/security/topics.htm>, Aug. 6, 2001, pp. 1-8. |
Novotny, Jason et al., “An Online Credential Repository for the Grid: MyProxy,” 2001, pp. 1-8. |
Pashalidis, Andreas et al., “A Taxonomy of Single Sign-On Systems,” 2003, pp. 1-16, Royal Holloway, University of London, Egham Surray, TW20, 0EX, United Kingdom. |
Pashalidis, Andreas et al., “Impostor: a single sign-on system for use from untrusted devices,” Global Telecommunications Conference, 2004, GLOBECOM '04, IEEE, Issue Date: Nov. 29-Dec. 3, 2004.Royal Holloway, University of London. |
Tulloch, Mitch, “Microsoft Encyclopedia of Security,” pp. 218, 300-301, Microsoft Press, 2003, Redmond, Washington. |
Gupta et al., “Algorithms for Packet Classification”, Computer Systems Laboratory, Stanford University, CA, Mar./Apr. 2001, pp. 1-29. |
Heinz Il G., “Priorities in Stream Transmission Control Protocol (SCTP) Multistreaming”, Thesis submitted to the Faculty of the University of Delaware, Spring 2003, pp. 1-35. |
Internet Protocol,“Darpa Internet Program Protocol Specification”, (RFC:791), Information Sciences Institute, University of Southern California, Sep. 1981, pp. 1-49. |
Ilvesmaki M., et al., “On the capabilities of application level traffic measurements to differentiate and classify Internet traffic”, Presented in SPIE's International Symposium ITcom, Aug. 19-21, 2001, pp. 1-11, Denver, Colorado. |
Modiano E., “Scheduling Algorithms for Message Transmission Over a Satellite Broadcast System,” MIT Lincoln Laboratory Advanced Network Group, Nov. 1997, pp. 1-7. |
Ott D., et al., “A Mechanism for TCP-Friendly Transport-level Protocol Coordination”, USENIX Annual Technical Conference, 2002, University of North Carolina at Chapel Hill, pp. 1-12. |
Padmanabhan V., et al., “Using Predictive Prefetching to Improve World Wide Web Latency”, SIGCOM, 1996, pp. 1-15. |
Rosen E., et al., “MPLS Label Stack Encoding”, (RFC:3032) Network Working Group, Jan. 2001, pp. 1-22, (http://www.ietrorg/rfc/rfc3032.txt). |
Wang B., “Priority and Realtime Data Transfer Over the Best-Effort Internet”, Dissertation Abstract, Sep. 2005, ScholarWorks©UMASS. |
Woo T.Y.C., “A Modular Approach to Packet Classification: Algorithms and Results”, Nineteenth Annual Conference of the IEEE Computer and Communications Societies 3(3):1213-22, Mar. 26-30, 2000, abstract only, (http://ieeexplore.ieee.org/xpl/freeabs—all.jsp?arnumber=832499). |
Number | Date | Country | |
---|---|---|---|
20090204649 A1 | Aug 2009 | US |
Number | Date | Country | |
---|---|---|---|
60987181 | Nov 2007 | US |