Parallel access virtual tape library and drives

Information

  • Patent Grant
  • 8706976
  • Patent Number
    8,706,976
  • Date Filed
    Tuesday, September 2, 2008
    16 years ago
  • Date Issued
    Tuesday, April 22, 2014
    10 years ago
Abstract
A system and method described herein allows a virtual tape library (VTL) to perform multiple simultaneous or parallel read/write or access sessions with disk drives or other storage media, particularly when subject to a sequential SCSI-compliant layer or traditional limitations of VTLs. In one embodiment, a virtualizing or transaction layer can establish multiple sessions with one or more clients to concurrently satisfy the read/write requests of those clients for physical storage resources. A table or other data structure tracks or maps the sessions associated with each client and the location of data on the physical storage devices.
Description
BACKGROUND

Virtual tape libraries (VTL) transform disk drives into virtual tape to improve backup/restore times, as well as to provide other benefits. A VTL operates by providing a layer or interface between a host or client computer and a disk drive system. This VTL layer acts like one or more tape drives with one or more tapes or pieces of media and mimics commands to and from the client computer to make the client believe that it is communicating with tape drives. Instead, however, the VTL actually exchanges data with one or more disk drives.


A Jan. 3, 2007 article by B. Pariseau in ComputerWeekly.com, entitled, “Storage Outlook '07: In Search of Better Data Management,” indicates that some consumers prefer tape and VTL because drives capable of up to 4 to 1 compression save storage resources. According to the article,

    • The VTL improves upon some of the traditional limitations with physical tape libraries. With VTLs [one] can create hundreds of virtual tape drives. Each server can have its own dedicated drives, reducing the complexity of shared resources. The SCSI tape protocol has some downsides shared by both VTL and physical tape . . . . Neither enables concurrent reads from the same volume by two separate hosts. This means tape replication for [data recovery] and restores cannot happen at the same time.


Tape by its nature is sequential and thus does not permit simultaneous read operations from different locations on a given tape. Disk drives, however, do permit simultaneous read and write commands. But, as noted above, a VTL forces the disk drive to operate like a tape drive and thus does not permit concurrent reads, or other capabilities incompatible with tape. Many limitations of VTLs exist because of a lack of disk management solutions or limitations in backup software applications that employ VTLs. For example, when a write command is provided to a drive, the VTL reserves that drive and no other commands can be performed.


The need exists for a system that overcomes the above problems, as well as providing additional benefits. Overall, the examples herein describe some prior or related systems, and their associated limitations are intended to be illustrative and not exclusive. Other limitations related to existing or prior systems will become apparent to those of skill in the art upon reading the following Detailed Description.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram illustrating an example of a VTL or parallel access virtual tape library and drives system.



FIG. 2A is a block diagram illustrating an alternate embodiment of the system shown in FIG. 1.



FIG. 2B is a block diagram illustrating components of an API or sequential SCSI-compliant layer shown in FIG. 2A.



FIG. 3 is a flow diagram illustrating an example of a routine for performing parallel access virtualization.



FIG. 4 is a virtual tape/disk table for use by the components of FIG. 2A.



FIG. 5 is a block diagram illustrating components of a data stream that may be used in a suitable data storage system.



FIG. 6 is a block diagram illustrating an example of a data storage system.



FIG. 7 is block diagram illustrating an example of components of a server used in data storage operations.





In the drawings, the same reference numbers and acronyms identify elements or acts with the same or similar functionality for ease of understanding and convenience. To easily identify the discussion of any particular element or act, the most significant digit or digits in a reference number generally refers to the Figure number in which that element is first introduced (e.g., element 310 is first introduced and discussed with respect to FIG. 3).


DETAILED DESCRIPTION

Various examples of the invention will now be described. The following description provides specific details for a thorough understanding and enabling description of these examples. One skilled in the art will understand, however, that the system may be practiced without many of these details. Additionally, some well-known structures or functions may not be shown or described in detail, so as to avoid unnecessarily obscuring the relevant description of the various examples.


The terminology used in the description presented below is intended to be interpreted in its broadest reasonable manner, even though it is being used in conjunction with a detailed description of certain specific examples of the system. Certain terms may even be emphasized below; however, any terminology intended to be interpreted in any restricted manner will be overtly and specifically defined as such in this Detailed Description.


Described in detail herein is a system that allows a VTL to perform multiple simultaneous or parallel read/write or access sessions with disk drives or other storage media, particularly when subject to a sequential SCSI-compliant layer or the traditional limitations of VTLs. In one embodiment described in detail herein, a virtualizing or transaction layer can establish multiple sessions with one or more clients to concurrently satisfy the read/write requests of those clients for physical storage resources. A table or other data structure then tracks or maps the sessions associated with each client and the location of data on the physical storage devices. The table may also identify virtual drives, virtual media, or other information, such as load data, content data, etc.


Referring to FIG. 1, an example of a parallel access virtual tape library and drive system 100 is shown. One or more client computers or hosts 102 communicate with a virtual tape or device library 103 that includes one or more disk drives 105. As explained in more detail below, the VTL 103 looks like one or more tape drives to the client, but in fact the VTL 103 employs disk drives and/or other non-sequential storage media. Indeed, the VTL 103 operates like existing virtual tape libraries, except as explained herein.


The VTL 103 causes a tape drive to show up at the operating system level of each client computer. The VTL 103 exposes multiple handles, addresses, or other APIs to the operating systems of each client computer, so that any application on those clients can use the (virtual) tape drive to perform any desired functions, including read or write commands. In this example, the VTL 103 employs one or more disk drives that effectively have a tape command front end that accepts commands through standard protocols, such as standard sequential SCSI protocols. The VTL 103 may employ standard file system-type disk management, such as Dynamic Disk-like functionality under Microsoft® Windows®, to dynamically grow or shrink allocated or available disk space for each session under this example (shown as file system 107).


Referring to FIG. 2A, an example of an alternative to the VTL 103 embodiment of FIG. 1 is shown as having an API or sequential SCSI-compliant layer 104 that communicates with both the clients and a lower layer 106, which includes a media agent/disk subsystem 108. The media agent/disk subsystem 108 may communicate with read and write processes 110 that receive and pass to drivers 112 read/write commands initiated by the clients. Data associated with these read/write commands may be cached in an optional intermediate cache 114, or may flow directly to one or more tape drives 116, one or more disk drives 118, one or more optical drives 120 (collectively, drives 116, 118, 120), or other data storage media and devices not shown. To the client computers 102, the system appears as mounted tape, but instead the media agent/disk subsystem 108 operates routines that receive blocks of data and store and manage them with respect to any form of secondary storage. This permits, for example, client 1 to write commands to a virtual tape drive at the same time that client 2 is reading from it.


The lower layer 106 operates, as known by those skilled in the relevant art, to implement communications and functions with data storage devices so as to satisfy read, write, or other data storage commands. Media agents may be employed to execute necessary subroutines for accessing appropriate drives, loading media, performing any necessary data manipulation (e.g., encryption/decryption, compression/decompression, error correction, de-duplication (often involving cryptographic hashing), etc.). The disk subsystem may include an associated file management system (not shown) that can include one or more file access tables (FAT) or other data structures for tracking locations of data stored on particular data storage drives 116, 118, 120. Note that the system may dynamically change the allocation of blocks of data on each drive 105, 116, 118, and/or 120, to permit greater flexibility and improve performance of the system. Note also that when multiple commands for the same tape, i.e., tape drive 116, are received, they may be stored temporarily in cache 114. The lower layer 106 can, for example, provide multiple concurrent lines of data flow to or from one or more disk drives 118. Note that the lower layer 106 provides access to multiple data storage media types, e.g., drives 116, 118, 120, even though the system receives tape drive commands from a client or host computer.


The media agent/disk subsystem 108 may communicate with other systems 113 so that data or commands from the clients receive further processing to provide further benefits. Such other operations can include snapshots, single instancing/de-duplication, data classification, load balancing, or other data management and data storage techniques, some of which are described in detail herein.


Referring to FIG. 2B, layer 104 includes a client interface 202 that receives read/write commands from the clients and passes them on to other components. For example, a session mapper 204 receives the read/write commands and responds by associating a session handle with each read/write request. The client interface 202 then passes the session handle to the appropriate client. The session mapper 204 also builds one or more data object mapping tables 206 to track read/write commands with session handles and physical locations of data written to and read from one or more drives or storage devices 105, 116, 118, and/or 120 (as described above with reference to FIGS. 1-2A). A data store interface 208 interfaces with these drives to implement the read/write commands.


The client interface 202 exposes a sequential SCSI interface (e.g., tape interface or API), or similar sequential media interface to any application running on the client computers. The client interface 202 thus accepts read/write commands like a typical tape drive. The data store interface 208 then passes standard VTL commands to any secondary storage device (via lower layer 106). Layer 104 of FIG. 2B acts as a standard VTL, but receive commands are based on sessions created and associated with received client commands via session mapper 204. In other words, an SCSI interface is provided on top of a file input/output handler.


Under an alternative implementation, components of FIG. 2A may be combined to form a storage appliance, similar to Network Attached Storage (NAS). Thus, within a single enclosure, components 104, 106, 113, 114, and one or more storage devices 116, 118, 120 can be combined to provide an appliance to client computers 102. Layer 104 can include a file system interface to provide uniform access to the appliance for the client's 102, while avoiding specific APIs, or sequential SCSI compliant commands. (The APIs of layer 104 can be specifically tailored for individual VTLs in the example above.) Thus, the file system of layer 104 will include uniform naming conventions and APIs through the file system, and provide client with a common or uniform interface (e.g., UNIX, Windows, etc.). The media agent/disc subsystem 106 is an independent component to help handle various processes that may be performed by the appliance, including snapshots, single instancing, data compression/decompression, data encryption/decryption, and other processes described herein.



FIG. 3 shows an example of a flow diagram illustrating VTL or virtualization routine 300. Beginning in block 302, the system receives read/write commands from a client, along with associated parameters. The parameters can indicate or specify a particular tape drive, media (typically identified by a barcode number), or possibly an offset, data size, etc. While barcodes are shown in this example, any identifier may be employed to identify, e.g., a desired disk. In block 304, the system returns a session label to the client, which the client then uses for subsequent association with that request. If the same media is requested by two different clients, then a different session handle is provided to each client.


While a session-based example is described herein, alternatively or additionally the system may employ a threaded architecture. Under a threaded architecture, the system could employ one or more threads to handle different requests from the same client. A multi-threaded architecture may be implemented at the operating system level to cater to multiple client requests simultaneously, where, e.g., Windows provides such access through APIs. The system described herein, in this example, creates threads via the specific Windows APIs. Each client may still be given a client session, where a thread is a parallel request. So multiple sessions can run in parallel, as assisted by the use of multiple threads under Windows.


In block 306, the system maps the received request and parameters to storage devices, as described below with respect to FIG. 4. In block 307, the system performs the read/write commands with respect to one or more drives and performs any additional operations, such as snapshotting, single instancing, data classification, load balancing, or other processes, described herein.


In decision block 308, the system determines whether a new request has been received. If so, the routine returns to block 302. If not, then in decision block 310 the system determines whether the read/write operation has been complete. If not, the routine moves back to 308; else, the session is released in block 312.


Referring to FIG. 4, an example of a virtual tape/disk table is shown. While a table is shown, any data structure may be employed. Here, the system provides a mapping for received read/write commands to associated physical data storage resources. For example, a “Client 1” provides a read/write command to the system and receives in return a session handle “Handle 1.” Parameters received from Client 1 include identification of a tape drive and media, “Drive 1” and “Barcode 1.” The system then maps that command to a particular storage device and block ID, which in this example is “Drive 1” and “Block ID 1,” and which can refer to a particular magnetic disk drive and one or more blocks on that drive.


Similar examples are evidenced from the table of FIG. 4. Note that Client 1 provides two additional requests, which are associated with “Handle 3” and “Handle 5,” where Handle 3 requests (tape) Drive 1 and media or Barcode 3, but the system maps that request to Drive 1 and Block ID 3. The system can provide simultaneous read/write commands to multiple clients and for different identified drives and media, or even for the same drives and media, despite the limitations of prior VTL technology. Of course, many more entries may exist in the table, and the table grows and shrinks dynamically as sessions are created and released.


As noted above, the system can provide other processing of the data under block 307. This can include, for example, performing characteristic analysis, such as data classification for data passing through the system. Further details regarding data classification may be found in the assignee's U.S. Patent Application No. 60/752,203, entitled “Systems and Methods for Classifying and Transferring Information in a Storage Network”. The system can also perform content indexing, which is described in the assignee's U.S. patent application Ser. No. 11/694,869, filed 30 Mar. 2007, entitled “Method and System for Offline Indexing of Content and Classifying Stored Data”. Further, the system can perform single instancing or deduplication, such as is described in the assignee's U.S. Patent Application No. 60/871,737, filed 22 Dec. 2006, entitled “System and Method for Storing Redundant Information”. Further, the system can provide load balancing operations among the one or more secondary storage drives 116, 118, 120, and dynamically allocate streams of data to and from such drives, as is described in further detail in the assignee's copending U.S. patent application Ser. No. 11/615,800, filed 22 Dec. 2006, entitled “Systems and Methods of Data Storage Management, Such as Dynamic Data Stream Allocation”. Note that the present system allows multiple data processing techniques to be employed in a virtual tape library. The system can perform many of various data processing and data storage management techniques before or during the writing (or reading) of data to the secondary storage drives 116, 118, 120.


One example of the type of further processing that may be utilized is the implementation of snapshots. Snapshots are a type of data backup known to those skilled in the relevant art. A snapshot driver, in this example, creates a snapshot of a disk drive, such as “Drive 1” referred to in FIG. 4, which has data from Clients 1 and 2 written to it. Multiple snapshots may be created over a period of time, with each snapshot associated with a particular time they were created. A user may then access a list or table of snapshots and select a point in time to which the user may wish to roll back and view an image of the disk at that specific time (based on when the selected snapshot was created). Each of the snapshots would store not only a bitmap image of the disk, but also the table that indicates block identifications or locations where data for each client is stored. Therefore, the system can identify those blocks associated with Client 1 and re-create an image of the data or disk based on the user-selected snapshot. Note that this snapshot represents (virtual tape. Thus, snapshots may be applied to a VTL under this example.


As long as snapshots are taken sequentially over a period of time and associated with the time at which those snapshots were taken, a user may be able to roll back and view the state of a virtual tape drive for a particular client at any time that a snapshot was created. The system can provide a user interface that allows a user to browse through the available snapshots. Alternatively or additionally, the system can store copies of each data object mapping table, such as the table of FIG. 4. If the time associated with each snapshot is added as another column to the table, the user can simply use such stored tables to identify a desired snapshot. Furthermore, if the table tracks individual files and times associated with snapshots, then a user may be able to obtain earlier versions of particular files by obtaining the appropriate bitmap associated with a client and file from the snapshot and pulling up the desired file copy. In other words, the system has granularity down to the file level assuming that the locations of individual files for each client on the secondary storage (e.g., disk 118) are tracked in the table.


Other operations, such as additional reads or writes, may be performed simultaneously on Drive 1. Thus, one operation may be performed while another is simultaneously performed. A read operation, for example, may be performed simultaneously with a replication of that same data, or a write operation for a particular data object being done simultaneously with replication of that data (albeit to a different portion of a particular drive or other drive of the system).


Another alternative is to employ an effectively unmodified VTL, and instead allow two drives to mount the same media. Traditionally, a VTL allows only one client to read or access a given tape drive and a given piece of media at one point of time. In this alternative, one or more clients may specify two or more drives. However, those clients would also specify the same media to be accessed. The VTL would then permit two or more simultaneous reads of that data back to the clients. In other words, under this alternative, an existing VTL would expose the same drive twice, such as to two different clients.


Under this alternative, the VTL itself will be able to do parallel or non-sequential SCSI commands, where two clients may request, e.g., reads at the same time for the same media. Again, sessions would be employed, but within the VTL. So, in one example each client/user session has a read request in each session, which may be mapped to non-sequential access on the same disk. Each client gains access to a tape drive and issues sequential reads in each session, which then become multiple random accesses on the disk. A middle layer converts these multiple sessions sequentially into random access on the disk and returns the clients responses as sequential reads. (This middle layer is in the VTL in this alternative, as opposed to that described in the system above, but similar operations occur.) Overall, each client believes it has requested and received a sequential read, but internally it is random access to the disk.


Such an alternative would require only minor or no modifications to existing VTLs to permit such functionality. For example, changes may be required to client-side applications that use the VTL. The client-side application using the VTL will have to manage its own requests to the VTL, e.g., requests to mount media in a different drive and to know that the drive is actually using that media, regardless of requests by other clients. With an unmodified VTL, the application must be modified e.g., to know that it is reading media already mounted in another drive.


In this alternative, a drawback would be that the client would have to do some management on its own to track the media. For example, the application's regular operation or logic requires that if one tape is already being used by a drive, then it cannot be used by another drive. Here, the application would need to modify that logic so that even if a certain tape is being used by one client, it could mount the same tape in a different drive.


Alternatively or additionally, minor modifications to the VTL may be implemented. Normally, if a given piece of media is already mounted in a drive, the VTL does not allow that media to be mounted to a different drive. So, here, a minor modification to the VTL would be to simply allow it to mount the same media to a different drive.


Yet another alternative is to expose the file system employed by the VTL. By providing access to the VTL file system (via appropriate APIs), simultaneous reads of data identified in the VTL file system may be performed. VTLs typically do not expose their underlying file system, i.e., how they store data on the disk, etc., but instead just provide a sequential access interface (e.g., sequential SCSI interface), which does not describe exactly how files are laid on disk by the VTL file system. In other words, the VTL completely manages, and completely hides, how files are managed and distributed by the VTLs internal disk drive(s).


If the VTL file system instead were opened up (via appropriate APIs, like in Microsoft Windows), simultaneous reads, writes, and other access may be provided, as described herein. However, such an alternative would require a VTL manufacturer or VTL subsystem to open up the underlying file system, which may require the manufacturer to manage multiple, concurrent write sessions, etc., thus potentially expanding the required scope of modification.


Suitable System


The above parallel access VTL system may be incorporated within a data storage system and may be subjected to a data stream during a data copy operation. Referring to FIG. 5, a block diagram illustrating components of a data stream 510 utilized by a suitable data storage and recovery system is shown. The data stream 510 may include client 102, a media agent 512, and a secondary storage device, such as VTL 103. For example, in storage operations, the system may store, receive, and/or prepare data to be stored, copied, or backed up at a server or client. The system may then transfer the data to be stored to media agent 512, which may then refer to storage policies, schedule policies, and/retention policies (and other policies) to choose a secondary storage device. The media agent 512 may include or be associated with an intermediate component, to be discussed herein.


The secondary storage device receives the data from the media agent 512 and stores the data as a secondary copy, such as a backup copy. Secondary storage devices may be magnetic tapes, optical disks, USB and other similar media, disk, and tape drives, and so on. Of course, the system may employ other configurations of data stream components not shown in FIG. 6.


Referring to FIG. 6, a block diagram illustrating an example of a data storage and recovery system 600 is shown. Data storage systems may contain some or all of the following components, depending on the needs of the system. FIG. 6 and the discussion herein provide a brief, general description of a suitable computing environment in which the system can be implemented.


For example, the data storage system 600 contains a storage manager 610, one or more clients 102, one or more media agents 512, and one or more storage devices 103, 613. Storage manager 610 controls media agents 512, which may be responsible for transferring data to storage devices. Storage manager 610 includes a jobs agent 611, a management agent 612, a database 613, and/or an interface module 614. Storage manager 610 communicates with client(s) 102. One or more clients may access data that is stored by the system from database 622 via a data agent 621. The system uses media agents 512, which contain databases 631, to transfer and store data into storage devices. Client databases 622 may contain data files and other information, while media agent databases 631 may contain indices and other data structures that assist and implement the storage of data in secondary storage devices, for example.


The data storage and recovery system may include software and/or hardware components and modules used in data storage operations. The components may be storage resources that function to copy data during storage operations. The components may perform other storage operations (or storage management operations) other that operations used in data stores. For example, some resources may create, store, retrieve, and/or migrate primary or secondary data copies. Additionally, some resources may create indices and other tables relied upon by the data storage system and other data recovery systems. The secondary copies may include snapshot copies and associated indices, but may also include other backup copies such as HSM copies, archive copies, and so on. The resources may also perform storage management functions that may communicate information to higher level components, such as global management resources.


In some examples, the system performs storage operations based on storage policies, as mentioned above. For example, a storage policy includes a set of preferences or other criteria to be considered during storage operations. The storage policy may determine or define a storage location and/or set of preferences about how the system transfers data to the location and what processes the system performs on the data before, during, or after the data transfer. In some cases, a storage policy may define a logical bucket in which to transfer, store, or copy data from a source to a data store, such as storage media. Storage policies may be stored in storage manager 610, or may be stored in other resources, such as a global manager, a media agent, and so on. Further details regarding storage management and resources for storage management will now be discussed.


Referring to FIG. 7, a block diagram illustrating an example of components of a server or system 700 used in data storage operations is shown. A server, such as storage manager 610, may communicate with clients 102 to determine data to be copied to storage media. As described above, the storage manager 610 may contain a jobs agent 611, a management agent 612, a database 613, an interface module 614, and/or other agents 720. Jobs agent 611 may manage and control the scheduling of jobs (such as copying data files) from clients 102 to media agents 512 (not shown). Management agent 612 may control the overall functionality and processes of the data storage system or may communicate with global managers. Database 613 or another data structure may store storage policies, schedule policies, retention policies, or other information, such as historical storage statistics, storage trend statistics, and so on. Interface module 615 may interact with a user interface, enabling the system to present information to administrators and receive feedback or other input from the administrators or with other components of the system (such as via APIs).


CONCLUSION

Systems and modules described herein may comprise software, firmware, hardware, or any combination(s) of software, firmware, or hardware suitable for the purposes described herein. Software and other modules may reside on servers, workstations, personal computers, computerized tablets, PDAs, and other devices suitable for the purposes described herein. Modules described herein may be executed by a general-purpose computer, e.g., a server computer, wireless device, or personal computer. Those skilled in the relevant art will appreciate that aspects of the invention can be practiced with other communications, data processing, or computer system configurations, including: Internet appliances, hand-held devices (including personal digital assistants (PDAs)), wearable computers, all manner of cellular or mobile phones, multi-processor systems, microprocessor-based or programmable consumer electronics, set-top boxes, network PCs, mini-computers, mainframe computers, and the like. Indeed, the terms “computer,” “server,” “host,” “host system,” and the like, are generally used interchangeably herein and refer to any of the above devices and systems, as well as any data processor. Furthermore, aspects of the invention can be embodied in a special purpose computer or data processor that is specifically programmed, configured, or constructed to perform one or more of the computer-executable instructions explained in detail herein.


Software and other modules may be accessible via local memory, a network, a browser, or other application in an ASP context, or via another means suitable for the purposes described herein. Examples of the technology can also be practiced in distributed computing environments where tasks or modules are performed by remote processing devices, which are linked through a communications network, such as a Local Area Network (LAN), Wide Area Network (WAN), or the Internet. In a distributed computing environment, program modules may be located in both local and remote memory storage devices. Data structures described herein may comprise computer files, variables, programming arrays, programming structures, or any electronic information storage schemes or methods, or any combinations thereof, suitable for the purposes described herein. User interface elements described herein may comprise elements from graphical user interfaces, command line interfaces, and other interfaces suitable for the purposes described herein.


Examples of the technology may be stored or distributed on computer-readable media, including magnetically or optically readable computer disks, hard-wired or preprogrammed chips (e.g., EEPROM semiconductor chips), nanotechnology memory, biological memory, or other data storage media. Indeed, computer-implemented instructions, data structures, screen displays, and other data under aspects of the invention may be distributed over the Internet or over other networks (including wireless networks), on a propagated signal on a propagation medium (e.g., an electromagnetic wave(s), a sound wave, etc.) over a period of time, or they may be provided on any analog or digital network (packet switched, circuit switched, or other scheme).


Unless the context clearly requires otherwise, throughout the description and the claims, the words “comprise,” “comprising,” and the like are to be construed in an inclusive sense, as opposed to an exclusive or exhaustive sense; that is to say, in the sense of “including, but not limited to.” As used herein, the terms “connected,” “coupled,” or any variant thereof, means any connection or coupling, either direct or indirect, between two or more elements; the coupling or connection between the elements can be physical, logical, or a combination thereof. Additionally, the words “herein,” “above,” “below,” and words of similar import, when used in this application, refer to this application as a whole and not to any particular portions of this application. Where the context permits, words in the above Detailed Description using the singular or plural number may also include the plural or singular number respectively. The word “or,” in reference to a list of two or more items, covers all of the following interpretations of the word: any of the items in the list, all of the items in the list, and any combination of the items in the list.


The above Detailed Description is not intended to be exhaustive or to limit the invention to the precise form disclosed above. While specific examples for the invention are described above for illustrative purposes, various equivalent modifications are possible within the scope of the invention, as those skilled in the relevant art will recognize. For example, while processes or blocks are presented in a given order, alternative implementations may perform routines having steps, or employ systems having blocks, in a different order, and some processes or blocks may be deleted, moved, added, subdivided, combined, and/or modified to provide alternative or subcombinations. Each of these processes or blocks may be implemented in a variety of different ways. Also, while processes or blocks are at times shown as being performed in series, these processes or blocks may instead be performed or implemented in parallel, or may be performed at different times. Further any specific numbers noted herein are only examples: alternative implementations may employ differing values or ranges.


The teachings of the invention provided herein can be applied to other systems, not necessarily the system described above. The elements and acts of the various examples described above can be combined to provide further implementations of the invention.


Any patents and applications and other references noted above, including any that may be listed in accompanying filing papers, are incorporated herein by reference. Aspects of the invention can be modified, if necessary, to employ the systems, functions, and concepts of the various references described above to provide yet further implementations of the invention.


These and other changes can be made to the invention in light of the above Detailed Description. While the above description describes certain examples of the invention and describes the best mode contemplated, no matter how detailed the above appears in text, the invention can be practiced in many ways. Details of the system may vary considerably in its specific implementation, while still being encompassed by the invention disclosed herein. As noted above, particular terminology used when describing certain features or aspects of the invention should not be taken to imply that the terminology is being redefined herein to be restricted to any specific characteristics, features, or aspects of the invention with which that terminology is associated. In general, the terms used in the following claims should not be construed to limit the invention to the specific examples disclosed in the specification, unless the above Detailed Description section explicitly defines such terms. Accordingly, the actual scope of the invention encompasses not only the disclosed examples, but also all equivalent ways of practicing or implementing the invention under the claims.


While certain aspects of the invention are presented below in certain claim forms, the applicant contemplates the various aspects of the invention in any number of claim forms. For example, while only one aspect of the invention is recited as a means-plus-function claim under 35 U.S.C. §112, ¶6, other aspects may likewise be embodied as a means-plus-function claim, or in other forms, such as being embodied in a computer-readable medium. (Any claims intended to be treated under 35 U.S.C. §112, ¶6, will begin with the words “means for”.) Accordingly, the applicant reserves the right to add additional claims after filing the application to pursue such additional claim forms for other aspects of the invention.

Claims
  • 1. A computer-implementable method of processing commands for a virtual tape library (VTL), wherein the VTL receives sequential access write commands from two or more host computers, and wherein the VTL reads data from or writes data to at least two random access data storage devices within the VTL, the method comprising: intercepting sequential access write commands from the two or more host computers, wherein the sequential access write commands from the two or more host computers are commands to write data to sequential data storage media, wherein each sequential access write command includes associated parameters, and wherein the associated parameters specify a particular sequential access data storage drive and a piece of removable, sequential access, data storage media;in response to each intercepted sequential access write command from the two or more host computers, returning a session identifier to each host computer, wherein if the same piece of removable, sequential access, data storage media is requested by two different host computers, then a different session handle is provided to each of the two different host computers, andwherein if the same sequential access data storage drive is requested by two different host computers, then a different session handle is provided to each of the two different host computers;tracking, via a data structure, the session identifiers, the parameters, the at least two random access data storage devices within the VTL, and each intercepted sequential access write command from the two or more host computers;in response to each intercepted sequential access write command from the two or more host computers, performing write commands with respect to the at least two random access data storage devices within the VTL;determining whether any of the intercepted sequential access write commands have been completed;if any one of the intercepted sequential access write commands has been completed, then releasing a session associated with a completed sequential access write command; andgenerating multiple snapshots of the data structure, wherein each of the multiple snapshots is generated at a different time; andproviding a user interface for displaying and permitting user access to data associated with each of the multiple snapshots.
  • 2. The computer-implementable method of claim 1, further comprising: in response to intercepted sequential access write commands, assigning one or more threads to manage different write requests from the same host computer.
  • 3. The computer-implementable method of claim 1, further comprising: performing at least one additional data storage operation on data to be written to the at least two random access data storage devices within the VTL, wherein the additional data storage operation is single instancing, classifying data, load balancing, or generating snapshots.
  • 4. The computer-implementable method of claim 1, further comprising: generating multiple snapshots of the at least two random access data storage devices within the VTL, wherein each of the multiple snapshots is generated at a different time; andproviding a user interface for displaying and permitting user access to data associated with each of the multiple snapshots.
  • 5. A non-transitory computer-readable medium containing code, which when executed by a computer causes a virtual tape library (VTL) to perform a method to permit parallel access to two or more drives associated with the VTL, the method comprising: intercepting sequential access write commands from two or more host computers, wherein the sequential access write commands from the two or more host computers are commands to write data to sequential data storage media, wherein each sequential access write command includes associated parameters, and wherein the associated parameters specify a particular sequential access data storage drive and a piece of removable, sequential access, data storage media;in response to each intercepted sequential access write command from the two or more host computers, returning a session identifier to each host computer, wherein if the same piece of removable, sequential access, data storage media is requested by two different host computers, then providing a different session handle to each of the two different host computers, andwherein if the same sequential access data storage drive is requested by two different host computers, then providing a different session handle to each of the two different host computers;tracking, via a data structure, the session identifiers, the parameters, the at least two random access data storage devices within the VTL, and each intercepted sequential access write command from the two or more host computers;in response to each intercepted sequential access write command from the two or more host computers, performing write commands with respect to the at least two random access data storage devices within the VTL;determining whether any of the intercepted sequential access write commands have been completed;if any one of the intercepted sequential access write commands has been completed, then releasing a session associated with a completed sequential access write command; andgenerating multiple snapshots of the data structure, wherein each of the multiple snapshots is generated at a different time; andproviding a user interface for displaying and permitting user access to data associated with each of the multiple snapshots.
  • 6. The computer-readable medium of claim 5, wherein the method further comprises: in response to intercepted sequential access write commands, assigning one or more threads to manage different write requests from the same host computer.
  • 7. The computer-readable medium of claim 5, wherein the method further comprises: performing at least one additional data storage operation on data to be written to the at least two random access data storage devices within the VTL, wherein the additional data storage operation is single instancing, classifying data, load balancing, or generating snapshots.
CROSS-REFERENCE TO RELATED APPLICATION(S)

This application claims the benefit of the assignee's U.S. Provisional Patent Application Nos. 60/969,103 and 60/969,105, both filed Aug. 30, 2007.

US Referenced Citations (315)
Number Name Date Kind
4686620 Ng Aug 1987 A
4995035 Cole et al. Feb 1991 A
5005122 Griffin et al. Apr 1991 A
5093912 Dong et al. Mar 1992 A
5133065 Cheffetz et al. Jul 1992 A
5193154 Kitajima et al. Mar 1993 A
5212772 Masters May 1993 A
5226157 Nakano et al. Jul 1993 A
5239647 Anglin et al. Aug 1993 A
5241164 Pavlidis et al. Aug 1993 A
5241668 Eastridge et al. Aug 1993 A
5241670 Eastridge et al. Aug 1993 A
5265159 Kung Nov 1993 A
5276860 Fortier et al. Jan 1994 A
5276867 Kenley et al. Jan 1994 A
5287500 Stoppani, Jr. Feb 1994 A
5321816 Rogan et al. Jun 1994 A
5333315 Saether et al. Jul 1994 A
5347653 Flynn et al. Sep 1994 A
5410700 Fecteau et al. Apr 1995 A
5412668 Dewey May 1995 A
5448724 Hayashi et al. Sep 1995 A
5455926 Keele et al. Oct 1995 A
5491810 Allen Feb 1996 A
5495457 Takagi et al. Feb 1996 A
5495607 Pisello et al. Feb 1996 A
5499364 Klein et al. Mar 1996 A
5504873 Martin et al. Apr 1996 A
5506986 Healy Apr 1996 A
5544345 Carpenter et al. Aug 1996 A
5544347 Yanai et al. Aug 1996 A
5548521 Krayer et al. Aug 1996 A
5559957 Balk Sep 1996 A
5619644 Crockett et al. Apr 1997 A
5638509 Dunphy et al. Jun 1997 A
5673381 Huai et al. Sep 1997 A
5677900 Nishida et al. Oct 1997 A
5699361 Ding et al. Dec 1997 A
5729743 Squibb Mar 1998 A
5751997 Kullick et al. May 1998 A
5758359 Saxon May 1998 A
5761677 Senator et al. Jun 1998 A
5764972 Crouse et al. Jun 1998 A
5778395 Whiting et al. Jul 1998 A
5812398 Nielsen Sep 1998 A
5813009 Johnson et al. Sep 1998 A
5813017 Morris Sep 1998 A
5815662 Ong Sep 1998 A
5832522 Blickenstaff et al. Nov 1998 A
5860068 Cook Jan 1999 A
5875478 Blumenau Feb 1999 A
5875481 Ashton et al. Feb 1999 A
5887134 Ebrahim Mar 1999 A
5893139 Kamiyama et al. Apr 1999 A
5898593 Baca et al. Apr 1999 A
5901327 Ofek May 1999 A
5924102 Perks Jul 1999 A
5950205 Aviani, Jr. Sep 1999 A
5958005 Thorne et al. Sep 1999 A
5974563 Beeler, Jr. Oct 1999 A
5978577 Rierden et al. Nov 1999 A
6021415 Cannon et al. Feb 2000 A
6026398 Brown et al. Feb 2000 A
6026414 Anglin Feb 2000 A
6052735 Ulrich et al. Apr 2000 A
6076148 Kedem et al. Jun 2000 A
6088694 Burns et al. Jul 2000 A
6094416 Ying Jul 2000 A
6131095 Low et al. Oct 2000 A
6131099 Johnson et al. Oct 2000 A
6131147 Takagi Oct 2000 A
6131190 Sidwell Oct 2000 A
6137864 Yaker Oct 2000 A
6148412 Cannon et al. Nov 2000 A
6154738 Call Nov 2000 A
6154787 Urevig et al. Nov 2000 A
6161111 Mutalik et al. Dec 2000 A
6167402 Yeager Dec 2000 A
6195794 Buxton Feb 2001 B1
6212512 Barney et al. Apr 2001 B1
6223205 Harchol-Balter et al. Apr 2001 B1
6246882 Lachance Jun 2001 B1
6260069 Anglin Jul 2001 B1
6266784 Hsiao et al. Jul 2001 B1
6269382 Cabrera et al. Jul 2001 B1
6269431 Dunham Jul 2001 B1
6275953 Vahalia et al. Aug 2001 B1
6301592 Aoyama et al. Oct 2001 B1
6304880 Kishi Oct 2001 B1
6308245 Johnson et al. Oct 2001 B1
6324581 Xu et al. Nov 2001 B1
6328766 Long Dec 2001 B1
6330570 Crighton et al. Dec 2001 B1
6330572 Sitka Dec 2001 B1
6330642 Carteau Dec 2001 B1
6338006 Jesionowski et al. Jan 2002 B1
6343324 Hubis et al. Jan 2002 B1
RE37601 Eastridge et al. Mar 2002 E
6353878 Dunham Mar 2002 B1
6356801 Goodman et al. Mar 2002 B1
6356901 MacLeod et al. Mar 2002 B1
6366900 Hu Apr 2002 B1
6374336 Peters et al. Apr 2002 B1
6389432 Pothapragada et al. May 2002 B1
6418441 Call Jul 2002 B1
6418478 Ignatius et al. Jul 2002 B1
6421711 Blumenau et al. Jul 2002 B1
6434682 Ashton et al. Aug 2002 B1
6457017 Watkins et al. Sep 2002 B2
6484166 Maynard Nov 2002 B1
6487561 Ofek et al. Nov 2002 B1
6490666 Cabrera et al. Dec 2002 B1
6496744 Cook Dec 2002 B1
6519679 Devireddy et al. Feb 2003 B2
6538669 Lagueux, Jr. et al. Mar 2003 B1
6542972 Ignatius et al. Apr 2003 B2
6564228 O'Connor May 2003 B1
6616047 Catan Sep 2003 B2
6658436 Oshinsky et al. Dec 2003 B2
6658526 Nguyen et al. Dec 2003 B2
6662281 Ballard et al. Dec 2003 B2
6669832 Saito et al. Dec 2003 B1
6674924 Wright et al. Jan 2004 B2
6721334 Ketcham Apr 2004 B1
6757794 Cabrera et al. Jun 2004 B2
6771595 Gilbert et al. Aug 2004 B1
6785078 Basham et al. Aug 2004 B2
6789161 Blendermann et al. Sep 2004 B1
6802025 Thomas et al. Oct 2004 B1
6820035 Zahavi Nov 2004 B1
6851031 Trimmer et al. Feb 2005 B2
6909356 Brown et al. Jun 2005 B2
6922687 Vernon Jul 2005 B2
6934879 Misra et al. Aug 2005 B2
6941370 Boies et al. Sep 2005 B2
6950723 Gallo et al. Sep 2005 B2
6968351 Butterworth Nov 2005 B2
6968479 Wyatt et al. Nov 2005 B2
6973369 Trimmer et al. Dec 2005 B2
6973553 Archibald, Jr. et al. Dec 2005 B1
6983351 Gibble et al. Jan 2006 B2
7006435 Davies et al. Feb 2006 B1
7010387 Lantry et al. Mar 2006 B2
7012529 Sajkowsky Mar 2006 B2
7034683 Ghazarian Apr 2006 B2
7035880 Crescenti et al. Apr 2006 B1
7058649 Ough et al. Jun 2006 B2
7069466 Trimmer et al. Jun 2006 B2
7082441 Zahavi et al. Jul 2006 B1
7085786 Carlson et al. Aug 2006 B2
7085904 Mizuno et al. Aug 2006 B2
7093089 de Brebisson Aug 2006 B2
7096269 Yamagami Aug 2006 B2
7096315 Takeda et al. Aug 2006 B2
7103619 Rajpurkar et al. Sep 2006 B1
7103731 Gibble et al. Sep 2006 B2
7103740 Colgrove et al. Sep 2006 B1
7107298 Prahlad et al. Sep 2006 B2
7107395 Ofek et al. Sep 2006 B1
7118034 Baldassari et al. Oct 2006 B2
7120823 Foster et al. Oct 2006 B2
7130970 Devassy et al. Oct 2006 B2
7136720 Deckers Nov 2006 B2
7146377 Nowicki et al. Dec 2006 B2
7155465 Lee et al. Dec 2006 B2
7155486 Aoshima et al. Dec 2006 B2
7162496 Amarendran et al. Jan 2007 B2
7162604 Nourmohamadian et al. Jan 2007 B1
7162693 Yamanaka et al. Jan 2007 B2
7191283 Amemiya et al. Mar 2007 B2
7197490 English Mar 2007 B1
7200621 Beck et al. Apr 2007 B2
7203944 van Rietschote et al. Apr 2007 B1
7209949 Mousseau et al. Apr 2007 B2
7213118 Goodman et al. May 2007 B2
7216244 Amano May 2007 B2
7246140 Therrien et al. Jul 2007 B2
7246207 Kottomtharayil et al. Jul 2007 B2
7246258 Chen et al. Jul 2007 B2
7277246 Barbian et al. Oct 2007 B2
7277953 Wils et al. Oct 2007 B2
7281032 Kodama Oct 2007 B2
7287047 Kavuri Oct 2007 B2
7293133 Colgrove et al. Nov 2007 B1
7343356 Prahlad et al. Mar 2008 B2
7343453 Prahlad et al. Mar 2008 B2
7343459 Prahlad et al. Mar 2008 B2
7346623 Prahlad et al. Mar 2008 B2
7346751 Prahlad et al. Mar 2008 B2
7379850 Sprogis et al. May 2008 B2
7395282 Crescenti et al. Jul 2008 B1
7395387 Berkowitz et al. Jul 2008 B2
7401728 Markham et al. Jul 2008 B2
7421312 Trossell Sep 2008 B2
7434090 Hartung et al. Oct 2008 B2
7447907 Hart, III et al. Nov 2008 B2
7454569 Kavuri et al. Nov 2008 B2
7467167 Patterson Dec 2008 B2
7472238 Gokhale et al. Dec 2008 B1
7500053 Kavuri et al. Mar 2009 B1
7539702 Deshmukh et al. May 2009 B2
7539783 Kochunni et al. May 2009 B2
7581011 Teng Aug 2009 B2
7584227 Gokhale et al. Sep 2009 B2
7584298 Klinker et al. Sep 2009 B2
7587749 Leser et al. Sep 2009 B2
7596586 Gokhale et al. Sep 2009 B2
7603518 Kottomtharayil Oct 2009 B2
7644245 Prahlad et al. Jan 2010 B2
7657666 Kottomtharayil et al. Feb 2010 B2
7659820 Schnee et al. Feb 2010 B2
7660812 Findlay et al. Feb 2010 B2
7680843 Panchbudhe et al. Mar 2010 B1
7689510 Lamkin et al. Mar 2010 B2
7702659 Ban et al. Apr 2010 B2
7707060 Chainer et al. Apr 2010 B2
7712094 Shapiro May 2010 B2
7739450 Kottomtharayil Jun 2010 B2
7748610 Bell et al. Jul 2010 B2
7765167 Prahlad et al. Jul 2010 B2
7765369 Prahlad et al. Jul 2010 B1
7805416 Compton et al. Sep 2010 B1
7809699 Passmore et al. Oct 2010 B2
7809914 Kottomtharayil et al. Oct 2010 B2
7818417 Ginis et al. Oct 2010 B2
7822715 Petruzzo Oct 2010 B2
7831566 Kavuri et al. Nov 2010 B2
7840537 Gokhale et al. Nov 2010 B2
7849266 Kavuri et al. Dec 2010 B2
7861011 Kottomtharayil et al. Dec 2010 B2
7873802 Gokhale et al. Jan 2011 B2
7877351 Crescenti et al. Jan 2011 B2
7877362 Gokhale et al. Jan 2011 B2
7889847 Gainsboro Feb 2011 B2
7890796 Pawar et al. Feb 2011 B2
7904350 Ayala et al. Mar 2011 B2
7917473 Kavuri et al. Mar 2011 B2
7917695 Ulrich et al. Mar 2011 B2
7934071 Abe et al. Apr 2011 B2
7937365 Prahlad et al. May 2011 B2
7945810 Soran et al. May 2011 B2
7953802 Mousseau et al. May 2011 B2
7969306 Ebert et al. Jun 2011 B2
7975061 Gokhale et al. Jul 2011 B1
7987319 Kottomtharayil Jul 2011 B2
8051043 Young Nov 2011 B2
20020010661 Waddington et al. Jan 2002 A1
20020049778 Bell et al. Apr 2002 A1
20020069324 Gerasimov et al. Jun 2002 A1
20020087950 Brodeur et al. Jul 2002 A1
20030055671 Nassar Mar 2003 A1
20030065759 Britt et al. Apr 2003 A1
20030101155 Gokhale et al. May 2003 A1
20030134619 Phillips et al. Jul 2003 A1
20030220901 Carr et al. Nov 2003 A1
20030227707 Kokami et al. Dec 2003 A1
20040054607 Waddington et al. Mar 2004 A1
20040073677 Honma et al. Apr 2004 A1
20040083202 Mu et al. Apr 2004 A1
20040107199 Dalrymple et al. Jun 2004 A1
20040122832 Heil Jun 2004 A1
20040193953 Callahan et al. Sep 2004 A1
20040204949 Shaji et al. Oct 2004 A1
20050008163 Leser et al. Jan 2005 A1
20050021524 Oliver Jan 2005 A1
20050033913 Kottomtharayil et al. Feb 2005 A1
20050039069 Prahlad et al. Feb 2005 A1
20050102203 Keong May 2005 A1
20050174869 Kottomtharayil et al. Aug 2005 A1
20050177828 Graham et al. Aug 2005 A1
20050187992 Prahlad et al. Aug 2005 A1
20050195660 Kavuri et al. Sep 2005 A1
20050246342 Vernon Nov 2005 A1
20060004639 O'Keefe Jan 2006 A1
20060004675 Bennett et al. Jan 2006 A1
20060011720 Call Jan 2006 A1
20060095385 Atkinson et al. May 2006 A1
20060161879 Lubrecht et al. Jul 2006 A1
20060169769 Boyarsky et al. Aug 2006 A1
20060224846 Amarendran et al. Oct 2006 A1
20060248165 Sridhar et al. Nov 2006 A1
20060282194 Schaefer et al. Dec 2006 A1
20060285172 Hull et al. Dec 2006 A1
20070130105 Papatla Jun 2007 A1
20070156897 Lim Jul 2007 A1
20070185912 Gupta et al. Aug 2007 A1
20070198722 Kottomtharayil et al. Aug 2007 A1
20070198802 Kavuri Aug 2007 A1
20080059704 Kavuri Mar 2008 A1
20080141242 Shapiro Jun 2008 A1
20080229037 Bunte et al. Sep 2008 A1
20080243420 Gokhale et al. Oct 2008 A1
20080243754 Gokhale et al. Oct 2008 A1
20080243795 Prahlad et al. Oct 2008 A1
20080243870 Muller et al. Oct 2008 A1
20080244177 Crescenti et al. Oct 2008 A1
20080249656 Gokhale et al. Oct 2008 A1
20080250076 Muller et al. Oct 2008 A1
20080320319 Muller et al. Dec 2008 A1
20090113056 Tameshige et al. Apr 2009 A1
20090313448 Gokhale et al. Dec 2009 A1
20090319534 Gokhale Dec 2009 A1
20090319585 Gokhale Dec 2009 A1
20100070466 Prahlad et al. Mar 2010 A1
20100070474 Lad Mar 2010 A1
20100070726 Ngo et al. Mar 2010 A1
20100082672 Kottomtharayil et al. Apr 2010 A1
20100138393 Crescenti et al. Jun 2010 A1
20100293112 Prahlad et al. Nov 2010 A1
20110087807 Kottomtharayil et al. Apr 2011 A1
20110093672 Gokhale et al. Apr 2011 A1
20110213755 Kavuri et al. Sep 2011 A1
20110231852 Gokhale et al. Sep 2011 A1
20110270859 Kottomtharayil Nov 2011 A1
20120084523 Littlefield et al. Apr 2012 A1
Foreign Referenced Citations (15)
Number Date Country
0259912 Mar 1988 EP
0405926 Jan 1991 EP
0467546 Jan 1992 EP
0620553 Oct 1994 EP
0757317 Feb 1997 EP
0774715 May 1997 EP
0809184 Nov 1997 EP
0899662 Mar 1999 EP
0981090 Feb 2000 EP
7254204 Oct 1995 JP
9044381 Feb 1997 JP
9081424 Mar 1997 JP
WO-9513580 May 1995 WO
WO-9912098 Mar 1999 WO
WO-2005024573 Mar 2005 WO
Non-Patent Literature Citations (20)
Entry
U.S. Appl. No. 10/655,764; Inventor: Nourmohamadian et al; filed Sep. 5, 2003.
Armstead et al., “Implementation of a Campus-wide Distributed Mass Storage Service: The Dream vs. Reality,” IEEE, 1995, pp. 190-199.
Arneson, “Mass Storage Archiving in Network Environments,” Digest of Papers, Ninth IEEE Symposium on Mass Storage Systems, Oct. 31, 1988-Nov. 3, 1988, pp. 45-50, Monterey, CA.
Cabrera et al., “ADSM: A Multi-Platform, Scalable, Backup and Archive Mass Storage System,” Digest of Papers, Compcon '95, Proceedings of the 40th IEEE Computer Society International Conference, Mar. 5, 1995-Mar. 9, 1995, pp. 420-427, San Francisco, CA.
Eitel, “Backup and Storage Management in Distributed Heterogeneous Environments,” IEEE, 1994, pp. 124-126.
Jander, M., “Launching Storage-Area Net,” Data Communications, US, McGraw Hill, NY, vol. 27, No. 4 (Mar. 21, 1998), pp. 64-72.
Jason Gait, “The Optical File Cabinet: A Random-Access File System for Write-Once Optical Disks,” IEEE Computer, vol. 21, No. 6, pp. 11-22 (1988) (see in particular figure 5 in p. 15 and recitation in claim 5).
Rosenblum et al., “The Design and Implementation of a Log-Structured File System,” Operating Systems Review SIGOPS, vol. 25, No. 5, New York, US, pp. 1-15 (May 1991).
U.S. Appl. No. 11/269,513, filed Nov. 7, 2005, Prahlad.
Ashton et al., “Two Decades of policy-based storage management for the IBM mainframe computer”, www.research.ibm.com, 19 pages, published Apr. 10, 2003, printed Jan. 3, 2009.
Campbell C: “Linux and Windows NT 4.0: Basic Administration—Part III” Internet Publication, [Online] Oct. 5, 2000, Retrieved from the Internet: URL: <http://linux.omnipotent.net/article.php?article—id=10933> [retrieved on Aug. 22, 2006], 6 pages.
Carrington D: “Backups Using the “at” Command”, Internet Publication, [Online] May 4, 1999, Retrieved from the Internet: URL: <http://groups.google.de/group/microsoft.public.windowsnt.misc/browse—thread/thread/d1406a9a8391afea/48bac300a0adcc7a?lnk=st&q=&rnum=12&h1=de#48bac300a0adcc7a> [retrieved on Aug. 2, 2006], 1 page.
Cook P: “ntbackup: eject tape at end of backup?” Internet Publication, [Online] Oct. 18, 2000, Retrieved from the Internet: URL: <http://groups.google.de/group/microsoft.public.windowsnt.misc/browse—thread/thread/8f67f0cc96df42b7/0ab1d93a6f91b511?lnk=st&g=%22ntbackup+eject%22+at&rnum=1&hl=de#0ab1d93a6f91b511> [retrieved on Aug. 22, 2006], 1 page.
Gonzalez-Seco, Jose, “A Genetic Algorithm as the Learning Procedure for Neural Networks,” International Joint Conference on Neural Networks, Jun. 1992, 356 pages.
MDM: “Automatically eject tape”, Internet Publication, [Online] Jun. 7, 1999, Retrieved from Internet: URL: <http://groups.google.de/group/microsoftpublic.windowsnt.misc/browse—thread/thread/66537271a88cebda/2f8b1b96dfc5f102?lnk=st&q=&rnum=11&hl=de#2f8b1b96dfc5f102> [retrieved on Jun. 22, 2006], 1 page.
Recycle Bin (Windows), Aug. 2007, Wikipedia, pp. 1-3.
Savill J: “Windows NT FAQ Single File Version—Section Backup's” Internet Publication, [Online] 2000, Retrieved from Internet: URL: <http://burks.bton.ac.uk/burks/pcinfo/osdocs/ntfag/ntfaq—09.htm> [retrieved on Aug. 22, 2006], 8 pages.
U.S. Appl. No. 13/432,852, filed Mar. 28, 2012, Gokhale.
U.S. Appl. No. 09/609,977, filed Jul. 5, 2000, Prahlad.
Allen, “Probablility, Statistics and Queuing Theory,” (1978), p. 370, col. 19, Lines 3-33, 1 page.
Related Publications (1)
Number Date Country
20090063765 A1 Mar 2009 US
Provisional Applications (2)
Number Date Country
60969103 Aug 2007 US
60969105 Aug 2007 US