Embodiments of the invention relate to problem reporting in a computer network and, in particular, pertain to remote monitoring of a data storage system.
Data migration systems are routinely utilized in computer networks to perform data migration operations on electronic data stored within a network. In general, primary data, comprising a production copy or other “live” version in a native format, is generally stored in local memory or another high-speed storage device that allows for relatively fast access. Such primary data is generally intended for short term retention, on the order of hours or days. After this retention period, some or all of the data is stored as one or more secondary copies, for example, to prevent loss of data in the event that a problem occurs with the data stored in primary storage.
Secondary copies are generally intended for longer-term storage, on the order of weeks to years, prior to being moved to other storage or discarded. Secondary copies may be indexed so that a user may browse and restore the data at a later point in time. In some embodiments, application data moves from more expensive, quick access storage to less expensive, slower access storage, over its lifetime.
While data migration systems function to preserve data in the event of a problem with the computer network, the data migration systems themselves may encounter difficulties in storing data. For this reason, human monitors may be used to observe the data migration system and intervene to resolve problems that arise. Often, these monitors are experts, employed by the provider of the data migration system, conversant in the operation of the data migration system and capable of gathering information from the system, diagnosing problems, and implementing solutions.
This conventional monitoring is problematic, though. In one aspect, problem resolution requires laborious, manual gathering of information necessary to diagnose and troubleshoot problems, increasing the time and cost associated with problem resolution. This information gathering is also complicated by the fact that much of the information gathered is often for points in time that are not required for problem resolution.
In another aspect, the monitors of the data migration system are typically employees of the data migration system provider, located remotely from the system, rather than locally based employees of the owner of the data migration system. The remote monitors therefore, must remotely access the network in order to gather information for problem resolution. Security measures against unauthorized intrusion, such as firewalls and other technologies, though, restrict remote access privileges to the data migration system. Lowering or reducing these defenses to allow remote monitors the access necessary to gather troubleshooting information may compromise the security of the data migration system and the computer network it serves.
It is also undesirable to allow individuals who are not employed and supervised by the owner of the data migration system access to the archived data within the data migration system. For example, a medical or financial institution may possess confidential information about its clients that, if accessed by unauthorized individuals, even inadvertently, may open the institution to significant liability. Conversely, however, without sufficient access privileges, the monitors' ability to obtain the information required for problem resolution is limited, prolonging the time required to resolve problems as a result.
The aforementioned deficiencies in the current monitoring of data migration systems illustrate the need for improved systems and methods for storage monitoring, in particular remote monitoring, and other improvements. In certain embodiments, one of more of these needs is satisfied by the automated problem reporting system and methods disclosed herein. In one embodiment, the invention provides a method of problem reporting in a computer network, such as a tiered data storage network. The method comprises monitoring a plurality of elements that perform data migration operations, detecting a problem that occurs during the data migration operation, requesting information from the elements, assembling the requested information into a report; and providing the report to a monitor (e.g., a human monitor) that does not possess access privileges to the elements.
In another embodiment, the invention provides a method of remotely monitoring the data migration operations within a computer network. The method comprises providing a plurality of elements, comprising at least one of hardware, software, and firmware components that perform data migration operations. The method also comprises monitoring at least one of log files generated by the elements, communications links between the elements, and configurations of the elements during the data migration operations to detect errors in the data migration operations. The method further comprises gathering and analyzing selected information from the monitored elements automatically in response to the detection of an error in a data migration operation. The method additionally comprises communicating the selected information to a remote monitor.
In a further embodiment, the invention provides a system for remote monitoring of a data migration operation occurring within a computer network. The system comprises a plurality of elements that perform data migration operations and a monitor agent which communicates with the elements to detect problems occurring within data migration operations. The monitor agent gathers information from the elements in response to a detected problem, where at least a portion of the gathered information is provided to a remote monitor that does not possess access privileges to the elements.
In an additional embodiment, the invention provides an automated problem reporting data migration system. The system comprises a client computer containing data, a plurality of storage media for storing the data, a storage manager which coordinates data migration between any of the client computers and storage media, a media agent which performs data migration operations in response to instructions from the storage manager, a monitor agent which monitors data migration operations and communicates errors occurring during data migration; and a remote listener which receives the errors communicated by the monitor agent and generates reports containing selected information regarding at least one of the system hardware, software, and firmware. The reports are provided to a remote monitor which does not possess access privileges to the data migration system.
These and other objects and advantages will become more apparent from the following description taken in conjunction with the accompanying drawings.
Embodiments of the present invention relate to systems and methods of automated, remote monitoring, problem reporting, and problem resolution in data migration systems for use with a computer network. However, embodiments of the invention can be applied to monitoring, problem reporting, and problem resolution in any suitable network environment, whether the monitor is remotely or locally based. Examples include, but are not limited to, monitoring of network communication failures and hardware, software, and firmware failures.
In one embodiment, data migration systems include combinations of hardware, software, and firmware programs, as well communications links, necessary for performing data migration operations on electronic data within a computer network. Certain embodiments of a data migration system are provided in U.S. patent application Ser. No. 11/120,619, entitled “HIERARCHICAL SYSTEMS AND METHODS FOR PROVIDING A UNIFIED VIEW OF STORAGE INFORMATION”, which is incorporated herein by reference in its entirety.
Based on the nature of the problem, the reporting manager 100 can determine which elements are involved in the failed data migration operation, where the elements may comprise hardware, software, or firmware components within the system 102. For example, the data migration of a Microsoft Exchange server can involve an Exchange server, a computer that manages the data migration hardware, and the reporting manager 100 itself. The reporting manager 100 may subsequently request information from these elements for analysis to ascertain the nature of the problem and extract of at least a portion of the received information pertinent to the failed process. Based on this pertinent information, the reporting manager 100 generates a report. Copies of the report are subsequently made available to a monitor 104, which in certain embodiments is a remote monitor 104.
In an alternative embodiment, as illustrated in
Based upon the nature of the problem, the monitor agent 100A can be instructed by the remote listener 132 to gather information about the elements involved in the failed data migration process. The remote listener 132 receives the requested information from the monitor agent 100A for preparation of an error report. Copies of the report are subsequently made available, through an operational dashboard, to the remote monitor 104 and, optionally, the customer 134. In certain embodiments, the remote listener 132 may further analyze the error report prior to delivery to the remote monitor 104 and/or and customer 134 in order to ascertain the source of the error and possible solutions.
Beneficially, no intervention is required on the part of the remote monitor 104 or the administrator of the data migration systems 102, 102A (referred to herein as the customer administrator) in generation of the error report. In one aspect, this feature reduces the costs associated with problem resolution, as by automatically determining the information necessary for problem resolution, gathering the information, and analyzing the information, the systems 102, 102A perform tasks that would otherwise be performed by the monitor 104 and/or customer administrator. This allows problems to be identified and remedied more quickly than if the problems were manually identified, reducing downtime of the systems 102, 102A. Furthermore, a greater portion of the remote monitor's time may be spent developing solutions to problems, rather than gathering and analyzing the information. Additionally, by reducing the time necessary to identify and resolve problems, fewer monitors 104 may be necessary to support the systems 102, 102A, reducing support costs.
In another aspect, the automated reporting capability enhances the security of the data migration systems 102, 102A. As the systems 102, 102A provide the information necessary for troubleshooting, the monitor 104 is not required to access to the computer network or the data migration systems 102, 102A to obtain the information. This configuration significantly reduces the need for remote access to potentially sensitive information regarding the systems 102, 102A, reducing vulnerabilities that an unauthorized user may exploit to gain access to the systems 102, 102A. Furthermore, this setup ensures that the monitor 104 does not access the data stored within the computer network, preserving the confidentiality of the data stored within the systems 102, 102A.
In a further aspect of the systems 102, 102A, discussed in greater detail below, the customer administrator may select the information that is provided in the report to the monitor 104. Thus, information regarding selected elements, log files, configurations, and other information may be omitted from the report. As a result, the monitor 104 may be provided with limited information for initial problem solving and, at the administrator's discretion, provided additional information, as necessary.
Embodiments of the storage operation cells 106 of the systems 102, 102A are illustrated in
In one embodiment, the storage operation cells 106 may comprise a plurality of elements such as monitor agents 100, storage managers 110, client computers 112, media agents 114, and primary and secondary storage devices 116A, B (collectively, 116), as discussed in greater detail below. It should be understood that this list is not exhaustive and that the number of these and other elements present or absent within the cell 106 may be provided as necessary for the data migration operations performed by the cell 106. In some embodiments, certain elements reside and execute on the same computer, while in alternate embodiments, the some or all of the elements reside and execute on different computers.
The storage manager 110 comprises a software module or other application which coordinates and controls data migration operations performed by the storage operation cell 106. These operations may include, but are not limited to, initiation and management of production data copies, production data migrations, and production data recovery. To perform these operations, the storage manager 110 may communicate with some or all elements of the storage operation cell 106. The storage manager 110 may also maintain a database 120 or other data structure to indicate logical associations between elements of the cell 106, for example, the logical associations between media agents 114 and storage devices 116 as discussed below.
The monitor agent 100A comprises a software module or other application that is in communication with at least the storage manager 110 and a remote listener 132. The monitor agent 100A resides on the same device as the storage manager 110. As discussed in greater detail below, the monitor agent 100A is configured to receive alerts from the storage manager 110 regarding errors in data migration operations and communicate the alerts to the remote listener 132. The monitor agent 100A can comprise a passive agent and/or can further receive instructions from the remote listener 132, which the monitor agent 100A communicates to the storage manager 110 for implementation.
In one embodiment, the media agent 114 is an element that instructs a plurality of associated storage devices 116 to perform operations which subsequently archive, migrate, or restore data to or from the storage devices 116 as directed by the storage manager 110. For example, the media agent 114 can be implemented as a software module that conveys data, as directed by the storage manager 110, between a client computer 112 and one or more storage devices 116, such as a tape library, a magnetic media storage device, an optical media storage device, or other suitable storage device. In one embodiment, media agents 114 can be communicatively coupled with, and control, a storage device 116 associated with that particular media agent 114. A media agent 114 can be considered to be associated with a particular storage device 116 if that media agent 114 is capable of routing and storing data to that storage device 116.
In operation, the media agent 114 associated with a particular storage device 116 can instruct the storage device 116 to use a robotic arm or other retrieval mechanism to load or eject certain storage media, and to subsequently archive, migrate, or restore data to or from that media. Media agents 114 can communicate with a storage device 116 via a suitable communications link 130, such as a Small Computer System Interface (SCSI) or fiber channel communication.
The media agent 114 can also maintain an index cache, database, or other data structure 120, 120A that stores index data generated during data migration, migration, and restore and other data migration operations that can generate index data. The data structure 120, 120A provides the media agent 114 with a fast and efficient mechanism for locating data stored or archived. Thus, in some embodiments, the storage manager database 120, 120A can store data that associates a client 112 with a particular media agent 114 or storage device 116. The database 120, 120A, associated with the media agent 114, can indicate specifically where client 112 data is stored in the storage device 116, what specific files are stored, and other information associated with the storage of client 112 data.
In one embodiment, a first storage operation cell 106A can be configured to perform a particular type of data migration operation, such as storage resource management operations (SRM). SRM can comprise operations include monitoring the health, status, and other information associates with primary copies of data (e.g., live or production line copies). Thus, for example, the storage operation cell 106A can monitor and perform SRM related calculations and operations associated with primary copy data. The first storage operation cell 106A can include a client computer 112 in communication with a primary storage device 116A for storing data directed by the storage manager 110 associated with the cell 106A.
For example, the client 112 can be directed using Microsoft Exchange data, SQL data, oracle data, or other types of production data used in business applications or other applications stored in the primary volume. The storage manager 110 can contain SRM modules or other logic directed to monitor or otherwise interacting with the attributes, characteristics, metrics, and other information associated with the data stored in the primary volume.
In another implementation, a storage operation cell 106B can also contain a media agent 114 and secondary storage volume 116B configured to perform SRM related operations on primary copy data The storage manager 110 can also track and store information associated with primary copy migration. In some embodiments, the storage manager 110 can also track where primary copy information is stored, for example in secondary storage.
In alternative implementations, the storage operation cell 106B can be directed to another type of data migration operation, such as hierarchical storage management (HSM) data migration operations. For example, the HSM storage cell can perform production data migrations, snapshots or other types of HSM-related operations known in the art. For example, in some embodiments, data is migrated from faster and more expensive storage such as magnetic storage (i.e., primary storage) to less expensive storage such as tape storage (i.e., secondary storage).
The storage manager 110 can further monitor the status of some or all data migration operations previously performed, currently being performed, or scheduled to be performed by the storage operation cell 106. In one embodiment, the storage manager 110 can monitor the status of all jobs in the storage cells 106 under its control as well as the status of each component of the storage operation cells 106. The storage manager can monitor SRM or HSM operations as discussed above to track information which can include, but is not limited to: file type distribution, file size distribution, distribution of access/modification time, distribution by owner, capacity and asset reporting (by host, disk, or partition), availability of resources, disks, hosts, and applications. Thus, for example, the storage manager 110 can track the amount of available space, congestion, and other similar characteristics of data associated with the primary and secondary volumes 116A, B, and issue appropriate alerts to the reporting manager 100 or monitor agent 100A when a particular resource is unavailable or congested.
The storage manager 110 of a first storage cell 106A can also communicate with a storage manager 110 of another cell, such as 106B. In one example, a storage manager 110 in a first storage cell 106A communicates with a storage manager 110 in a second cell 1068 to control the storage manager 110 of the second cell 1068. Alternatively, the storage manager 110 of the first cell 106A can bypass the storage manager 110 of the second cell 1068 and directly control the elements of the second cell 1068.
In further embodiments, the storage operation cells 106 can be hierarchically organized such that hierarchically superior cells control or pass information hierarchically to subordinate cells and vice versa. In one embodiment, a master storage manager 122 can be associated with, communicate with, and direct data migration operations for a plurality of storage operation cells 106. In some embodiments, the master storage manager 122 can reside in its own storage operation cell 128. In other embodiments, the master storage manager 122 can itself be part of a storage operation cells 106.
In other embodiments, the master storage manager 122 can track the status of its associated storage operation cells 106, such as the status of jobs, system elements, system resources, and other items by communicating with its respect storage operation cells 106. Moreover, the master storage manager 122 can track the status of its associated storage operation cells 106 by receiving periodic status updates from the cells 106 regarding jobs, elements, system resources, and other items. For example, the master storage manager 122 can use methods to monitor network resources such as mapping network pathways and topologies to, among other things, physically monitor the data migration operations.
The master storage manager 122 can contain programming or other logic directed toward analyzing the storage patterns and resources of its associated storage cells 106. Thus, for example, the master storage manager 122 can monitor or otherwise keep track of the amount of resources available such as storage media in a particular group of cells 106. This allows the master storage manager 122 to determine when the level of available storage media, such as magnetic or optical media, falls below a selected level, so that an alert can be issued to the reporting manager 110 or monitor agent 100A, that additional media can be added or removed as necessary to maintain a desired level of service.
In one embodiment, the monitor agents 100A can also be hierarchically organized (
At block 204, the reporting manager 100 or monitor agent 100A detects at least one failure occurring in the data migration process. At block 206, information is requested and obtained pertaining to all elements involved in a failed data migration process. At block 210, the information is analyzed in order ascertain the nature of the problem. At block 212, the problem report is generated, based upon selection criteria provided by the customer administrator. At block 214, the problem report is disseminated to at least the remote monitor 104 and, optionally, the customer 134.
At block 202, the automated reporting manager 100 or monitor agent 100A monitors data migration operations performed within the entire system 102, 102A, as well as the status of the system elements. In one embodiment, this monitoring can comprise active monitoring, by the reporting manager 100, of data migration operations. These data migration operations may include, but are not limited to, SRM operations on primary copy information, HSM operations on secondary copy information, communication between storage cells 106 and, if hierarchically organized, between the cells 106 and the master storage manager 122. In an alternative embodiment, this monitoring may comprise the monitor agent 100A waiting to receive an alert generated during data migration operations.
An example data migration operation can be one performed according data migration protocols 304 (
At block 204, the reporting manager 100 or monitor agent 100A detects an error which has previously occurred, is currently occurring, or is expected to occur in one or more data migration operations or elements of the system. In one embodiment, as depicted in
In an alternative embodiment, as depicted in
At block 206, the reporting manager 100, or monitor agent 100A, gathers the relevant information from the elements upon detection of an error. In the embodiment of
At block 210, the reporting manager 100, or remote listener 132, determines whether a human actionable problem report should be generated. In certain embodiments, the reporting manager 100 or remote listener 132 can utilize programming or other logic to perform content-based analysis on the gathered information to make this determination. For example, the reporting manager 100 may retrieve logs 300 of the system elements for analysis, such as by parsing the logs 300, to determine the presence of selected key strings, such as error codes. In alternative embodiments of the system 102A, the remote listener 132 may instruct the monitor agent 100A to return this information for analysis. In either case, the data structure 302 can further comprise instructions regarding a course of action for each error. For example, the data structure 302 can provide instructions for the reporting manager 100 or remote listener 132. When detecting an error code in the logs 300, the reporting manager 100 or remote listener 132 can review the data structure 302 in light of the error codes to determine the appropriate course of action.
In one embodiment, the reporting manager 100 or remote listener 132 may determine that a problem report should not be generated. For example, the data structure 302 may indicate that the error is correctable by the system 102. In this case, the reporting manager 100 or monitor agent 100A can be instructed to ignore the error and return to monitoring the data migration process from block 202. Alternatively, the reporting manager 100, or the remote listener 132 via the monitor agent 100A, can provide the appropriate storage manager 110 with the instructions contained within the data structure 302 to resolve the problem.
In an alternative embodiment, the reporting manager 100, or the remote listener 132, can determine that a problem report should be generated upon consultation of the data structure 302. For example, the data structure 302 may indicate that an error may not be corrected by the system 102 at block 210A. In another example, the data structure 302 can indicate that a problem report should be generated when an error repeatedly occurs over a selected time window. Advantageously, in this manner, reports are generated on true failure problems that require the attention of the monitor 104, rather than routine errors that are readily resolved by the system 102 itself.
Use of the data structure 302 by the reporting manager 100 or remote listener 132 can also advantageously allow for the prioritization of reports. For instance, the data structure 302 can further contain a selected priority rating for each error it contains. For example, serious errors can be assigned a high priority and trivial errors can be assigned a low priority. Thus, when the monitor 104 receives a report, the report can be sorted into an ordered queue for resolution based on its priority. Beneficially, this priority rating ensures that the most serious reported problems are highlighted for attention, based on their severity, and not ignored during the resolution of less severe problems.
At block 212 of the method 200, the report can be generated according to selection criteria provided by the customer administrator. As discussed in greater detail below, the reporting manager 100 and remote listener 132 provide a graphical user interface, also referred to as a dashboard, which allows the customer administrator to select the portions of the collected information provided in the report. Filtering based upon a job ID, the relevant elements, and a selected time period, as well as element error logs, crash dumps, and configurations and other criteria may be utilized.
In one embodiment, the report comprises a plurality of files that provide the information selected by the customer administrator, as discussed below. In general, the report can comprise combinations of text files, xml, and html files, cabinet files, and other file types appropriate for providing the information requested by the reporting manager 100 or remote listener 132 in conjunction with problem resolution. Alternatively, the customer administrator can also initiate the generation of a problem report at his or her discretion.
In one embodiment, the report can contain a text file or other appropriate file which provides a summary of the collected information. The summary can include the job ID and failure reason of the failed process, if a job ID option is selected for reporting along with a subject, as discussed in greater detail below. The summary can additionally comprise the cell ID (such as a CommCell ID for a cell within the CommVault GALAXY system), element name, operating system, platform, time zone, version of the data migration system software, IP address, combinations of the same or the like.
Another portion of the report can comprise a collection of files pertaining to each client 112. In a non-limiting example, the files can include combinations of one or more of the following: data migration system logs (such as those provided by the CommVault GALAXY system), element hardware, software, and firmware configurations, system logs, crash dumps, and registries (such as those provided by the CommVault GALAXY system). In a preferred embodiment, the GALAXY registries are included by default, with other information provided optionally, at the administrator's discretion. In one embodiment, if the customer administrator selects to report the job ID or filter the information presented in the report by time, as discussed below, the reported log lines can be sent by the clients 112 to the reporting manager 100 or remote listener 132 in plain text, using encryption, such as through secure sockets, or combinations thereof. The received information can be combined into a single file for inclusion in the report. Optionally, a separate log file for each client 112 can be provided, rather than combined into a single file.
Another portion of the report can optionally comprise a fingerprint, in an xml format. The fingerprint provides a unique identifier that allows the system to distinguish between the machines that are being reported on. Any generally understood fingerprint technology can be utilized, including, but not limited to, the serial numbers of hardware or software present in the machines (e.g., CPU, hard disk drive, volume creation date, or operating system), addresses (e.g., MAC address of the network adapter of the machines, network address), or combinations thereof.
An additional component of the log bundle can optionally comprise database dumps. In general, a database dump contains a record of the table structure and/or the data from a database. In one embodiment, the database dump can be in the form of list of SQL queries. The database dump can be utilized in order to restore the contents of a database in the event of data loss. For example, corrupted databases can often be recovered by analysis of the dump.
A further component of the report can optionally comprise SQL_ERROR_LOGS.CAB, a cabinet file that contains all files with the name ERRORLOG.<NUM> as discussed above.
At block 214 of the method 200, the report is issued to the remote monitor 104 and, optionally, the customer 134. The remote monitor 104, in one embodiment, comprises a plurality of computer professionals capable of troubleshooting problems arising in the data migration system who reside in one or more locations removed from the physical location of the data migration system. As discussed in greater detail below in
For example, in the case of a relatively minor problem, the report can be provided through electronic mechanisms, such as one or more of e-mail, ftp, local upload, and the remote listener 132. In the case of more severe problems, the report can be followed up by telephone messages. The monitor 104 can read at least a portion the report to ascertain the nature of the problems that triggered the report or utilize another program to analyze the report in part or in total. Upon ascertaining possible causes for the problems, appropriate actions can be suggested and/or implemented for problem resolution.
In one embodiment, the monitor 104 does not possess access privileges to the data migration system. The monitor 104 thus operates in a support capacity, analyzing the problem report and suggesting possible courses of action to those locally who possess access privileges and/or physical access to the system. Advantageously, this system design allows the customer administrator to employ the remote monitor 104 for support without compromising the security of the data migration system or computer network by allowing remote access. Furthermore, as discussed in greater detail below, the report does not contain any information on the data within the computer network, and the customer administrator can limit information the reporting manager 100 or monitor agent 100A provides regarding the data migration system 102 in the report, further enhancing the security of the system.
In an alternative embodiment, the monitor 104 can possess a selected level of remote access privileges to the data migration system. This access allows the monitor to use the report as a starting point for problem resolution, isolating possible causes, and allowing the monitor 104 to execute solutions remotely. Advantageously, this setup can be appropriate for systems requiring only low security. For example, in a small business without a local computer professional, the automated report could assist a remote monitor 104 in identifying problems that they could subsequently fix, without the need for the small business owner to contract for a local computer professional, reducing the cost of maintaining the data migration system.
In one embodiment, the monitor agent 100A determines that the alert should not be passed along to the remote listener 132 (block 406). For example, the alert can comprise notification of a trivial error that is not harmful. In another example, the alert can comprise errors that the customer has specifically asked not to be notified of. In this event, the alert can be saved for archival purposes.
In another embodiment, the monitor agent 100A determines that the alert should be passed along to the remote listener 132 (block 410). The alert is passed to the remote listener 132 through a secure communications protocol. For example, the alert can be transmitted using a secure socket layer (SSL). Advantageously, the use of a secure communications protocol significantly reduces the likelihood of eavesdropping, tampering, and/or message forgery. The remote listener 132 determines the requirements to process the alert or to catalogue the alert as benign and not requiring additional activity.
Upon receipt of the alert, the remote listener 132 communicates the alert to the monitor 104 and, optionally, the customer 134. In one embodiment, the alert can be provided to respective graphical user interfaces, or dashboards, of the customer 134 (block 412) and/or monitor 104 (block 420). When received by the customer dashboard, the customer administrator can view and, optionally, act upon the alert (block 414). When received by the monitor dashboard, the alert begins an automated process of diagnosis and resolution, discussed in more detail below.
In one embodiment, the customer can close and/or tag the alert (blocks 414 and 416). For example, the alert can comprise a reoccurring problem already known to the customer administrator. Alternatively, the error can comprise an advisory alert that is not actionable by the customer administrator. Closure and/or tagging of an alert is subsequently communicated to the monitor 104 (block 422), which also tags the alert (block 424). Advantageously, this particular action “closes the loop” on communications with the customer, providing the monitor 104 with explicit confirmation that the customer administrator has received the alert.
Tagging is a workflow process associated with the triage and redress of the alert (block 424) dependent on the service level associated with the affected system. The customer is provided with the capability to tag an alert in the event that the alert is an anticipated event. For example, the customer might upgrade the firmware of a tape library and fail to suspend the data migration process, which would utilize this library during the time which the library is offline. The remote monitor 104 tags the alert (block 422) as part of the service workflow associated with the ROMS process.
In another embodiment, the customer 134 can choose not to close the alert, instead requesting more information on the alert. The customer dashboard can interface with a knowledge base maintained remote listener 132 for this purpose. In certain embodiments, the knowledge base comprises a specialized database for knowledge management that is human searchable and readable. Through the customer dashboard and knowledge base, the customer 134 can search and retrieve requisite information on the alert (blocks 426 and 430). Beneficially, this link between the customer 134 and knowledge base can allow the customer administrator to gain desired information regarding alerts, and their underlying errors, without the labor of independently searching for such information.
The knowledge base can further comprise a machine-readable portion that can be searched by the remote listener 132 to determine possible causes and solutions for the error. For example, the remote listener 132 can employ automated deductive reasoning in conjunction with the knowledge base to determine possible causes and solutions for the error.
At block 432, the remote listener 132, using the knowledge base and the alert, determines whether sufficient information is available to make a conclusion regarding the error. In one embodiment, a conclusion can comprise determining at least one possible source of the error. Optionally, the conclusion can further comprise one or more solutions. In another embodiment, the conclusion can comprise a determination that the knowledge base cannot reach a conclusion. If sufficient information is available to make a conclusion, the workflow 400 continues with block 436.
If the remote listener 132 determines that a conclusion can be reached with additional information, the workflow moves to block 434, where the remote listener 132 requests the additional information. In one embodiment, such information can be requested from the monitor agent 100A. In alternative embodiments, the information can be requested from a human monitor 104. In further embodiments, information can be requested from both the monitor agent 100A and human monitors 104. Upon receipt of the new, requested information, the workflow 400 returns to block 426, searching the knowledge base. The blocks of requesting and analyzing information (blocks 426, 432, 434) can be performed iteratively until the remote listener 132 reaches a conclusion. If a conclusion is reached by the remote listener 132, an e-mail message containing the conclusion and relevant analysis is sent to the appropriate customer contacts.
At block 436, it is determined whether a trouble report is opened. In one embodiment, the trouble report comprises a human actionable report that both instructs the monitor 104 to solve the problem, as well as disseminates information gathered to date by the remote listener 132. In one embodiment, the report can include, but is not limited to, a summary of the alert, the information gathered by the monitor agent 100A from the elements of the system 102, and the conclusions gathered by the remote listener 132 from the knowledge base. Based upon the information provided by the remote listener 132, the human monitor can decide to open a trouble report or not. If a trouble report is not opened, the alert is tagged (block 440), and the workflow 400 ends. If a trouble report is opened, the workflow 400 continues to blocks 442 and 444, where a trouble report is issued and tagged.
In alternative embodiments, the remote listener 132 can make the trouble report decision automatically. For instance, the remote listener 132 can determine from the knowledge base that the error is correctible without human intervention. The remote listener 132 then provides the monitor agent 100A with instructions drawn from the knowledge base to remedy the error. Alternatively, the error can be trivial and, therefore, ignored.
In another example, the remote listener 132 can decide to open a trouble report. For instance, the remote listener 132 can contain explicit instructions that errors of a certain nature are to be handled by human monitors 104. As above, if no trouble report is issued, then the alert is tagged at block 440 and the workflow 400 ends, while if a trouble report is issued, the workflow 400 moves to blocks 442 and 444, where a trouble report is opened and a tag is made.
In one embodiment, after opening the trouble report, a decision is made at block 446 regarding contacting the customer 134 through mechanisms other than the customer dashboard (block 446). In one embodiment, the customer is contacted at block 452 if they have contracted for a level of service that includes this option. Customer contact can include any known communication mechanisms, including, but not limited to, site visits, person to person telephone calls, computer automated telephone calls, voice mail messages, instant messaging, text messaging, electronic mail, combinations of the same and the like. In an alternative embodiment, the customer 134 is not contacted, other than through the dashboard, and the trouble reporting workflow ends at block 450.
It may be further understood that the report 500 can also be arbitrarily generated by the administrator's discretion. For example, the administrator can schedule periodic report generation in the absence of detected errors in order to provide selected information regarding the hardware, software, and firmware of the data migration system to the remote monitor.
In one embodiment, the interface 508 includes tabbed windows, dividing the selectable report parameters into broad sections. Advantageously, this interface 508 enhances the ease with which the administrator can customize the report. In a non-limiting embodiment, the sections, discussed in greater detail below, can comprise: an overview 502, a log summary 504, cell information 506, a time range filter 510, and element information 512, and, optionally, an output selector 414. In the discussion, below, the sections of the report 500 and the tabbed windows of the interface 508 are referred to interchangeably, as the selections within the interface 508 give rise to the sections presented in the report 500.
The overview 502 of the report 500 provides the monitor 104 with a summary of the problems which prompted the generation of the report 500. The overview 502 can include a subject that comprises a unique ticket number or job ID that identifies the particular data migration process which failed. The overview 502 can further comprise a description of the problem, as determined by analysis of the information received from the cells 106. The description can stress specific information needed for troubleshooting, which can include, but is not limited to, combinations of specific hardware, software, and firmware involved in the data migration problem, the specific data migration process which has failed, and communication link problems within the system. Advantageously, the overview 502 allows the monitor 104 to ascertain, quickly, the specific reasons for the problem report 500 rather than laboriously analyzing the log files 300 generated by the selected elements, hastening problem resolution.
The log window 504 provides the customer administrator control over the logs provided to the monitor 104. These logs 300 can comprise any of the logs 300 generated by the elements during data migration operations. In general, the logs 300 comprise lists of data migration operations performed, containing information which can include, but is not limited to, a job ID for the operation, a cell ID for the cell in which the operation was performed, a element ID for the elements on which the operation was performed, and acknowledgement that the job was completed. In one embodiment, the logs 300 can comprise logs generated by the CommVault GALAXY system.
In one embodiment, the customer administrator can use the log window 504 to filter the logs 300 provided to the monitor 104 in the report 500. For a monitor 104 to review all the logs 300 of all the elements involved in the data migration system for problem resolution would be a significant, time consuming task, as much of the content of the logs 300 may not be relevant to the problem at hand. Furthermore, the logs 300 can reveal information about the data migration system or computer network that the customer administrator may not wish to be disseminated. Thus, to save time and resources, as well as improve the security of the data migration system, the customer administrator can select from several options for how the logs 300 are filtered when reporting to the monitor 104.
In one embodiment, the customer administrator can select which elements are included in the report 500. For example, the customer administrator may wish to omit information regarding a particular computer for security reasons. Alternatively, the administrator may generally have reason to believe that logs 300 from certain elements do not need to be reported. Choosing this option, all of the log files 300 generated by the data migration system from the selected elements will be provided, such as GALAXY logs.
In further embodiments, the logs 300 can be provided based on the job IDs they contain. When this option is selected, the reporting manager 100 or monitor agent 100A searches the logs 300 of the elements for specific job ID numbers. Then, the reporting manager 100 or remote listener 132 includes only the log lines related to the job ID in the report.
Advantageously, the job ID and element filters allow the customer administrator significant flexibility in tailoring the logs 300 provided to the monitor 104. For example, if problems that occur throughout the data migration system 102 are a concern, the customer administrator can select to allow all logs 300 from elements involved in the failure process. Alternatively, if security is a primary concern, the customer administrator can select to allow only log fragments from certain computers to be viewed by the monitor 104. The administrator can further loosen these restrictions in subsequent reports, as necessary, should the monitor 104 require more information than provided. This flexibility allows the customer administrator to balance the amount of information released to facilitate problem evaluation and problem solving with security concerns.
In one embodiment, the cell window 506 allows the customer administrator to permit the reporting manager 100 or monitor agent 100A to provide information regarding a disaster recovery database in the report 500. At least a portion of this database can comprise meta-data regarding the client environment, or data regarding the data contained within the client environment. When the client environment sufferers a problem, this database can be utilized to recreate the client environment in a properly operating state.
The cell window 506 can further allow the customer administrator the option to include SQL error logs in the report. The errors logged may generally comprise system and user-defined events that occur on an SQL server, and more specifically, errors in data retrieval operations in SQL Server. In one non-limiting example, a Microsoft SQL server using the CommVault SIMPANA system can provide all files with the name ERRORLOG.<NUM>, where <NUM> is the number of the selected error log, under SQL path retrieved by the registry SOFTWARE\\Microsoft\\Microsoft SQL Server\\COMMVAULTQINETIX\\Setup\\SQLPath.
The cell information 506 can further contain fingerprints, as discussed above, for the machines discussed in the report 500.
The time range filter 510 allows the customer administrator to filter the report 508 based on a selected time period. In one embodiment, the time range filtering is optional, and can be disabled when the customer administrator elects to provide logs 300 by job ID, as discussed above. In another embodiment, the time range can comprise a selected time period prior to generation of the report 500, such as the last 24 hours. In an alternative embodiment, the customer administrator can provide information in the report over a selected, arbitrary time range.
Time filtering allows the customer administrator further control over the information provided to the remote monitor 104. In one embodiment, this mechanism of filtering can be useful when problems are most easily tracked and solved chronologically. In an alternative embodiment, the customer administrator can allow access to logs 300 relevant to a particular time to a monitor 104 experienced in solving the type of problem occurring over that time period. Dividing the logs 300 in this manner allows troubleshooting resources to be allocated by the customer administrator where they are needed. In a further embodiment, in the case where multiple monitors 104 work on a problem, time filtering can be used to divide the problem report 500 into sections based on a time period such that monitors 104 may only be provided pieces of the problem, giving the customer administrator greater control over security of the report information.
The element information section 512 of the interface 508 further allows the customer administrator to provide information specific to the elements involved in the failed data migration operation such as element hardware, software and firmware configurations, system logs, and crash dumps. Non-limiting examples of the element hardware, software, and firmware configurations are: processor type, processor speed, operating system, physical memory, available memory, available virtual memory, element name, IP address, time zone, and the version of the data migration software operating on the element. Non-limiting examples of system logs are: System/Application Event Logs (Microsoft Windows), /var/adm/messages* and /etc/system (Sun Microsystems Solaris), “errpt -a” output (IBM AIX), files similar to /etc/system (Linux and HP-UX) and abend logs (Novell Netware). Non-limiting examples of the crash dump information are the Dr. Watson log (Microsoft Windows) and a list of core files and the name of the executables which caused the core (Unix). Advantageously, this element information allows the monitor 104 to determine if hardware or software associated with the element operation, as separate from the data migration process, can be responsible for data migration problems.
The output selector 514 allows the administrator to determine the manner in which the report 508 is provided to the remote monitor. In one embodiment, the output can comprise at least one of upload to an FTP location, an electronic mail message with the subject line of the job ID or ticket number, and saving to a local directory. Advantageously, this flexibility in the delivery mechanism of the report 500 allows the report 500 to be provided in the manner that is most appropriate to the circumstances of the data migration system. For example, if one line of communication is unavailable, inaccessible, or insecure, the report may still be provided, enhancing the robustness of the problem reporting manager.
In one embodiment, the output selector 514 further allows the customer administrator to select a size limit for the e-mail which is sent containing the report 500. Often e-mail systems are limited from sending or receiving e-mail messages over a certain size. Further, depending on the nature of the problem within the system, the report 500 can be relatively large. Thus, when a limit is specified, the reporting manager can check the final report 500 size against the selected limit. If the size of the report 500 exceeds the limit, the report 500 can be split into multiple CAB files, each with a size less than the limit. In this case, multiple e-mails are then sent containing the smaller CAB files. Optionally, a utility can be provided to the remote monitor for re-assembly of the CAB files. Advantageously, this size flexibility enhances the robustness of the reporting system, ensuring that the e-mails are not delayed or rejected because of their size during their transmission or receipt.
In alternative embodiments, the selections available through the output selector 514 can be limited, for security reasons. For example, the output selector may be limited to transmissions through secure sockets.
In the following examples, circumstances in which problem reports 500 may be generated are discussed. In general, the examples illustrate the wide range of problems which may be automatically identified and reported through embodiments of the automated problem reporting systems 102, 102A and further illustrate how the problem report 500 may be utilized by computer professionals to identify and resolve problems more quickly and easily than through conventional, manual problem resolution. These examples are discussed for illustrative purposes and should not be construed to limit the embodiments of the invention.
In one embodiment, the reporting manager 100 or monitor agent 100A may monitor or be alerted to the physical status of the elements of the data migration system and issue a problem report 500 when a mechanical failure occurs. For example, media agents 114 perform copy or restore operations in response to instructions from storage managers 110. The data to be archived or recovered may reside on media such as a tape or optical disk that is mechanically retrieved, such as with a mechanical arm, and loaded into a storage volume 116 for access. This mechanical operation, however, may fail if the mechanical arm fails to actuate properly.
Should the mechanical arm fail to operate properly, the media agent 114 alerts one or more associated storage managers 110, which in turn provides the alert to the reporting manager 100 or monitor agent 100A, which triggers generation of a problem report. For example, the reporting manager 100 or the monitor agent 100A, under instructions from the remote listener 132, may gather information regarding the storage volume 116 and monitor agent 110, the machine and cell containing the storage volume 116 and storage manager 110 (or monitor agent 100A), as well as associated logs 300. The reporting manager 100 or remote listener 132 may then apply the reporting selections entered in the interface 508 for use in preparing the trouble report 500. The filtering options selected by the customer administrator will determine the information provided in the report. In one example, the summary of the report 500 may contain the job ID for the data migration function that has failed and a description stating that the storage volume at issue experienced a hardware problem.
The report 500 can be communicated to the monitor and/or the customer 134. In one embodiment, such communication may comprise at least one of the mechanisms selected within the output selector 514. For example, the customer dashboard 508 can be updated with this information, along with suggested remedies that the customer administrator may execute. Upon review of the report, the customer administrator and/or monitor 104 may determine that a mechanical failure has occurred in one or more storage volumes 116 by review of the summary and bundled files.
The report 500 can also comprise suggested remedies for the problem. Examples of remedies can include scheduling the data migration operation to be performed on another storage volume, repairing or replacing the mechanical system that has failed, cancelling the data migration operation, combinations of the same or the like. In alternative embodiments, the monitor 104 may additionally initiate further communication with the customer administrator for discussion of additional remedies. In further alternative embodiments, if the monitor possesses sufficient access privileges, the monitor may perform problem resolution themselves.
In one embodiment, the reporting manager 100 or monitor agent 100A may monitor or be configured to receive alerts regarding errors which occur during use of the communications links 130. For example, when a client 112 requests archived files, the client 112 communicates with a storage manager 110, which then issues instructions to the appropriate media agent 114 to retrieve the requested data and transmit the data to the client 112. Often, these various functions are performed on different machines. Thus, when disruptions occur in the communication links 130 between the client 112 and storage manager 110, the storage manager 110 and media agent 114, or the media agent 114 and client 112 due to hardware or software problems, data migration operation may not be performed correctly.
Depending on the severity of the connectivity problem, the reporting manager 100 or remote listener 132 may trigger the generation of the problem report 500. The reporting manager 100 or monitor agent 100A gathers information which may include, but is not limited to, information regarding hardware, software, and firmware settings, as well as appropriate logs 300, of the remote monitor 104 or monitor agent 100A itself, the media agent 114, storage manager 110, and client computer 112. The reporting manager 100 or remote listener 132 may then apply the reporting selections entered in the interface 508 and issue the report 500. In certain embodiments, the filtering options selected by the customer administrator determine the information provided in the report. For example, the summary of the report 500 may contain the job ID for the scheduled retrieval function and a description stating that a network connectivity problem is at issue.
The report 500 may be communicated to the monitor 104 and/or the customer 134. In one embodiment, such communication may comprise at least one of the mechanisms selected within the output selector 514.
The report 500 can also comprise suggested remedies for the problem. Examples of remedies may include checking the network configuration within the operating system and data migration software of the elements involved in the failed process. Remedies may further include checking the status of the network hardware and physical network connections of the elements involved in the failed process. In alternative embodiments, the monitor 104 may additionally initiate further communication with the customer administrator to discuss further remedies. In further alternative embodiments, if the monitor 104 possesses sufficient access privileges, the monitor 104 may perform problem resolution.
In one embodiment, the reporting manager 100 or monitor agent 100A may be alerted to errors occurring in data migration operations conducted between cell elements. As described above, agents, such as the media agents 114, are responsible for executing data migration operations designated by the storage manager 110. When data is migrated under normal operations, the relevant agent receives instruction from the storage manager 110, identifies the location of the data from the relevant database 120A, performs the designated migration operation, updates the location of the migrated data in the agent database 120 for later reference, and provides an acknowledgement of the operation to the storage manager 110.
In the event that one or more blocks in this process are not successfully completed, the media agent 114 may fail to acknowledge the completion of the data migration operation. In one embodiment, this error initiates the problem reporting process, where the reporting manager 100 or remote listener 132 may request more information to determine whether to generate a problem report 500. For example, the remote monitor 100 or monitor agent 100A may contact the storage manager 110 to obtain log files and hardware, software, and firmware configurations for the machines containing the storage managers 110 and media agents 114 involved in the failed process. Similar information may also be gathered for the reporting manager 100 or monitor agent 100A itself.
Upon receipt of this information, the reporting manager 100 or remote listener 132 applies programming or other logic to the received information to determine the problem, applies the selection criteria entered in the interface 508 for reporting, and issues the problem report 500. For example, the remote listener 132 can utilize the knowledge base in conjunction with discovered error codes, as discussed above. The filtering options selected by the customer administrator will determine the information provided in the report. In one example, the summary of the generated report 500 contains the job ID for the scheduled data migration operation and a description stating that an acknowledgement failure is at issue.
The report 500 may be communicated to the monitor 104 and/or the customer 134. In one embodiment, such communication may comprise at least one of the mechanisms selected within the output selector 514. For example, the customer dashboard 508 may be updated with this information, along with suggested remedies that the customer administrator may execute. For example, the received information on the media agent 114 and storage device 116 may be reviewed in greater depth to determine if an identifiable hardware or software failure has occurred in either element. Examples of checking hardware errors may include examining the network connectivity of the media agent 114 and storage device 116 and the mechanical status of the storage device 116 as discussed above. Examples of checking software errors may include examining the file system for problems, such as corrupted databases, a file pathway that cannot be determined, or other problems opening or writing files and directories, as well as incompatibilities between the server a restore is attempted on and the server from which the files originated. In alternative embodiments, the monitor 104 can additionally initiate further communication with the customer administrator to discuss further remedies. In further alternative embodiments, if the monitor 104 possesses sufficient access privileges, the monitor 104 can perform problem resolution actions.
In one embodiment, the problem reporting systems 102, 102A may also issue problem reports 500 based upon predicted problems. For example, an element of the data migration system 102, such as a storage manager 110 or master storage manager 122, may record the rate at which data is stored on a storage volumes 116A, 1168 and/or have access to historical records of the same, as well as monitor the capacity of storage volumes 116A, 1168 within the data migration system. For instance, the system may predict, based upon trends in storage usage and the present capacity of the storage volume, the available storage capacity on the selected day and generate an alert if the size of the scheduled backup exceeds the space predicted to be available.
In response to the alert, the problem evaluation and reporting process is initiated by the reporting manager 100 or monitor agent 100A. The reporting manager 100, or monitor agent 100A at the instruction of the remote listener 132, may subsequently gather information from the storage volume 120, the machine containing the storage volume 120, the cell 106 containing the storage volume 120, and/or associated logs 300 for each, as well as information and/or log files 300 for the machine containing the reporting manager 100 or monitor agent 100A. The reporting manager 100 or remote listener 132 may then apply the selections entered in the interface 508 for reporting and issue the report 500. The summary of the report 500 may contain the job ID for the scheduled data migration function and a description stating that the storage volume 120 at issue may not possess sufficient capacity for the data migration.
Advantageously, this predictive capability allows problems to be prevented before they occur. The report 500 can be communicated to the monitor 104 and/or the customer 134. In one embodiment, such communication may comprise at least one of the mechanisms selected within the output selector 514. For example, the customer dashboard 508 may be updated with this information, along with suggested remedies that the customer administrator may execute. The summary description may allow the monitor 104 and/or customer administrator to quickly determine that the storage capacity of one or more storage volumes 116A, 1168 is the cause of the problem report 500, rather than reviewing a large amount of log files 300 to determine the same.
The report 500 may also comprise suggested remedies for the problem. Examples of remedies may include scheduling the data migration operation to be performed on another storage volume 116A, 1168, installing a new storage volume 120, deleting unnecessary files on the storage volume 116A, 1168 to provide additional capacity, or cancelling the data migration operation. In alternative embodiments, the monitor 104 may additionally initiate further communication with the customer administrator to discuss further remedies. In further alternative embodiments, if the monitor 104 possesses sufficient access privileges, the monitor may perform problem resolution themselves.
Although the foregoing description has shown, described, and pointed out the fundamental novel features of the present teachings, it will be understood that various omissions, substitutions, and changes in the form of the detail of the apparatus as illustrated, as well as the uses thereof, may be made by those skilled in the art, without departing from the scope of the present teachings. Consequently, the scope of the present teachings should not be limited to the foregoing discussion, but should be defined by the appended claims.
This application is a continuation of U.S. patent application Ser. No. 15/685,958, entitled “SYSTEMS AND METHODS FOR REMOTE MONITORING IN A COMPUTER NETWORK” and filed on Aug. 24, 2017, which is a continuation of U.S. patent application Ser. No. 14/826,958, entitled “SYSTEMS AND METHODS FOR REMOTE MONITORING IN A COMPUTER NETWORK” and filed on Aug. 14, 2015, which is a continuation of U.S. patent application Ser. No. 14/139,241, entitled “SYSTEMS AND METHODS FOR REMOTE MONITORING IN A COMPUTER NETWORK” and filed on Dec. 23, 2013, which is a continuation of U.S. patent application Ser. No. 13/615,231, entitled “SYSTEMS AND METHODS FOR REMOTE MONITORING IN A COMPUTER NETWORK” and filed on Sep. 13, 2012, which is a continuation of U.S. patent application Ser. No. 11/963,732, entitled “SYSTEMS AND METHODS FOR REMOTE MONITORING IN A COMPUTER NETWORK AND REPORTING A FAILED MIGRATION OPERATION WITHOUT ACCESSING THE DATA BEING MOVED” and filed on Dec. 21, 2007, which claims priority to U.S. Provisional Patent Application No. 61/190,057, entitled “REMOTE MONITORING IN A COMPUTER NETWORK” and filed on Dec. 22, 2006, each of which are hereby incorporated herein by reference in its entirety. Any and all applications for which a foreign or domestic priority claim is identified in the Application Data Sheet, or any correction thereto, are hereby incorporated by reference into this application under 37 CFR 1.57
Number | Name | Date | Kind |
---|---|---|---|
4296465 | Lemak | Oct 1981 | A |
4686620 | Ng | Aug 1987 | A |
4995035 | Cole et al. | Feb 1991 | A |
5005122 | Griffin et al. | Apr 1991 | A |
5062104 | Lubarsky et al. | Oct 1991 | A |
5093912 | Dong et al. | Mar 1992 | A |
5133065 | Cheffetz et al. | Jul 1992 | A |
5193154 | Kitajima et al. | Mar 1993 | A |
5212772 | Masters | May 1993 | A |
5226157 | Nakano et al. | Jul 1993 | A |
5239647 | Anglin et al. | Aug 1993 | A |
5241668 | Eastridge et al. | Aug 1993 | A |
5241670 | Eastridge et al. | Aug 1993 | A |
5276860 | Fortier et al. | Jan 1994 | A |
5276867 | Kenley et al. | Jan 1994 | A |
5287500 | Stoppani, Jr. | Feb 1994 | A |
5301351 | Jippo | Apr 1994 | A |
5311509 | Heddes et al. | May 1994 | A |
5321816 | Rogan et al. | Jun 1994 | A |
5333315 | Saether et al. | Jul 1994 | A |
5347653 | Flynn et al. | Sep 1994 | A |
5388243 | Glider et al. | Feb 1995 | A |
5410700 | Fecteau et al. | Apr 1995 | A |
5448724 | Hayashi et al. | Sep 1995 | A |
5465359 | Allen et al. | Nov 1995 | A |
5487160 | Bemis | Jan 1996 | A |
5491810 | Allen | Feb 1996 | A |
5495607 | Pisello et al. | Feb 1996 | A |
5504873 | Martin et al. | Apr 1996 | A |
5515502 | Wood | May 1996 | A |
5544345 | Carpenter et al. | Aug 1996 | A |
5544347 | Yanai et al. | Aug 1996 | A |
5555404 | Torbjornsen et al. | Sep 1996 | A |
5559957 | Balk | Sep 1996 | A |
5559991 | Kanfi | Sep 1996 | A |
5592618 | Micka et al. | Jan 1997 | A |
5598546 | Blomgren | Jan 1997 | A |
5606359 | Youden et al. | Feb 1997 | A |
5615392 | Harrison et al. | Mar 1997 | A |
5619644 | Crockett et al. | Apr 1997 | A |
5638509 | Dunphy et al. | Jun 1997 | A |
5642496 | Kanfi | Jun 1997 | A |
5673381 | Huai et al. | Sep 1997 | A |
5675511 | Prasad et al. | Oct 1997 | A |
5680550 | Kuszmaul et al. | Oct 1997 | A |
5682513 | Candelaria et al. | Oct 1997 | A |
5687343 | Fecteau et al. | Nov 1997 | A |
5699361 | Ding et al. | Dec 1997 | A |
5719786 | Nelson et al. | Feb 1998 | A |
5737747 | Vishlitsky et al. | Apr 1998 | A |
5751997 | Kullick et al. | May 1998 | A |
5758359 | Saxon | May 1998 | A |
5761677 | Senator et al. | Jun 1998 | A |
5761734 | Pfeffer et al. | Jun 1998 | A |
5764972 | Crouse et al. | Jun 1998 | A |
5778395 | Whiting et al. | Jul 1998 | A |
5790828 | Jost | Aug 1998 | A |
5805920 | Sprenkle et al. | Sep 1998 | A |
5812398 | Nielsen | Sep 1998 | A |
5813008 | Benson et al. | Sep 1998 | A |
5813009 | Johnson et al. | Sep 1998 | A |
5813017 | Morris | Sep 1998 | A |
5829023 | Bishop | Oct 1998 | A |
5829046 | Tzelnic et al. | Oct 1998 | A |
5860104 | Witt et al. | Jan 1999 | A |
5875478 | Blumenau | Feb 1999 | A |
5875481 | Ashton et al. | Feb 1999 | A |
5887134 | Ebrahim | Mar 1999 | A |
5890159 | Sealby et al. | Mar 1999 | A |
5897643 | Matsumoto | Apr 1999 | A |
5901327 | Ofek | May 1999 | A |
5924102 | Perks | Jul 1999 | A |
5926836 | Blumenau | Jul 1999 | A |
5933104 | Kimura | Aug 1999 | A |
5950205 | Aviani, Jr. | Sep 1999 | A |
5956519 | Wise et al. | Sep 1999 | A |
5958005 | Thorne et al. | Sep 1999 | A |
5970233 | Liu et al. | Oct 1999 | A |
5970255 | Tran et al. | Oct 1999 | A |
5974563 | Beeler, Jr. | Oct 1999 | A |
5987478 | See et al. | Nov 1999 | A |
5995091 | Near et al. | Nov 1999 | A |
5999629 | Neer et al. | Dec 1999 | A |
6003089 | Shaffer et al. | Dec 1999 | A |
6009274 | Fletcher et al. | Dec 1999 | A |
6012090 | Chung et al. | Jan 2000 | A |
6021415 | Cannon et al. | Feb 2000 | A |
6026414 | Anglin | Feb 2000 | A |
6041334 | Cannon | Mar 2000 | A |
6052735 | Ulrich et al. | Apr 2000 | A |
6058494 | Gold et al. | May 2000 | A |
6076148 | Kedem et al. | Jun 2000 | A |
6094416 | Ying | Jul 2000 | A |
6094684 | Pallmann | Jul 2000 | A |
6105129 | Meier et al. | Aug 2000 | A |
6105150 | Noguchi et al. | Aug 2000 | A |
6112239 | Kenner et al. | Aug 2000 | A |
6122668 | Teng et al. | Sep 2000 | A |
6131095 | Low et al. | Oct 2000 | A |
6131190 | Sidwell | Oct 2000 | A |
6137864 | Yaker | Oct 2000 | A |
6148412 | Cannon et al. | Nov 2000 | A |
6154787 | Urevig et al. | Nov 2000 | A |
6154852 | Amundson et al. | Nov 2000 | A |
6161111 | Mutalik et al. | Dec 2000 | A |
6167402 | Yeager | Dec 2000 | A |
6175829 | Li et al. | Jan 2001 | B1 |
6212512 | Barney et al. | Apr 2001 | B1 |
6230164 | Rikieta et al. | May 2001 | B1 |
6260069 | Anglin | Jul 2001 | B1 |
6269431 | Dunham | Jul 2001 | B1 |
6275953 | Vahalia et al. | Aug 2001 | B1 |
6292783 | Rohler | Sep 2001 | B1 |
6295541 | Bodnar et al. | Sep 2001 | B1 |
6301592 | Aoyama et al. | Oct 2001 | B1 |
6304880 | Kishi | Oct 2001 | B1 |
6324581 | Xu et al. | Nov 2001 | B1 |
6328766 | Long | Dec 2001 | B1 |
6330570 | Crighton | Dec 2001 | B1 |
6330572 | Sitka | Dec 2001 | B1 |
6330642 | Carteau | Dec 2001 | B1 |
6343324 | Hubis et al. | Jan 2002 | B1 |
6350199 | Williams et al. | Feb 2002 | B1 |
RE37601 | Eastridge et al. | Mar 2002 | E |
6353878 | Dunham | Mar 2002 | B1 |
6356801 | Goodman et al. | Mar 2002 | B1 |
6374266 | Shnelvar | Apr 2002 | B1 |
6374336 | Peters et al. | Apr 2002 | B1 |
6381331 | Kato | Apr 2002 | B1 |
6385673 | DeMoney | May 2002 | B1 |
6389432 | Pothapragada et al. | May 2002 | B1 |
6418478 | Ignatius et al. | Jul 2002 | B1 |
6421711 | Blumenau et al. | Jul 2002 | B1 |
6438586 | Hass et al. | Aug 2002 | B1 |
6446224 | Chang et al. | Sep 2002 | B1 |
6487561 | Ofek et al. | Nov 2002 | B1 |
6487644 | Huebsch et al. | Nov 2002 | B1 |
6505307 | Stell et al. | Jan 2003 | B1 |
6519679 | Devireddy et al. | Feb 2003 | B2 |
6538669 | Lagueux, Jr. et al. | Mar 2003 | B1 |
6542909 | Tamer et al. | Apr 2003 | B1 |
6542972 | Ignatius et al. | Apr 2003 | B2 |
6564228 | O'Connor | May 2003 | B1 |
6571310 | Ottesen | May 2003 | B1 |
6577734 | Etzel et al. | Jun 2003 | B1 |
6581143 | Gagne et al. | Jun 2003 | B2 |
6604149 | Deo et al. | Aug 2003 | B1 |
6622264 | Bliley et al. | Sep 2003 | B1 |
6631442 | Blumenau | Oct 2003 | B1 |
6631493 | Ottesen et al. | Oct 2003 | B2 |
6647396 | Parnell et al. | Nov 2003 | B2 |
6654825 | Clapp et al. | Nov 2003 | B2 |
6658436 | Oshinsky et al. | Dec 2003 | B2 |
6658526 | Nguyen et al. | Dec 2003 | B2 |
6675177 | Webb | Jan 2004 | B1 |
6732124 | Koseki et al. | May 2004 | B1 |
6742092 | Huebsch et al. | May 2004 | B1 |
6757794 | Cabrera et al. | Jun 2004 | B2 |
6763351 | Subramaniam et al. | Jul 2004 | B1 |
6772332 | Boebert et al. | Aug 2004 | B1 |
6785786 | Gold et al. | Aug 2004 | B1 |
6789161 | Blendermann et al. | Sep 2004 | B1 |
6791910 | James et al. | Sep 2004 | B1 |
6859758 | Prabhakaran et al. | Feb 2005 | B1 |
6871163 | Hiller et al. | Mar 2005 | B2 |
6880052 | Lubbers et al. | Apr 2005 | B2 |
6886020 | Zahavi et al. | Apr 2005 | B1 |
6909722 | Li | Jun 2005 | B1 |
6928513 | Lubbers et al. | Aug 2005 | B2 |
6952758 | Chron et al. | Oct 2005 | B2 |
6965968 | Touboul et al. | Nov 2005 | B1 |
6968351 | Butterworth | Nov 2005 | B2 |
6973553 | Archibald, Jr. et al. | Dec 2005 | B1 |
6983351 | Gibble et al. | Jan 2006 | B2 |
7003519 | Biettron et al. | Feb 2006 | B1 |
7003641 | Prahlad et al. | Feb 2006 | B2 |
7035880 | Crescenti et al. | Apr 2006 | B1 |
7039810 | Nichols | May 2006 | B1 |
7062761 | Slavin et al. | Jun 2006 | B2 |
7069380 | Ogawa et al. | Jun 2006 | B2 |
7085904 | Mizuno et al. | Aug 2006 | B2 |
7103731 | Gibble et al. | Sep 2006 | B2 |
7103740 | Colgrove et al. | Sep 2006 | B1 |
7107298 | Prahlad et al. | Sep 2006 | B2 |
7107395 | Ofek et al. | Sep 2006 | B1 |
7117246 | Christenson et al. | Oct 2006 | B2 |
7120757 | Tsuge | Oct 2006 | B2 |
7130970 | Devassy et al. | Oct 2006 | B2 |
7155465 | Lee et al. | Dec 2006 | B2 |
7155633 | Tuma et al. | Dec 2006 | B2 |
7159110 | Douceur et al. | Jan 2007 | B2 |
7174433 | Kottomtharayil et al. | Feb 2007 | B2 |
7209972 | Ignatius et al. | Apr 2007 | B1 |
7237232 | Smith | Jun 2007 | B2 |
7246140 | Therrien et al. | Jul 2007 | B2 |
7246207 | Kottomtharayil et al. | Jul 2007 | B2 |
7246272 | Cabezas et al. | Jul 2007 | B2 |
7257743 | Glerum et al. | Aug 2007 | B2 |
7269612 | Devarakonda et al. | Sep 2007 | B2 |
7272606 | Borthakur et al. | Sep 2007 | B2 |
7277941 | Ignatius et al. | Oct 2007 | B2 |
7278142 | Bandhole et al. | Oct 2007 | B2 |
7284165 | Auvenshine et al. | Oct 2007 | B2 |
7287047 | Kavuri | Oct 2007 | B2 |
7287252 | Bussiere et al. | Oct 2007 | B2 |
7293133 | Colgrove et al. | Nov 2007 | B1 |
7298846 | Bacon et al. | Nov 2007 | B2 |
7302618 | Shafer et al. | Nov 2007 | B1 |
7315923 | Retnamma et al. | Jan 2008 | B2 |
7340649 | Angamuthu | Mar 2008 | B2 |
7346623 | Prahlad et al. | Mar 2008 | B2 |
7359917 | Winter et al. | Apr 2008 | B2 |
7380072 | Kottomtharayil et al. | May 2008 | B2 |
7398429 | Shaffer et al. | Jul 2008 | B2 |
7401154 | Ignatius et al. | Jul 2008 | B2 |
7409509 | Devassy et al. | Aug 2008 | B2 |
7448079 | Tremain | Nov 2008 | B2 |
7454569 | Kavuri et al. | Nov 2008 | B2 |
7457933 | Pferdekaemper et al. | Nov 2008 | B2 |
7467167 | Patterson | Dec 2008 | B2 |
7472238 | Gokhale | Dec 2008 | B1 |
7484054 | Kottomtharayil et al. | Jan 2009 | B2 |
7490207 | Amarendran | Feb 2009 | B2 |
7500053 | Kavuri et al. | Mar 2009 | B1 |
7500150 | Sharma et al. | Mar 2009 | B2 |
7509019 | Kaku | Mar 2009 | B2 |
7519726 | Palliyll et al. | Apr 2009 | B2 |
7523483 | Dogan | Apr 2009 | B2 |
7529748 | Wen et al. | May 2009 | B2 |
7529976 | Beard | May 2009 | B2 |
7536291 | Retnamma et al. | May 2009 | B1 |
7546324 | Prahlad et al. | Jun 2009 | B2 |
7546482 | Blumenau | Jun 2009 | B2 |
7581077 | Ignatius et al. | Aug 2009 | B2 |
7596586 | Gokhale et al. | Sep 2009 | B2 |
7613748 | Brockway et al. | Nov 2009 | B2 |
7627598 | Burke | Dec 2009 | B1 |
7627617 | Kavuri et al. | Dec 2009 | B2 |
7631194 | Wahlert et al. | Dec 2009 | B2 |
7653633 | Villella et al. | Jan 2010 | B2 |
7685126 | Patel et al. | Mar 2010 | B2 |
7765369 | Prahlad et al. | Jul 2010 | B1 |
7809914 | Kottomtharayil et al. | Oct 2010 | B2 |
7831553 | Prahlad et al. | Nov 2010 | B2 |
7840537 | Gokhale et al. | Nov 2010 | B2 |
7840809 | Bell | Nov 2010 | B2 |
7861050 | Retnamma et al. | Dec 2010 | B2 |
7870355 | Erofeev | Jan 2011 | B2 |
7962642 | Ignatius et al. | Jun 2011 | B2 |
7975061 | Gokhale et al. | Jul 2011 | B1 |
8166263 | Prahlad | Apr 2012 | B2 |
8219524 | Gokhale | Jul 2012 | B2 |
8312323 | Gokhale et al. | Nov 2012 | B2 |
8539253 | Sobel et al. | Sep 2013 | B2 |
8650445 | Gokhale et al. | Feb 2014 | B2 |
9122600 | Gokhale et al. | Sep 2015 | B2 |
10671472 | Gokhale et al. | Jun 2020 | B2 |
20010034812 | Ignatius et al. | Oct 2001 | A1 |
20020004883 | Nguyen et al. | Jan 2002 | A1 |
20020029281 | Zeidner et al. | Mar 2002 | A1 |
20020040405 | Gold | Apr 2002 | A1 |
20020042869 | Tate et al. | Apr 2002 | A1 |
20020042882 | Dervan et al. | Apr 2002 | A1 |
20020049778 | Bell et al. | Apr 2002 | A1 |
20020069369 | Tremain | Jun 2002 | A1 |
20020107877 | Whiting et al. | Aug 2002 | A1 |
20020129203 | Gagne et al. | Sep 2002 | A1 |
20020133755 | Lopke | Sep 2002 | A1 |
20020188691 | Ignatius et al. | Dec 2002 | A1 |
20020194340 | Ebstyne et al. | Dec 2002 | A1 |
20020198983 | Ullmann et al. | Dec 2002 | A1 |
20030014607 | Slavin et al. | Jan 2003 | A1 |
20030016609 | Rushton et al. | Jan 2003 | A1 |
20030014433 | Teloh et al. | Mar 2003 | A1 |
20030051097 | Ottesen et al. | Mar 2003 | A1 |
20030061491 | Jaskiewicz et al. | Mar 2003 | A1 |
20030169733 | Gurkowski et al. | Sep 2003 | A1 |
20040073716 | Boom et al. | Apr 2004 | A1 |
20040088432 | Hubbard et al. | May 2004 | A1 |
20040107199 | Dairymple et al. | Jun 2004 | A1 |
20040193953 | Callahan et al. | Sep 2004 | A1 |
20040210796 | Largman et al. | Oct 2004 | A1 |
20040230829 | Dogan et al. | Nov 2004 | A1 |
20050033756 | Kottomtharayil et al. | Feb 2005 | A1 |
20050086443 | Mizuno et al. | Apr 2005 | A1 |
20050114406 | Borthakur et al. | May 2005 | A1 |
20050114477 | Willging et al. | May 2005 | A1 |
20050144520 | Tuma et al. | Jun 2005 | A1 |
20050166011 | Burnett et al. | Jul 2005 | A1 |
20050172093 | Jain | Aug 2005 | A1 |
20050246510 | Retnamma et al. | Nov 2005 | A1 |
20050246568 | Davies | Nov 2005 | A1 |
20050256972 | Cochran et al. | Nov 2005 | A1 |
20050262296 | Peake | Nov 2005 | A1 |
20050268068 | Ignatius et al. | Dec 2005 | A1 |
20060005048 | Osaki et al. | Jan 2006 | A1 |
20060010227 | Atluri | Jan 2006 | A1 |
20060044674 | Martin et al. | Mar 2006 | A1 |
20060149889 | Sikha | Jul 2006 | A1 |
20060152754 | Boyd et al. | Jul 2006 | A1 |
20060224846 | Amarendran et al. | Oct 2006 | A1 |
20060230244 | Amarendran et al. | Oct 2006 | A1 |
20060294431 | Bader | Dec 2006 | A1 |
20070183224 | Erofeev | Aug 2007 | A1 |
20070186042 | Kottomtharayil et al. | Aug 2007 | A1 |
20070264956 | Bicker et al. | Nov 2007 | A1 |
20070288536 | Sen et al. | Dec 2007 | A1 |
20080059515 | Fulton | Mar 2008 | A1 |
20080091894 | Retnamma et al. | Apr 2008 | A1 |
20080229037 | Bunte et al. | Sep 2008 | A1 |
20080243914 | Prahlad et al. | Oct 2008 | A1 |
20080243957 | Prahlad et al. | Oct 2008 | A1 |
20080243958 | Prahlad et al. | Oct 2008 | A1 |
20080256173 | Ignatius et al. | Oct 2008 | A1 |
20090138763 | Arnold | May 2009 | A1 |
20090319534 | Gokhale | Dec 2009 | A1 |
20100131461 | Prahlad et al. | May 2010 | A1 |
20100131467 | Prahlad et al. | May 2010 | A1 |
20160041867 | Gokhale et al. | Feb 2016 | A1 |
20180089018 | Gokhale et al. | Mar 2018 | A1 |
20190294492 | Gokhale et al. | Sep 2019 | A1 |
Number | Date | Country |
---|---|---|
0259912 | Mar 1988 | EP |
0405926 | Jan 1991 | EP |
0467546 | Jan 1992 | EP |
0774715 | May 1997 | EP |
0809184 | Nov 1997 | EP |
0862304 | Sep 1998 | EP |
0899662 | Mar 1999 | EP |
0981090 | Feb 2000 | EP |
1174795 | Jan 2002 | EP |
1115064 | Dec 2004 | EP |
2366048 | Feb 2002 | GB |
WO 1995013580 | May 1995 | WO |
WO 1998039707 | Sep 1998 | WO |
WO 1998039709 | Sep 1998 | WO |
WO 1999012098 | Mar 1999 | WO |
WO 1999014692 | Mar 1999 | WO |
WO 1999017204 | Apr 1999 | WO |
WO 2004090788 | Oct 2004 | WO |
WO 2005055093 | Jun 2005 | WO |
Entry |
---|
Armstead et al., “Implementation of a Campus-Wide Distributed Mass Storage Service: The Dream vs. Reality,” IEEE, 1995, pp. 190-199. |
Arneson, “Development of Omniserver; Mass Storage Systems,” Control Data Corporation, 1990, pp. 88-93. |
Arneson, “Mass Storage Archiving in Network Environments” IEEE, 1998, pp. 45-50. |
Ashton, et al., “Two Decades of policy-based storage management for the IBM mainframe computer”, www.research.ibm.com, 19 pages, published Apr. 10, 2003, printed Jan. 3, 2009., www.research.ibm.com, Apr. 10, 2003, pp. 19. |
Cabrera, et al. “ADSM: A Multi-Platform, Scalable, Back-up and Archive Mass Storage System,” Digest of Papers, Compcon '95, Proceedings of the 40th IEEE Computer Society International Conference, Mar. 5, 1995-Mar. 9, 1995, pp. 420-427, San Francisco, CA. |
Commvault Systems, Inc., Continuous Data Replicator 7.0, Product Data Sheet, 2007. |
Eitel, “Backup and Storage Management in Distributed Heterogeneous Environments,” IEEE, 1994, pp. 124-126. |
Farley, “Building Storage Networks,” pp. 328-331, Osborne/McGraw-Hill, 2000. |
Gait, “The Optical File Cabinet: A Random-Access File system for Write-Once Optical Disks,” IEEE Computer, vol. 21, No. 6, pp. 11-22 (1988). |
Gibson, “Network Attached Storage Architecture,” pp. 37-45, ACM, Nov. 2000. |
http://en.wikipedia.org/wiki/Naive_Bayes_classifier, printed on Jun. 1, 2010, in 7 pages. |
Jander, “Launching Storage-Area Net,” Data Communications, US, McGraw Hill, NY, vol. 27, No. 4(Mar. 21, 1998), pp. 64-72. |
Recycle Bin (Windows), Aug. 2007, Wikipedia, pp. 1-3. |
Number | Date | Country | |
---|---|---|---|
20200264947 A1 | Aug 2020 | US |
Number | Date | Country | |
---|---|---|---|
61190057 | Dec 2006 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15685958 | Aug 2017 | US |
Child | 16857012 | US | |
Parent | 14826958 | Aug 2015 | US |
Child | 15685958 | US | |
Parent | 14139241 | Dec 2013 | US |
Child | 14826958 | US | |
Parent | 13615231 | Sep 2012 | US |
Child | 14139241 | US | |
Parent | 11963732 | Dec 2007 | US |
Child | 13615231 | US |