This application is based upon and claims the benefit of priority of the prior Japanese Patent Application No. 2010-212166, filed on Sep. 22, 2010, the entire contents of which are incorporated herein by reference.
The embodiments discussed herein are directed to a measure presentation device, a non-transitory computer readable storage medium.
Monitoring has been conventionally performed on various types of devices that constitute an IT (information technology) system. For example, an IP (internet protocol) network may be provided with a network monitor that monitors a router, a switch, and the like as a monitoring target device. The network monitor informs a network administrator or the like of a warning when the failure of the monitoring target device is detected, for example.
There has been recently known a measure presentation device that presents measures against a failure to the network administrator when the network monitor detects that the monitoring target device has the failure. For example, the measure presentation device presents the measures on the basis of information on the failure received from the network monitor, and presents the next measures on the basis of the execution result of the measures when the measures are executed by the network administrator. In other words, the network administrator sequentially executes the measures presented by the measure presentation device to deal with the failure of the monitoring target device. The technique has been known as disclosed in, for example, Japanese Laid-open Patent Publication No. 6-119174.
However, the conventional measure presentation device may make a network administrator select the measures for an execution target. Specifically, the conventional measure presentation device may present a plurality of measures without narrowing down measures to be next presented into one, depending on the failures and the measures of a monitoring target device. In this case, the network administrator selects the measures for an execution target from the plurality of measures presented by the conventional measure presentation device on the basis of the own capability and experience. This causes a problem that effective measures may not be performed on a failure because only individual measures are performed on the failure of the monitoring target device.
The problem may be also caused when the network monitor detects a possibility of the failure of the monitoring target device. Furthermore, the problem may be caused also when the network monitor and the measure presentation device are integrated with each other.
According to an aspect of an embodiment of the invention, a measure presentation device includes a measure storage unit that stores therein measure contents that are sequentially performed on a phenomenon of a device in association with an execution result of one measure content and a measure content performed next to the measure content; a history storage unit that stores therein measure procedure histories indicating the measure contents sequentially performed in past times against the phenomenon of the device and successes or failures of the measure procedure histories; an evaluating unit that evaluates, when the phenomenon occurs from the device, which of measure procedures including measure contents that are split from and associated with one execution result is effective among measure procedures determined from the measure contents stored in the measure storage unit on the basis of the successes or the failures of the measure procedure histories stored in the history storage unit; and a presenting unit that presents the measure procedure that is evaluated to be effective by the evaluating unit.
The object and advantages of the embodiment will be realized and attained by means of the elements and combinations particularly pointed out in the claims.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are not restrictive of the embodiment, as claimed.
Preferred embodiments of the present invention will be explained with reference to accompanying drawings.
The present invention is not limited to the embodiments explained below.
It will be explained about an IP network that includes a measure presentation device according to the first embodiment with reference to
The monitoring target device 10 is various types of devices included in the IP network 1. For example, the monitoring target device 10 is a router, a switch, a server, and the like. The monitoring target device 10 is monitored by the network monitor 30.
The state management device 20 manages various states of the monitoring target device 10. Specifically, the state management device 20 acquires various types of information from the monitoring target device 10 and saves the acquired information. For example, the state management device 20 transmits ping to the monitoring target device 10 to save information on the conduction state of the monitoring target device 10. Moreover, the state management device 20 acquires various logs from the monitoring target device 10 and saves the acquired logs. Furthermore, the state management device 20 saves information on the operating state of communication ports of the monitoring target device 10 when the monitoring target device 10 is a router, a switch, and the like.
The network monitor 30 monitors whether the monitoring target device 10 operates normally. For example, the network monitor 30 performs polling on the monitoring target device 10 to monitor the operating state of the monitoring target device 10. Moreover, when the monitoring target device 10 autonomously reports a warning, the network monitor 30 monitors the operating state of the monitoring target device 10 on the basis of the warning received from the monitoring target device 10.
Then, when it is detected that the monitoring target device 10 has a phenomenon such a failure, the network monitor 30 informs a network administrator or the like of the warning. In the following embodiments, a “phenomenon” indicates, for example, a failure that is caused by the monitoring target device 10, an event in which there is a possibility that the monitoring target device 10 has a failure, or the like. As an example, the “phenomenon” includes an event in which a response to ping is not output from the monitoring target device 10, an event in which the monitoring target device 10 has a heavy load, and the like.
When it is detected that the monitoring target device 10 has a phenomenon, the network monitor 30 transmits a new incident notification that indicates the generation of the phenomenon to the measure presentation device 100. At this time, the network monitor 30 transmits a new incident notification that includes phenomenon information indicative of the contents of the phenomenon, attribute information on the monitoring target device 10, and the like. As an example, phenomenon information included in the new incident notification includes information that indicates an event in which a response to ping is not output from the monitoring target device 10, like the example. Moreover, as an example, attribute information on the monitoring target device 10 included in the new incident notification includes the device name, the maker, the model name, and the like of the monitoring target device 10.
When the new incident notification is received from the network monitor 30, the measure presentation device 100 presents a measure procedure that is performed on the phenomenon. Moreover, the term “measure procedure” indicates a combination of measures that are sequentially performed on the phenomenon. For example, the “measure procedure” presented by the measure presentation device 100 includes information that includes a measure A, a measure B, and a measure C and indicates the process of the measures in order of the measure A, the measure B, and the measure C.
Herein, the measure presentation device 100 stores, every phenomenon of the monitoring target device 10 that can occur, a measure procedure candidate that is performed on the phenomenon. In some cases, the measure procedure stored in the measure presentation device 100 can include a measure that is split into next several measures with respect to one execution result. In other words, in the case of a measure included in the measure procedure stored in the measure presentation device 100, the next measures may not be uniquely determined by the execution result of the measure in some cases.
When a new incident notification is received from the network monitor 30, the measure presentation device 100 that stores the measure procedures presents a measure procedure that is effective against the failure of the monitoring target device 10, among the measure procedures that are saved in the device itself. Specifically, the measure presentation device 100 performs the next process.
The measure presentation device 100 stores, as history information, measure procedures executed in past times and the execution results of the measure procedures. Then, when a new incident notification is received, the measure presentation device 100 evaluates, with respect to one measure result, effectiveness for a split destination measure of a split measure that is split into a plurality of measures on the basis of the history information. In other words, the measure presentation device 100 evaluates which route's measure procedure is effective with respect to a measure procedure including a split measure on the basis of the history information, that is, which route's measure procedure has a high possibility for solving a phenomenon.
Then, the measure presentation device 100 presents a measure procedure that goes through a split measure and a split destination measure that are effective. As a result, the measure presentation device 100 according to the first embodiment can present an effective measure procedure against the phenomenon of the monitoring target device 10.
The configuration of the IP network 1 in which the measure presentation device 100 according to the first embodiment is placed is not limited to the example illustrated in
It will be below explained in detail about the measure presentation device 100 according to the first embodiment. Hereinafter, a measure procedure can be referred to as a “scenario pattern” and one measure included in a measure procedure can be referred to as a “scenario part”. A scenario pattern can be a plurality of scenario parts that is arranged in accordance with a certain sequence.
Configuration of Measure Presentation Device of First Embodiment
Next, it will be explained about the measure presentation device 100 according to the first embodiment with reference to
The scenario storage unit 110 stores therein scenario parts that are sequentially performed against the phenomenon of the monitoring target device 10, by using an association between the execution result of one scenario part and a scenario part performed next to the one scenario part. In other words, the scenario storage unit 110 can be called a measure storage unit. Herein, the plurality of scenario parts stored in the scenario storage unit 110 includes scenario parts that are associated in accordance with one execution result. Hereinafter, scenario parts that are associated in accordance with one execution result can be referred to as “split scenario parts” in some cases.
It will be explained about a relationship between scenario parts stored in the scenario storage unit 110 with reference to
In an example illustrated in
Specifically, the scenario part PA1 indicates a phenomenon “node uncertainty”. The “node uncertainty” indicates, for example, a phenomenon in which a response of ping is not output from the monitoring target device 10. Attribute information indicated in the scenario part PA1 will be later described. Moreover, the scenario part PA1 is the first scenario part of the scenario pattern and does not have a measure. Hereinafter, a scenario part that does not have a measure like the scenario part PA1 can be referred to as an “introduction scenario part” in some cases.
Each of the scenario parts PA2 to PA9 indicates a measure performed against the phenomenon indicated by the scenario part PA1. Specifically, the scenario part PA2 indicates a measure “acquisition of the state of X”, the scenario part PA3 indicates a measure “acquisition of the state of Y”, and the scenario part PA6 indicates a measure “acquisition of the state of Z”. The “acquisition of the state” indicates, for example, that various states of the monitoring target device 10 are acquired from the state management device 20.
The scenario part PA4 indicates a measure “problem solving procedure SP1” and the scenario part PA5 indicates a measure “problem solving procedure SP2”. Moreover, the scenario part PA7 indicates a measure “problem solving procedure SP3” and the scenario part PA8 indicates a measure “problem solving procedure SP4”. The “problem solving procedure” indicates, for example, a measure for “rebooting the monitoring target device 10”, a measure for “contacting a network administrator”, and the like.
The example illustrated in
The scenario storage unit 110 stores the scenario parts PA1 to PA9 in association with the execution results of the scenario parts. Specifically, the scenario storage unit 110 stores the scenario part PA3 and the scenario part PA6 in association with the execution result “NG” of the scenario part PA2. Moreover, the scenario storage unit 110 stores the scenario part PA9 in association with the execution result “OK” of the scenario part PA2. Moreover, the scenario storage unit 110 stores the scenario part PA4 in association with the execution result “NG” of the scenario part PA3 and stores the scenario part PA5 in association with the execution result “OK” of the scenario part PA3. Moreover, the scenario storage unit 110 stores the scenario part PA7 in association with the execution result “NG” of the scenario part PA6 and stores the scenario part PA8 in association with the execution result “OK” of the scenario part PA6.
In other words, when the execution result of the measure “acquisition of the state of X” of the scenario part PA2 is “NG” in the example illustrated in
In this way, scenario parts stored in the scenario storage unit 110 include split scenario parts that are split into several scenario parts from one execution result. Specifically, as illustrated in
Next, it will be explained in detail about the configuration of scenario parts stored in the scenario storage unit 110 with reference to
As illustrated in
In the example illustrated in
In other words, the introduction scenario part has items such as for example “scenario part ID”, “phenomenon ID”, and “attribute information”. The scenario parts other than the introduction scenario part have items such as for example “scenario part ID”, “phenomenon ID”, “measure”, “rule”, “explanation”, “result”, “simulation permission”, and “termination flag”. Moreover, when the “measure” of the split scenario part (the scenario part PA2) of
The “scenario part ID” indicates identification information that identifies a scenario part. In the example illustrated in
The “attribute information” indicates device information of the monitoring target device 10.
The “attribute information ID” of the attribute information is an identifier of attribute information and the “attribute information Value” is an attribute corresponding to the identifier of the attribute information. In the example of
It should be noted that attribute information is not limited to the example illustrated in
The “measure” indicates a content that is performed against the phenomenon of the monitoring target device 10. The content of “measure” is a content that should be performed against the content of “phenomenon ID” of the scenario part. The “rule” indicates association information between scenario parts, and determines whether its own scenario part is a scenario part that is performed next to another scenario part. Specifically, the “rule” includes a description on “phenomenon”, another scenario part ID, and an execution result thereof.
For example, “phenomenon=node uncertainty” is described in the “rule” of the scenario part PA2 illustrated in
“The scenario part PA2=NG” is described in the “rule” of the scenario part PA3. In this way, the scenario part in which the execution result of another scenario part is described in the “rule” is associated with the other scenario part. Specifically, the scenario part PA3 is associated with the scenario part PA2, and becomes the candidate of a scenario part to be referred to next when the result of the measure “acquisition of the state of X” of the scenario part PA2 is “NG”.
The “explanation” is information for a network administrator, and is the explanation for “measure”. For example, a network administrator can refer to information stored in “explanation” to perform a measure in some cases.
The “result” is information that may be the execution result of “measure”. For example, the execution result of the measure “acquisition of the state of X” of the scenario part PA2 illustrated in
Herein, the execution result of a measure is illustrated when the “result” is “OK” or “NG” and the execution result of a measure cannot be determined when the “result” is “ERROR”. For example, it is assumed that “measure” is “to confirm whether an error log is output”. At this time, when it can be confirmed that an error log is not output from the monitoring target device 10, the “result” becomes “OK” because the monitoring target device 10 does not have an error. Meanwhile, when it can be confirmed that an error log is output from the monitoring target device 10, the “result” becomes “NG” because the monitoring target device 10 has an error. On the other hand, when it cannot be confirmed whether an error log is output from the monitoring target device 10, the “result” becomes “ERROR” because the measure cannot be performed.
The “simulation permission” is information that indicates whether its scenario part is a scenario part that may be automatically executed by a system. In the example illustrated in
The “termination flag” is information that indicates whether its own scenario part is a scenario part to be finally executed in the scenario pattern. The example illustrated in
Moreover, the scenario parts PA2 to PA9 that are associated with the introduction scenario part PA1 that stores the phenomenon ID “2” are illustrated in the examples illustrated in
Returning to
The incident information 121 stores the past phenomenon of the monitoring target device 10 and the scenario pattern performed against the phenomenon in association with each other. Hereinafter, a combination of the phenomenon and the scenario pattern stored in the incident information 121 can be described as an “incident”.
The “incident ID” is identification information that identifies an incident. The “phenomenon ID” corresponds to the phenomenon ID illustrated in
A scenario pattern performed in past times is stored in the “history”. Specifically, as illustrated in
Herein, the “result” of a scenario part stored in the “history” indicates the execution result of the scenario part. In this case, the “result” of the scenario part illustrated in
In the example illustrated in
The phenomenon history information 122 stores a past phenomenon of the monitoring target device 10 and an incident obtained by performing a scenario pattern against the phenomenon, in association with each other.
The attribute history information 123 stores attribute information of the monitoring target device 10 from which a phenomenon has occurred in past times and an incident that is obtained by performing a scenario pattern on the monitoring target device 10 that has the attribute information, in association with each other.
The “attribute information Hash value” is a hash value of attribute information. For example, the hash value is computed by MD5 (Message Digest Algorithm 5) or the like. For example, it is assumed that the attribute information of the monitoring target device 10 from which a phenomenon has occurred in past times is information illustrated in
The example illustrated in
The scenario part statistical information 124 stores, every scenario part stored in the scenario storage unit 110, statistical information on the scenario part.
The “scenario part ID” corresponds to the scenario part ID illustrated in
The example illustrated in
Returning to
When the monitoring target device 10 has a phenomenon, the candidate extracting unit 131 acquires the plurality of scenario parts corresponding to this phenomenon from the scenario storage unit 110, and extracts a scenario pattern associated with the acquired scenario part.
Specifically, when the monitoring target device 10 has a phenomenon, the candidate extracting unit 131 receives a new incident notification from the network monitor 30. Then, the candidate extracting unit 131 acquires, from the scenario storage unit 110, an introduction scenario pattern for which phenomenon information included in the new incident notification and attribute information of the monitoring target device 10 are identical to. Then, the candidate extracting unit 131 extracts the extracted introduction scenario pattern and a scenario pattern associated with the introduction scenario pattern as a scenario pattern candidate. Moreover, the candidate extracting unit 131 virtually executes a measure content that is described in the “measure” of each the scenario part included in the scenario pattern candidate. At this time, when “1 (automatic execution permission)” is described in the “simulation permission” of each the scenario part included in the scenario pattern candidate, the candidate extracting unit 131 actually executes the measure content described in the “measure” of each the scenario part.
In general, the measure for a phenomenon is changed depending on a device name, a maker, or a model name of the monitoring target device 10 from which a phenomenon occurs. However, a measure may not be changed depending on the device name or the like. Therefore, the candidate extracting unit 131 may acquire, from the scenario storage unit 110, an introduction scenario pattern for which only the phenomenon information included in the new incident notification is identical to.
It will be explained about a candidate extraction process that is performed by the candidate extracting unit 131 by using an example illustrated in
The new incident notification illustrated in
When the new incident notification illustrated in
Next, the candidate extracting unit 131 extracts a scenario part associated with the scenario part PA1 acquired from the scenario storage unit 110. In the example illustrated in
Then, the candidate extracting unit 131 aligns the scenario parts in execution order on the basis of the information stored in the “rule” of each of the scenario parts PA1 to PA9, and virtually executes a measure content described in the “measure” of each of the scenario parts in sequence from the introduction scenario part. At this time, when “1 (automatic execution permission)” is described in the “simulation permission”, the candidate extracting unit 131 actually executes a measure content described in the “measure” of the scenario part. Then, when it reaches a split scenario part as the execution result of the measure content of each the scenario part, the candidate extracting unit 131 extracts a scenario pattern candidate on the basis of the measure result.
It is specifically explained by using an example illustrated in
First, it is decided that the scenario part PA2 is referred to next to the scenario part PA1 that is an introduction scenario part. Therefore, as illustrated in the first line of
Herein, it is assumed that the execution result of the measure “acquisition of the state of X” is “NG”. In this case, the candidate extracting unit 131 determines that a scenario part that is performed next to the scenario part PA2 is the scenario part PA3 or PA6. In other words, the candidate extracting unit 131 determines that the scenario part PA9 is not performed next to the scenario part PA2. Herein, the candidate extracting unit 131 cannot uniquely specify a scenario part that is performed next to the scenario part PA2. Therefore, the candidate extracting unit 131 extracts, as a scenario pattern candidate, all the scenario patterns in which the scenario part PA3 or PA6 is performed next to the scenario parts PA1 and PA2.
Specifically, as illustrated in the second line of
Returning to
Specifically, when the monitoring target device 10 has a phenomenon, the history extracting unit 132 receives the new incident notification transmitted by the network monitor 30 from the candidate extracting unit 131. Then, the history extracting unit 132 acquires an incident ID corresponding to phenomenon information included in the new incident notification from the phenomenon history information 122 of the history storage unit 120.
Next, the history extracting unit 132 computes, every incident acquired from the phenomenon history information 122, a similarity between the attribute information of the incident and the attribute information included in the new incident notification. In other words, the history extracting unit 132 can be called a computing unit. Hereinafter, a similarity between the attribute information of an incident and the attribute information included in the new incident notification may be described as “incident similarity”.
Then, the history extracting unit 132 extracts a scenario pattern that has been executed in past times, on the basis of the scenario part described in the “history” of the incident acquired from the phenomenon history information 122. Hereinafter, a scenario pattern that has been executed in past times may be described as a “scenario pattern history”. Then, the history extracting unit 132 extracts a scenario pattern history, which is identical to the scenario pattern candidate extracted by the candidate extracting unit 131, among scenario pattern histories extracted from the phenomenon history information 122.
Hereinafter, it will be explained about an incident similarity computation process that is performed by the history extracting unit 132. First, the history extracting unit 132 acquires an incident ID identical to all the attribute items of the attribute information included in the new incident notification from the attribute history information 123 of the history storage unit 120. Next, the history extracting unit 132 excludes an attribute item one-by-one from the attribute information included in the new incident notification, and acquires an incident ID identical to the attribute information except for the attribute item from the attribute history information 123.
For example, it is assumed that the new incident notification transmitted from the network monitor 30 is an example illustrated in
Next, the history extracting unit 132 excludes the attribute item “KIND=Type A” from the attribute information “HARD=router, MAKER=AAA, and KIND=Type A” included in the new incident notification. Then, the history extracting unit 132 computes a hash value of the attribute information “HARD=router, MAKER=AAA” except for the attribute item “KIND=Type A”, and extracts an incident ID stored in association with the computed hash value from the attribute history information 123.
The history extracting unit 132 excludes the attribute items “MAKER=AAA, KIND=Type A” from the attribute information “HARD=router, MAKER=AAA, and KIND=Type A” included in the new incident notification. Then, the history extracting unit 132 computes a hash value of the attribute information “HARD=router” except for the attribute items “MAKER=AAA, KIND=Type A”, and extracts an incident ID stored in association with the computed hash value from the attribute history information 123.
Then, the history extracting unit 132 gives the higher incident similarity to an incident that has more attribute items that are identical to the attribute information included in the new incident notification among the incidents acquired from the phenomenon history information 122. Specifically, the history extracting unit 132 gives the highest incident similarity to an incident that is identical to the hash value of all the attribute items of the attribute information included in the new incident notification among the incidents acquired from the phenomenon history information 122. Moreover, the history extracting unit 132 gives the second high incident similarity to an incident that is identical to the hash value of the attribute information except for one attribute item from the attribute information included in the new incident notification among the incidents acquired from the phenomenon history information 122. Then, the history extracting unit 132 gives the lowest incident similarity to an incident that is not identical to the attribute information included in the new incident notification among the incidents acquired from the phenomenon history information 122.
This example indicates that the history extracting unit 132 excludes an attribute item from attribute information in order of “KIND (model name)”, “MAKER (manufacturer)”, and “HARD (device name)”. This reason is that “HARD (device name)” of attribute information is information of specifying a device and has a higher level of importance than that of the other “KIND (model name)” and “MAKER (manufacturer)”. However, the history extracting unit 132 is not limited to the example. For example, the history extracting unit 132 may exclude an attribute item from attribute information in order of “MAKER (manufacturer)”, “HARD (device name)”, and “KIND (model name)”. Moreover, the history extracting unit 132 may compute a hash value for all combinations of attribute items and acquire an incident identical to the computed hash value from the phenomenon history information 122.
In the example illustrated in
Moreover, the attribute information of “HASH value 2” is identical with “HARD” and “MAKER” included in the new incident attribute information but is not identical with “KIND”. In other words, the attribute information of “HASH value 2” and the new incident attribute information are identical with each other with respect to items other than one item “KIND”. In this case, the history extracting unit 132 gives an incident similarity “1.0” to an incident that stores the attribute information of “HASH value 2”.
The attribute information of “HASH value 3” is identical with “HARD” included in the new incident attribute information but is not identical with “MAKER” and “KIND”. In this case, the history extracting unit 132 gives an incident similarity “0.9” to an incident that stores the attribute information of “HASH value 3”. Moreover, the attribute information of “HASH value 4” is not identical to all the items of the new incident attribute information. In this case, the history extracting unit 132 gives an incident similarity “0.8” to an incident that stores the attribute information of “HASH value 4”.
In this way, the history extracting unit 132 acquires an incident identical with the phenomenon information included in the new incident notification from the phenomenon history information 122. Then, the history extracting unit 132 gives a higher incident similarity to an incident that has more attribute items identical with the attribute information included in the new incident notification, among incidents identical with the phenomenon information included in the new incident notification.
Next, the history extracting unit 132 extracts a scenario pattern that has been executed in past times on the basis of the scenario parts stored in the “history” of the incident acquired from the phenomenon history information 122. For example, it is assumed that the history extracting unit 132 extracts an incident ID “1” from the phenomenon history information 122. Moreover, it is assumed that the incident indicated by the incident ID “1” is the incident 121a illustrated in
Then, the history extracting unit 132 extracts a scenario pattern history that is identical with the scenario pattern candidate extracted by the candidate extracting unit 131, among the scenario pattern histories extracted from the phenomenon history information 122.
The execution result applying unit 133 executes each scenario part included in the scenario pattern histories extracted by the history extracting unit 132. Specifically, the execution result applying unit 133 executes each the scenario part included in the scenario pattern histories, and narrows down a scenario pattern history on the basis of the execution result. In other words, the execution result applying unit 133 applies the present state of the monitoring target device 10 to the scenario pattern histories to narrow down a scenario pattern history.
Now, it will be explained about a narrowing down process that is performed by the history extracting unit 132 and the execution result applying unit 133 with reference to
The upper stage of
Information may be stored by a network administrator in the “result” of the scenario part that is finally executed in the scenario pattern history. For example, when the final scenario part is executed among the scenario patterns included in the scenario pattern history and thus the phenomenon is solved, it is considered that the network administrator registers “OK” in the “result” of the scenario part that is finally executed in the scenario pattern history. On the other hand, when the final scenario part is executed and the phenomenon is not solved, it is considered that the network administrator registers “NG” in the “result” of the scenario part that is finally executed in the scenario pattern history.
In the example illustrated in
In the example illustrated in the upper stage of
Herein, it is assumed that the scenario pattern candidates illustrated in
Then, the execution result applying unit 133 executes a scenario part for which the execution is permitted among the scenario parts included in the scenario pattern history illustrated in the middle stage of
Specifically, the execution result applying unit 133 executes a measure content stored in the “measure” of the scenario part PA3 corresponding to the scenario part ID “3” among the scenario parts included in the history IDs “2” to “5”. As illustrated in
Herein, it is assumed that the execution result of the measure “acquisition of the state of Y” of the scenario part PA3 is “OK”. As illustrated in
Moreover, it is assumed that the execution result of the measure “acquisition of the state of Z” of the scenario part PA6 is “NG”. As illustrated in
In other words, the execution result applying unit 133 narrows down the scenario pattern histories corresponding to the history IDs “2” to “7” illustrated in the middle stage of
In this way, the execution result applying unit 133 executes the scenario pattern histories extracted by the history extracting unit 132 to narrow down a scenario pattern history. In other words, the execution result applying unit 133 narrows down the scenario pattern histories extracted by the history extracting unit 132 by using the present state of the monitoring target device 10 that has a phenomenon.
When the scenario parts included in the scenario pattern history are executed, the execution result applying unit 133 may store the execution result in the phenomenon history information 122. At this time, the execution result applying unit 133 stores the execution result of the scenario pattern history in the phenomenon history information 122 in such a manner that it can be determined that it is not the actually-performed incident but is the temporarily-executed incident. For example, a “temporary history flag” indicating whether an incident is a temporary incident may be provided in an incident ID, and whether an incident is a temporarily-executed incident may be determined by the “temporary history flag”.
The filter unit 134 selects a scenario pattern history for which the “result” of the scenario part that is finally executed in the scenario pattern history is a success, among the scenario pattern histories narrowed down by the execution result applying unit 133. In other words, the filter unit 134 executes the scenario pattern to select a scenario pattern history for which a phenomenon is solved. In other words, the filter unit 134 can be called a selecting unit.
It is explained by using the example illustrated in
The priority processing unit 135 gives a priority to the scenario pattern history selected by the filter unit 134 on the basis of the incident similarity, the occurrence number of scenario patterns, the occurrence frequency of scenario patterns, a productive time, and the like. Moreover, the priority processing unit 135 does not perform the process when one scenario pattern history is selected by the filter unit 134. For example, when one scenario pattern history corresponding to the history ID “5” is selected by the filter unit 134 like the example illustrated in the lower stage of
Now, it will be explained about a priority process that is performed by the priority processing unit 135 with reference to
First, it will be explained about a scenario pattern history illustrated in
When the plurality of scenario pattern histories is selected by the filter unit 134 like the example illustrated in
The “incident similarity” indicates an incident similarity that is computed by the history extracting unit 132. For example, the priority processing unit 135 gives a higher priority to a scenario pattern history that has a larger incident similarity. This reason is that a scenario pattern that has a larger incident similarity is a scenario pattern that is performed on a phenomenon similar to the phenomenon of the monitoring target device 10.
The “scenario pattern occurrence number” indicates a total selection number by which the scenario parts included in the scenario pattern history have been selected in past times. Specifically, the scenario part statistical information 124 is associated with each scenario part included in the scenario pattern history. Like the example illustrated in
The “scenario pattern occurrence frequency” indicates a ratio of the number of executions of the scenario pattern to the “scenario pattern occurrence number”. Specifically, the incident information 121 of the history storage unit 120 stores the scenario pattern that has been executed in past times. In other words, the number of times of the scenario pattern that has been actually executed in past times can be calculated by referring to the incident information 121. The “scenario pattern occurrence frequency” is a value that is obtained by dividing the number of executions of the scenario pattern by the scenario pattern occurrence number. The priority processing unit 135 gives a higher priority to a scenario pattern history that has a larger scenario pattern occurrence frequency. This reason is that a scenario pattern that has a larger scenario pattern occurrence frequency is a scenario pattern that is more frequently executed in operation and thus has a higher reliability.
The “productive time” indicates a total execution time when the scenario pattern history is executed. The priority processing unit 135 gives a higher priority to a scenario pattern history that has a smaller productive time. This reason is that a scenario pattern that has a smaller productive time is a scenario pattern that can quickly respond to the phenomenon of the monitoring target device 10.
Meanwhile, the priority processing unit 135 may not give a priority like the example. For example, the priority processing unit 135 may give a higher priority to a scenario pattern history that has a smaller scenario pattern occurrence number. The setting method of a priority performed by the priority processing unit 135 can be changed by tuning up the system.
Moreover, the priority processing unit 135 may give a priority on the basis of information other than the items illustrated in
Furthermore, for example, the priority processing unit 135 may give a higher priority to a scenario pattern history that is more frequently selected by the filter unit 134. For example, in the example illustrated in
Moreover, for example, the priority processing unit 135 may give a higher priority to a scenario pattern history that includes scenario parts that are more frequently selected from split scenario parts as a split-destination scenario part among the scenario pattern histories selected by the filter unit 134. For example, in the example illustrated in
Moreover, the priority processing unit 135 may give a priority to each weighted item illustrated in
Returning to
Meanwhile, when a plurality of scenario patterns is selected by the filter unit 134, the presenting unit 141 presents the scenario patterns and the scheduled execution times of the scenario patterns in descending order of priorities given by the priority processing unit 135. At this time, the presenting unit 141 may present a scenario pattern of which the priority is higher than a predetermined threshold value or may present only one scenario pattern of which the priority is the highest.
For example, the presenting unit 141 may present a scenario pattern on a display device such as a display (not illustrated). Moreover, for example, the presenting unit 141 may transmit a scenario pattern to the network monitor 30 to present the scenario pattern to the network administrator.
When the scenario pattern presented by the presenting unit 141 is executed by the network administrator or the like, the updating unit 142 updates the history storage unit 120. Specifically, when the scenario pattern is executed, the updating unit 142 registers the scenario pattern in the incident information 121. At this time, the updating unit 142 takes out a new incident ID and generates a new incident corresponding to the incident ID. Then, the updating unit 142 stores a phenomenon ID corresponding to a phenomenon included in the new incident notification in the phenomenon ID of the new incident information. Moreover, the updating unit 142 stores attribute information included in the new incident notification in the attribute information of the incident information. Furthermore, the updating unit 142 sequentially stores the executed scenario parts in the history of the incident information.
Meanwhile, when the scenario pattern is executed, the updating unit 142 registers the newly taken-out incident ID in the phenomenon history information 122 corresponding to the phenomenon included in the new incident notification. Moreover, when the scenario pattern is executed, the updating unit 142 registers the newly taken-out incident ID in the attribute history information 123 corresponding to the hash value of the attribute information included in the new incident notification. Moreover, when the scenario pattern is executed, the updating unit 142 increments the number of selections of the scenario part statistical information 124, and registers the newly taken-out incident ID in the incident list of the scenario part statistical information 124. Moreover, when the phenomenon is solved by executing the scenario pattern, the updating unit 142 increments the number of problem solutions of the scenario part statistical information 124.
The updating unit 142 according to the first embodiment may automatically execute the scenario pattern presented by the presenting unit 141. For example, when the number of the scenario patterns presented by the presenting unit 141 is one, the updating unit 142 may sequentially and automatically execute scenario parts up to the scenario part in which “1 (automatic execution permission)” is stored in the simulation permission among the scenario parts included in the scenario pattern.
The scenario storage unit 110 and the history storage unit 120 described above are, for example, a semiconductor memory device such as a RAM (random access memory), a ROM (read only memory), and a flash memory, or a storage device such as a hard disk and an optical disc. Moreover, the evaluating unit 130, the presenting unit 141, and the updating unit 142 described above may be realized by, for example, an integrated circuit such as ASIC (application specific integrated circuit).
Processing Procedures by Measure Presentation Device of First Embodiment
Next, it will be explained about the processing procedures that are performed by the measure presentation device 100 according to the first embodiment with reference to
As illustrated in
Next, the history extracting unit 132 performs a history extraction process (Step S103). It will be below described about the history extraction process that is performed by the history extracting unit 132 with reference to
Next, the execution result applying unit 133 performs an execution result application process (Step S104). It will be below described about the execution result application process that is performed by the execution result applying unit 133 with reference to
Next, the filter unit 134 and the priority processing unit 135 perform a filter priority process (Step S105). Specifically, the filter unit 134 performs a filtering process and the priority processing unit 135 performs a priority process. It will be below described about the filter priority process that is performed by the filter unit 134 and the priority processing unit 135 with reference to
Then, the presenting unit 141 presents a scenario pattern having a high priority that is given by the priority processing unit 135, among the scenario patterns selected by the filter unit 134 (Step S106).
History Extraction Processing Procedures by History Extracting Unit
Next, it will be explained about the procedures of the history extraction process illustrated at Step S103 of
As illustrated in
Next, the history extracting unit 132 acquires an incident ID identical to all attribute information included in the new incident notification from the attribute history information 123 (Step S202). Next, the history extracting unit 132 excludes one attribute item from the attribute information included in the new incident notification (Step S203). Then, the history extracting unit 132 acquires an incident ID identical with the attribute information from which the attribute item is excluded from the attribute history information 123 (Step S204).
Next, when the number of attribute items of the attribute information included in the new incident notification is not zero (Step S205: NO), the history extracting unit 132 performs the processing procedures of Steps S203 and S204.
On the other hand, when the number of attribute items of the attribute information included in the new incident notification is zero (Step S205: YES), the history extracting unit 132 gives an incident similarity to the incident indicated by the incident ID acquired at Step S201. Specifically, the history extracting unit 132 gives a higher incident similarity to an incident that has more attribute items that are identical with the attribute information included in the new incident notification (Step S206).
Next, the history extracting unit 132 extracts scenario pattern histories that have been executed in past times on the basis of the scenario parts stored in the “history” of the incident acquired at Step S201 (Step S207). Then, the history extracting unit 132 extracts, among the scenario pattern histories, a scenario pattern history that is identical with the scenario pattern candidate extracted by the candidate extracting unit 131 (Step S208).
In this case, the incident similarity given at Step S206 is used when a priority is given to a scenario pattern history by the priority processing unit 135. It will be below described about a process that is performed by the priority processing unit 135 with reference to FIG. 19.
Execution Result Application Processing Procedures by Execution Result Applying Unit
Next, it will be explained about the procedures of the execution result application process illustrated at Step S104 of
As illustrated in
Next, the execution result applying unit 133 sets a scenario part of which the execution sequence is first as a processing target among the scenario parts included in the scenario pattern history selected at Step S301 (Step S302). Next, the execution result applying unit 133 determines whether the processing-target scenario part can be automatically executed on the basis of the information stored in the simulation permission of the processing-target scenario part (Step S303).
Then, when the processing-target scenario part can be automatically executed (Step S303: YES), the execution result applying unit 133 executes the measure content stored in the “measure” of the processing-target scenario part (Step S304). Next, the execution result applying unit 133 acquires the execution result of the measure content from the state management device 20 (Step S305).
Then, when the execution result can be acquired from the state management device 20 (Step S306: YES), the execution result applying unit 133 registers a temporary incident in the phenomenon history information 122 on the basis of the execution result (Step S307). On the other hand, when the execution result cannot be acquired from the state management device 20 (Step S306: NO), the execution result applying unit 133 returns the process control to Step S302. Specifically, the execution result applying unit 133 sets a scenario part to be next executed as a processing target (Step S302).
Then, when the process is not performed on all the scenario parts included in the scenario pattern history selected at Step S301 (Step S308: NO), the execution result applying unit 133 returns the process control to Step S302. On the other hand, when the process is performed on all the scenario parts included in the scenario pattern history (Step S308: YES), the execution result applying unit 133 determines whether the execution result application process is performed on all the scenario pattern histories (Step S309).
Then, when the execution result application process is not performed on all the scenario pattern histories (Step S309: NO), the execution result applying unit 133 returns the process control to Step S301 and selects one scenario pattern history on which the execution result application process is not performed. On the other hand, when the execution result application process is performed on all the scenario pattern histories (Step S309: YES), the execution result applying unit 133 terminates the process. In addition, when the processing-target scenario part cannot be automatically executed (Step S303: NO), the execution result applying unit 133 performs the process of Step S309. In this way, the execution result applying unit 133 narrows down the scenario pattern histories extracted by the history extracting unit 132 by using the latest state of the monitoring target device 10 that has the phenomenon.
In the example illustrated in
Filter Priority Processing Procedures by Filter Unit and Priority Processing Unit
Next, it will be explained about the procedures of the filter priority process illustrated at Step S105 of
As illustrated in
Next, the priority processing unit 135 gives a priority to the scenario pattern history selected by the filter unit 134 on the basis of the incident similarity (Step S402). For example, the priority processing unit 135 gives a higher priority to a scenario pattern history that has a larger incident similarity.
Next, the priority processing unit 135 gives a priority to the scenario pattern history selected by the filter unit 134 on the basis of the scenario pattern occurrence number (Step S403). For example, the priority processing unit 135 gives a higher priority to a scenario pattern history that has a larger scenario pattern occurrence number.
Next, the priority processing unit 135 gives a priority to the scenario pattern history selected by the filter unit 134 on the basis of the scenario pattern occurrence frequency (Step S404). For example, the priority processing unit 135 gives a higher priority to a scenario pattern history that has a larger scenario pattern occurrence frequency.
Next, the priority processing unit 135 gives a priority to the scenario pattern history selected by the filter unit 134 on the basis of the productive time (Step S405). For example, the priority processing unit 135 gives a higher priority to a scenario pattern history that has a smaller productive time.
As described above, the measure presentation device 100 according to the first embodiment stores a scenario part group including a split scenario part associated with several other scenario parts with respect to one execution result in the scenario storage unit 110, like the scenario part PA2 illustrated in
As a result, the measure presentation device 100 according to the first embodiment can present an effective measure for the failure of the monitoring target device 10. In other words, even if there is a split scenario part associated with several scenario parts with respect to one execution result, the measure presentation device 100 can predict a scenario pattern that has a high problem solution possibility and present the scenario pattern to the network administrator. As a result, because the selection of a measure content dependent on the person can be removed, the measure presentation device 100 can perform efficient and appropriate measures on the phenomenon without relying on the capability and experience of the network administrator. In other words, if the measure presentation device 100 according to the first embodiment is used, errors in judgment of the network administrator can be prevented. As a result, the replay of a measure and the occurrence of a new failure can be prevented.
Moreover, because the executed scenario patterns are accumulated in the history storage unit 120, the measure presentation device 100 according to the first embodiment can save, as history information, a scenario pattern that has a higher reliability as the duration of use is longer. Because a scenario pattern candidate is evaluated on the basis of a scenario pattern having a high reliability, the measure presentation device 100 can present a scenario pattern that has a high problem solution possibility as the duration of use gets longer.
Moreover, because the measure presentation device 100 according to the first embodiment acquires the present state of the monitoring target device 10 that has the phenomenon and narrows down scenario pattern candidates, the measure presentation device 100 can present an appropriate scenario pattern for the present phenomenon.
Moreover, the measure presentation device 100 according to the first embodiment computes an incident similarity that is a similarity between the present phenomenon and the past phenomenon and preferentially presents a scenario pattern that has a high incident similarity. As a result, the measure presentation device 100 according to the first embodiment can present a scenario pattern that has a high problem solution possibility on the basis of the scenario pattern performed on the past phenomenon similar to the present phenomenon.
Moreover, as illustrated in
As illustrated in
Configuration of Measure Presentation Device by Second Embodiment
First, it will be explained about a measure presentation device 200 according to the second embodiment with reference to
Herein, it will be explained about a process that is performed by the execution result applying unit 233 and the history extracting unit 232 with reference to
The execution result applying unit 233 performs the execution result application process on the scenario pattern candidates illustrated in the upper stage of
In the example illustrated in
Herein, it is assumed that the execution result of the measure “acquisition of the state of Y” of the scenario part PA3 is “OK” and the execution result of the measure “acquisition of the state of Z” of the scenario part PA6 is “NG”. In this case, as illustrated in the lower stage of
The history extracting unit 232 performs the history extraction process on the scenario pattern candidates narrowed down by the execution result applying unit 233. Specifically, similarly to the history extracting unit 132 illustrated in
As described above, the measure presentation device 200 according to the second embodiment narrows down scenario pattern candidates by using the present state of the monitoring target device 10 and evaluates the scenario pattern candidates by using the scenario pattern histories that have been executed in past times. As a result, because the measure presentation device 200 previously narrows down scenario pattern candidates that are an evaluation target even if enormous amount of scenario pattern histories are saved, the measure presentation device 200 can present an effective measure at high speed in accordance with a low-load process. In other words, when the measure presentation device 200 is used, a more prompt action can be performed on a phenomenon.
Each component of each device illustrated in the embodiments is a functional concept. Therefore, these components are not necessarily constituted physically as illustrated in the drawings. In other words, the specific configuration of dispersion/integration of each device is not limited to the illustrated configuration. Therefore, all or a part of each device can dispersed or integrated functionally or physically in an optional unit in accordance with various types of loads or operating conditions. For example, the filter unit 134 and the priority processing unit 135 illustrated in
A program can be created that is obtained by describing the measure presentation process performed by the measure presentation device according to the embodiments in a language that can be executed by a computer. In this case, the computer executes the program to obtain the same effects as those of the embodiments. Furthermore, the same measure presentation process as that of the embodiments may be realized by recording the program in a computer-readable recording medium and making the computer read and execute the program recorded in the recording medium.
The CPU 1010 reads the program recorded in the recording medium 1100 via the reader 1050 and then executes the program to realize the measure presentation process. As an example, the recording medium 1100 includes an optical disc, a flexible disk, CD-ROM, a hard disk, and the like. The program may be introduced into the computer 1000 via the network 1200. At this time, the network 1200 may be a wireless network or a wired network.
As described above, according to an aspect of the present invention, effective measure against the failure of the monitoring target device can be presented.
All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the invention and the concepts contributed by the inventor to furthering the art, and are to be construed as being without limitation to such specifically recited examples and conditions, nor does the organization of such examples in the specification relate to a showing of the superiority and inferiority of the invention. Although the embodiments of the present invention have been described in detail, it should be understood that the various changes, substitutions, and alterations could be made hereto without departing from the spirit and scope of the invention.
Number | Date | Country | Kind |
---|---|---|---|
2010-212166 | Sep 2010 | JP | national |