The present disclosure relates to the field of data processing, and more particularly to an apparatus and method for detecting an anomaly in a dataset by using two or more anomaly detection algorithms, as well as to a corresponding computer program product.
Anomaly detection refers to identifying data items that do not conform to an expected behavior pattern or do not correspond to other (e.g., normal) data items in a dataset. Anomaly detection algorithms are being currently used for a variety of purposes, such, for example, as fraud detection in stock markets, malicious activity detection in computer or communication networks, malfunction detection in software or hardware, disease detection in medicine, etc.
Anomalies may be conveniently divided into those which are relevant to an event of interest, and those which are irrelevant to the event of interest. The latter anomalies, also known as spurious anomalies, may have a negative impact on user experience, resulting in false alarms, and therefore have to be excluded from consideration when searching for the former anomalies in the dataset. To this end, a particular anomaly detection algorithm may be applied to calculate a specified number of top anomalies and visualize the top anomalies in the descending order of anomaly importance, thereby allowing a user to manually filter out the spurious anomalies. However, such manual work is time consuming and requires solid knowledge in a certain usage domain.
To reduce a false alarm rate, two or more anomaly detection algorithms may be used in concert with each other to provide an average anomaly score for each data item in a dataset of interest. As for the manual work, it may be avoided, at least partly, by combining the anomaly detection algorithms with conventional machine learning techniques, such as unsupervised learning and supervised learning. In the meantime, all known anomaly detection systems do not provide a sufficient accuracy, and still rely on user-defined rules which may vary depending on a certain usage domain.
Therefore, there is still a need for a new solution that allows mitigating or even eliminating the above-mentioned drawbacks peculiar to the prior approaches.
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the detailed description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
It is an object of the present disclosure to provide a technical solution for improving the accuracy of anomaly detection, and minimizing user involvement.
The object above is achieved by the features of the independent claims in the appended claims. Further embodiments and examples are apparent from the dependent claims, the detailed description and the accompanying drawings.
According to a first aspect, an apparatus for detecting an anomaly in a dataset is provided. The apparatus comprises at least one processor, and a storage coupled to the at least one processor and storing executable instructions. The instructions, when executed, cause the at least one processor to receive the dataset comprising multiple data items among which at least one data item is anomalous, and select at least two anomaly detection algorithms. The at least one processor is then instructed, by using each of the at least two anomaly detection algorithms, to: calculate an anomaly score for each of the data items; based on the anomaly scores, obtain a partial ranking of the data items, the partial ranking causing the data items to be divided into subsets each corresponding to a different interval of intermediate ranks; based on the partial ranking, select a probabilistic model describing the intermediate ranks of the data items in each subset; and based on the probabilistic model, assign a degree of belief to the intermediate rank of each of the data items in each subset. The at least one processor is next instructed to obtain a total degree of belief for the intermediate rank of each of the data items by combining the degrees of belief obtained, for this intermediate rank, by using all of the at least two anomaly detection algorithms in accordance with a predefined combination rule. After that, the at least one processor is instructed to convert the total degrees of belief for the intermediate ranks of the data items to a probability distribution function describing expected ranks of the data items. The at least one processor is further instructed to sort the data items according to the expected ranks of the data items, and find the at least one anomalous data item among the sorted data items. By doing so, it is feasible to detect anomalies in more accurate and robust manner, without having to use expert rules specific to a particular knowledge domain.
In an embodiment form of the first aspect, the at least one processor is configured to select the at least two anomaly detection algorithms based on a usage domain which the data items belong to. This provides flexibility in use because the apparatus according to the first aspect can equally operate in different usage domains.
In a further embodiment form of the first aspect, each of the at least two anomaly detection algorithms is provided with a different weight coefficient, and the at least one processor is further configured to assign the degree of belief based on the probabilistic model in concert with the weight coefficient of the anomaly detection algorithm. By assigning the different weight coefficients to the anomaly detection algorithms, one can obtain a more objective degree of belief for the intermediate rank of each of the data items in each subset.
In a further embodiment form of the first aspect, the at least two anomaly detection algorithms are unsupervised learning based anomaly detection algorithms, and the different weight coefficients of the at least two anomaly detection algorithms are specified based on user preferences such that the sum of the weight coefficients is equal to 1. By doing so, it is feasible to minimize the user involvement in anomaly detection, i.e. to make the apparatus according to the first aspect more automatic.
In a further embodiment form of the first aspect, the at least two anomaly detection algorithms are supervised learning based anomaly detection algorithms, and the weight coefficients of the at least two anomaly detection algorithms are adjusted by using a pre-arranged training set comprising different previous datasets and target rankings each corresponding to one of the previous datasets. By doing so, it is feasible to minimize the user involvement in anomaly detection.
In a further embodiment form of the first aspect, when the supervised learning based anomaly detection algorithms are used, the weight coefficients of the at least two anomaly detection algorithms are further adjusted based on the Kendall tau distance. The Kendall tau distance serves a measure of distance between the combined partial rankings obtained by the at least two anomaly detection algorithms and respective one of the target rankings from the training set. With the Kendall tau distance, the contribution of each anomaly detection algorithm is adjusted more efficiently.
In a further embodiment form of the first aspect, the subsets obtained based on the partial ranking of the data items comprises at least two first subsets each comprising the data items having the same anomaly scores. This allows the data items to be separated into multiple anomaly classes in a simple and efficient manner.
In a further embodiment form of the first aspect, the intervals of intermediate ranks of the at least two first subsets are non-overlapping. This allows making the separation of the data items into the anomaly classes more explicit.
In a further embodiment form of the first aspect, the subsets obtained based on the partial ranking of the data items further comprises a second subset comprising the data items falling outside of the at least two first subsets, and the at least one processor is further configured to select the probabilistic model taking into account the second subset. This makes the apparatus according to the first aspect more flexible in the sense that it can take account of the different anomaly classes when detecting one or more anomalies in the dataset.
In a further embodiment form of the first aspect, the data items of the second subset may be erroneously missed data items or data items having the anomaly scores differing from those of the data items belonging to the at least two first subsets. By doing so, it is feasible to provide the proper accuracy and robustness of anomaly detection even if there are data items mistakenly unranked or missed during the operation of the apparatus according to the first aspect.
In a further embodiment form of the first aspect, the interval of intermediate ranks of the second subset covers the intervals of intermediate ranks of the at least two first subsets. This means that the apparatus according to the first aspect is able to operate successfully even if the intermediate ranks of some data items are dispersed accidentally and arbitrarily in the whole interval of intermediate ranks.
In a further embodiment form of the first aspect, the predefined combination rule comprises the Dempster's rule of combination. This allows combining the degrees of belief entirely based on a statistical fusion approach rather than on the expert rules, thereby minimizing the user involvement to a greater extent and making the apparatus according to the first aspect easy to use.
In a further embodiment form of the first aspect, the at least two anomaly detection algorithms comprises any combination of the following algorithms: a nearest neighbor-based anomaly detection algorithm, a clustering-based anomaly detection algorithm, a statistical anomaly detection algorithm, a subspace-based anomaly detection algorithm, and a classifier-based anomaly detection algorithm. This provides additional flexibility in use because each of the algorithms listed above gives advantages when being applied in a certain usage domain.
In a further embodiment form of the first aspect, the degree of belief for the intermediate rank comprises a basic belief assignment. This allows increasing the accuracy of anomaly detection to a greater extent.
In a further embodiment form of the first aspect, the at least one processor is further configured to convert the total degrees of belief for the intermediate ranks of the data items to the probability distribution function by using a pignistic transformation, and the probability distribution function is a pignistic probability function. This allows increasing the accuracy of anomaly detection to a greater extent.
In a further embodiment form of the first aspect, the data items comprise network flow data, and the at least one anomalous data item relates to abnormal network flow behavior. This allows one to quickly detect and respond to a malicious activity or a device fault in a computer network.
According to a second aspect, a method for detecting an anomaly in a dataset is provided. The method is performed as follows. The dataset is received, which comprises multiple data items with at least one anomalous data item. Next, at least two anomaly detection algorithms are selected. By using each of the at least two anomaly detection algorithms, the following steps are performed: calculating an anomaly score for each of the data items; based on the anomaly scores, obtaining a partial ranking of the data items, the partial ranking causing the data items to be divided into subsets each corresponding to a different interval of intermediate ranks; based on the partial ranking, selecting a probabilistic model describing the intermediate ranks of the data items in each subset; and based on the probabilistic model, assigning a degree of belief to the intermediate rank of each of the data items in each subset. After that, a total degree of belief for the intermediate rank of each of the data items is obtained by combining the degrees of belief obtained, for this intermediate rank, by using all of the at least two anomaly detection algorithms in accordance with a predefined combination rule. Further, the total degrees of belief for the intermediate ranks of the data items are converted to a probability distribution function describing expected ranks of the data items. The data items are then sorted according to the expected ranks of the data items, and the at least one anomalous data item is eventually found among the sorted data items. By doing so, it is feasible to detect anomalies in more accurate and robust manner, without having to use expert rules specific to a particular knowledge domain.
According to a third aspect, a computer program product comprising a computer-readable storage medium storing a computer program is provided. The computer program, when executed by at least one processor, causes the at least one processor to perform the method according to the second aspect. Thus, the method according to the second aspect can be embodied in the form of the computer program, thereby providing flexibility in use thereof.
Other features and advantages of the present disclosure will be apparent upon reading the following detailed description and reviewing the accompanying drawings.
The essence of the present disclosure is explained below with reference to the accompanying drawings in which:
Various embodiments of the present disclosure are further described in more detail with reference to the accompanying drawings. However, the present disclosure can be embodied in many other forms and should not be construed as limited to any certain structure or function disclosed in the following description. In contrast, these embodiments are provided to make the description detailed and complete.
According to the detailed description, it will be apparent to ones skilled in the art that the scope of the present disclosure encompasses any embodiment disclosed herein, irrespective of whether this embodiment is implemented independently or in concert with any other embodiment. For example, the apparatus and method disclosed herein can be implemented in practice by using any numbers of the embodiments provided herein. Furthermore, it should be understood that any embodiment can be implemented using one or more of the elements or steps presented in the appended claims.
As used herein, the term “anomaly” and its derivatives, such as “anomalous”, “abnormal”, etc., refer to something that deviates from what is standard, normal, or expected. In particular, the term “anomalous data item” also used herein means a data item in a dataset, which falls outside the ranges of the standard deviation of data items in the dataset. An anomaly may be characterized by two or more neighboring or close anomalous data items, and is called a collective anomaly in this case. The anomaly may relate to an event of interest, i.e. a problem to be detected and solved, or be irrelevant to the event of interest. In the latter case, the anomaly is called a spurious anomaly. One example of the anomaly includes a suspiciously large (i.e., non-typical) network flow which may be caused by malicious software. Although references are hereby made to network flow data, it should be apparent to those skilled in the art that this is done only by way of example but not limitation. In other words, the embodiments disclosed herein may be equally applied in other usage domains where anomaly detection is required, such, for example, as the detection of fraudulent pump and dump on stocks, the detection of excessive scores mistakenly issued in figure skating or other kinds of sports, etc.
The term “combination rule” used herein refers to an analytical rule or condition that may be applied to output data of multiple data sources to integrate the output data into more consistent, accurate, and useful information than the output data of any individual data source. The data sources are presented herein as anomaly detection algorithms, and their output data to be integrated or combined comprise degrees of beliefs. One example of the combination rule includes the Dempster's rule of combination.
The term “degree of belief” used herein refers to a mathematical object called a belief function that is used in the theory of belief functions, also known as the evidence theory or Dempster-Shafer theory. The theory of belief functions allows one to combine evidence from different data sources to arrive at a degree of belief that takes into account all the available evidence. As will be shown later, the degree of belief is applied herein to intermediate ranks of data items obtained by using the anomaly detection algorithms. One example of degrees of belief are basic belief assignments (bbas) which will be discussed later in context of the embodiments disclosed herein. By definition, assuming that θ represents a set of hypotheses H (for example, all possible states of a system under consideration), which is called a frame of discernment, the basic belief assignment represents a function assigning a belief mass m to each data element of a power set 2θ which is a set of all subsets of θ, including an empty set Ø, such that m: 2θ→[0,1]. The basic belief assignment has the following two main properties:
where the subsets Hn of θ are called focal elements of m (non-zero masses).
As used herein, the term “rank” refers to a numerical parameter used to classify data items into different anomaly classes. Each anomaly class is represented by a certain interval of ranks. An intermediate rank discussed herein is obtained by using any one anomaly detection algorithm. An expected rank also discussed herein is a more valid rank resulted from using the intermediate ranks obtained by multiple anomaly detection algorithms.
Generally speaking, the absolute values of the anomaly scores themselves are meaningless—they are used solely for establishing the ordering relationship among the data items. Therefore, the accuracy of anomaly detection is low in cases of using only one anomaly detection algorithm.
The aspects of the present disclosure discussed below take into account the above-mentioned drawbacks, and are directed to improving the accuracy and robustness of anomaly detection, particularly, in the network flow data.
The storage 302 may be implemented as a volatile or nonvolatile memory used in modern electronic computing machines. Examples of the nonvolatile memory include Read-Only Memory (ROM), flash memory, ferroelectric Random-Access Memory (RAM), Programmable ROM (PROM), Electrically Erasable PROM (EEPROM), solid state drive (SSD), magnetic disk storage (such as hard drives and magnetic tapes), optical disc storage (such as a compact disc (CD), digital vide disc (DVD) and Blu-ray discs), etc. As for the volatile memory, examples thereof include Dynamic RAM, Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDR SDRAM), Static RAM, etc.
Relative to the processor 304, it may be implemented as a central processing unit (CPU), general-purpose processor, single-purpose processor, microcontroller, microprocessor, application specific integrated circuit (ASIC), field programmable gate array (FPGA), digital signal processor (DSP), complex programmable logic device, etc. It should be also noted that the processor 304 may be implemented as any combination of one or more of the aforesaid. As an example, the processor 304 may be a combination of two or more microprocessors.
The executable instructions 306 stored in the storage 302 may be configured as a computer executable code which causes the processor 304 to perform the aspects of the present disclosure. The computer executable code for carrying out operations or steps for the aspects of the present disclosure may be written in any combination of one or more programming languages, such as Java, C++ or the like. In some examples, the computer executable code may be in the form of a high level language or in a pre-compiled form, and be generated by an interpreter (also pre-stored in the storage 302) on the fly.
Being caused with the executable instructions 306, the processor 304 first receives the dataset comprising multiple data items among which the at least one data item is anomalous, as noted above. After that, the processor 304 selects at least two anomaly detection algorithms based on the usage domain which the data items belong to. The reason for using two or more anomaly detection algorithms is a synergic effect consisting in that the accuracy of anomaly detection provided by the two or more anomaly detection algorithms is higher than that provided by any single anomaly detection algorithm. More specifically, if a user of the apparatus 300 is absolutely sure that one of the anomaly detection algorithms provides 100% accuracy, he or she will not combine it with any other of the anomaly detection algorithms. However, in practice, any anomaly detection algorithm is prone to errors, which forces the user to decide which of the anomaly detection algorithms has to be selected and under what circumstances. That is why the aggregated accuracy provided by the two or more anomaly detection algorithms is more preferable and useful in the process of anomaly detection.
In one embodiment, the at least two anomaly detection algorithms comprise any combination of the following algorithms: a nearest neighbor-based anomaly detection algorithm, a clustering-based anomaly detection algorithm, a statistical anomaly detection algorithm, a subspace-based anomaly detection algorithm, and a classifier-based anomaly detection algorithm. Some examples of such anomaly detection algorithms are described by Goldstein M. and Uchida S. in their work “A Comparative Evaluation of Unsupervised Anomaly Detection Algorithms for Multivariate Data”, PLoS ONE 11(4): e0152173 (2016). Moreover, the at least two anomaly detection algorithms may be unsupervised or supervised learning based anomaly detection algorithms, thereby making the apparatus 300 more automatic and flexible in use. As should apparent to those skilled in the art, unsupervised or supervised learning may involve using neural networks, decision trees, and/or other artificial intelligence techniques, depending on particular applications.
Once the at least two anomaly detection algorithms are selected, the processor 304 uses them to calculate the anomaly score for each of the data items. The anomaly scores are then used by the processor 304 to obtain a partial ranking of the data items. The partial ranking causes the data items to be divided into subsets each corresponding to a different interval of intermediate ranks, as schematically shown in
By using the partial ranking, the processor 304 further selects a probabilistic model describing the intermediate ranks of the data items in each of the subsets. In general, the probabilistic model defines a probability distribution of the intermediate ranks among the data items in each subset.
However, if there are not all the data items put in the non-overlapping subsets either mistakenly or due to the presence of the data items having the anomaly scores other than those of the data items put in the non-overlapping subsets, the uniform probability distributions for the non-overlapping subsets will be violated. This situation is schematically shown in
To calculate the probability distribution of the intermediate ranks in the subset of interest in the presence of the unranked data items, the processor 304 may be configured to perform the following procedure. At first, let us assume that, as a result of the partial ranking, there are an arbitrary number of ranked subsets (i.e., buckets), like the subsets 600a and 600b in
where B1 denotes the j-th subset, y and z are the left and right boundary data items, respectively, in the middle group, and x is the unranked data item;
Further, the processor 304 uses a pseudo code for computing the probability distribution P1 of the intermediate ranks of the data items in B1, which is given below as Algorithm 1. It is assumed that Pj is the |X|-component vector such that Pj(r)=Pr(rank(x)=r) for any x∈Bj and r∈{1, . . . , |X|}. By definition, Σr=1|X|Pj(r)=1.
In Algorithm 1, pdecomp is the probability of the decomposition of the unranked data items, which is defined by the parameters kmiddle, kbottom, ktop, the sign “←” is the value assignment operator, and the function Hyp( ) is the hypergeometric distribution. In particular, the function Hyp( ) describes the probability of obtaining the total number of k black circles in the sample of length n without replacement, starting out with N circles among which K circles are black. In other words,
where CKk is the binomial coefficient.
Thus, by using Algorithm 1, the processor 304 calculates the probability distribution Pj of the intermediate ranks of the data items in Bj in case of using each of the at least two anomaly detection algorithms. In other words, if the processor 304 uses L anomaly detection algorithms, it will be required for the processor 304 to calculate the probability distributions Pj(1), . . . , Pj(L) respectively, for the intermediate ranks of the data items in Bj.
When the probabilistic model, or, in other words, the probability distribution Pj, is calculated, the processor 304 further assigns, the based on Pj, a degree of belief to the intermediate rank of each of the data items in B1. Further, the degree of belief is exemplified by the basic belief assignment (bba). However, the degree of belief is not limited to the bba, and may be presented as any other belief functions specific to the Dempster-Shafer theory.
In one embodiment, the processor 304 is configured to provide each of the at least two anomaly detection algorithms with a different weight coefficient and assign the bba based on the probabilistic model in concert with the weight coefficient of the anomaly detection algorithm. This allows adjusting the contribution of each anomaly detection algorithm into the aggregated accuracy of anomaly detection.
In one embodiment, in case of the unsupervised learning based anomaly detection algorithms, the processor 304 is configured to specify the different weight coefficients of the at least two anomaly detection algorithms based on user preferences such that the sum of the weight coefficients is equal to 1, i.e. Σi=1wi=1, where L is the number of the anomaly detection algorithms used. This allows the user of the apparatus 300 to prioritize the anomaly detection algorithms according to his or her experience.
In another embodiment, in case of the supervised learning based anomaly detection algorithms, the processor 304 is configured to adjust the weight coefficients of the at least two anomaly detection algorithms by using a pre-arranged training set comprising different previous datasets and target rankings each corresponding to one of the previous datasets. The training set may be stored in the storage 302 in advance, i.e. before the operation of the apparatus 300. In this case, the processor 304 first searches for the previous dataset similar to that of interest, and then changes the weight coefficient of each anomaly detection algorithm until the partial ranking coincides with the target ranking of this previous dataset. The weight coefficients of the at least two anomaly detection algorithms may be further adjusted by the processor 304 based on the Kendall tau distance serving a measure of distance between the combined partial rankings obtained by the at least two anomaly detection algorithms and respective one of the target rankings from the training set. In this case, the Kendall tau distance, which exploits a probability distribution similar to Pj calculated earlier, for a pair of partial rankings σ and τ are expressed as follows (here the signs “∨” and “∧” represent the grouping and intersection signs, respectively):
and its normalized analogue is given by
Being governed by M training sets, the weight coefficient adaptation procedure strives to find non-negative weight coefficients w1, . . . , wL which minimize the following loss function:
and satisfy the condition Σl=1Lwl=1. Here σgr.truthi is the partial ranking that is known to be true for the data items in the i-th training set, τli is the partial ranking computed by the l-th anomaly detection algorithm for the data items in the i-th training set, w1τ1i+ . . . +wLτLi is the partial ranking obtained by the processor 304, i.e. by combining the partial rankings τ1i, . . . , τLi with the weight coefficients w1, . . . , wL.
Turning now back to the assignment of the bbas, it should be noted that the processor 304 may use Algorithm 2 for this purpose, which is given below and takes into account the weight coefficients of the anomaly detection algorithms.
In other words, by using Algorithm 2, the processor 304 considers the following frame of discernment 0={rank(x)=1, . . . , rank(x)=|X|} for each data item, and computes (|X|+1)-component bbas, with the components corresponding to the following outcomes rank(x)=1, . . . , rank(x)=|X|, rank(x)=Θ. The last outcome, i.e. rank(x)=Θ, means that x may have any intermediate rank. By construction, Σlml=1.
When the bbas for all the anomaly detection algorithms are obtained, the processor 304 then obtains a total degree of belief, i.e. a total bba, for the intermediate rank of each of the data items. To do this, the processor 304 combines the bbas obtained for the intermediate rank in accordance with a predefined combination rule. Algorithm 3 given below describes this operation, taking the Dempster's rule of combination as one example of the predefined combination rule.
In Algorithm 3, A, B, C are the indices that can take on any value from 1 to |X|+1, and m1,2, m1, and m2 are the vectors of length |X|+1, with m1, and m2 corresponding to the first and the second anomaly detection algorithms, respectively, the results of which are subjected to combination, and m1,2 being the result of this combination. Since the Dempster's rule of combination is both commutative and associative, it can combine all L bbas (according to the number of the anomaly detection algorithms) in a single total bba m.
After that, the processor 304 converts the total bbas for the intermediate ranks of the data items to a probability distribution function describing expected ranks of the data items. This may be done in one embodiment by using a pignistic transformation, and the probability distribution function is a pignistic probability function betP in such case. The pignistic transformation performed by the processor 304 is generalized below as Algorithm 4.
Next, the processor 304 computes the expected rank of each data item x∈X by using the pignistic probability betP and sorts all the data items in the dataset X by their expected ranks according to the following formula:
E[rank(x)]=Σr=1|X|r·betP(r).
Finally, the processor 304 finds the at least one anomalous data item among the sorted data items. Thus, by using the above-described procedure comprising Algorithms 1-4, the processor 304 is able to detect the anomaly of interest in the dataset, and even filter out the spurious anomalies if they are present in the dataset.
In one embodiment, the processor 304 may further convert the expected ranks to the partial ranking in the same way as the original anomaly scores are converted to the partial rankings but with the reverse order of the subsets because, by convention, the smaller ranks should correspond to the higher anomaly scores.
With reference to
The method 800 starts up in step 802, in which the dataset comprising at least one anomalous data item is received. As noted earlier, the dataset may relate to different usage domains. Once the dataset is received, the method proceeds to step 804, in which the at least two anomaly detection algorithms are selected based on the usage domain which the dataset belongs to. Further, steps 806-812 are carried out by using each of the at least two anomaly detection algorithms independently.
In particular, an anomaly score for each of the data items is calculated in the step 806. In the step 808, a partial ranking of the data items is obtained based on the anomaly scores. The partial ranking represents the division of the data items into subsets each corresponding to a different interval of intermediate ranks and, consequently, a different anomaly class. The examples of such subsets have been discussed above with reference to
Once the degrees of belief for each intermediate rank are obtained by using each of the at least two anomaly detection algorithms, the method 800 proceeds to step 814, in which the degrees of belief are combined in accordance with the combination rule to obtain a total degree of belief. This may be done by using Algorithm 3 discussed above, in which the combination rule is exemplified by the Dempster's rule of combination. Further, in step 816, the total degrees of belief for the intermediate ranks of the data items are converted to a probability distribution function describing expected ranks of the data items. Such conversion may be implemented by using the pignistic transformation described above with reference to Algorithm 4. After that, the data items are sorted, in step 818, according to the expected ranks of the data items. Finally, in step 820, the at least one anomalous data items is found among the sorted data items.
It should be noted that some approaches suggest an alternative solution for the same problem which is addressed by the method 800 using the Dempster's rule of combination. In particular, the alternative solution involves adopting a median rank aggregation to partial rankings. However, the median rank aggregation method provides a lower accuracy of anomaly detection compared to the accuracy of the method 800. This has been proved by numerical experiments, the results of which are shown in
Those skilled in the art should understand that each step of the method 800, or any combinations of the steps, can be implemented by various means, such as hardware, firmware, and/or software. As an example, one or more of the steps described above can be embodied by computer or processor executable instructions, data structures, program modules, and other suitable data representations. Furthermore, the computer executable instructions which embody the steps described above can be stored on a corresponding data carrier and executed by at least one processor like the processor 304 included in the apparatus 300. This data carrier can be implemented as any computer-readable storage medium configured to be readable by said at least one processor to execute the computer executable instructions. Such computer-readable storage media can include both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, the computer-readable media comprise media implemented in any method or technology suitable for storing information. In more detail, the practical examples of the computer-readable media include, but are not limited to information-delivery media, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, DVD, holographic media or other optical disc storage, magnetic tape, magnetic cassettes, magnetic disk storage, and other magnetic storage devices.
Although the exemplary embodiments are disclosed herein, it should be noted that any various changes and modifications could be made in these embodiments, without departing from the scope of legal protection which is defined by the appended claims. In the appended claims, the mention of elements in a singular form does not exclude the presence of the plurality of such elements, if not explicitly stated otherwise.
This application is a continuation of International Application No. PCT/CN2018/096425, filed on Jul. 20, 2018, the disclosure of which is hereby incorporated by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2018/096425 | Jul 2018 | US |
Child | 17152019 | US |