This U.S. patent application claims priority under 35 U.S.C. § 119 to: Indian Patent Application No. 201721028875, filed on 14 Aug. 2017. The entire contents of the aforementioned application are incorporated herein by reference.
The embodiments herein generally relate to data classification and more particularly to systems and methods for synthetic rare class generation by preserving morphological identity for facilitating data classification.
Data-driven computational method is a challenging task in a scenario wherein rare class examples are scarce. For instance, examples or training datasets of disease class is very less in number compared to examples or training datasets of normal class. Again, fraud credit card events available for a certain type of transaction is very less in number compared to normal transaction events. Existing supervised learning methods perform poorly when one of the class examples is rare in number.
Embodiments of the present disclosure present technological improvements as solutions to one or more of the above-mentioned technical problems recognized by the inventors in conventional systems.
In an aspect, there is provided a processor implemented method comprising: analyzing a labeled abundant training dataset and labeled rare class training dataset; generating an extended synthetic rare class super dataset based on the analysis using an extended oversampling method; and extracting a subset of the extended synthetic rare class super dataset to obtain a synthetic rare class dataset by maximizing similarity and diversity in the synthetic rare class dataset, wherein maximizing similarity ensures maximizing mutual information and maximizing diversity ensures minimum redundancy in the synthetic rare class dataset such that morphological identity of the synthetic rare class dataset is preserved with respect to the labeled rare class training dataset while maintaining diversity.
In another aspect, there is provided a system comprising: one or more data storage devices operatively coupled to the one or more processors and configured to store instructions configured for execution by the one or more processors to: analyze a labeled abundant training dataset and labeled rare class training dataset; generate an extended synthetic rare class super dataset based on the analysis using an extended oversampling method; and extract a subset of the extended synthetic rare class super dataset to obtain a synthetic rare class dataset by maximizing similarity and diversity in the synthetic rare class dataset, wherein maximizing similarity ensures maximizing mutual information and maximizing diversity ensures minimum redundancy in the synthetic rare class dataset such that morphological identity of the synthetic rare class dataset is preserved with respect to the labeled rare class training dataset while maintaining diversity.
In yet another aspect, there is provided a computer program product comprising a non-transitory computer readable medium having a computer readable program embodied therein, wherein the computer readable program, when executed on a computing device, causes the computing device to: analyze a labeled abundant training dataset and labeled rare class training dataset; generate an extended synthetic rare class super dataset based on the analysis using an extended oversampling method; and extract a subset of the extended synthetic rare class super dataset to obtain a synthetic rare class dataset by maximizing similarity and diversity in the synthetic rare class dataset, wherein maximizing similarity ensures maximizing mutual information and maximizing diversity ensures minimum redundancy in the synthetic rare class dataset such that morphological identity of the synthetic rare class dataset is preserved with respect to the labeled rare class training dataset while maintaining diversity.
In an embodiment of the present disclosure, the one or more hardware processors are further configured to perform extended oversampling method based on a Markov chain model.
In an embodiment of the present disclosure, the labeled abundant training dataset comprises labeled non-anomalous examples and the rare class training dataset comprises labeled anomalous examples.
In an embodiment of the present disclosure, the one or more hardware processors are further configured to perform the step of extracting a subset of the extended synthetic rare class super dataset to obtain a synthetic rare class dataset by: determining a similarity function pertaining to the extended synthetic rare class super dataset to obtain an extended rare class similar dataset; generating a ranked extended rare class similar dataset by ranking elements of the extended rare class similar dataset based on a similarity index associated thereof and sorting in descending order; and determining a diversity function for the ranked extended rare class similar dataset to obtain the synthetic rare class dataset with elements that are top ranked in the ranked extended rare class similar dataset and satisfies the diversity function.
In an embodiment of the present disclosure, the diversity function is based on l-diversity.
In an embodiment of the present disclosure, the synthetic rare class dataset is independent of dimensionality and is signal space rare class dataset.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the embodiments of the present disclosure, as claimed.
The embodiments herein will be better understood from the following detailed description with reference to the drawings, in which:
It should be appreciated by those skilled in the art that any block diagram herein represent conceptual views of illustrative systems embodying the principles of the present subject matter. Similarly, it will be appreciated that any flow charts, flow diagrams, state transition diagrams, pseudo code, and the like represent various processes which may be substantially represented in computer readable medium and so executed by a computing device or processor, whether or not such computing device or processor is explicitly shown.
The embodiments herein and the various features and advantageous details thereof are explained more fully with reference to the non-limiting embodiments that are illustrated in the accompanying drawings and detailed in the following description. The examples used herein are intended merely to facilitate an understanding of ways in which the embodiments herein may be practiced and to further enable those skilled in the art to practice the embodiments herein. Accordingly, the examples should not be construed as limiting the scope of the embodiments herein.
The words “comprising,” “having,” “containing,” and “including,” and other forms thereof, are intended to be equivalent in meaning and be open ended in that an item or items following any one of these words is not meant to be an exhaustive listing of such item or items, or meant to be limited to only the listed item or items.
It must also be noted that as used herein and in the appended claims, the singular forms “a,” “an,” and “the” include plural references unless the context clearly dictates otherwise. Although any systems and methods similar or equivalent to those described herein can be used in the practice or testing of embodiments of the present disclosure, the preferred, systems and methods are now described.
Some embodiments of this disclosure, illustrating all its features, will now be discussed in detail. The disclosed embodiments are merely exemplary of the disclosure, which may be embodied in various forms.
Before setting forth the detailed explanation, it is noted that all of the discussion below, regardless of the particular implementation being described, is exemplary in nature, rather than limiting.
Clinical decision making in data-driven computational methods is a challenging task due to scarcity of negative examples. The main drawback of prior-art is that simple over-sampling of available rare class examples are performed to generate synthetic rare class, which does not ensure diversity in the generated examples. Also, prior-art does not consider preserving morphological identities between the available rare class examples and generated rare class examples, thereby ignoring balancing of performance of the learning method. Systems and methods of the present disclosure ensure diversity (by not merely cloning as in the prior art) in the generated rare class examples while preserving morphological identity to overcome the class imbalance issue of the prior art.
Referring now to the drawings, and more particularly to
The I/O interface device(s) 106 can include a variety of software and hardware interfaces, for example, a web interface, a graphical user interface, and the like and can facilitate multiple communications within a wide variety of networks N/W and protocol types, including wired networks, for example, LAN, cable, etc., and wireless networks, such as WLAN, cellular, or satellite. In an embodiment, the I/O interface device(s) can include one or more ports for connecting a number of devices to one another or to another server.
The memory 102 may include any computer-readable medium known in the art including, for example, volatile memory, such as static random access memory (SRAM) and dynamic random access memory (DRAM), and/or non-volatile memory, such as read only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes. In an embodiment, one or more modules (not shown) of the system 100 can be stored in the memory 102.
In an embodiment, the system 100 comprises one or more data storage devices or memory 102 operatively coupled to the one or more processors 104 and is configured to store instructions configured for execution of steps of the method 200 by the one or more processors 104.
+
−
−−−
−−
In accordance with an embodiment of the present disclosure, the one or more processors 104 are configured to analyze, at step 202, a labeled abundant training dataset (+) and labeled rare class training dataset (−). In the context of the present disclosure, the expression “abundant training dataset” refers to a dataset for a class that is available in large numbers. For instance in case of a transaction scenario, positive training dataset in the form of normal transaction events may be available in large numbers. On the contrary, negative training dataset in the form of fraud credit card events may be available in small numbers only and may be referred to as “rare class training dataset”. Again in a clinical decision making scenario, disease class examples may be referred to as “rare class training dataset” while normal class examples may be referred to as “abundant training dataset”. Accordingly, in an embodiment, the labeled abundant training dataset may refer to labeled non-anomalous examples and the rare class training dataset may refer to labeled anomalous examples. Although this is a generally observed scenario, it may be true otherwise.
In accordance with an embodiment of the present disclosure, the one or more processors 104 are configured to generate, at step 204, an extended synthetic rare class super dataset (−−−) based on the analysis using an extended oversampling method. As explained above, a critical problem in anomaly detection is the limited availability of labeled rare class training dataset, which in an embodiment may be labeled anomalous examples. Let Π, π be the cardinality of the labeled abundant training dataset (+) and the labeled rare class training dataset (−) respectively and Π>>π.
Let =(+, −),
+={xi+}i=1Π, −={xi−}i=1π, where xi+, xi−∈d, where d represents training instances.
At step 204, a generation function generates the extended synthetic rare class super dataset (−−−) represented
One example of generation function is that of permutated data generation in Markov chain model, an extended oversampling method. Alternatively, any deterministic model with a known function may be employed. Given the labeled rare class training dataset (−−), some predicted number of states and associated state transition probabilities, the extended synthetic rare class super dataset X−−−={xi−−−}i=1π++ is generated, where length of [length({xi−−−}i=1π++), ∀i]≤median(−)∓3 σ(−). Let the cardinality Π of the labeled abundant training dataset (+) be 500 and the cardinality π of the labeled rare class training dataset be −={xi−}i=1π=20. In accordance with the present disclosure, the extended synthetic rare class super dataset (−−−) of 10000 instances is firstly generated by the generation function . From (−−−), the synthetic rare class dataset −−={xi−−}i=1π having 500 examples are extracted. Here, in the exemplary embodiment, Π=500, π=20 and Π++=10000.
In physiological signal space, typically a noisy signal consists of four segments: clean segment, motion artifact, random noise and power line interference segment, which correspond to measurement, instrumentation and interference plane respectively. In an embodiment, a Markov model based synthetic signal generation is provided, wherein the Markov model provides a systematic and stochastic method to model the time varying signals. In the Markov model, a future state only depends upon current states, not on predecessor states. This assumption and property makes the Markov model best suited for the generation of noisy/anomalous synthetic data or the rare class training dataset.
A stochastic process {Xn} is called a Markov chain if for all times n≥0 and all states i0, i1, . . . j∈S.
wherein Pij denotes a probability of moving from one state to another state, subject to ΣPij=1 and is known as one state Markov chain.
In accordance with an embodiment, generating the extended synthetic rare class super dataset (−−−) using the Markov model may be represented as given below.
Input:
It may be noted that the method described herein above, imposes restriction on the signal length (step 4) and arbitrary length is not permitted as in the case of the art.
In accordance with an embodiment of the present disclosure, the one or more processors 104 are configured to extract, at step 206, a subset of the extended synthetic rare class super dataset (−−−) to obtain the synthetic rare class dataset (−−) by maximizing similarity and diversity in the synthetic rare class dataset, wherein maximizing similarity ensures maximizing mutual information (explained hereinafter) and maximizing diversity ensures minimum redundancy in the synthetic rare class dataset such that morphological identity of the synthetic rare class dataset is preserved with respect to the labeled rare class training dataset while maintaining diversity. In the context of the present disclosure, morphological identity refers to characteristics associated with the available rare class dataset that are preserved in the synthetic rare class dataset while preserving diversity.
Dissimilarity property with + may be represented as given below.
((−),(−),(+)) (1)
with the assumption that the generated synthetic rare class dataset (−−) would be similar yet not redundant with the labeled rare class training dataset (−) but distinct from the labeled abundant training dataset (+). In accordance with the present disclosure, it is further assumed that the labeled abundant training dataset (+) and the labeled rare class training dataset (−) are independent and subsequently, similarity among − and dissimilarity between (+) and (−) are equivalent. For simplicity of explanation, it is assumed that satisfying similarity is practically sufficient to satisfying dissimilarity and accordingly, the condition may be omitted from further consideration. The problem addressed in the present disclosure is therefore to generate the synthetic rare class dataset −−={xi−−}i=1π from the labeled rare class training dataset −={xi−}i=1π, where Π>>π. Let, the extended synthetic rare class super dataset −−−={xi−−−}i=1π++, Π++>>Π be the universe of the synthetic rare class dataset (−−) generated. The problem is to find (−−) from the universe (−−−) such that equation (1) is satisfied.
In an embodiment, the step of extracting a subset of the extended synthetic rare class super dataset (−−−) to obtain a synthetic rare class dataset (−−) comprises firstly determining a similarity function pertaining to the extended synthetic rare class super dataset (−−−) to obtain an extended rare class similar dataset and then generating a ranked extended rare class similar dataset by ranking elements of the extended rare class similar dataset based on a similarity index associated thereof and sorting in descending order. In accordance with the present disclosure, one example of the similarity function may be constructed as: Find mutual information
for each of −−− with each of − that spawns Π++×π in (−−−;−). There are π number of (xi−−−; −) for each i. Then, find the cluster centroid i−−− that contains higher number of cluster elements when performing k-means (k=2) clustering on (xi−−−;−). For example, let there be 20 number of labeled rare class training dataset (−), π=20: −={xi−}i=1π=20. The generation function generates 10000 extended synthetic rare class super dataset, where: −−−={xi−−−}i=1π=10000, Π++×π=200000. For each computed (xi−−−; −), ∀i={1, 2, . . . , π=20} total 20 mutual information values for each of the generated: −−−={xi−−−}i=1π=10000 are 2-means clusterd. The cluster centroid i−−− that contains higher number of cluster elements of each (xi−−−;−) set is marked as the similarity index of i−−− on −. i−−−s are ranked in descending order sorting of the similarity indices. Let, the sorted order of −−− be −−−similar.
After the ranked extended rare class similar dataset is generated, a diversity function is determined for the ranked extended rare class similar dataset to obtain the synthetic rare class dataset with elements that are top ranked in the ranked extended rare class similar dataset and satisfies the diversity function. Accordingly, the diversity function is determined on −−−similar and −−−similar+diversed is identified that are both ranked high in −−−similar and significantly diverse. Thus, −−−similar+diversed⊆−−−similar and −−−similar+diversed=−−={xi−−}i=1π, the generated synthetic rare class dataset. One example of the diversity function may be constructed as: Find X−−={xi−−}i=1π set from −−−similar which are l-diverse (Machanavajjhala; 2007) and X−− contains the top-ranked in −−−similar in each of the l-diverse groups. Definition l-diversity: A group is l-diverse if each of the l different group 's entropy≥log2 l:
After constructing l-diverse groups from −−−similar, where l=Π1/4, −−−similar+diversed {xi−−}i=1π is formed taking the top ranked corresponds −−−similar from each of the l groups such that total number of elements chosen is Π.
Thus in accordance with the present disclosure, systems and methods of the present disclosure facilitate addressing the class imbalance problem in applications such as identifying noisy phonocardiogram (PCG) signals. The present disclosure deals with signal (time-series) space rare-class dataset as compared to prior art that deal with feature space rare-class dataset. The subset of the extended synthetic rare class super dataset is extracted without reducing the dimensionality (feature space) of the dataset, thereby making the step of obtaining the synthetic rare class dataset independent of the dimensionality properties of the datasets and non-parametric.
The written description describes the subject matter herein to enable any person skilled in the art to make and use the embodiments of the present disclosure. The scope of the subject matter embodiments defined here may include other modifications that occur to those skilled in the art. Such other modifications are intended to be within the scope if they have similar elements that do not differ from the literal language of the claims or if they include equivalent elements with insubstantial differences from the literal language.
The scope of the subject matter embodiments defined here may include other modifications that occur to those skilled in the art. Such other modifications are intended to be within the scope if they have similar elements that do not differ from the literal language of the claims or if they include equivalent elements with insubstantial differences from the literal language.
It is, however to be understood that the scope of the protection is extended to such a program and in addition to a computer-readable means having a message therein; such computer-readable storage means contain program-code means for implementation of one or more steps of the method, when the program runs on a server or mobile device or any suitable programmable device. The hardware device can be any kind of device which can be programmed including e.g. any kind of computer like a server or a personal computer, or the like, or any combination thereof. The device may also include means which could be e.g. hardware means like e.g. an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), or a combination of hardware and software means, e.g. an ASIC and an FPGA, or at least one microprocessor and at least one memory with software modules located therein. Thus, the means can include both hardware means and software means. The method embodiments described herein could be implemented in hardware and software. The device may also include software means. Alternatively, the embodiments of the present disclosure may be implemented on different hardware devices, e.g. using a plurality of CPUs.
The embodiments herein can comprise hardware and software elements. The embodiments that are implemented in software include but are not limited to, firmware, resident software, microcode, etc. The functions performed by various modules comprising the system of the present disclosure and described herein may be implemented in other modules or combinations of other modules. For the purposes of this description, a computer-usable or computer readable medium can be any apparatus that can comprise, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. The various modules described herein may be implemented as software and/or hardware modules and may be stored in any type of non-transitory computer readable medium or other storage device. Some non-limiting examples of non-transitory computer-readable media include CDs, DVDs, BLU-RAY, flash memory, and hard disk drives.
Further, although process steps, method steps, techniques or the like may be described in a sequential order, such processes, methods and techniques may be configured to work in alternate orders. In other words, any sequence or order of steps that may be described does not necessarily indicate a requirement that the steps be performed in that order. The steps of processes described herein may be performed in any order practical. Further, some steps may be performed simultaneously.
The preceding description has been presented with reference to various embodiments. Persons having ordinary skill in the art and technology to which this application pertains will appreciate that alterations and changes in the described structures and methods of operation can be practiced without meaningfully departing from the principle, spirit and scope.
Number | Date | Country | Kind |
---|---|---|---|
201721028875 | Aug 2017 | IN | national |
Entry |
---|
Cao, H. et al. (2013). “Integrated Oversampling for Imbalanced Time Series Classification,” IEEE Transactions on Knowledge and Data Engineering, vol. 25, issue 12; 14 pages. |
Cao, H. et al. (2011). “SPO: Structure Preserving Oversampling for Imbalanced Time Series Classification,” ICDM 2011 11th IEEE International Conference on Data Mining, Vancouver, Canada; pp. 1008-1013. |
Das, B. et al. (2015). “RACOG and wRACOG: Two Probabilistic Oversampling Techniques,” IEEE Transactions on Knowledge and Data Engineering, vol. 27, No. 1; pp. 222-234. |
Dangi, A. P. et al. (2015). “Privacy Preservation Measurement through Diversity and Anonymity Using Closeness,” JETIR, vol. 2, issue 11; pp. 35-40. |
Tang, S. et al. (2008). “The Generation Mechanism of Synthetic Minority Class Examples,” 2008 International Conference on Information Technology and Applications in Biomedicine, Shenzhen, China; pp. 444-447. |
Number | Date | Country | |
---|---|---|---|
20190050673 A1 | Feb 2019 | US |