This is a National Phase Application under 35 USC 371 of PCT/EP2014/052494 filed Feb. 7, 2014 (published on Aug. 14, 2014 as WO 2014/122295); which claims priority to PCT/EP2013/052476 filed Feb. 7, 2013 (published on Aug. 14, 2014 as WO 2014/121839) and Great Britain Application No. 1302415.3 filed Feb. 12, 2013; all of which are incorporated by reference herein in their entirety.
The present invention relates to methods and systems for data cleaning and more particularly relates to methods and systems for repairing errors in attribute values in a database.
There are numerous known methods and systems for cleaning data in a database. The term “cleaning” is used herein to mean correcting or repairing errors in values or attribute values which are stored as information in a database.
The following examples illustrate the drawbacks of the state-of-the-art work in the area of data cleaning.
Consider a database D of travel records. The database is specified by the following schema:
Here a travel tuple specifies a person, identified by name, has travelled to conference (conf), held at the city of the country and its capital. Example instances of travel are shown in
The following four techniques may be used to detect and repair errors in the database of
(1) Integrity Constraints
A functional dependency (FD) is used to specify the consistency of travel data D as:
where φ1 asserts that country uniquely determines capital.
The FD φ1 detects that in the database in
Other constraints, such as conditional functional dependencies (CFDs) or denial constraints may also be introduced to detect various errors. However, these other constraints are also not able to repair data.
Using such integrity constraints, existing heuristic based approaches may choose any of the three values, Beijing, Shanghai, or Tokyo to update r1[capital]r3[capital].
(2) User Guided Repairs
It is known to clean data using repairs which are guided by users. Assuming that the three violations among tuples r1-r3 have been detected as in (1), a typical user guided repair raises a question to users such as: Which is the capital of China: Beijing, Shanghai, or Tokyo?
One can assume that the users pick Beijing as the capital of China. This corrects the erroneous value r2[capital], from Shanghai to Beijing. However, the error in r3 should be r3[country], which should be Japan instead of China. The response from the users is therefore not helpful to fix the error in r3. Worse still, the change prompted by the uses will introduce a new error as it changes r3[capital] from Tokyo to Beijing.
(3) Editing Rules
Editing rules can be used to capture and repair errors. Master data stores correct information about countries and their capitals. The schema of the master data is:
A master relationship between the attributes in the data in
A conventional editing rule ψ1 is defined on two relations (travel,cap) as:
The editing rule ψ1 states that: for a tuple r in the travel database of
For instance, to repair r2 in the database of
(4) Extract Transform Load (ETL) Rules
A typical task in an ETL rule is a lookup operation, assuming the presence of a dictionary (e.g., the master data Dm in
The above four repair examples illustrate the following problems with such conventional techniques:
There is therefore a need for improved data cleaning rules which seek to overcome the above problems.
According to one aspect of the present invention, there is provided, a method for cleaning data stored in a database, the method comprising providing a set of fixing rules, each fixing rule incorporating a set of attribute values that capture an error in a plurality of semantically related attribute values, and a deterministic correction which is operable to replace one of the set of attribute values with a correct attribute value to correct the error, wherein the method further comprises comparing at least two of the fixing rules with one another to check that the error correction carried out by one fixing rule is consistent with the error correction carried out by another fixing rule.
Preferably, the method comprises comparing all fixing rules in the set of fixing rules pairwise with one another.
Conveniently, the method comprises applying at least two of the fixing rules to a tuple of attribute values to check whether the at least two fixing rules apply different corrections to the tuple, thereby indicating that the at least two fixing rules are not consistent with one another.
Advantageously, the method comprises identifying a tuple of attribute values that satisfies two of the fixing rules and applying the two fixing rules to the tuple alternately in different orders to determine if the two fixing rules apply different error corrections to the tuple when the fixing rules are applied to the tuple in different orders, thereby indicating that the fixing rules are not consistent with one another.
Preferably, the method further comprises combining at least part of two inconsistent fixing rules with one another to produce one or more modified fixing rules which are consistent with one another.
Conveniently, the method comprises repeating the comparison between at least two of the fixing rules until the method identifies that all of the fixing rules in the set of fixing rules are consistent with one another.
Advantageously, the method comprises outputting at least two fixing rules that are not consistent with one another to a user so that the user can amend or delete at least one of the fixing rules to remove the inconsistency.
Preferably, the method further comprises applying at least one of the fixing rules to a plurality of tuples stored in a database to detect if at least one of the tuples comprises the respective set of attribute values that captures the error and, if the respective set of attribute values is detected, applying the deterministic correction to correct the error in the at least one tuple.
According to another aspect of the present invention, there is provided a method for providing a set of fixing rules, each fixing rule incorporating a set of attribute values that capture an error in a plurality of semantically related attribute values, and a deterministic correction which is operable to replace one of the set of attribute values with a correct attribute value to correct the error, wherein the method comprises applying at least one of the fixing rules to a plurality of tuples stored in a database to detect if at least one of the tuples comprises the respective set of attribute values that captures the error and, if the respective set of attribute values is detected, applying the deterministic correction to correct the error in the at least one tuple.
Preferably, the method comprises applying a plurality of the fixing rules to the tuples stored in the database, the method applying each fixing rule only once to a respective tuple.
Conveniently, the method comprises allocating an attribute to each tuple which indicates each fixing rule that has been applied to the tuple.
Advantageously, the method comprises incrementing at least one counter to record when a fixing rule is applied to a tuple.
Preferably, each counter is a hash counter which records the number of tuples that correspond to each fixing rule.
Conveniently, the method further comprises generating an inverted list of a plurality of fixing rules, the inverted list comprising the plurality of fixing rules indexed according to at least one attribute value of each respective fixing rule.
Advantageously, the method comprises generating the inverted list only once during the operation of the method.
Preferably, the fixing rule comprises at least one similarity operator which is operable to detect variants of attribute values.
Conveniently, the fixing rule is operable to use a wildcard attribute value in the set of attribute values.
Advantageously, the fixing rule is operable to detect the negation of an attribute value.
Preferably, the method comprises providing a plurality of fixing rules and applying at least one of the plurality of fixing rules to the database.
According to another aspect of the present invention, there is provided a system for cleaning data stored in a database, the system being operable to perform the method of any one of claims 1 to 19 defined hereinafter.
According to a further aspect of the present invention, there is provided a tangible computer readable medium storing instructions which, when executed, perform the method of any one of claims 1 to 19 defined hereinafter.
So that the present invention may be more readily understood, embodiments of the present invention will now be described, by way of example, with reference to the accompanying drawings, in which:
An embodiment of the present invention utilises a set of data cleaning rules that not only detect errors from semantically related attribute values, but also automatically correct these errors without necessarily using any heuristics or interacting with users.
A data fixing rule of an embodiment of the invention contains an evidence pattern, a fact and a set of negative patterns. When a given tuple matches both the evidence pattern and the negative pattern of the rule, it is identified as an error, and the fixing rule will use the fact to correct the tuple.
This is possible by combining an evidence pattern, negative patterns and a fact into a single data fixing rule. The evidence pattern is a set of values with each value for one attribute. The negative patterns are a set of attribute values that capture an error on one attribute from semantically related values. The fact specifies a deterministic way to correct the error.
Consider a tuple t in relation travel, an example fixing rule φ1 is: for t, if its country is China and its capital is Shanghai or Hong Kong, t[capital] should be updated to Beijing.
This rule makes corrections to attribute t[capital], by taking the value from φ1, if t is identified by φ1 that current value t[capital] is wrong.
Another fixing rule φ2 is: for t in travel, if its country is Canada and its capital is Toronto, t[capital] should be updated to Ottawa.
Consider the database in
Fixing rules φ1 and φ2 are summarised in
After applying φ1-φ2, two errors (r2[capital], r4[capital]) have been fixed, while one remains (r3[capital]).
The above example indicates that:
(a) Fixing rules make dependable fixes, which do not introduce errors as in the heuristics rule in case (1) described above.
(b) Fixing rules do not claim to correct all errors, e.g., the combination (China, Tokyo). This combination may even be difficult for users to correct.
(c) Fixing rules neither require master data (3,4), or assume some attributes to be correct (2,4), nor interact with the users (2,3).
Fixing Rules—Syntax
A fixing rule φ defined on a relation R is of the form (((X,tp[X]),(B,−Tp[B]))→+tp[B]) where:
1. X is a set of attributes in attr(R), and B is an attribute in attr(R)\X. Here, the symbol ‘\’ represent set minus;
2. tp[X] is a set of attribute values in X, referred to as the evidence pattern. For each A∈X, tp[A] is a constant in dom(A);
3. −Tp[B] is a finite set of constant values in dom(B), referred to as the negative patterns of B; and
4. +tp[B] is a constant value in dom(B)\−Tp[B], referred to as the fact of B.
Intuitively, the evidence pattern tp[X] of X, together with the negative patterns −Tp[B] of B impose the condition to determine whether a tuple contains an error on attribute B, and the fact +tp[B] of B indicates how to correct the error on attribute B.
Note that the above condition 4 enforces that the correct value (i.e., the fact) is different from any known wrong values (i.e., negative patterns).
A tuple t of R matches a rule φ: (((X,tp[X]),(B,−Tp[B]))→+tp[B]), if
Consider the fixing rules described in the above example. The rules can be formally expressed as follows:
φ1: ((([country],[China]),(capital,{Shanghai,Hong Kong}))→Beijing)
φ2: ((([country],[Canada]),(capital,{Toronto}))→Ottawa)
In both φ1 and φ2, X consists of country, B is capital. The pattern of φ1 states that, for a tuple, if its country is China and its capital value is in the set {Shanghai, Hong Kong}, its capital value should be updated to Beijing.
Consider the database D in
Fixing Rules—Semantics
A fixing rule φ applies to a tuple t, denoted by t→φt′, if
That is, if t[X] agrees with tp[X] and t[B] appears in the set −Tp[B], then +tp[B] is assigned to t[B]. Intuitively, if t[X] matches tp[X] and t[B] matches some value in −Tp[B], it is dependable to judge that t[B] is erroneous and hence, it is reliable to update t[B] to +tp[B]. This yields an updated tuple t′ with t′[B]=+tp[B] and t′[R\{B}]=t[R\{B}].
Fixing rules are quite different from integrity constraints, such as CFDs. Integrity constraints have static semantics: they only detect data violations for given constraints, but they do not tell how to change resolve them. In contrast, a fixing rule φ specifies an action: applying φ to a tuple t yields an updated t′.
Editing rules have a dynamic semantics. In contrast to them, fixing rules (a) neither require the presence of master data or confidence values placed on attributes, and (b) nor interact with the users.
Fixing rules are different from Extract Transform Load (ETL) rules which refer to a process in database usage and especially in data warehousing that involves: (a) Extracting data from outside sources, (b) Transforming it to fit operational needs (which can include quality levels), and (c) Loading it into the end target e.g., database. Fixing rules, on the other hand, focus on detect errors from attribute values that depend on each other. Fixing rules can capture errors that ETL rules fail to detect.
In one embodiment, ETL rules are used to extract data from a source and fixing rules are then used to clean the extracted data.
Heuristic solutions, which use integrity constraints, may be used in addition to fixing rules. That is, fixing rules can be used initially to find dependable fixes and then heuristic solutions can be used to compute a consistent database.
Editing rules and fixing rules should be used for different targets. Editing rules are used for critical data, which needs heavy involvement of experts to ensure, for each tuple, that the attributes are correct. Fixing rules, on the other hand, can be used for more general data cleaning applications that cannot afford to involve users to clean each tuple.
Fixing Rule Algorithm
Recall that when applying a fixing rule φ to a tuple t, t[B] is updated with the value +tp[B]. To ensure that the change makes sense, the values that have been validated to be correct should remain unchanged in the following process. That is, after applying φ to t, the set X∪{B} of attributes should be marked as correct for tuple t.
The following algorithm is based on the above observation.
Note that the above algorithm will terminate, since the number of validated attributes in V will increase monotonically, up to the total number of attributes in relation R.
Data Fixing Rule Extensions
(1) Similarity Operators
Domain-specific similarity functions are used in one embodiment to replace all equality comparisons. This makes it easier to capture typographical errors (e.g., Ottawo) and different spelling variants (e.g., Hong Kong and Peking), as opposed to including them as negative patterns in fixing rules.
(2) Wildcard
The wildcard ‘*’ may be allowed in the pattern. For instance, a fixing rule can be extended as:
Intuitively, the rule φ′ assumes that for a tuple t, t[country] is correct, if t[country] is China. No matter what value that t[capital] takes, φ′ will update t[capital] to Beijing. This is equivalent to the ETL lookup operations.
(3) Negation
In one embodiment, negations are added to the match conditions. Intuitively, a tuple can match a rule only when certain conditions are not satisfied. For instance, certain fixing rules can be applied when the country is not China.
The clear advantage of fixing rules, compared with the prior art, is that they can automatically detect errors and derive dependable repairs without interacting with the users, and without the assumption that some values have been validated to be correct. In contrast, all conventional techniques either (1) use heuristic approaches to compute a consistent database by making minimum number of changes, or (2) to consult the users, or use master data, or assume some attributes are correct, in order to derive dependable fixes.
Data fixing rules can be employed easily in many products to detect errors and perform dependable data repairing. Data fixing rules can be used to carry out more dependable data repairs than tools that are currently widely employed in industry (i.e., ETL tools) for name standardization, address check, etc.
Data has become an important asset in today's economy. Extracting values from large amounts of data to provide services and to guide decision making processes has become a central task in all data management stacks. The quality of data becomes one of the differentiating factors among businesses and the first line of defence in producing value from raw input data. As data is born digitally and is fed directly into stacks of information extraction, data integration, and transformation tasks, ensuring the quality of the data with respect to business and integrity constraints have become more important than ever.
2.2 Repairing Semantics with Fixing Rules
We next describe in more detail the semantics of applying a set of fixing rules.
Notations. For convenience, we use the following notations. Given fixing rule φ: ((X, tp[X]), (B,Tp−[B]))→tp+, [B], we denote by Xφ the set X of attributes in φ. Similarly, we write tp[Xφ],Bφ,Tp−[Bφ,] and tp+[Bφ], relative to φ.
Note that when applying a fixing rule φ to a tuple t, we update tp[Bφ] with tp+[Bφ]. To ensure that the change makes sense, the values that have been used and corrected should remain unchanged in the following process. That is, after applying φ to t, the set Xφ∪{Bφ} of attributes should be marked as correct for tuple t.
In order to keep track of the set of attributes that has been marked correct, we introduce the notion assured attributes to represent them, denoted by t relative to tuple t. We simply write when t is clear from the context. Consider a fixing rule φ. We say that φ is properly applied to a tuple t w.r.t. the assured attributes , denoted by tt′, if (i) t matches φ, and (ii) Bφ∉.
That is, it is justified that to apply φ to t, for those t match φ, is correct. As has been assured, we do not allow it to be changed by enforcing Bφ∉ (the (ii) above).
Consider the fixing rule φ1 in
Observe that if t′, then Xφ and Bφ will also be marked correct. Thus, the assured attributes should be extended as well, to become ∪Xφ∪{Bφ}.
Consider Example 1. After φ1 is applied to r2, the assured attribute r
We write tt if φ cannot be properly applied to t, i.e., t is unchanged by φ relative to , if either t does not match φ, or Bφ∉.
Consider a set Σ of fixing rules defined on R. Given a tuple t of R, we want a unique fix of t by using Σ. That is, no matter in which order the fixing rules of Σ are properly applied, Σ yields a unique by t′ updating t.
To formalize the notion of unique fixes, we first recall the repairing semantics of fixing rules. Notably, if φ is properly applied to t via tt′ w.r.t. assured attributes , it yields an updated t′ where t[Bφ]∈Tp−[Bφ] and t′[Bφ]=tp+[Bφ]. More specifically, the fixing rule φ first identifies t[Bφ] as incorrect, and as a logical consequence of the application of φ, t[Bφ] will be updated to tp+[Bφ], as a validated correct value in t′. Once an attribute value t′[B] is validated, we do not allow it to be changed, together with the attributes Xφ that are used as the evidence to assert that t[Bφ] is incorrect.
Fixes.
We say that a tuple t′ is a fix of t w.r.t. a set Σ of fixing rules, denoted by tt′, if there exists a finite sequence t=t0, t1, . . . , tk=t′ of tuples of R such that for each i∈[1, k], there exists a φi∈Σ such that
1. ti−1ti, where 1=Ø, i=i−1∪Xφ
2. for any φ∈Σ, t′t′.
Condition (1) ensures that each step of the process is justified, i.e., a fixing rule is properly applied. Condition (2) ensures that t′ is a fixpoint and cannot be further updated.
Unique Fixes.
We say that an R tuple t has a unique fix by a set Σ of fixing rules if there exists a unique t′ such that t*→(Ø,Σ)t′.
Consider Example 1. Indeed, r′2 is a fix of r w.r.t. properly applied to r2 rules φ1 and φ2 in Example 3, since no rule can be properly applied to r′2, given the assured attributes to be {country, capital}.
Moreover, r2 is also a unique fix, since one cannot get a tuple different from when trying to apply rules φ1 and φ2 on tuple r2 in other orders.
We next identify possible problems associated with fixing rules, and establish their complexity.
Termination.
One natural question associated with rule based data repairing processes is the termination problem. It is to determine that whether a rule-based process will stop. In fact, it is readily to verify that for the fix process (see Section 2.2) by applying fixing rules, it always terminates.
Consider the following. For a sequence of updates t0t1ti . . . , each time when a fixing rule φi(i≥1) is applied as ti−1ti, the number of validated attributes in is strictly increasing, up to |R|, the cardinality of schema R.
Consistency.
The problem is to decide whether a set Σ of fixing rules do not have conflicts. We say that Σ is consistent if for any input tuple t of R, t has a unique fix by Σ.
Consider a fixing rule φ′1 by adding a negative pattern to the φ1 in the fixing rules in
φ′1: (([country], [China]), (capital, {Shanghai, Hongkong, Tokyo}))→Beijing
The revised rule φ′1 states that, for a tuple, if its country is China and its capital value is Shanghai, Hongkong or Tokyo, its capital is wrong and should be updated to Beijing.
Consider another fixing rule φ3 as: for t in relation Travel, if the conf is ICDE, held at city Tokyo and capital Tokyo, but the country is China, its country should be updated to Japan. This fixing rule can be formally expressed below: φ3 (([capital, city, conf], [Tokyo, Tokyo, ICDE]), (country, {China}))→Japan We show that these two fixing rules, φ′1 and φ3, are inconsistent. Consider the tuple r3 in
(1) r3: it will change attribute r3 [capital] from Tokyo to Beijng. This will result in an updated tuple as: r′3: (Peter, China, Beijing Tokyo, ICDE).
It also marks attributes {country, capital} as assured, such that φ3 cannot be properly applied, i.e., r′3 is a fixpoint.
(2) r3→(Ø,φ
The attributes {country, capital, conf} will be marked as also a fixpoint. Observe that the above two fixes (i.e., r′3 and r″3) will lead to different fixpoints, where the difference is highlight above. Therefore, φ′1 and φ3 are inconsistent. Indeed, r′3 contains errors while r″3 is correct.
The consistency problem is to determine, given a set Σ of fixing rules defined on R, whether Σ is consistent. Intuitively, this is to determine whether the rules in Σ are dirty themselves. The practical need for the consistency analysis is evident: we cannot apply these rules to clean data before Σ is ensured consistent itself.
This problem has been studied for CFDs, MDs, and editing rules. It is known that the consistency problem for MDs is trivial: any set of MDs is consistent. They are NP-complete (resp. coNP-complete) for CFDs (resp. editing rules). We shall show that the problem for fixing rules is PTIME, lower than their editing rules counterparts.
Theorem 1:
The consistency problem of fixing rules is PTIME.
We prove Theorem 1 by providing a PTIME algorithm for determining whether a given set of fixing rules is consistent (see Section 4.2).
The low complexity from the consistency analysis tells us that it is feasible to efficiently find consistent fixing rules.
Implication.
Given a set Σ of consistent fixing rules, and another fixing rule φ that is not in Σ, we say that φ is implied by Σ, denoted by Σ|=φ, if:
(i) Σ∪{φ} is consistent; and
(ii) for any input t where t*→Σt′. and t*→Σ∪{φ})t″, t′,and t″ are the same.
Condition (i) says that Σ and φ must agree on each other.
Condition (ii) ensures that for any tuple t, applying Σ or Σ∪{φ} will result in the same updated tuple, which means that φ is redundant.
The implication problem is to decide, given a set Σ of consistent fixing rules, and another fixing rule φ, whether Σ implies φ.
Intuitively, the implication analysis helps us find and remove redundant rules from Σ, i.e., those that are a logical consequence of other rules in Σ, to improve performance.
No matter how desirable to remove redundant rules, unfortunately, the implication problem is coNP-complete.
Theorem 2:
The implication problem of fixing rules is coNP-complete. It is down to PTIME when the relation schema R is fixed.
Proof Sketch: (A) General Case. Lower Bound.
We show the implication problem is coNP-hard by reduction from the 3SAT problem, which is NP-complete [23], to the complement of the implication problem.
Upper Bound.
To show it is in coNP, we first establish a small model property: a set Σ of fixing rules is consistent if and only if for any tuple t of R consisting of values appeared in Σ, t has a unique fix by Σ. We then give an NP algorithm to its complement problem that first guesses a tuple t with values appear in Σ and then checks whether t has a unique fix by Σ in PTIME.
(B) Special Case: When R is Fixed.
We show that for fixed R, only polynomially number of tuples need to be guessed and checked with a PTIME algorithm. Thus it is down to PTIME in this special case.
Determinism.
The determinism problem asks whether all terminating cleaning processes end up with the same repair. From the definition of consistency of fixing rules, it is trivial to get that, if a set Σ of fixing rules is consistent, for any t of R, applying Σ to t will terminate, and the updated t′ is deterministic (i.e., a unique result).
The following description covers methods for identifying consistent rules. We first describe the workflow for obtaining a set of consistent fixing rules (Section 4.1). We then present algorithms to check whether a given set of rules is consistent (Section 4.2). We also discuss how to resolve inconsistent fixing rules, and ensure the workflow terminates (Section 4.3).
Overview
Given a set Σ of fixing rules, our workflow contains the following three steps to obtain a set Σ′ of fixing rules that is ensured to be consistent. The workflow is illustrated in
Step 1:
It checks whether the given Σ of fixing rules is consistent. If it is inconsistent, it goes to step (2). Otherwise, it goes to step (3).
Step 2:
We allow either an automatic algorithm or experts to examine and resolve inconsistent fixing rules. After some rules are revised, it will go back to step (1).
Step 3:
It terminates when the set Σ′ of (possibly) modified fixing rules is consistent.
It is desirable that the users are involved in step (2) when resolving inconsistent rules, in order to obtain high quality fixing rules.
4.2 Checking Consistency
We first present a proposition, which is the pivot of designing efficient algorithms for checking consistency.
Proposition 3:
For a set Σ of fixing rules, Σ is consistent, iff (if and only if) any two fixing rules φi and φj in π are consistent.
Proof Sketch:
Let n be the number of rules in Σ. When n=1, Σ is trivially consistent. When n=2, Σ is consistent is the same as φi and φj are consistent
i≠j. When n≥3, we prove by contradiction.
Suppose that although the fixing rules are pairwise consistent, when putting together, they are inconsistent. In other words, they may lead to (at least) two different fixes, i.e., the fixes are not unique. More concretely, there exist (at least) two non-empty sequences of fixes as follows:
(i) m∩′n=Ø
(ii) m∩′n≠Ø and t′[m∩′n]=t″[m∩′n]; and
(iii) m∩′n≠Ø and t′[m∩′n]≠t″[m∩′n], where
m=m−1∪Xφ
For cases (i)(ii), we prove that either S1 or S2 does not reach a fixpoint, i.e., it is not a fix. For case (iii), we show that there must exist a φ′i (in sequence S1) and a φ′j sequence S2) that are inconsistent.
Putting all contradicting cases (i,ii,iii) together, it suffices to see that we were wrong to assume that Σ is inconsistent.
Assume there exist inconsistent φi and φj. We show that for any tuple t that leads to different fixes by φi and φj, we can construct two fixes S′1 and S′2 on t by using the rules in Σ. In S′1, φi is applied first; while in S′2, φj is applied first. We prove that these two fixes must yield two different fixpoints. This suffices to show that we were wrong to assume that there exist inconsistent φi and φj.
The Appendix section below shows a detailed proof.
Proposition 3 tells us that to determine whether Σ is consistent, it suffices to only check them pairwise. This significantly simplifies the problem and complexity of checking consistency. Next, we describe two algorithms to check the consistency of two fixing rules, by using the result from Proposition 3. One algorithm is based on tuple enumeration, while the other is via rule characterization.
4.2.1 Tuple Enumeration
Consider that although there may exist infinitely many t, whether there exists a finite set of tuples such that it suffices to inspect those t only for two rules φi and φj. In other words, for the other tuples, neither φi and φj can be applied.
To begin an algorithm for tuple enumeration, we describe what tuples are necessary to be enumerated, and in which case that tuple enumeration can be avoided.
Lemma 4:
Fixing rules φi and φj are consistent, if there does not exist any tuple t that matches both φi and φj.
Proof. If t such that t├φi and ├φj, for any t, there are two cases: either no rule can be applied, or there exists a unique sequence of applying both rules. Either case will not cause different fixes, i.e., φi and φj are consistent.
Note that Lemma 4 is for “if” but not “iff”, which tells us that only tuples that draw values from evidence pattern and negative patterns can (possibly) match both rules at the same time. Next we illustrate the tuples that are needed to be generated by an example.
Consider rules φi and φj in shown in
Let {A1 . . . Am} be all attributes appearing in φi and appear either in evidence pattern or negative patterns of φj. Let Vφij(A) denote the set of constant values of A that φi and φj. The total number of tuples to be enumerated is Πl∈[1,m](|Vφij(Al)|), where Π indicates a product and |Vφij(Al)| denotes the cardinality of Vφij(Al).
Given a set Σ of fixing rules, we check them pairwise (see Example 4). If any pair of rules is inconsistent, we judge that Σ is inconsistent; otherwise, Σ is consistent. The algorithm is shown in
4.2.2 Rule Characterization
The following description covers analysis by characterizing the fixing rules.
Also based on Lemma 4, let us focus on the cases of φi and φj that there exists some t that can match both fixing rules, where these rules are represented as follows:
Note that a tuple t matching φi and φj implies that the following conditions hold: t[Xi]=tp
Case 1:
Bi=Bj. Let B=Bi=Bj. There is a conflict only when (i) there exists a tuple t that matches both φi and φj, and (ii) φi and φj will update t to different values. From (i) we have t[B]∈Tp
Case 2:
Bi≠Bj. Again, we consider four cases:
Consider φ′1 and φ3 in Example 8 and φ2 in the fixing rules shown in
Since φ′1 (resp. φ2) is only applied to a tuple whose country is China (resp. Canada), there does not exist any tuple that can match both rules at the same time. Therefore, based on Lemma 4, we have φ′1 and φ2 are consistent.
Also, it can be verified that φ′1 and φ3 are inconsistent Consider the following:
(i) Bφ
(ii) tp
(iii) Bφ′
(iv) tp
Hence, these two rules will lead to different fixes, which is captured by case 2(c).
Algorithm.
The algorithm to check whether a set of fixing rules is consistent via rule characterization, referred to as isConsistr, is given in
It enumerates all pairs of distinct rules (lines 1-11). If any pair is inconsistent, it returns false (lines 5,7,9,11); otherwise, it reports that Σ is consistent (line 12). It covers all cases that two rules can be inconsistent, i.e., case 1 (lines 2-5), case 2(a) (lines 6-7), case 2(b) (lines 8-9) and case 2(c) (lines 10-11). Note that in case 2(d), two rules are trivially consistent. Hence, there is no need to capture such case.
Correctness & Complexity.
Its correctness is ensured by Proposition 3 and Lemma 4. From the analysis above, Algorithm isConsistr covers all cases that two rules can be inconsistent. That is, the two rules φi and φj are consistent for all other cases. We use a hash table to check that whether a constant matches some negative pattern in constant time. To summarize, it enumerates all pairs of rules, and runs in O(size(Σ)2) time, where size(Σ) is the size of Σ.
4.3 Resolving Inconsistent Rules
When inconsistency of fixing rules is detected, this inconsistency has to be resolved, before these rules can be used.
Consider Example 2 that two rules φ′1 and φ3 are inconsistent, which is depicted in
Hence, in order to obtain high quality rules, the system requests an expert to examine rules that are in conflict manually. For example, the expert can naturally remove Tokyo from the negative patterns of φ′1, since one cannot judge, given (China, Tokyo), which attribute is wrong. This will result in a modified rule φ1 (see the fixing rules in
We now describe how to use the fixing rules to repair data.
In the following, we first present a chase-based algorithm to repair one tuple (Section 5.1), with time complexity in O(size(Σ)|R|). We also present a fast algorithm (Section 5.2) running in O(size(Σ)) time for repairing one tuple.
5.1 Chase-Based Algorithm
Note that, given a tuple t and a set Σ of fixing rules, if Σ is consistent, it has the Church-Rosser property, i.e., all the chase sequences using Σ lead to a unique fix, no matter in what orders these rules are applied.
We next present the algorithm, by using a chase process.
Algorithm.
The algorithm, referred to as cRepair, is shown in
The algorithm first initializes a set of assured attributes, a set of fixing rules that can be possibly applied, a tuple to be repaired, and a flag to indicate whether the tuple has been changed (line 1). It then iteratively examines and applies the rules to the tuple (lines 2-7). If there is a rule that can be properly applied (line 5), it updates the tuple (line 6), maintains the assured attributes and rules that can be used correspondingly, and flags this change (line 7). It terminates when no rule can be further properly applied (line 2), and the repaired tuple will be returned (line 8).
Correctness & Complexity.
The correctness of cRepair is inherently ensured by the Church-Rosser property, since Σ is consistent. For the complexity, observe the following. The outer loop (lines 2-7) iterates at most |R| times. For each loop, it needs to scan each unused rule, and checks whether it can be properly applied to the tuple. From these it follows that Algorithm 6 runs in O(size(Σ)|R|) time.
5.2 A Fast Repairing Algorithm
We now describe how to improve the chase-based procedure. One way is to consider how to avoid repeatedly checking whether a rule is applicable, after each update of the tuple being examined.
Note that a property of employing fixing rules is that, for each tuple, each rule can be applied only once. After a rule is applied, in consequence, it will mark the attributes associated with this rule as assured, and does not allow these attributes to be changed any more (see Section 2.2).
Hence, after each value update, to (i) efficiently identify the rules that cannot be applied, and (ii) determine unused rules that can be possibly applied.
We employ two types of indices in order to perform the above two targets. Inverted lists are used to achieve (i), and hash counters are employed for (ii). Before describing how to use these indices to design a fast algorithm, we shall define these indices to assist in understanding the algorithm.
Inverted Lists.
Each inverted list is a mapping from a key to a set Υ of fixing rules. Each key is a pair (A, a) where A is an attribute and a is a constant value. Each fixing rule φ in the set Υ satisfies A∈Xφ and tp[A]=a.
For example, an inverted list w.r.t. φ1 in
Intuitively, when the country of some tuple is China, this inverted list will help to identify that φ1 might be applicable.
Hash Counters.
It uses a hash map to maintain a counter for each rule. More concretely, for each rule and and φ, the counter c(φ) is a nonnegative integer, denoting the number of attributes that a tuple agrees with tp[Xφ].
For example, consider φ1 in Example 3 and r2 in
We now describe a fast algorithm by using the two indices introduced above. Note that inverted lists are built only once for a given Σ, and keep unchanged for all tuples. The hash counters will be initialized to zero for the process of repairing each new tuple.
Algorithm.
The algorithm IRepair is shown in
During the process (lines 8-16), it first randomly picks a rule that might be used (line 9). The rule will be applied if it is verified to be applicable (lines 10-11). The set of attributes that is assured correct is increased correspondingly (line 12). The counters will be recalculated (lines 13-14). Moreover, if new rules might be used due to this update, it will be identified (line 15). The rule that has been checked will be removed (line 16), no matter it is applicable or not. Observe the following two cases. (i) If a rule is removed after being applied at line 16 (i.e., line 10 gives a true), it cannot be used again and will not be checked at lines 13-15. (ii)) If a rule φ is removed without being applied at line 16 (i.e., line 10 gives a false), it cannot be used either at lines 13-15. The reason is that: for any rule φ, if φ cannot be properly applied to t′, any update on attribute Bφ will mark it as assured, such that q, cannot be properly applied afterwards. From the above (i) and (ii), it follows that it is safe to remove a rule from Γ, after it has been checked, once and for all.
Correctness.
Note that Σ is consistent, we only need to prove the repaired tuple t′ is a fix of t. This can be proved based on (1) at any point, Γ includes all fixing rules that might match the given tuple; and (2) each fixing rule is added into Γ at most once. Hence, the algorithm terminates until it reaches a fixpoint when Γ is empty.
Complexity.
It is clear that the three loops (line 2, lines 3-5 and lines 6-7) all run in time linear to size(Σ). Next let us consider the while loop (lines 8-16). Observe that each rule φ will be checked in the inner loop (lines 13-15) up to |Xφ| times, by using the inverted lists and hash counters, independent of the number of outer loop iterated. The other lines of this while loop can be done in constant time. Putting together, the total time complexity of the algorithm is O(size(Σ)).
We next show by example how Algorithm IRepair works.
Consider Travel data D in
φ4: (([capital, conf], [Beijing, ICDE]), (city, {Hongkong})→Shanghai
Rule φ4 states that: for t in relation to Travel, if the conf is ICDE, held at some country whose capital is Beijing, but the city is Hongkong, its city should be Shanghai. This holds since ICDE was held in China only once at 2009, in Shanghai but never in Hongkong.
Given the four fixing rules φ1-φ4, the corresponding inverted lists are given in
Now we show how the algorithm works over tuples r1 to r4, which is also depicted in
r1:
The algorithm initializes (lines 1-7) and finds that φ1 may be applied, maintained in Γ. In the first iteration (lines 8-16), it finds that φ1 cannot be applied, since r1 [capital] is Beijing, which is not in the negative patterns {Shanghai, Hongkong} of φ1. Also, no other rules can be applied. It terminates with tuple r1 unchanged. Actually, r1 is a clean tuple.
r2:
The algorithm initializes and finds that φ1 might be applied. In the first iteration (lines 8-16), rule φ1 is applied to r2 and updates r2 [capital] to Beijing. Consequently, it uses inverted lists (line 13) to increase the counter of φ4 (line 14) and finds that φ4 might be used (line 15). In the second iteration, rule φ1 is applied and updates r2 [city] to Shanghai. It then terminates since no other rules can be applied.
r3:
The algorithm initializes and finds that φ3 might be applied. In the first iteration, rule φ3 is applied and updates r3 [country] to Japan. It then terminates, since no more applicable rules.
r4:
The algorithm initializes and finds that φ4 might be applied. In the first iteration, rule φ2 is applied and updates r4 [capital] to Ottawa. It will then terminate.
At this point, we see that all the four errors shown in
We conducted experiments with both real-life and synthetic data to examine our algorithms. Specifically, we evaluated (1) the efficiency of consistency checking for fixing rules; (2) the accuracy of our data repairing algorithms with fixing rules; and (3) the efficiency of data repairing algorithms using fixing rules.
It is worth noting that the purpose of these experiments is to test, when given high quality fixing rules, how they can be used to automatically repair data with high dependability.
6.1 Experimental Setting
Experimental Data.
We used real-life and synthetic data. (1) HOSP was taken from us Department of Health & Human Services (http://www.hospitalcompare.hhs.gov/). It has 115K records with the following attributes: Provider Number (PN), Hospital Name (HN), address1, address2, address3, city, state, zip, county, Phone Number (phn), HospitalType (ht), HospitalOwner (ho), EmergencyService (es) Measure Code (MC), Measure Name (MN), condition, and stateAvg.
(2) UIS data was generated by a modified version of the UIS Database generator (http://www.cs.utexas.edu/users/ml/riddle/data.html). It produces a mailing list that has the following schema: RecordID, ssn, FirstName (fname), MiddleInit (minit), LastName (lname), stnum, stadd, apt, city, state, zip.
Dirty Data Generation.
We treated clean datasets as the ground truth. Dirty data was generated by adding noise only to the attributes that are related to some integrity constraints, which is controlled by noise rate (10% by default). We introduced two type of noises: typos and errors from the active domain.
Fixing Rules Generation from Samples.
Note that fixing rules are instance based, i.e., all values for identifying and correcting errors are encoded inside fixing rules, one natural question is that how fixing rules were obtained.
Sample Generation.
Since each fixing rule is defined on semantically related attributes, we start with known data dependencies (e.g., functional dependencies for our testing). We first detect violations of given functional dependencies (FDs), and present them to the experts. The experts produced several fixing rules as samples, based on their understanding of these violations.
Rule Generation.
Given sample fixing rules, we enrich them by only enlarging their negative patterns, via extracting new negative patterns from other tables in the same domain. For instance, consider the fixing rules shown in
We generated 100 fixing rules for HOSP data, and 1000 fixing rules for UIS data. Note that the purpose of fixing rules generation is not to use them for some specific dataset. It is, by collecting expert knowledge for specific errors, to learn high quality domain related rules that can be used to automatically detect and repair data for other datasets in the same domain.
Measuring Quality.
To assess the accuracy of data cleaning algorithms, we use precision and recall, where precision is the ratio of corrected attribute values to the number of all the attributes that are updated, and recall is the ratio of corrected attribute values to the number of all erroneous attribute values.
We mainly compare automated data cleaning techniques. Note that they are designed for a slightly different target: computing a consistent database. We consider it a relative fair comparison, since all fixing rules we generated are from FD violations. In other words, the fixing rules and the FDs used are defined on exactly the same set of attributes. We employed the FDs shown in
(2) Editing Rules.
We also compared our approach with editing rules. Although editing rules can repair data that is guaranteed to be correct, they are measured by the number of user interactions per tuple. That is, for each tuple and for each editing rule to be applied, the users have to be asked. To this purpose, we evaluated the number of errors that can be corrected by every fixing rule (see
Moreover, we encoded data values from master data into editing rules, to make it an automated rule. Note that error information is not in master data, e.g., the negative patterns in fixing rules, which cannot be encoded. Hence, we removed negative patterns in fixing rules, to simulate editing rules. Specifically, each time when seeing an evidence pattern, it simulated users by saying yes, and then updated the right hand side value to the fact. The experimental results are shown in
Algorithms.
We have implemented the following algorithms in C++: (1) isConsistt: the algorithm for checking consistency based on tuple enumeration (Section 4.2); (2) isConsistr: the algorithm for checking consistency based on rule characterization (
6.2 Experimental Results
We next report our findings from the experimental study.
Exp-1: Efficiency of Checking Consistency.
We evaluated the efficiency of checking consistency by varying the number of rules employed. The results for HOSP and UIS are shown in
For either isConsistt or isConsistr, we plotted its worst case, i.e., checking all pairs of rules, as well as its 10 real cases where it terminated when some pair was detected to be inconsistent. For example, in
These figures show that to check consistency of fixing rules, the algorithm with tuple enumeration (isConsistt) is slower, as expected. The reason is that enumerating tuples for two rules is more costly than characterizing two rules.
In addition, this set of experiment validated that the consistency of fixing rules can be checked efficiently. For example, it only needs 12 seconds to check the consistency of 1000*1000 pairs of rules, i.e., the top right point in
The results of this study indicate that it is feasible to check consistency for a reasonably large set of fixing rules.
Exp-2: Accuracy.
In this set of experiments, we will study the followings. (a) The effect of different data errors (i.e., typos or errors from active domain) for repairing algorithms. (b) The influence of fixing rules w.r.t. their sizes. We use Fix to represent repairing algorithms with fixing rules.
(a) Noise from the Active Domain.
Recall that noise was obtained by either introducing typos to an attribute value or changing an attribute value to another one from the active domain of that specific attribute. For example, an error for Ottawa could be Ottawo (i.e., a typo) or Beijing (i.e., a value from active domain).
Precision.
We fixed the noise rate at 10%, and varied the percentage of typos from 0% to 100% by a step of 10% (x-axis in both charts from
Note that fixing rules also made mistakes, e.g., the precision in
Recall.
In order to better understand the behaviour of these algorithms,
(b) Varying the Number of Fixing Rules.
We studied the accuracy of our repairing algorithms by varying the number of fixing rules. We fixed noise rate at 10% and half of them are typos. For HOSP, we varied the number of rules from 100 to 1000, and reported the recall and precision in
The experimental results indicate that when more fixing rules are available, our approach can achieve better recall, while keeping a good precision, as expected.
(c) Number of Tuples Corrected.
To further understand fixing rules, we next study how many errors each fixing rule can repair. We used the real-life data, i.e., HOSP, and calculated the number of errors that can be correctly repaired by each rule.
We see that a fixing rule can be used to repair multiple errors (e.g., 52 errors for the top left point), which shows its potential to be reused by other datasets in the same domain.
Exp-3: Efficiency of Repairing Algorithms.
In this last set of experiments, we study the efficiency of our data repairing algorithms. As they are linear in data size, we only evaluated their efficiency by varying the number of rules.
The results for HOSP and UIS are given in
Summary.
We find the followings from the above experiments. (a) It is efficient to detect whether a set of fixing rules is consistent (Exp-1). (b) Data repairing using fixing rules is dependable, i.e., they repair data errors with high precision (Exp-2). (c) The recall of using fixing rules can be improved when more fixing rules are available (Exp-2). (d) It is efficient to repair data via fixing rules, which reveals its potential to be used for large datasets (Exp-3).
We have proposed a novel class of data cleaning rules, namely, fixing rules, that (1) compared with data dependencies used in data cleaning, are able to find dependable fixes for input tuples, without using heuristic solutions; and (2) differ from editing rules, are able to repair data automatically without any user involvement. We have identified fundamental problems for deciding whether a set of fixing rules is consistent or redundant, and established their complexity bounds. We have proposed efficient algorithms for checking consistency, and discussed strategies to resolve inconsistent fixing rules. We have also presented dependable data repairing algorithms by capitalizing on fixing rules. Our experimental results with real-life and synthetic data have verified the effectiveness and efficiency of the proposed rules and the presented algorithms. These yield a promising method for automated and dependable data repairing.
When used in this specification and claims, the terms “comprises” and “comprising” and variations thereof mean that the specified features, steps or integers are included. The terms are not to be interpreted to exclude the presence of other features, steps or components.
Number | Date | Country | Kind |
---|---|---|---|
PCT/EP2013/052476 | Feb 2013 | WO | international |
1302415.3 | Feb 2013 | GB | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2014/052494 | 2/7/2014 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2014/122295 | 8/14/2014 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5770841 | Moed | Jun 1998 | A |
5918232 | Pouschine | Jun 1999 | A |
6131101 | Maitino | Oct 2000 | A |
6965888 | Cesare et al. | Nov 2005 | B1 |
7653753 | Chen et al. | Jan 2010 | B2 |
8515863 | Morejon | Aug 2013 | B1 |
20020103834 | Thompson et al. | Aug 2002 | A1 |
20030023593 | Schmidt | Jan 2003 | A1 |
20030088481 | Elias | May 2003 | A1 |
20030110192 | Valente | Jun 2003 | A1 |
20030131337 | Perumainar | Jul 2003 | A1 |
20040024790 | Everett | Feb 2004 | A1 |
20040133443 | Payne | Jul 2004 | A1 |
20040148284 | Baker | Jul 2004 | A1 |
20040267611 | Hoerenz | Dec 2004 | A1 |
20040267760 | Brundage | Dec 2004 | A1 |
20050022115 | Baumgartner | Jan 2005 | A1 |
20060195460 | Nori | Aug 2006 | A1 |
20060238919 | Bradley | Oct 2006 | A1 |
20090006282 | Ann et al. | Jan 2009 | A1 |
20090055206 | Orbke | Feb 2009 | A1 |
20100250596 | Fan | Sep 2010 | A1 |
20100318499 | Arasu | Dec 2010 | A1 |
20110055172 | Tan | Mar 2011 | A1 |
20110138312 | Yeh | Jun 2011 | A1 |
20120197868 | Fauser | Aug 2012 | A1 |
20120296879 | Yakout et al. | Nov 2012 | A1 |
20140222722 | Varma | Aug 2014 | A1 |
20160004743 | Tang et al. | Jan 2016 | A1 |
20160103867 | Papotti | Apr 2016 | A1 |
Entry |
---|
Wei, Weijie, et al. “A data cleaning method based on association rules.” International Conference on Intelligent Systems and Knowledge Engineering 2007. Atlantis Press, 2007. (Year: 2007). |
Diallo, Thierno, Jean-Marc Petit, and Sylvie Servigne. “Discovering editing rules for data cleaning.” Proceedings of AQB conference. 2012. (Year: 2012). |
K. Hima Prasad et al., “Data Cleansing Techniques for Large Enterprise Datasets,” 2011 Annual SRII Global Conference, 2011 IEEE, ISBN 978-0-7695-4371-0/11, DOI 10.1109/SRII.2011.26, pp. 135-144, 10 pages. |
Muller, Heiko, et al.; “Problems, Methods, and Challenges in Comprehensive Data Cleansing;” 2003; pp. 1-23. |
Wenfei Fan et al., “Towards certain fixes with editing rules and master data,” The VLDB Journal, The International Journal on Very Large Data Bases, Springer, Berlin, DE, vol. 21, No. 2, Oct. 30, 2011, pp. 213-238, 26 pages. |
Arenas et al., “Consistent Query Answers in Inconsistent Databases,” PODS '99 Philadelphia PA, ACM 1999, pp. 68-79, May 1999, 12 pgs. |
Bertossi et al., “Data Cleaning and Query Answering with Matching Dependencies and Matching Functions,” ICDT 2011, pp. 268-279, Mar. 21-23, 2011, 12 pgs. |
Beskales et al., “Modeling and Querying Possible Repairs in Duplicate Detection,” VLDB '09, pp. 598-609, Aug. 24-28, 2009, 12 pgs. |
Beskales et al., “Sampling the Repairs of Functional Dependency Violations under Hard Constraints,” Proceedings of the VLDB Endowment, vol. 3, No. 1, pp. 197-207, Sep. 13-17, 2010, 11 pgs. |
Bohannon et al., “A Cost-Based Model and Effective Heuristic for Repairing Constraints by Value Modification,” SIGMOD 2005, pp. 143-154, Jun. 14-16, 2005, 12 pgs. |
Bravo et al., “Extending Dependencies with Conditions,” VLDB '07, pp. 243-254, Sep. 23-28, 2007, 12 pgs. |
Chomicki et al., “Minimal-Change Integrity Maintenance Using Tuple Deletions,” pp. 1-42, Oct. 20, 2004, 42 pgs. |
Chu et al., “Holistic Data Cleaning: Putting Violations Into Context,” IEEE 29th International Conference on Data Engineering (ICDE), 2013, Mar. 2013, 12 pgs. |
Cong et al., “Improving Data Quality: Consistency and Accuracy,” VLDB '07, pp. 315-326, Sep. 23-28, 2007, 12 pgs. |
Fan, Wenfei, et al., “Conditional Functional Dependencies for Capturing Data Inconsistencies,” ACM Transactions on Database Systems, vol. 33, No. 2, Article 6, Jun. 2008, 48 pgs. |
Fan, Wenfei, “Dependencies Revisited for Improving Data Quality,” PODS'08, pp. 159-170, Jun. 9-12, 2008, 12 pgs. |
Fan, Wenfei, et al., “Interaction between Record Matching and Data Repairing,” SIGMOD'11, pp. 469-480, Jun. 12-16, 2011, 12 pgs. |
Fan, Wenfei, et al., “Reasoning about Record Matching Rules,” VLDB '09, pp. 407-418, Aug. 24-28, 2009, 12 pgs. |
Fellegi et al., “A Systematic Approach to Automatic Edit and Imputation,” Journal of the American Statistical Association, vol. 71, No. 353, pp. 17-35, Mar. 1976, 20 pgs. |
Geerts et al., “The LLUNATIC Data-Cleaning Framework,” Proceedings of the VLDB Endowment, vol. 6, No. 9, pp. 625-636, Aug. 26-30, 2013, 12 pgs. |
Herzog et al., “Data Quality and Record Linkage Techniques,” Springer, May 2007, 241 pgs. |
Kolahi et al., “On Approximating Optimum Repairs for Functional Dependency Violations,” ICDT 2009, pp. 53-62, Mar. 23-25, 2009, 10 pgs. |
Mayfield et al., “ERACER: A Database Approach for Statistical Inference and Data Cleaning,” SIGMOD'10, pp. 75-86, Jun. 6-11, 2010, 12 pgs. |
Naumann et al., “Data Fusion in Three Steps: Resolving Inconsistencies at Schema-, Tuple-, and Value-level,” Bulletin of the IEEE Computer Society Technical Committee on Data Engineering, pp. 1-11, Jan. 2006, 11 pgs. |
Raman et al., “Potter's Wheel: An Interactive Data Cleaning System,” Proceedings of the 27th VLDB Conference, Roma, Italy, Sep. 2001, 10 pgs. |
Singh et al., “Learning Semantic String Transformations from Examples,” Proceedings of the VLDB Endowment, vol. 5, No. 8, pp. 740-751, Aug. 27-31, 2012, 12 pgs. |
Wijsen, “Database Repairing Using Updates,” ACM Transactions on Database Systems, vol. 30, No. 3, Sep. 2005, pp. 722-768, 47 pgs. |
Yakout et al., “Guided Data Repair,” Proceedings of the VLDB Endowment vol. 4, No. 5, pp. 279-289, Aug. 29-Sep. 3, 2011,11 pgs. |
Number | Date | Country | |
---|---|---|---|
20150379051 A1 | Dec 2015 | US |