Computing correlated aggregates over a data stream

Information

  • Patent Grant
  • 8868599
  • Patent Number
    8,868,599
  • Date Filed
    Tuesday, August 28, 2012
    12 years ago
  • Date Issued
    Tuesday, October 21, 2014
    10 years ago
Abstract
Described herein are approaches for computing correlated aggregates. An aspect provides for receiving a stream of data elements at a device, each data element having at least one numerical attribute; maintaining in memory plurality of tree structures comprising a plurality of separate nodes for summarizing numerical attributes of the data elements with respect to a predicate value of a correlated aggregation query, said maintaining comprising: creating the plurality of tree structures in which each node implements one of: a probabilistic counter and a sketch, wherein said probabilistic counter and said sketch each act to estimate aggregated data element numerical attributes to form a summary of said numerical attributes; and responsive to a correlated aggregation query specifying said predicate value, using said plurality of tree structures as a summary of said data element numerical attributes to compute a response to said correlated aggregate query.
Description
FIELD OF THE INVENTION

The subject matter presented herein generally relates to efficiently computing correlated aggregates over a data stream.


BACKGROUND

Processing a massive data set presented as a large data stream is challenging. The data set may be, for example, network traffic data or streaming data from external memory. A difficulty encountered is that the data set is too large to store in cache for analysis. Because of the size of the data set, processing requires fast update time, use of very limited storage, and possibly only a single pass over the data.


In computing statistics for certain types of data set analysis, a problem is the correlated aggregate query problem. Here, unlike in traditional data streams, there is a stream of two dimensional data items (i, y), where i is an item identifier, and y is a numerical attribute. A correlated aggregate query requires first, applying a selection predicate along the y dimension, followed by an aggregation along the first dimension. An example of a correlated query is: “On a stream of IP packets, compute the k-th frequency moment of all the source IP address fields among packets whose length was more than 100 bytes”. Answering such a query may for example allow a network administrator to identify IP addresses using an excessive amount of network bandwidth.


BRIEF SUMMARY

One aspect provides a method for summarizing attributes of data elements of a data set for computing correlated aggregates, comprising: receiving a stream of data elements at a device, each data element having at least one numerical attribute; maintaining in memory a plurality of tree structures comprising a plurality of separate nodes for summarizing numerical attributes of the data elements with respect to a predicate value of a correlated aggregation query, said maintaining comprising: creating the plurality of tree structures in which each node implements one of: a probabilistic counter and a sketch, wherein said probabilistic counter and said sketch each act to estimate aggregated data element numerical attributes to form a summary of said numerical attributes; and responsive to a correlated aggregation query specifying said predicate value, using said plurality of tree structures as a summary of said data element numerical attributes to compute a response to said correlated aggregate query.


The foregoing is a summary and thus may contain simplifications, generalizations, and omissions of detail; consequently, those skilled in the art will appreciate that the summary is illustrative only and is not intended to be in any way limiting.


For a better understanding of the embodiments, together with other and further features and advantages thereof, reference is made to the following description, taken in conjunction with the accompanying drawings. The scope of the invention will be pointed out in the appended claims.





BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS


FIG. 1 illustrates an example operating environment for correlated aggregation queries.



FIGS. 2-4 illustrate an example of processing to provide an estimate for a COUNT correlated aggregation.



FIGS. 5-7 illustrate an example of processing to provide an estimate for general correlated aggregations.



FIG. 8 illustrates examples of sketch space v. relative error for example data sets.



FIG. 9 illustrates examples of sketch space v. stream size tor example data sets.



FIG. 10 illustrates examples of sketch space v. stream size for example data sets.



FIG. 11 illustrates examples of sketch space v. stream size for example data sets.



FIG. 12 illustrates examples of sketch space v. relative error for example data sets.



FIG. 13 illustrates examples of sketch space v. stream size for example data sets.



FIG. 14 illustrates an example computer system.





DETAILED DESCRIPTION

It will be readily understood that the components of the embodiments, as generally described and illustrated in the figures herein, may be arranged and designed in a wide variety of different configurations in addition to the described example embodiments. Thus, the following more detailed description of the example embodiments, as represented in the figures, is not intended to limit the scope of the claims, but is merely representative of those embodiments.


Reference throughout this specification to “embodiment(s)” (or the like) means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. Thus, appearances of the phrases “according to embodiments” or “an embodiment” (or the like) in various places throughout this specification are not necessarily all referring to the same embodiment.


Furthermore, the described features, structures, or characteristics may be combined in any suitable manner in different embodiments. In the following description, numerous specific details are provided to give a thorough understanding of example embodiments. One skilled in the relevant art will recognize, however, that aspects can be practiced without certain specific details, or with other methods, components, materials, et cetera. In other instances, well-known structures, materials, or operations are not shown or described in detail to avoid obfuscation.


On a stream of two dimensional data items (i, y) where i is an item identifier, and y is a numerical attribute, a correlated, aggregate query requires applying a selection predicate first along the second dimension, y, followed by an aggregation along the first dimension. For selection predicates of the form (y<c) or (y>c), where parameter c is provided at query time, embodiments provide streaming processing and lower bounds for estimating statistics of the resulting substream of elements that satisfy the predicate.


For example, an embodiment can answer the following query: “On a stream of IP packets, compute the k-th frequency moment Fk(k=0, 1, 2 or greater) of all the source IP address fields among packets whose length was more than 100 bytes”. A process according to an embodiment for F1 (basic count) improves previous approaches by roughly a logarithmic factor and is near-optimal. An approach provided according to embodiments for k≠1 provides the first sub-linear space algorithms in this model for these problems. Embodiments also estimate heavy hitters, rarity, and similarity in this model. The memory requirements are significantly smaller than existing linear storage schemes for large data sets, and embodiments simultaneously achieve last per-record processing time.


The description now turns to the figures. The illustrated example embodiments will be best understood by reference to the figures. The following description is intended only by way of example and simply illustrates certain example embodiments representative of the invention, as claimed.


Referring to FIG. 1, consider a data stream (data set/stream are used interchangeably herein) of tuples S=(xi, yi), i=1 . . . n, where xi is an item identifier, and yi a numerical attribute. A correlated aggregate query C(σ, AGG, S) specifies two functions, a selection predicate σ along the y dimension. It requires that the selection be applied first on stream S, followed by the aggregation. More formally,

C(σ, AGG, S)=AG{xi|σ(yi) is true}


Importantly, the selection predicate σ need only be completely specified at query time, and may not be known when the stream is being observed. For example, on a stream of IP packets, with x denoting the source IP address, and y denoting the number of bytes in the packet, on an observed packet stream, a network administrator may ask the query: “What is the second frequency moment (F2) of all IP addresses for those packets that are less than 1000 bytes”? The selection predicate σ=(y<1000) is completely specified only at query time, giving the network administrator flexibility to perform further queries based on the result of previous queries. For example, if the answer for the above query was “interesting” (however defined), the network administrator may then refine the next query further by asking for the same aggregate, but giving a different selection predicate, for example σ=(y<500). Note that with this framework, the network administrator could form the query: “What is the F2 of all IP addresses for those packets whose packet size is smaller than the mean packet size of the stream”, by first determining the mean packet size, say μ, using an aggregated query with section predicate y<μ.


A traditional data stream summary provides an accurate answer to an aggregate query, such as frequency moments, quantities, and heavy-hitters, over the entire data stream. In contrast, a summary for correlated aggregate provides accurate answers to a family of queries that can he posed on the stream, where different queries in the family specify the same aggregation function, but over different subsets of the input stream. These subsets are selected by the predicate along the primary dimension. Correlated aggregates arise naturally in analytics on multi-dimensional streaming data, and space-efficient methods for implementing such aggregates are useful in streaming analytics systems such as in IBM SYSTEM S.


The summaries for correlated aggregates provided by embodiments allow for more flexible interrogation of the data stream than is possible with a traditional stream summary. For example, consider a data stream of IP flow records, such as those output by network routers equipped with CISCO netflow. Suppose that one is interested in only two attributes per flow record, the destination address of the flow, and the size (number of bytes) of the flow. Using a summary for correlated aggregate AGG, along with a quantile summary for the y dimension (any well known stream quantity summaries may be used), it is possible for a network administrator to execute the following sequence queries on the stream:


(1) First, the quantile summary can be queried to find the median size of the flow. (2) Next, using the summary for correlated aggregates, the network administrator can query the aggregate AGG of all those flow records whose flow size was more than the median flow size. (3) If the answer to query (2) was abnormal in the network administrator's opinion, and the network administrator needed to find properties of the very high volume flows, this may be accomplished by querying for the aggregate of all those flow records whose flow size is in, for example, the 95 percent quantile or above (the 95 percent quantile may be found using the quantile summary).


Thus, a summary that allows for such queries can be used for deeper investigation of larger data streams. Important to this is the flexibility that the parameter for the selection predicate can be specified at query time, which allows the user to iteratively adapt his or her query based on the results of previous queries. For data-intensive analytics systems, such as network management systems and sensor data management, this can be a very powerful tool.


In this description, both processing and lower bounds for summarizing data streams for correlated aggregate queries are provided. Embodiments consider selection predicates of the form (y≦y0) or (y≧y0), where y0 is specified at query time.


Estimation of Basic Count (F1)


For estimating the basic count, the x values of the stream elements (x, y) do not matter, hence only considered is the stream Y=y1, y2, . . . , yn. The yi's may not be distinct, and do not necessarily arrive in increasing order. The problem is to design a summary that can answer the following query: given a parameter y provided at query time, how many yi's are less than or equal to y? In other words, estimate COUNT(y)=|{yiεY|yi<y}|. Given user parameters ε, 0<ε<1 and δ, 0<δ<1, an embodiment seeks an answer that is within an ε relative error of the exact answer with probability at least 1−δ.


The starting point for a solution is the “splittable histogram” data structure of prior work. This data structure stores histograms at multiple levels, i=0, 1, 2, . . . , such that the error due to using a data structure at level i is proportional to 2i, but as i increases, the set of possible queries that can be answered by the data structure at level i also become larger. The rough intuition here is that if a data structure is used at a large level (a large value of i), though the error is large (proportional to 2i), the answer must also be large, hence the relative error is controlled.


It is noted that that the data structures in prior work use many counters that may need to hold very large values. Observe that these counts need not be exact, and hence an embodiment uses probabilistic counters, which provide approximate counts of large numbers using significantly smaller space than exact counters. The error in the counters and the error in the data structures combine to give the total error in the estimate.


There is a technical difficulty in getting the above idea to work, which is that the states of the counters are used to trigger certain actions of the data structures, such as the formation of new nodes in the data structure. Replacing these counters with probabilistic counters leads to dealing with random variables and events that are not all independent of each other, and this dependency has to be handled carefully. Presented herein are further details of the process for doing this and the analysis.


Probabilistic Counters


A probabilistic counter is a data structure for approximate counting that supports two operations: increment and query. For positive integer M and γ>0, let PC(M, γ) denote a probabilistic counter with base (1+γ), that counts until a maximum estimated value of M. Let Cn denote the state of the counter after n increment operations. Using Cn, it is possible to return an estimate {circumflex over (n)} of n. If {circumflex over (n)}≧M, the counter cannot be incremented any further, and the counter is said to be “closed”. If {circumflex over (n)}<M, then the bucket is said to be “open”.


Theorem 1: Let {circumflex over (n)} denote the estimated count returned by the counter PC(M, γ) after n increments. Then E[{circumflex over (n)}]=n, and







VAR


[

n
^

]


=



γ






n


(

n
+
1

)



2

.






The counter can be stored in space no more than






(


log





log





M

+

log


1
γ



)





bits.


The above theorem concerns the estimates returned by the probabilistic counter. The number of increments needed for the counter to reach its maximum value need to be analyzed, so a slightly different type of guarantee is required.


Lemma 1: For C=PC(M, γ), let I(C) denote the number of insertions into C that causes it to be closed. Then E[I(C)]=M+1, and







VAR


[




(
C
)


]


=



γ






M


(

M
+
1

)



2

.






The proof of this lemma is in the appendix.


Data Structure


The data structure consists of histograms S0, S1, . . . , Slmax, where lmax=log(2εU), where U is an upper bound on the value of COUNT(y) for any value of y. Each histogram Si consists of no more than






α
=


4


(


log






y
max


+
1

)


ε






buckets.


Each bucket bεSi is a probabilistic counter







PC


(



2
i

+
1

,



ε
2


δ

α


)


.





The buckets form a rooted binary tree, where each internal node has exactly two children. The tree in Si is stored within an array by keeping the nodes of the tree in in-order traversal of the tree. This is unambiguous, by storing within each bucket an additional bit that says whether the bucket is a leaf in the tree, or an internal node.


Every bucket is responsible for a range of y values, which is represented implicitly by the position of the bucket in the tree. The “span” of a bucket b is the range of timestamps that the bucket b is responsible for. The span of the root of the tree is [0, ymax]. If the span of node v is [y1, y2], then the span of its left child is [y1, (y1+y2−1)/2], and the span of its right child is [(y1+y2+1)/2, y2]. In a given level l, the span of all buckets at the same depth in the tree are disjoint, but the span of buckets at different depth may overlap with each other. If they do overlap, then the span of one of the buckets must be a proper subset of the span of the other (overlapping) bucket. Each bucket has a probabilistic counter associated with it.


For bucket b, let span(b) denote the span of the bucket. Let left(b) denote the leftmost endpoint of span(b), and right(b) denote the rightmost endpoint of span(b). Let PC(b) to denote the counter associated with the bucket. The process for COUNT is illustrated in FIGS. 2-4.


Lemma 2: The space complexity of the data structure for F1 is






O


(


1
ε



(

log





ε





U

)



(

log






y
max


)



(


log





log





U

+

log


1
ε


+

log


1
δ



)


)






bits, where U is an upper bound on the value of COUNT(y). If log(ymax)=Θ(log(U)), then for constant ε and δ, the space complexity is O(log2 U log log U).


Proof: The space complexity is the space taken to store O(lmax, α) buckets. Each bucket b consists of a probabilistic counter PC(M, ε2δ/α), where the value of M is no more than U. The space complexity of the data structure follows from Theorem 1.


Definition 1: For l=0 . . . lmax, level l is defined to be incomplete for query y if Tl<y. If Tl≧y, then level l is said to be complete for query y.


For bucket b, let A(b) denote the number of stream items that were inserted into this bucket. Let EST(b) denote the current estimate of the number of items that were inserted into the bucket. This notation is extended to sets of buckets too. For any set of buckets B, all in the same level, let








A


(
B
)


=




b

B







A


(
b
)




,





and let







ℰ𝒮𝒯


(
B
)


=




b

B








ℰ𝒮𝒯


(
b
)


.






Lemma 3 Let Il be any set of internal nodes in Sl. Then EST(Il)=2l|Il|, E[A(Il)]=2l|Il|, and






Pr
[





A


(

I


)


-


EST


(

I


)






>

ε






EST


(

I


)




]



<


δ

2




I






.







Proof: For any bεIl, there is EST(b)=2l. Since








EST


(

I


)


=




b


I









EST


(
b
)




,





it follows that EST(Il)=2l|Il|. Further, A(b)=I(PC(2l−1, γ)). From Lemma 1, there is E[A(b)]=2l, and







VAR


[

A


(
b
)


]


=



γ






2




(


2


-
1

)


2

.










A


(

I


)


=




b


I









A


(
b
)







Using linearity of expectation yields:







E


[

A


(

I


)


]


=





b


I









E


[

A


(
b
)


]



=


2






I










For each bucket bεIl,






Pr
[





A


(
b
)


-


2






>

ε






2




]






VAR


[

A


(
b
)


]




ε
2



2

2







γ

2


ε
2









Using a union bound over all the buckets, and using







γ
=



ε
2


δ

α


,





the desired result is obtained.


Lemma 4: Let L be any set of leaves in Sl, Then E[EST(L)]=A(L), and

Pr[|EST(L)−A(L)|>εA(L)]<δ


Proof: For any bεL, there is E[EST(b)]=A(B), from Theorem 1. Using Linearity of Expectation,










E


[

EST


(
L
)


]


=



E
[




b

L




EST


(
b
)



]







=






b

L




E


[

EST


(
b
)


]









=






b

L




A


(
b
)









=



A


(
L
)









For the leaves, after the actual counts A(b)'s have been assigned, the random variables EST(b)'s are all independent. Hence:










VAR


[

EST


(
L
)


]


=






b

L




VAR


[

EST


(
b
)


]









=






b

L





γ






A


(
b
)




(


A


(
b
)


+
1

)


2








=




γ
2



(


A


(
L
)


+




b

L





A
2



(
b
)




)













γ
2



(


A


(
L
)


+


A
2



(
L
)



)




γ







A
2



(
L
)










Applying Chebyshev's inequality:







Pr


[





EST


(
L
)


-

A


(
L
)





>

ε






A


(
L
)




]





VAR


[

EST


(
L
)


]




ε
2




A
2



(
L
)






γ

ε
2


<
δ




Lemma 5: If level l is incomplete for query y, then with probability at least







1
-

δ

α
-
1



,





COUNT(y)≧(1−ε)2l−1(α−1).


Proof: Since Tl≦y, for each bucket bεSl, it must be true that right(b)<y. Thus each item that was inserted into bucket b in Sl must be contained in query y. Let Sl1 denote the internal nodes in Sl.

COUNT(y)≧A(Sl)≧A(Sl1)


Using Lemma 3, the following is true with probability at least






1
-


δ

2




S

1





.







A(Sl1)≧(1−ε)E[EST(Sl1)]=(1−ε)2l|Sl1|


Since Sl is a binary tree with α nodes, there are (α−1)/2 nodes in Sl1. Thus,







Pr
[


COUNT


(
y
)


>


(

1
-
ε

)



(

α
-
1

)



2


-
1




]

>

1
-

δ

α
-
1







Lemma 6: The estimate returned for COUNT(y) by process 3 has a relative error of less than 7ε with probability at least 1-4δ.


Proof: Suppose the process uses level l for its estimate. Let B1S denote the set of all nodes bεSl such that the span of b is a subset of [0, y]. The estimator uses the sum of the estimated counts of all nodes in B1. Let B1lB1 denote the internal nodes in B1, and B1LB1 denote the leaves in B1. Let COÛNT(y) denote the estimate returned by the process. Let B2Sl denote the set of all buckets bεSl such that the span of b intersects [0, y], but is not a subset of [0, y].

COU{circumflex over (N)}T(y)=EST(B1)=EST(B1l)+EST(B1L)
A(B1)≦COUNT(y)≦A(B1)+A(B2)  (1)


The following events are defined:

E1 is A(B2)≦(1+ε)(log ymax)2l
E2 is COUNT(y)≧(1−ε)2l−1(α−1)
E3 is EST(B1l)≧(1−ε)A(B1l)
E4 is EST(B1L)≧(1−ε)A(B1L)


First consider E1. Bounding A(B2), each bucket in B2 must span y. There can be no more than log ymax buckets in Sl that can span y, and each such bucket, except for the smallest bucket, must be an internal node of Sl. Let f2 denote the leaf bucket in B2, if it exists (note that f2 may not exist). Let B2′ denote the internal buckets in B2. If f2 does not exist, then A(f2) is defined to be 0.

A(B2)=A(f2)+A(B2′)


From Lemma 3, E[A(B2′)]=2l|B2′|, and










Pr


[


A


(

B
2


)


>


(

1
+
ε

)



2






B
2






]


<

δ

2




B
2










(
2
)







Next, it is shown that A(f2) is small, with high probability. If f2 does not exist, then A(f2)=0, so it can be assumed that f2 exists. In such a case, f2 cannot be a singleton bucket, since then its span would have been completely contained within [0, y]. Consider I(f2), which is the number of stream items that need to be inserted into f2 to close the counter. From Lemma 3: Pr[|Y(f2)−2l|>ε2l]≦δ. Thus, with probability at least (1−δ), it must be true that Y(f2)<(1+ε)2l. Consequently:

Pr[A(f2)>(1+ε)2l]≦δ  (3)


Let E5 denote the event A(B2′)≦(1+ε)2l|B2′| and Eεdenote the event A(f2)>(1+ε)2l. From Equations 2 and 3:







Pr


[



5




6


]




1
-

Pr


[



5

_

]


-

Pr


[



6

_

]





1
-
δ
-

δ

2




B
2









1
-

1.5

δ






Since |B2′|<log ymax, Pr[A(B2)>(1+ε)(log ymax)2l]<1.5δ. Thus:

Pr[E1]>1−1.5δ  (4)


From Lemma 5:










Pr


[


2

]




1
-

δ

α
-
1







(
5
)







Since B1l is a set of internal nodes in Sl, from Lemma 3,







Pr


[


3

]


=

Pr
[


ℰ𝒮𝒯


(

B
1
I

)





(

1
-
ε

)



A


(

B
1
I

)






A


(

B

3





I

)



1
+
ε




1
-


δ
2

.









Thus:










Pr


[


3

]




1
-

δ
2






(
6
)







From Lemma 4,

Pr[E4]>1−δ  (7)


Suppose that E1, E2, E3, E4 were true. From Equations 4-7, this is true with probability at least







1
-
δ
-

0.5

δ

-

1.5

δ

-

δ

α
-
1





1
-

4


δ
.


















COU


N
^



T
(
y
)


=




EST
(

B
1
I

)

+

EST
(

B
1
L

)














(

1
-
ε

)



A


(

B
1
I

)



+


(

1
-
ε

)



A


(

B
1
L

)




(
9
)














(

1
-
ε

)



A


(

B
1

)




(
10
)









COUNT
(
y
)






A


(

B
1

)


+

A


(

B
2

)















A


(

B
1

)


+


(

1
+
ε

)



(

log






y
max


)



2
l




,







(
8
)








since E1 is true. This is at most








A


(

B
1

)


+


COUNT


(
y
)






(

1
+
ε

)



2
l


log






y
max




(

1
-
ε

)



2

l
-
1




(

α
-
1

)





,





since E2 is true. This equals








A


(

B
1

)


+



ε


(

1
+
ε

)



2


(

1
-
ε

)





COUNT


(
y
)




,





using






α
=


4
ε




(


log






y
max


+
1

)

.







If ε≦½, then








1

1
-
ε




(

1
+

2

ɛ


)


,





and thus








1
+
ε


1
-
ε





(

1
+

2

ε


)



(

1
+
ε

)




(

1
+

4

ε


)











COUNT


(
y
)







A


(

B
1

)


+



(

ε
+

4


ε
2



)

2



COUNT


(
y
)
















A


(

B
1

)


+



3

ε

2



COUNT


(
y
)





,








where ε≦½ was used.










COUNT


(
y
)





A


(

B
1

)



1
-

1.5

ε






(

1
+

6

ε


)



A


(

B
1

)







(
11
)








From Equations 11 and 8:










COU


N
^



T


(
y
)









1
-
ε


1
+

6

ε





COUNT


(
y
)














(

1
-
ε

)



(

1
-

6

ε


)



COUNT


(
y
)














(

1
-

7

ε


)



COUNT


(
y
)










This proves that the estimate is within a 7ε relative error of COUNT. The other direction is proved similarly, by noting the following.










COU


N
^



T


(
y
)



=




EST


(

B
1
I

)


+

EST


(

B
1
L

)















B
1
I


1
-
ε


+


(

1
+
ε

)



B
1
L















(

1
+

2

ε


)



B
1
I


+


(

1
+
ε

)



B
1
L






Using





ε




1
/
2












(

1
+

2

ε


)



B
1













(

1
+

2

ε


)



COUNT


(
y
)







Using





Equation





1








Thus, with probability at least 1-4δ, the estimated count is within a 7ε relative error of the actual.


Theorem 2: processes 1-3 yield an (ε, δ)-approximation for the correlated basic count, using space (in bits)






O
(


1
ε



(


log
2


U

)




(


log





log





U

+

log


(

1
/
ε

)


+

log


(

1
/
δ

)



)

.







The space complexity is within a factor of O(log log U+log(1/ε)+log(1/δ))) of the optimal. The amortized processing time per record can be made O(log n). In the above expression, it is assumed that log(ymax)−O(log U).


Proof: The space complexity and correctness follow from Lemma 2 and Lemma 6, respectively. In prior work, it was shown that any data structure for basic counting for a (synchronous) sliding window with an ε relative error needs






Ω
(



log
2


U

ε

)





bits of space. Since basic counting over a synchronous sliding window is a special case of correlated basic count estimation, the space lower bound of






Ω
(



log
2


U

ε

)





holds here too. Hence, the achieved space bound is within a factor of O(log log U+log(1/ε)+log(1/δ))) of the optimal. To achieve O(log n) amortized processing time, observe that there are O(log n) data structures S0, S1, . . . , Slmax, each containing O(ε−1 log n) buckets, each containing a probabilistic counter. These probabilistic counters are packed into words of size Θ(log n), that is, the standard RAM model. Processing a batch of O(ε−1 log n) counts together as follows. First, the batch is sorted in order of non-decreasing y-coordinate. This can be done in O(ε−1 log n(log 1/ε+log log n)) time. Then, the following for each Si is done, sequentially, so as to keep within the space bound. For data structure Si, unpack it, putting each probabilistic counter into a word. This can be done in O(ε−1 log2 n) bits of space and O(ε−1 log n) time. Further, the data structure is unpacked and represented as a list in a pre-order traversal of the tree that Si represents. Then, an embodiment walks through the list and updates the appropriate buckets. Since the updates are sorted in increasing y-value and the list is represented as a pre-order traversal, this can be done in O(ε1 log n) time. Hence, as an embodiment ranges over all Si, the total processing time is O(ε−1 log n(log 1/ε+log log n+log n))=O(ε−1 log2 n), or equivalently, the amortized update time is O(log n).


General Scheme for Correlated Aggregation


A general scheme for constructing a summary for correlated aggregation for any aggregation function that satisfies a certain set of properties is described. For any such aggregation function, it is described how to reduce the construction of a sketch for correlated aggregation to the construction of a sketch for aggregation over the entire stream. This reduction allows the use of previously known stream summaries for aggregation over streams in constructing summaries for correlated aggregation.


An embodiment employs this scheme to construct small space summaries for estimating correlated frequency moments over a data stream. For k>0, the k-th frequency moment of a stream of identifiers, each assumed to be an integer from {1, . . . , m}, is defined as Fki=1mfik, where fi is the number of occurrences of the i-th item. The estimation of the frequency moments over a data stream has been the subject of much study; however, embodiments provide processing for estimating the correlated frequency moments with provable guarantees on the relative error. The memory requirements are optimal up to small factors, namely factors that are logarithmic in m and the error probability δ, and factors that are polynomial in the relative error parameter ε.


The technique for general correlated aggregation described herein builds on prior work for the correlated estimation of the basic count. While the basic data structure used in prior work is an exact counter, embodiments rather employ a “sketch”, which can accurately estimate an aggregate on a stream, and provide a (possibly) probabilistic guarantee on the correctness. The error due to the randomized sketches employed by embodiments is analyzed, as well as the combination of different sketches. A central new issue is that of dependency between random variables.


Also presented in detail are approaches for the first space-efficient processing for correlated aggregates of the number of distinct elements (F0), and other aggregates related to frequency moments, such as the Fk-heavy hitters and rarity. A technique for achieving lower amortized update time is provided. For correlated F2, the update time is Õ(log n) per stream update.


General Streaming Models and Lower Bounds


The case of streams where items can have an associated positive or negative integer weight is described. Allowing negative weights is useful for analyzing the symmetric difference of two data sets, since the items in the first data set can be inserted into the stream with positive weights, while the items from the second data set can be inserted into the stream with negative weights.


In this model, each stream element is a 3-tuple (xi, yi, zi), where zi specifies the weight of the item. It is shown that, even if ziε{−1, 1} for all i, then for a general class of functions that includes the frequency moments, any summary that can estimate correlated aggregates accurately and that is constructed in a single pass must use memory that is linear in the size of the stream. This is to be contrasted with the estimation of frequency moments over a stream in the non correlated case, where it is known how to estimate these aggregates in sublinear space even in the presence of positive and negative weights.


Also described is a model with arbitrary positive and negative weights in which multiple passes over the stream are allowed. This more general model allows the processing to make a small (but more than one) number of passes over the stream and store a small summary of what is seen. At a later time a query is made and must be answered using only the summary. Such a setting arises if data is collected and stored from previous days for the purpose of creating a summary, but is then deleted or archived while only the summary is retained.


In this case, a smooth pass-space tradeoff is obtained for these problems, in that with a logarithmic number of passes there are space-efficient algorithms for a large class of correlated aggregates even with negative weights, but with fewer passes no such space-efficient algorithms can exist.


Aggregation Over Asynchronous Streams


A closely related problem is that of computing aggregates over a sliding window on an asynchronous stream. In this scenario, a stream of (v1, t2) tuples, where v is a data item, and t is the timestamp at which it was generated. Due to asynchrony in the transmission medium, it is possible that the stream elements do not arrive in the order of the time-stamps. In other words, it is possible that t1<t2, but that (v1, t1) is received later than (v2, t2). This was not possible in the traditional definition of count-based or time-based sliding windows. There is a straightforward reduction from the problem of aggregating over asynchronous streams to that of computing correlated aggregates, and vice versa, from previous work. Hence, all of the results for correlated aggregates described herein apply also to aggregation over a sliding window on an asynchronous stream with essentially the same space and time bounds. Thus, embodiments achieve the first low-space solutions for asynchronous streams for a wide class of statistics.


Correlated Estimation of a Class of Statistics


Embodiments provide for the estimation of correlated, aggregates for any aggregation function that satisfies a set of properties. Consider an aggregation function ƒ that takes as input a multi-set of real numbers R and returns a real number ƒ(R). In the following, the term “set of real numbers” is used to mean a “multi-set of real numbers”. Also, union of sets implies a multi-set union, when the context is clear.


For any set of tuples of real numbers T={(xi, yi)|1≦i≦n} and real number y, let ƒ(T, c) denote the correlated aggregate ƒ({xi|((xi, yi)εT)custom character(yi≦c)}). For any function ƒ satisfying the following properties, it is shown herein how a reduction from the problem of space-efficient estimation of correlated aggregate ƒ(T, c) to the problem of space-efficient estimation of ƒ in an uncorrected manner on the entire set R.


The following definition of an (ε, δ) estimator is used herein. Definition 2: Given parameters ε, δ, where 0<ε<1, and 0<δ<1, an (ε, δ) estimator for a number Y is a random variable X such that, with probability at least 1−δ, the following is true:

(1−ε)Y≦X≦(1+ε)Y.


In the following description, the term “sketching function” denotes a compression of the input set with certain properties. More precisely, ƒ has a sketching function skƒ(v, γ, R) if:


Using skƒ(v, γ, R) it is possible to get an (v, γ)-estimator of ƒ(R); and


For two sets R1and R2, given sk(v, γ, R1) and sk(v, γ, R2), it is possible to compute sk(v, γ, R1∪R2).


Many functions ƒ have sketching functions. For instance, the second frequency moment F2 and the k-th frequency moment Fk have sketches due to prior work. In these two examples, the sketching function is obtained by taking random linear combinations of the input.


Embodiments require the following conditions from ƒ in order to construct a sketch for estimating correlated aggregates. These conditions intuitively correspond to “smoothness” conditions of the function ƒ, bounding how much ƒ can change when new elements are inserted or deleted from the input multi-set. Informally, the less the function value is sensitive to small changes, the easier it is to apply to estimating correlated aggregates.

    • I. ƒ(R) is bounded by a polynomial in |R|
    • II. For sets R1 and R2, ƒ(R1∪R2)≧ƒ(R1)+ƒ(R2)
    • III. There exists a function c1ƒ(·) such that for sets R1, . . . Rj, if ƒ(Ri)≦α for all i=1 . . . j, then: ƒ(∪i=1jRi)≦c1ƒ(j)·α
    • IV. For ε<1, there exists a function c2ƒ(ε) with the following properties. For two sets A and B such that BA, if ƒ(B)≦c2ƒ(ε)·ƒ(A), then ƒ(A−B)≧(1−ε)ƒ(A)
    • V. ƒ has a sketching function skƒ(γ, v, R) where γε(0, 1) and v ε(0, 1).


For any function ƒ with a sketch skƒ with the above properties, it is shown herein how to construct a sketch skƒcor(ε, δ, T) for estimating the correlated aggregate ƒ(T, c) with the following properties:

    • A. Using skjcor(ε, δ, T), it is possible to get an (ε, δ)-estimator of ƒ(T, c) for any real c>0
    • B. For any triple (x, y), using skƒcor(γ, ε, T), it is possible to construct skƒcor(γ, ε, T∪{(x, y)}).


Processing Description


Referring to FIG. 5, let ƒmax denote an upper bound on the value of ƒ(·,·) over all input streams considered. The process (FIGS. 5-7) uses a set of levels l=0, 1, 2, . . . , lmax, where lmax is such that 2lmaxmax for input stream T and real number c. From Property 1, it follows that lmax is logarithmic in the stream size.


Parameters α, γ, v are chosen as follows:







α
=


64



c
1
f



(

log






y
max


)





c
2
f



(

ε
/
2

)




,

v
=

ε
2


,

γ
=

δ

4



y
max



(

l
+
1

)





,





where ymax is the largest possible y value.


Without loss of generality, assume that ymax is of the form 2β−1 for some integer β. The dyadic intervals within [0, ymax] are defined inductively as follows:

    • (1) [0, ymax] is a dyadic interval;
    • (2) If [a, b] is a dyadic interval and a≠b, then [a, (a+b−1)/2] and [(a+b+1)/2, b] are also dyadic intervals.


Within each level l, from 0 to lmax, there is a “bucket” for each dyadic interval within [0, ymax]. Thus, there are 2ymax−1 buckets in a single level. Each bucket b is a triple custom characterk(b), l(b), r(b)custom character, where [l(b), r(b)] is a dyadic interval, that, corresponds to a range of y values that this, bucket is responsible for, and k(b) is defined below.


When a stream element (x, y) arrives, it is inserted into each level l=0, . . . , lmax. Within level l, it is inserted into exactly one bucket, as described in Process 5. For a bucket b in level l, let S(b) denote the (multi-)set of stream items that were inserted into b. Then, k(b)=skƒ(v, γ, S(b)) is a sketch of S(b).


Within each level, no more than α of the 2ymax−1 buckets are actually stored. In the process, Sl denotes the buckets that are stored in level l. The level S0 is a special level that just consists of singletons. Among the buckets that are not stored, there are two types of buckets, those that were discarded in Process 5 (see the “Check for overflow” comment in Process 5), and those that were never used by the process. The above three types of buckets are termed “stored”, “discarded”, and “empty”, respectively. Note that S(b) is defined for each of these three types of buckets (if b is an empty bucket, then S(b) is defined as the null set φ).


The buckets in Sl are organized into a tree, induced by the relation between the dyadic intervals that these buckets correspond to.


The initialization for the process for a general function is described in Process 4, illustrated in FIG. 5. The update and query processing are described in Processes 5 and 6, illustrated in FIGS. 6(A-B) and 7, respectively.


Theorem 3 (Space Complexity): The space complexity of the sketch for correlated estimation is







O


(





c
1
f



(

log






y
max


)


·

(

log






f
max


)




c
2
f



(

ε
2

)



·
len

)


,





where






len
=




sk
f



(


ε
2

,

δ

4



y
max



(

2
+

log






f
max



)




,
S

)









is the number of bits needed to store the sketch.


Proof: There are no more than 2+log fmax levels and in each level l, Sl stores no more than α buckets. Each bucket b contains skƒ (v, γ, S(b)). The space complexity is α(2+log ƒmax) times the space complexity of sketch skf. Here it is assumed that the space taken by skƒ is larger than the space required to store l(b) and r(b).


Process Correctness


Let S denote the stream of tuples observed up to a given point. Suppose the required correlated aggregate is ƒ(S, c). Let A be the set

{xi|((xi, yiS)custom character(yi≦c)}.


Thus, ƒ(S, c)=ƒ(A).


For level l, 0≦l≦lmax, B1l and B2l are defined as follows. Let B11 denote the set of buckets b in level l such that span(b)[0, c]. Let B2l denote the set of buckets b in level l such that span(b)⊂/[0, c], but span(b) has a non-empty intersection with [0, c].


Note that for each level l, B1l and B2l are uniquely determined once the query ƒ(S, c) is fixed. These do not depend on the actions of the process. This is an important property used, which allows the choice of which buckets to use during estimation to be independent of the randomness in the data structures. Further, note that only a subset of B1l and B2l is actually stored in Sl.


Consider any level l, 0≦l≦lmax. For bucket b, recall that S(b) denotes the set of stream items inserted into the bucket until the time of the query. For bucket bεSl, let f(b) denote f(S(b)). Let estf(b) denote the estimate of f(b) obtained using the sketch k(b). If S(b)=φ, then f(b)=0 and estf(b)=0. Thus note that f (b) and estf(b) are defined no matter whether b is a stored, discarded, or an empty bucket. Further, for a set of buckets B in the same level, let S(B)=∪bεBS(b), and let f(B)=f(S(B)). Let estf(B) be the estimate for f(B) obtained through the composition of all sketches in ∪bεBk(b) (by property V, sketches can be composed with each other).


Definition 3: Bucket b is defined to be “good” if (1−v)f(b)≦estf(b)≦(1+v)f(b). Otherwise, b is defined to be “bad”.


Let G denote the following event: each bucket b in each level 0 . . . lmax is good.


Lemma 7:







Pr


[
G
]




1
-

δ
2






Proof: For each bucket b, note that estf(b) is a (v, γ)-estimator for f(b). Thus, the probability that b is bad is no more than γ. Noting that there are less than 2ymax buckets in each level, and lmax+1 levels in total, and applying a union bound, yields:








Pr




G
_






2



y
max



(


l





max


+
1

)



γ


=

δ
2





Lemma 8: For any level l, S(B1l)AS(B1l∪B2l).


Proof: Every bucket bεB1l must satisfy span(b)ε[0, c]. Thus every element inserted into B1l must belong in A. Hence S(B1l)A. Each element in A has been inserted into some bucket in level l (it is possible that some of these buckets have been discarded). By the definitions of B1l and B2l, an element in A cannot be inserted into any bucket outside of B1l∪B2l. Thus AS(B1l∪B2l).


Using Lemma 8 and Condition II on f yields the following for any level l:

f(B1l)≦f(A)≦f(B1l∪B2l).  (12)


Lemma 9: Conditioned on event G, process 6 does not output FAIL in step 1.


Proof: Consider lmax. Ylmax>c if event G occurs. Observe that Ylmax is initialized to ∞ in Process 4. Its value can only change if the root b of Slmax closes. For this to happen, there must be est(k(b))≧2lmax+1. But 2lmax+1>2fmax, which means that est(k(b)) does not provide a (1+ε)-approximation. This contradicts the occurrence of event G. Hence, Ylmax>c and so Process 6 does not output FAIL in step 1.


Let l* denote the level used by Process 6 to answer the query f(S, c).


Lemma 10: If l*≧1 and G is true, then

f(B2l*)≦c1f(log ymax)2l*+2.


Proof: First, note that there can be no singleton buckets in B2l* by definition of B2l for a level l. Thus, for each bucket bεB2l*, estf (b)≦2l*+1. Because G is true, for every bucket bεB2l*, b is good, so that







f


(
b
)






2


l
*

+
1



1
-
v


.





Next, note that there are no more than log ymax buckets in B2l*, since there can be only one dyadic interval of a given size that intersects [0, c] but is not completely contained within [0, c].


From Property III:







f


(

B
2

l
*


)


=


f


(




b


B
2

l
*






S


(
b
)



)






c
1
f



(

log






y

m





ax



)


·



2


l
*

+
1



1
-
v


.








Since v≦½, the desired result is obtained.


Lemma 11: If l*≧1 and G is true, then:

f(A)≧α2l*−4.


Proof: Since the process used level l* for answering the query, it must be the case that there are buckets in Sl*−1 that had an intersection with [0, c] but were discarded from the data structure. It follows that there are at most log ymax buckets bεSl*−1 such that span(b)⊂/[0, c]. For the remaining buckets bεSl*−1, it must be true that span(b)⊂[0, c]. If Sl*−1 is viewed as a binary tree with α nodes, according to the ordering between the different dyadic intervals, then Sl*−1 must have (α−1)/2 internal nodes. Suppose I denoted the set of buckets in bεSl*−1 such that b is an internal node, and span(b)⊂[0, c]. Thus |I|≧(α−1)/2−log ymax. Since G is true, for any bucket







b

I

,


f


(
b
)






2


l
*

-
1



1
+
v


.







Using property II repeatedly yields:







f


(
A
)




f


(
I
)







I




2


l
*

-
1




1
+
v






Using v<1, and for an appropriately large value of α, ((α−1)/2−log ymax)≧α/4. Combining the above yields the following:








f


(
A
)





α






2

l
*




2
·
4
·
2



=


2


l
*

-
4




α
.






Theorem 4: When presented with a query for f(S, c), let est denote the estimate returned by the process. Then, with probability at least 1−δ:

(1−ε)f(S, c)≦est≦(1+ε)f(S, c).


Proof: If l*=0, then all elements (x, y)εS such that y≦c are stored in S0. In this case, the theorem follows by the definition of event G and Lemma 1. Otherwise, est=estf(B1l*), and f(S, c)=f(A). First note that in level l*, none of the buckets in B1l* have been discarded. Thus each bucket bεB1l* is either empty or is stored. Thus, it is possible to execute line 7 in Process 6 correctly to construct a sketch of S(B1l*). From property (b) of sketching functions yields a sketch sk(v, γ, S(B1l*)).


Let ε1 denote the event (1−v)ƒ(B1l*)≦estƒ(B1l*)≦(1+v)ƒ(B1l*). Thus:

Pr[ε1]≧1−γ


The following is conditioned on both ε1 and G occurring. From Equation 1: ƒ(A)≦ƒ(B1l*∪B2l*). From Lemmas 4 and 5:








f


(

B
2

l
*


)



f


(
A
)








c
1
f



(

log






y

m





ax



)




2


l
*

+
2




α






2


l
*

-
4









c
1
f



(

log






y

m





ax



)




2
6


α




c
2
f



(

ε
2

)







where the value of α has been substituted.


Since (A−B1l*)B2l*, the following is true:








f


(

B
2

l
*


)



f


(
A
)







c
2
f



(

ε
2

)


.






Using Property IV yields:







f


(

B
1

l
*


)


=


f


(

A
-

(

A
-

B
1

l
*



)


)





(

1
-

ε
2


)



f


(
A
)








Conditioned on ε1 and G both being true:

estƒ(B1l*)≧(1−ε/2)(1−v)ƒ(A)≧(1−ε)ƒ(A)


This proves that conditioned on G and ε1, the estimate returned is never too small. For the other direction, note that conditioned on ε1 being true:

estƒ(B1l*)≦(1+v)ƒ(B1l*)≦(1+v)ƒ(A)≦(1+ε)ƒ(A)

where ƒ(B1l*)≦ƒ(A) has been used, and v<ε.


To complete the proof of the theorem, note that:








1
-

Pr


[

G
_

]


-

Pr


[


ɛ
1

_

]









1
-

δ
2

-

γ





using





Lemma





7








1
-

δ





using





γ


<

δ
/
2.





Frequency Moments Fk


The general technique presented herein can yield a data structure for the correlated estimation of the frequency moments Fk, k≧2.


Fact 1: (Hölder's Inequality) For vectors a and b of the same dimension, and any integer k≧1, custom charactera, bcustom character≦∥a∥k·∥b∥k/(k−1).


Lemma 12: For sets Si, i=1 . . . j, if Fk(Si)≦β for each i=1 . . . j, then Fk(∪i=1jSi)≦jkβ.


Proof: Fact 1, j-dimensional vectors a and b implies that |custom charactera, bcustom character|k≦∥a∥kk·∥b∥k/(k−1)k. Setting B=(1, 1, . . . , 1), it follows that (a1+ . . . +aj)k≦jk−1(a1k+ . . . +ajk). Hence, it follows that








F
k



(




i
=
1

j





S
i



)





j

k
-
1







i
=
1

j




F
k



(

S
i

)







j
k



β
.






Lemma 13: If Fk(B)≦(ε/(3k))kFk(A), then Fk(A∪B)≦(1+ε)Fk(A).


Proof: Suppose A and B have support on {1, 2, . . . , n}. Let a and b be the characteristic vectors of sets A and B, respectively. Using Fact 1 yields:











F
k



(

A

B

)


=






i
=
1

n




(


a
i

+

b
i


)

k








=





F
k



(
A
)


+


F
k



(
B
)


+




i
=
1

n






j
=
1


k
-
1





(



k




j



)



a
i
j



b
i

k
-
j












=





F
k



(
A
)


+


F
k



(
B
)


+




j
=
1


k
-
1





(



k




j



)






i
=
1

n




a
i
j



b
i

k
-
j



















F
k



(
A
)


+


F
k



(
B
)


+




j
=
1


j
-
1





(



k




j



)




(




i
=
1

n




(

a
i
j

)


k
/
j



)


j
k














(




i
=
1

n




(

b
i

k
-
j


)


k
/

(

k
-
j

)




)







=





F
k



(
A
)


+


F
k



(
B
)


+




j
=
1


k
-
1





(



k




j



)





F
k



(
A
)



j
k






F
k



(
B
)




k
-
j

k

















F
k



(
A
)


+


F
k



(
B
)


+




j
=
1


k
-
1





(



k




j



)




F
k



(
A
)





(

ε

3

k


)


k
-
j

















(

1
+

ε
/
3


)




F
k



(
A
)



+



F
k



(
A
)







j
=
1


k
-
1





(



k




j



)




(

ε
/

(

3

k

)


)


k
-
j


















(

1
+

ε
/
3


)




F
k



(
A
)



+



F
k



(
A
)





(

1
+

ε
/

(

3

k

)



)

k


-


F
k



(
A
)















(

1
+

ε
/
3


)




F
k



(
A
)



+



F
k



(
A
)




(

1
+

2


ε
/
3



)


-


F
k



(
A
)















(

1
+
ε

)




F
k



(
A
)




,








where (1+x)y≦exy was used for all x and y, and ez≦1+2z for z≦½. This completes the proof.


Lemma 14: If C⊂D, and Fk(C)≦(ε/(9k))kFk(D), then Fk(D−C)≧Fk(D).


Proof: It is know that for any two sets A and B, Fk(A∪B)≦2k(Fk(A)+Fk(B)).

Fk(D)=Fk((D−C)∪(C)≦2k(Fk(D−C)+Fk(C)), which leads to:
Fk(D−C)≧Fk(D)/2k−Fk(C)≧(9k/ε)kk)k−1)Fk(C)≧(3k/ε)kFk(C). Thus, Fk(C)≦(ε/3k)kFk(D−C). Applying Lemma 7 yields
Fk(C∪(D−C))≦(1+ε)Fk(D−C). Thus,
Fk(D−C)≧Fk(D)/(1+ε)≧(1−ε)Fk(D).


Theorem 5: For parameters 0<ε<1 and 0<δ<1, there is a sketch for an (ε, δ)-estimation of the correlated aggregate Fk on a stream of tuples of total length n, using spacer n1−2/kpoly(ε−1 log(n/δ)).


Proof: From Lemma 12, c1Fk(j)=jk. From Lemma 8, c2Fk(ε)=(ε/(9k))k. Using these in Theorem 3 yields c1Fk(log ymax)=(log ymax)k, and c2ƒ(ε/2)=(ε/(18k))k. Using the sketches for F2 and Fk, k>2 from prior work, the above result is obtained.


The space can be improved to r1−2/kpoly(ε−1 log(n/δ)), where r is the number of distinct xi-values in the stream. In the worst-case, though, r could be Θ(n). The dependence is made more explicit for the case of F2.


Lemma 15: For parameters 0<ε<1 and 0<δ<1, there is a sketch for (ε, δ) error correlated estimation of F2 on a stream of tuples of total length n, using space O(ε−4(log(1/δ)+log ymax)(log2 ymax)(log2 ƒmax)) bits. The amortized update time is O(log ƒmax·log ymax).


Proof: The space taken by a sketch for an (ε, δ) estimator for F2 on a stream is O((log ƒmax)(1/ε2)log(1/δ)) bits. From the proof of Theorem 5, c1F2(j)=j2, and c2F2(ε)=(ε/18)2.


Using the above in Theorem 3, the space is O(ε−4 log2 ƒmax log2 ymax(log 1/δ+log ymax)) bits.


To get O(log ƒmax(log 1/δ+log ymax)) amortized processing time, observe that there are O(log ƒmax) data structures Si, each containing O(ε−2 log2 ymax) buckets, each holding a sketch of O(ε−2 log ƒmax(log 1/δ+log ymax)) bits. A batch of O(ε−2 log2 y max) updates is processed at once. The batch is first sorted in order of non-decreasing y-coordinate. This can be done in O(ε−2 log2 ymax(log 1/ε+log log ymax)) time. Then, the following processing is done for each Si. A pre-order traversal of the buckets in Si is conducted and the appropriate buckets are updated. Importantly, each bucket maintains an update-efficient AMS sketch (shown in prior work), which can be updated in time O(log 1/δ+log ymax). Since the updates are sorted in increasing y-value and the list is represented as a pre-order traversal, the total time to update Si is O(ε−2 log2 ymax(log 1/δ+log ymax)). The time to update all the Si is O(log ƒmax) times this. So the amortized time is O(log ƒmax(log 1/δ+log ymax)).


Other Useful Statistics


While many aggregation functions satisfy the properties described above, some important ones do not. However, in many important remaining cases, these aggregation functions are related to aggregation functions that do satisfy these properties, and the mere fact that they are related in the appropriate way enables efficient estimation of the corresponding correlated aggregate.


The correlated F2-heavy hitters can be computed, as well as the rarity (defined below) by relating these quantities to F2 and F0, respectively.


For example, in the correlated F2-heavy hitters problem with y-bound of c and parameters ε, φ, 0<ε<φ<1, letting F2(c) denote the correlated F2-aggregate with y-bound of c, to return all x for which |{xi, yi)|xi=xcustom characteryi≦c}|2≧φF2(c), and no x for which |{(xi, yi)|xi=xcustom characteryi≦c}|2≦(φ−ε)F2(c). To do this, the same data structures Si may be used as used for estimating the correlated aggregate F2. However, for each Si and each bucket in Si maintain a process is maintained for estimating the squared frequency of each item inserted into the bucket up to an additive (ε/10)·2i. To estimate the correlated F2-heavy hitters, for each item an additive (ε/10)·F2(c) approximation to its squared frequency is obtained by summing up the estimates provided for it over the different buckets contained in [0, c] in the data structure Si used for estimating F2(c). Since only an ε/10 fraction of F2(c) does not occur in such buckets, the list of all heavy hitters is obtained this way, and no spurious ones.


In the rarity problem, the problem is to estimate the fraction of distinct items that occur exactly once in the multi-set. The ideas for estimating rarity are similar, where the same data structures Si are maintained for estimating the correlated aggregate F0, but in each bucket maintain data structures for estimating the rarity of items inserted into that bucket.


Number of Distinct Elements


The number of distinct elements in a stream, also known as the zenith frequency moment F0, is a fundamental and widely studied statistic of the stream. Here, the correlated estimation of the number of distinct elements in a stream is described. Consider a stream of (x, y) tuples, where xε{1, . . . , m} and yε{1, ymax}. The process employed by an embodiment is an adaptation of the process for estimating the number of distinct elements within a sliding window of a data stream, due to prior work. Similar to prior work, the process for correlated estimation of F0 is based on “distinct sampling”, or sampling based on the hash values of the item identifiers. Multiple samples, S0, S1, . . . , Sk, are maintained, where k=log m. Suppose that for simplicity, there is a hash function h that maps elements in {1, . . . , m} to the real interval [0, 1]. This assumption of needing such a powerful hash function can be removed, as shown in prior work.


The process of prior work proceeds as follows. Stream items are placed in these samples Si in the following manner. (A) Each item (x, y) is placed in S0. (B) For i>0, an item (x, y) is placed in level i if







h


(
x
)


<=


1

2
i


.






Note that if an item x is placed in level i, it must have been placed in level i−1 also.


Since each level has a limited space budget, say α, a way to discard elements from each level is needed. The process according to an embodiment differs from prior work in the following aspect of how to discard elements from each level. For correlated aggregates, maintained in Si are only those items (x, y) that (1) have an x value that is sampled into Si, and (2) have the smallest values of y among all the elements sampled into Si. In other words, it is a priority queue using the y values as the weights, whereas in prior work, each level was a simple FIFO (first-in-first-out) queue.


It can be shown that the above scheme of retaining those elements with a smaller value of y, when combined with the sampling scheme in prior work, yields an (ε, δ) estimator for the correlated distinct counts.


Theorem 6: Given parameters 0<ε<1 and 0<δ<1, there is a streaming process that can maintain a summary of a stream of tuples (x, y), where at xε{1, . . . , m} and yε{1, ymax} such that: (1) The space of the summary is











O


(


log





m

+

log






y

ma





x




)









log





m


ε
2



log






1
/
δ






bits

;




(
2
)








The summary can be updated online as stream elements arrive; and (3) Given a query y0, the summary can return, an (ε, δ)-estimator of the number of distinct x values among all tuples (x, y) with y values satisfying y≧y0.


Example Implementations


Correlated F2, Second Frequency Moment


An example process for correlated F2 estimation was implemented in Python. The following four data sets were used for example evaluations. (1) Packet traces of Ethernet traffic on a LAN and a WAN. The relevant data here is the number of bytes in the packet, and the timestamp on the packet (in milliseconds). This data set is referred to as the “Ethernet” data set. This data set has 2 million packets, and was constructed by taking two packet traces and combining them by interleaving. (2) The Uniform data set, which is a sequence of tuples (x, y) where x is generated uniformly at random from the set {0, . . . , 10000} and y is generated uniformly at random from the set {0, . . . , 213−1}. This maximum size of this data set is 107. (3) The Zipfian data set, with α=1. Here the x values are generated according to the Zipfian distribution with α=1, from the domain {0, . . . , 10000}, and the y values are generated uniformly at random from the set {0, . . . , 213−1}. The maximum size of this dataset is 107. (4) The Zipfian data set as described above, with α set to 2.


For the sketch for F2, the implementation used a variant of a process developed in prior work. An observation was that for all cases tested, the relative error of the process was almost always within the desired approximation error ε, for δ<0.2.


Space Usage as a Function of ε


The space consumption of the process was studied. The space depends on a number of factors, including the values of ε, and δ. The space also depends on ƒmax, since ƒmax determines the maximum size of the data structure at each level. But the most obvious dependence is on ε, and this dependence is explored further.


In FIG. 8, the space taken for the summary for F2 is plotted as a function of ε. This is illustrated for all the data sets described above, with each dataset of size 2 million tuples. It is noted that the space taken by the sketch increases rapidly with decreasing ε, and the rate of the growth is similar for all four datasets. For smaller values of ε (ε≦0.1), the summary data structure took even more space than the input stream itself. This is not surprising since the space increases as the fourth power of







1
ε

.




It can be observed that the space taken for the Ethernet dataset is greater than the space for the Uniform and Zipfian datasets. The reason for this is that the range of the y values in the Ethernet dataset was much larger (0 to 3200000) than in the Uniform and the Zipfian datasets, where the y values ranged from 0 to 8191. The larger range of y values meant that the number of nodes at each level is larger. In addition, a larger value of ymax also leads to a larger value of ƒmax, for the Ethernet dataset than for the other datasets, and hence also to a larger number of levels of data structures. However, note that the rate of growth of the space with respect to ε remains similar for all data sets. For smaller values of ε (ε≦0.15), the summary data structure often took even more space than the input stream itself. This is not surprising since the space increases as the fourth power of







1
ε

.





As described herein, the space savings of the sketch improves as the stream size increases.


Space Usage as a Function of the Stream Size


Example results are illustrated in FIG. 9 (for ε=0.15), FIG. 10 (for ε=0.2), and FIG. 11 (for ε=0.25). In all cases, as predicted by theory, the space taken by the sketch does not change much, and increases only slightly as the stream size increases. This shows that the space savings of the process is much larger with streams that are larger in size.


The time required for processing the stream of 2 million elements was nearly the same (about 5 minutes) for the Uniform, Zipfian, as well as Ethernet datasets. The processing rate can be improved by using the C/C++ language, and by using a more optimized implementation.


These example results illustrate that a reasonable processing rate can be achieved for the data structure for F2, and that correlated query processing is indeed practical, and provides significant space savings, especially for large data streams (of the order of 10 million tuples or larger).


Correlated F0, Number of Distinct Elements


An example of the process for F0 in Python and four data sets, as described in the case of F2, was also implemented. The only difference was that in the Uniform and Zipfian datasets, the range of x values was much larger (0 . . . 1000000) than in the case of F2, where the range of x values was 0 . . . 10000. The reason for this change is that there are much simpler approaches for correlated F0 estimation when the domain size is small: simply maintain the list of all distinct elements seen so far along the x dimension, along with the smallest value associated with it in the y dimension. Note that such a simplification is not (easily) possible for the case of F2.


The variation of the sketch size with ε is shown in FIG. 12. Note that while the sketch size decreases with increasing ε, the rate of decrease was not as fast as in the case of F2. Further, note that the sketch size for comparable values of ε was much smaller than the sketch for correlated F2. Another point is that, the space taken by the sketch for the Ethernet dataset was significantly smaller than the sketch for the other datasets. This is due to the fact that the range of x values in the Ethernet dataset was much smaller (0 . . . 2000) than for the other datasets (0 . . . 1000000). The number of levels in the data structure is proportional to the logarithm of the number of possible values along the x dimension. Note that as explained above, the process of choice for correlated F0 estimation for the Ethernet-type datasets (where the x range is small) will be different, as explained herein. Thus, the process is particularly useful for datasets where the x range is much larger.


The size of the sketch as a function of the stream size is shown in FIG. 13, for ε=1. It can be seen that the sketch size hardly changed with the stream size. Note however, that for much smaller streams, the sketch will be smaller, since some of the data structures at different levels have not reached their maximum size yet. The results for other values of ε are similar, and are not shown here.


Accordingly, embodiments provide for computing correlated aggregates over a data stream. Referring to FIG. 14, it will be readily understood that embodiments may be implemented using any of a wide variety of devices or combinations of devices. An example device that may be used in implementing embodiments includes a computing device in the form of a computer 1410. In this regard, the computer 1410 may execute program instructions configured to compute correlated aggregates over a data stream, and perform other functionality of the embodiments, as described herein.


Components of computer 1410 may include, but are not limited to, at least one processing unit 1420, a system memory 1430, and a system bus 1422 that couples various system components including the system memory 1430 to the processing unit(s) 1420. The computer 1410 may include or have access to a variety of computer readable media. The system memory 1430 may include computer readable storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) and/or random access memory (RAM). By way of example,, and not limitation, system memory 1430 may also include an operating system, application programs, other program modules, and program data.


A user can interface with (for example, enter commands and information) the computer 1410 through input devices 1440. A monitor or other type of device can also be connected to the system bus 1422 via an interface, such as an output interface 1450. In addition to a monitor, computers may also include other peripheral output devices. The computer 1410 may operate in a networked or distributed environment using logical connections (network interface 1460) to other remote computers or databases (remote device(s) 1470). The logical connections may include a network, such local area network (LAN) or a wide area network (WAN), but may also include other networks/buses.


As will be appreciated by one skilled in the art, aspects may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, et cetera) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in at least one computer readable medium(s) having computer readable program code embodied thereon.


Any combination of at least one computer readable medium(s) may be utilized. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having at least one wire, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible or non-signal medium that can contain or store a program for use by or in connection with an instruction execution system, apparatus, or device.


Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.


Computer program code for carrying out operations for embodiments may be written in any combination of at least one programming language, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).


Embodiments are described with reference to figures of methods, apparatus (systems) and computer program products according to embodiments. It will be understood that portions of the figures can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified.


These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified. The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified.


This disclosure has been presented for purposes of illustration and description but is not intended to be exhaustive or limiting. Many modifications and variations will be apparent to those of ordinary skill in the art. The example embodiments were chosen and described in order to explain principles and practical application, and to enable others of ordinary skill in the art to understand the disclosure for various embodiments with various modifications as are suited to the particular use contemplated.


Although illustrated example embodiments have been described herein with reference to the accompanying drawings, it is to be understood that embodiments are not limited to those precise example embodiments, and that various other changes and modifications may be affected therein by one skilled in the art without departing from the scope or spirit of the disclosure.

Claims
  • 1. A method for summarizing attributes of data elements of a data set for computing correlated aggregates, comprising: receiving a stream of data elements at a device, each data element having at least one numerical attribute;maintaining in memory a plurality of tree structures comprising a plurality of separate nodes for summarizing numerical attributes of the data elements with respect to a predicate value of a correlated aggregation query, said maintaining comprising:creating the plurality of tree structures in which each node implements one of: a probabilistic counter and a sketch, wherein said probabilistic counter and said sketch each act to estimate aggregated data element numerical attributes to form a summary of said numerical attributes, said probabilistic counter providing an estimated count via incrementing by a predetermined weight in response to an increment operation; andresponsive to a correlated aggregation query specifying said predicate value, using said plurality of tree structures as a summary of said data element numerical attributes to compute a response to said correlated aggregate query,
  • 2. The method of claim 1, wherein estimation of aggregated data element numerical attributes to form said summary of said data element numerical attributes is within a predetermined relative error.
  • 3. The method of claim 2, wherein a memory size requirement for maintaining the summary of said data element numerical attributes is a function of the predetermined relative error.
  • 4. The method of claim 1, wherein each node implements a probabilistic counter, and further wherein said correlated aggregation query is a count query with respect to said predicate value.
  • 5. The method of claim 4, wherein nodes of said plurality of tree structures are organized into levels of probabilistic counters.
  • 6. The method of claim 5, wherein each level of nodes provides a more refined interval for counting said numerical attributes of said data elements.
  • 7. The method of claim 6, wherein each node is capped to a predetermined size, and further wherein, responsive to the predetermined size being reached, two child nodes are formed for a parent node.
  • 8. The method of claim 1, wherein each node implements a sketch function suitable for a statistical inquiry specified in said correlated aggregation query.
  • 9. The method of claim 8, wherein said statistical inquiry is computation of a frequency moment Fk of said data stream.
  • 10. The method of claim 1, wherein said stream of data elements comprises a stream of IP packets, wherein said data elements comprise IP packets, and further wherein said numerical attributes comprise sizes of said IP packets.
CROSS REFERENCE TO RELATED APPLICATION

This application is a continuation of U.S. patent application Ser. No. 13/278,469, entitled COMPUTING CORRELATED AGGREGATES OVER A DATA STREAM, filed on Oct. 21, 2011, which is incorporated by reference in its entirety.

STATEMENT OF GOVERNMENT INTEREST

This invention was made with government support under grant numbers: CNS0834743 and CNS0831903, awarded by the National Science Foundation. The government has certain rights in the invention.

US Referenced Citations (5)
Number Name Date Kind
7437397 Koudas et al. Oct 2008 B1
20080016031 Miao et al. Jan 2008 A1
20100030741 Johnson et al. Feb 2010 A1
20100067535 Ma et al. Mar 2010 A1
20110153554 Cohen et al. Jun 2011 A1
Non-Patent Literature Citations (8)
Entry
Gehrke et al., “On Computing Correlated Aggregates Over Continual Data Streams”, May 21-24, 2011, ACM Sigmond 2001, Santa Barbara, California, USA, 12 pages.
Cohen et al., “Bottom-k Sketches: Better and More Efficient Estimation of Aggregates”, Jun. 12-16, 2007, ACM, SIGMETRICS'07, San Diego, California, USA, 2 pages.
Indyk et al., “Optimal Approximations of the Frequency Moments of Data Streams”, May 22-24, 2005, ACM, STOC'05, Baltimore, Maryland, USA, 8 pages.
Lahiri, B., et al., “Finding Correlated Heavy-Hitters over Data Streams”, Dec. 14-16, 2009, pp. 307-314, IEEE 28th International Performance Computing and Communications Conference, Scottsdale, Arizona, USA.
Gehrke, J., et al., “On Computing Correlated Aggregates Over Continual Data Streams”, May 21-24, 2001, 12 pages, ACM SIGMOND 2001, Santa Barbara, California, USA.
Anathakrishna, R., et al., “Efficient Approximation of Correlated Sums on Data Streams”, IEEE Transactions on Knowledge and Data Engineering, May-Jun. 2003, pp. 576-572, vol. 15, No. 3, IEEE Computer Society Press, Los Alamitos, California, USA.
Cohen, Edith et al., “Bottom-k Sketches: Better and More Efficient Estimation of Aggregates,” Jun. 12-16, 2007, ACM, SIGMETRICS'07, San Diego, California, USA, 2 pages.
Indyk, Piotr et al., “Optimal Approximations of the Frequency Moments of Data Streams,” May 22-24, 2005, ACM, STOC'05, Baltimore, Maryland, USA, 8 pages.
Related Publications (1)
Number Date Country
20130103713 A1 Apr 2013 US
Continuations (1)
Number Date Country
Parent 13278469 Oct 2011 US
Child 13596514 US