An index is an ordered collection of numbers or character strings, or both, such that some numbers or character strings represent objects and other numbers or character strings represent information about these objects. For example, one form of index can be viewed as a table whose rows represent documents and whose columns represent attributes of these documents. Such a column in a table T can be referred to as T. attribute.
Joins are a class of data-processing operations that may be performed on indexes. Joins are used to match documents from different indexes by finding matching values of certain attributes of these documents. As an illustrative example,
SELECTA.surname, A.first_name, B.company FROM A, B where A.city=B.city
To evaluate this join, these two indexes are joined by join attributes A.city and B.city. The corresponding documents from the two indexes are merged by matching their values of the join attributes. This yields the table of values shown in
Consider a distributed landscape in which indexes are hosted on separate machines. One problem is that the amount of network traffic required to compute such a distributed join may be the main factor limiting the performance achievable with a given join process. Conventional processes for computing the join may require network traffic proportional to the size of the join table. If, for two indexes, one index has N rows and the other index has M rows, then the join table may consist of as many as N*M rows, and has all the requested attributes from both tables.
The problem is exacerbated in situations where indexes are too large for a single machine. Such indexes may be split up and stored on different servers. To process join queries over such distributed indexes, it may be necessary to transfer even more data over the network than in the case where each index has its own host. What is needed is a join method that minimizes network traffic.
This document presents methods and systems for calculating distributed joins, and for accelerating the execution of join queries in a distributed network landscape where the relevant indexes are stored on different hosts. The disclosed techniques minimize the amount of network traffic required. The methods and systems offer the benefits that only relatively small lists of integer values are transferred over the network, and value dictionaries from different indexes are fetched only once.
In one embodiment, a method of executing a distributed join query for a set of documents includes communication between a first server and a second server. In the first server, the method includes generating a first tuple list from a first list of documents matching at least part of the query, and extracting, from the first tuple list, a first set of value identifiers of attributes associated with the first list of documents. The method further includes generating a first set of dictionary keys from the set of value identifiers, and sending the first set of dictionary keys with a join condition attribute to a second server. In the second server, the method includes converting the first set of value identifiers to a second set of value identifiers of attributes associated with the second server based on the set of dictionary keys, and performing a lookup of documents based on the second set of value identifiers.
The details of one or more embodiments are set forth in the accompanying drawings and the description below. Other features, objects, and advantages will be apparent from the description and drawings, and from the claims.
These and other aspects will now be described in detail with reference to the following drawings.
Like reference symbols in the various drawings indicate like elements.
A method for matching documents in a distributed data processing system is presented. The method includes a technique for calculating distributed joins for matching documents that minimizes the amount of network traffic required. In accordance with an embodiment, attribute values for a document are stored in a dictionary, as shown in
Equi-joins are a special type of join in which join attributes are compared for equality (i.e., “A.x=B.y”). For the sake of simplicity, only equi-joins are described herein. Provided that the value dictionaries are constructed such that the values are in ascending order before the keys are assigned (so that k1<k2 if v1<v2 for any two key/value pairs k1/v1 and k2/v2), those types of joins that involve comparing the join attributes using operators “not equal”, “less than”, etc. (e.g. “A.x<B.x” or “A.x>=B.x”) can also be evaluated using methods described herein, such as in steps 410, 506 and 506 described below. The relevant indexes should remain unmodified while an executed query is in progress. For each attribute used as a join attribute, there is a corresponding value dictionary, as described above, and joins are AND-connected to the other query parts. In an exemplary embodiment, each of the relevant indexes resides on a different index server A, B, C, . . . etc., and the final result is created on the logical server L.
Attributes within an index are denoted as follows: Attribute x in index A is A.x, attribute y in index B is B.y . . . etc. Documents within an index are uniquely identified by internal document IDs: idA, idB, . . . which denote the document IDs of indexes A, B, . . . etc. Each dictionary key in a value dictionary is a value ID: for example, VA.x denotes the value ID of attribute x in index A. Square brackets denote a list of values as follows: [idA] stands for a list of document IDs from index A, for example. Square brackets enclosing multiple values denote tuples of values: [idA, idB, idC] denotes a list of ordered triples (3-tuples) of document IDs from indexes A, B, and C, respectively. The elements can be reordered at no cost; for example, the tuple list [a, b] is obtained from [b, a] by swapping the columns.
The function [VA.x]→[VB.y] converts value IDs for attribute A.x to value IDs for attribute B.y, using the respective value dictionaries. Value IDs from A.x that correspond to values missing in value dictionary for B.y are discarded. The function uniq(x) takes as its argument a list of values and removes from this list all duplicates, so that only unique values are present in the result. The function rfl(a, b) takes two lists as arguments and removes from list <a> those entries that are not present in list <b>. If one or both of these lists are tuples, then the lists are compared on the basis of the values in the respective first elements of the tuples.
Server B performs the conversion [VA.x]→[VB.y] using the value dictionary of A.x, at 408 (
At 414 (
At 508, server C performs [VB.z]→[Vcq] and creates and retains [VB.z, idC]. From this list, uniq([VB.z]) is sent back to server B. At 509, server B uses this list to perform rfl([VB.z, idB], [VB.z]) to remove the value IDs that were not found on server C. The resulting [VB.z, idB] is then used at 510 to perform rfl([idB, VA.x], [idB, VB.z]). Steps 507 through 510 are repeated (properly substituting the indexes and attributes) when queries with more indexes are to be evaluated.
At 511, the resulting list [idB, VA.x] is used to send uniq([VA.x]) back to server A. [VA.x, idB] and [idB, VB.z] are retained. At 512, server A uses this list to perform rfl([VA.x, idA], [VA.x]) and retains the list [VA.x, idA]. To assemble the final result, at 514 server L retrieves list [idA, VA.x] from server A, lists [VA.x, idB] and [idB, VB.z] from server B, and list [VB.z, idC] from server C. From these lists, server L creates the result tuples [idA, idB, idC] along the chain [idA, VA.x][VA.x, idB][idB, VB.z][VB.z, idC] at 516. Server L then retrieves the requested attributes A.a, B.b, and C.c from the respective indexes using the document Ids created in step 516.
Only relatively small lists of integer values are transferred over the network, such as [VA.x], [idA, VA.x], . . . etc. Value dictionaries from different indexes are fetched only once. They may then be retained locally (ideally in a least-recently-used (LRU) cache to limit the amount of memory used).
Although a few embodiments have been described in detail above, other modifications are possible. The sequential order of execution shown in
Number | Name | Date | Kind |
---|---|---|---|
4811207 | Hikita et al. | Mar 1989 | A |
6003040 | Mital et al. | Dec 1999 | A |
6513041 | Tarin | Jan 2003 | B2 |
6567802 | Popa et al. | May 2003 | B1 |
Number | Date | Country | |
---|---|---|---|
20060136388 A1 | Jun 2006 | US |