Hash joins are a common operator in business intelligence (BI) queries. Hash joins involve many random accesses to hash tables. Conventional systems do partitioned joins by dividing the join hash table into partitions, so that lookups are focused on small data structures, which will fit better in cache. But below a threshold on the hash table size, a non-partitioned join is more efficient because it avoids the costs of partitioning.
Embodiments relate to building hash tables. One embodiment includes a method for creating a compact hash table includes a thread executing using a processor for scanning a subset of keys and values, and for each key, performing a hash operation and inserting hashed keys into a bitmap structure. Cumulative population counts of keys and values are determined within the bitmap. The method repeats scanning the subset of the keys and values. The keys and values are inserted into a compacted array using the cumulative population counts.
These and other features, aspects and advantages of the embodiments will become understood with reference to the following description, appended claims and accompanying figures.
Aspects of the embodiments are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products, according to the embodiments. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
In use, the gateway 101 serves as an entrance point from the remote networks 102 to the proximate network 108. As such, the gateway 101 may function as a router, which is capable of directing a given packet of data that arrives at the gateway 101, and a switch, which furnishes the actual path in and out of the gateway 101 for a given packet.
Further included is at least one data server 114 coupled to the proximate network 108, which is accessible from the remote networks 102 via the gateway 101. It should be noted that the data server(s) 114 may include any type of computing device/groupware. Coupled to each data server 114 is a plurality of user devices 116. Such user devices 116 may include a desktop computer, laptop computer, handheld computer, printer, and/or any other type of logic-containing device. It should be noted that a user device 111 may also be directly coupled to any of the networks in some embodiments.
A peripheral 120 or series of peripherals 120, e.g., facsimile machines, printers, scanners, hard disk drives, networked and/or local storage units or systems, etc., may be coupled to one or more of the networks 104, 106, 108. It should be noted that databases and/or additional components may be utilized with, or integrated into, any type of network element coupled to the networks 104, 106, 108. In the context of the present description, a network element may refer to any component of a network.
According to some approaches, methods and systems described herein may be implemented with and/or on virtual systems and/or systems, which emulate one or more other systems, such as a UNIX system that emulates an IBM z/OS environment, a UNIX system that virtually hosts a MICROSOFT WINDOWS environment, a MICROSOFT WINDOWS system that emulates an IBM z/OS environment, etc. This virtualization and/or emulation may be implemented through the use of VMWARE software in some embodiments.
In other examples, one or more networks 104, 106, 108, may represent a cluster of systems commonly referred to as a “cloud.” In cloud computing, shared resources, such as processing power, peripherals, software, data, servers, etc., are provided to any system in the cloud in an on-demand relationship, therefore allowing access and distribution of services across many computing systems. Cloud computing typically involves an Internet connection between the systems operating in the cloud, but other techniques of connecting the systems may also be used, as known in the art.
A hash table (HT) is made up of two parts: an array (the actual table where the data to be searched is stored) and a mapping function, known as a hash function. With a hash table, any value may be used as an index, such as a floating-point value, a string, another array, or even a structure as the index. This index is called the key, and the contents of the array element at that index is called the value. Therefore, an HT is a data structure that stores key/value pairs and can be quickly searched by the key. The hash function is a mapping from the input space to the integer space that defines the indices of the array. The hash function provides a way for assigning numbers to the input data such that the data can then be stored at the array index corresponding to the assigned number.
A hash join is an example of a join process and is used in the implementation of a relational database management system (DMS). The task of a join process is to find, for each distinct value of the join attribute, the set of tuples in each relation which have that value. Hash joins require an equijoin predicate (a predicate comparing values from one table with values from the other table using the equals operator ‘=’).
An inner join creates a new result table by combining column values of two tables (e.g., table1 and table2) based upon the join-predicate. The query compares each row of table1 with each row of table2 to find all pairs of rows which satisfy the join-predicate. When the join-predicate is satisfied, column values for each matched pair of rows of A and B are combined into a result row.
Non-partitioned hash joins perform a join between two inputs by building a single HT on one of the inputs, and performing lookups into this HT with the join columns of the other input. Conventionally, the HT for a non-partitioned join is typically built in one of two ways: built serially with a single thread, and built in parallel using latches on the HT buckets. Both ways result in slow HT builds. Further, conventional methods for building HTs focus on the case of provably n:1 HTs (when there is an index enforcing that each key has at most 1 matching payload), and require a large memory allocation for the case that such an index is not present, due to fill factor overheads from linear probing.
A classic hash join algorithm for an inner join of two relations proceeds as follows. First prepare an HT of the smaller relation. The HT entries consist of the join attribute and its row. Because the HT is accessed by applying a hash function to the join attribute, it is quicker to find a given join attribute's rows by using this table than by scanning the original relation. Once the HT is built, the larger relation is scanned and the relevant rows from the smaller relation are found by looking in the HT. The first phase is usually called the “build” phase, while the second is called the “probe” phase. Similarly, the join relation on which the HT is built is called the “build” input, whereas the other input is called the “probe” input. The process requires that the smaller join relation fits into memory, which is sometimes not the case. A simple approach to handling this situation proceeds as follows:
In one or more embodiments, the system 100 utilizes a process for providing a parallel build of a compact, non-partitioned join HT, without any latches. One or more embodiments perform partitioning on the join inner, according to the higher order bits of the hash function, and then performing builds of each partition in parallel, with a variant form of linear probing.
One or more embodiments provide building HTs. One embodiment includes a method for building an HT that includes mapping HT keys to values using multiple parallel computation threads. Each thread scans a subset of the HT keys and values and partitions the subset of the HT keys and values into multiple partitions. A cumulative count for HT keys and values in each partition is determined. An HT with space reserved for each partition is formed based on the determined cumulative count. Each thread selects one or more partitions and inserts HT keys and values belonging to the selected one or more partitions into the HT in the reserved space for those partitions.
In one example, the workstation may have resident thereon an operating system, such as the MICROSOFT WINDOWS Operating System (OS), a MAC OS, a UNIX OS, etc. It will be appreciated that other examples may also be implemented on platforms and operating systems other than those mentioned. Such other examples may include operating systems written using JAVA, XML, C, and/or C++ language, or other programming languages, along with an object oriented programming methodology. Object oriented programming (OOP), which has become increasingly used to develop complex applications, may also be used.
In one embodiment, to keep the HT size small, a linear probing HT that has 100% fill factor in the array of keys and values (payloads) is created, and uses a sparse bitmap to almost entirely avoid collisions. The net result is 2× to 3× smaller table space consumption.
Conventional HT creation has focused on n−1 joins, usually with single-column keys. However, n-m joins show up in many customer queries. Further, even when data almost (or fully) meets n−1 constraints, the constraints are not always enforced via indexes. In one embodiment, generalizations of CHT and a join for n-m joins, ‘near’ n−1 joins, and left-outer joins are provided by the system 300. In one embodiment, a variant of CHT referred to as a compact array table (CAT) that is collision-free, avoids storing join keys in the HT, and still has nearly 100% fill factor, even without dense join keys is provided by the system 300.
Query predicates are usually on dimension tables, therefore joins are usually highly filtering processes. Conventional joins assume that filtering is performed during the join, and so advocate ‘late materialization’ (i.e. join just maps key to RID, and only at the end maps row identifier (RID) to payload columns). In one embodiment, a CHT bitmap is used as a Bloom filter to eliminate most non-matching join outers—thus processing cost paid is only for 1 bit per inner in this data structure (vs width of key+width of RID in late materialization).
In one embodiment, the mapping processor 315 maps HT keys to values using multiple parallel computation threads. In one embodiment, the scanning processor 310 provides for each thread to scan a subset of one input to the join, and partitions it, using the high order bits of a hash computed on the join key. In one embodiment, after all of the threads complete the scan, the threads exchange among themselves the number of keys in each partition in the subset they scanned, and the counting processor 320 sums up the number of keys in each partition. In one embodiment, the counting processor 320 provides a total count for each partition by summing up the number of keys received for that partition, from each of the scanning threads. In one embodiment, the counting processor 320 computes a cumulative form of the count—that is, the number of keys in all partitions 0, 1, . . . up to each partition. In one embodiment, the counting processor 320 is embedded in the scanning threads itself, and is not a separate processor.
In one embodiment, the HT creation processor provides for each thread to take ownership of one or more partitions, and builds a sub-range of the HT for that partition. In the HT, the sub-range of the array holding keys and payloads is indicated by a starting position given by the cumulative count of the partitions up to (and excluding) the one being built, and extends to a length equal to the number of keys in this partition. In one embodiment, any linear probing (or quadratic probing or other collision resolution technique) is cycled around within this partition alone.
In one embodiment, the system 300 provides for building a CHT as follows. In one embodiment, the scanning processor 310 scans the input (e.g., partitioned as above), and for each key, a hash function is performed on the input and the bit map processor 306 inserts the result into a bitmap. In one embodiment, if there is a collision, it is not known if it is due to a duplicate key or just a hash collision. In the case of duplicate keys, it is desired to eliminate them (and thus save memory), rather than placing them into the CHT. In one embodiment, the duplicate keys are treated as a hash collision until the collision chain length grows beyond a threshold (e.g., 2). At that point, in one embodiment the bitmap processor 305 begins tracking the actual keys that are seen, so that in the future if that same hash bucket is hit the system 300 can detect if it is a collision or not. In one embodiment, two processes may be used to perform the tracking.
In one embodiment, tracking is provided as follows. In one embodiment, a small local HT or similar data structure is used by the bitmap processor 306 to detect and eliminate highly frequent duplicate keys. When tracking starts, a key is placed into the data structure because of high collision chain length. In one embodiment, any key that is not detected as a duplicate in this structure is placed into the CHT bitmap by the bitmap processor 306 and is treated as collisions.
Another embodiment for tracking is provided as follows. In one embodiment, all keys that collide to a chain length of, for example, >2, are taken and placed into an ‘overflow’ HT. The crucial difference compared to the first tracking embodiment is that these overflow keys are never inserted into the HT. This has huge performance benefits during probe time because it is known that the linear probe chain will never be >2.
In one embodiment, the CHT creation process continues as follows. In one embodiment, the counting processor 320 computes cumulative population counts for the bitmap provided by the bitmap processor 306. In one embodiment, the scanning processor 310 scans the input again, inserts keys and values/payloads into a compacted array, using the population counts from the counting processor 320. In one example, a few duplicate keys may be encountered. In this case, in one embodiment, the CHT creation processor marks the key specially as a duplicate so that the cost of returning the marked keys as join matches is avoided, and also potentially for avoiding the processing cost of copying them into the CHT (although the space the keys take up in the CHT cannot be reclaimed).
In one embodiment, the CHT 420 includes an array of keys (indicated by Kn, where n is a positive integer) and payloads or values (indicated by pn), whosesize is the cardinality of the input on which the CHT is built. The Bitvector has a size equal to the size of the HT in the CHT, and it indicates occupied buckets of that HT. In one example, for every 64-bits, a pre-computed 32-bit prefix population count is stored.
It is noted that if keys are all guaranteed unique (i.e. unique keys with an enforced uniqueness constraint), linear probing may be performed without inserting keys: the position of a key in a linear probe HT can be inferred by just performing linear probing on the CHT bitmap.
But, enforced PK constraints are not very popular with database applications (DBAs) because they impose a unique index maintenance cost during inserts. However, queries sometimes care only for the first match for a join outer—this commonly arises with existential sub-queries and also with group-by. In this case, the input could have duplicate keys, but these should be eliminated, since the join doesn't need them.
In one embodiment, process 800 may choose the subset of the keys and values from a subset of a join input. In one embodiment, partitioning of the subset of the keys and values is performed using bits (e.g., high order) of a hash value computed on a column used in a join predicate. In one embodiment, process 800 may include that determining the cumulative count for keys and values in each partition includes that each thread exchanges with other threads a number of keys in each partition in its subset, and determines the number of keys for each of the partitions by summing the values received from other threads for those partitions, and sums the determined number of keys for each of the partitions for determining the cumulative count up to each partition. For example, the cumulative count up to partition 5 is the sum of the number of keys in partitions 1, 2, 3, and 4.
In one embodiment, process 800 may include that each thread takes ownership of one or more partitions and builds a sub-range of the hash table for that partition. The sub-range includes a sub-range of an array holding keys and payloads and is indicated by a starting position given by the cumulative count of the partitions up to (and excluding) a current partition being built, and extends to a length equal to a number of keys in the current partition. In one embodiment, process 800 may further include creating a compact hash table by: scanning the subset of the hash table keys and values, and for each key, performing a hash operation and inserting hashed keys into a bitmap; determining cumulative population counts for hash table keys and values for the bitmap; repeating scanning of the subset of the hash table keys and values; and inserting the hash table keys and values into a compacted array using the cumulative population counts. In one embodiment, process 800 may provide that the compacted array for the compact hash table contains one entry for each distinct key.
In one embodiment, process 800 may include treating duplicate keys as a hash collision until a collision chain length reaches a predetermined threshold, and then tracking of hash table keys. In one embodiment, tracking hash table keys in process 800 may include: inserting the duplicate keys in a local data structure; and using the local data structure for detecting and eliminating frequent duplicate keys. In one embodiment, the un-detected keys in the local data structure are treated as collisions and placed in the bitmap. In another embodiment, tracking in process 800 may include placing the duplicate keys into an overflow hash table. In one embodiment, the duplicate keys are marked for avoidance of returning the marked duplicate keys as join matches and avoiding copying the marked duplicate keys into the hash table.
In one embodiment, process 900 may include that duplicate keys are treated as a hash collision until a collision chain length reaches a predetermined threshold (e.g., 2), and then tracking of hash table keys is performed. In one embodiment, in process 900, tracking of keys may include inserting the duplicate keys in a local data structure, and using the local data structure for detecting and eliminating frequent duplicate keys. In one embodiment, un-detected keys in the local data structure are treated as collisions and placed in the bitmap. In one embodiment, tracking of hash table keys in process 900 may include placing the duplicate keys into an overflow hash table. In one embodiment, duplicate keys are marked for avoidance of returning the marked duplicate keys as join matches and avoiding copying the marked duplicate keys into the hash table.
As will be appreciated by one skilled in the art, aspects of the embodiments may be embodied as a system, method or computer program product. Accordingly, aspects of the embodiments may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the embodiments may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the embodiments may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
Aspects of the embodiments are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to the embodiments. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.
References in the claims to an element in the singular is not intended to mean “one and only” unless explicitly so stated, but rather “one or more.” All structural and functional equivalents to the elements of the above-described exemplary embodiment that are currently known or later come to be known to those of ordinary skill in the art are intended to be encompassed by the present claims. No claim element herein is to be construed under the provisions of 35 U.S.C. section 112, sixth paragraph, unless the element is expressly recited using the phrase “means for” or “step for.”
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the embodiments. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the embodiments has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the embodiments in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the embodiments. The embodiments were chosen and described in order to best explain the principles of the embodiments and the practical application, and to enable others of ordinary skill in the art to understand the various embodiments with various modifications as are suited to the particular use contemplated.
Number | Name | Date | Kind |
---|---|---|---|
5455826 | Ozveren et al. | Oct 1995 | A |
5598559 | Chaudhuri | Jan 1997 | A |
5706495 | Chadha et al. | Jan 1998 | A |
5740440 | West | Apr 1998 | A |
5794229 | French | Jun 1998 | A |
5893086 | Schmuck et al. | Apr 1999 | A |
5930785 | Lohman et al. | Jul 1999 | A |
6026394 | Tsuchida et al. | Feb 2000 | A |
6052697 | Bennett | Apr 2000 | A |
6134601 | Spilo et al. | Oct 2000 | A |
6247014 | Ladwig et al. | Jun 2001 | B1 |
6292795 | Peters et al. | Sep 2001 | B1 |
6505189 | On Au et al. | Jan 2003 | B1 |
6609131 | Zait et al. | Aug 2003 | B1 |
6757677 | Pham et al. | Jun 2004 | B2 |
6775681 | Ballamkonda et al. | Aug 2004 | B1 |
6937652 | Gold et al. | Aug 2005 | B2 |
6941432 | Ronstrom | Sep 2005 | B2 |
6954776 | Cruanes et al. | Oct 2005 | B1 |
7062481 | Pham et al. | Jun 2006 | B2 |
7136883 | Flamma et al. | Nov 2006 | B2 |
7177883 | Yagawa | Feb 2007 | B2 |
7287131 | Martin et al. | Oct 2007 | B1 |
7293028 | Cha et al. | Nov 2007 | B2 |
7308539 | Fuhs et al. | Dec 2007 | B2 |
7343363 | Parker | Mar 2008 | B1 |
7412439 | Bossman et al. | Aug 2008 | B2 |
7499960 | Dageville et al. | Mar 2009 | B2 |
7653670 | Hasan et al. | Jan 2010 | B2 |
7688758 | Denby et al. | Mar 2010 | B2 |
7716180 | Vermeulen et al. | May 2010 | B2 |
7827182 | Panigrany | Nov 2010 | B1 |
7827218 | Mittal | Nov 2010 | B1 |
7868789 | Binnig et al. | Jan 2011 | B1 |
8078593 | Ramarao et al. | Dec 2011 | B1 |
8145642 | Cruanes et al. | Mar 2012 | B2 |
8195644 | Xu | Jun 2012 | B2 |
8271564 | Dade | Sep 2012 | B2 |
8321385 | Burroughs et al. | Nov 2012 | B2 |
8346810 | Beaverson et al. | Jan 2013 | B2 |
8370316 | Bensberg et al. | Feb 2013 | B2 |
8438574 | Lyle et al. | May 2013 | B1 |
8443155 | Adams et al. | May 2013 | B2 |
8661005 | McKenney et al. | Feb 2014 | B2 |
8692695 | Fallon et al. | Apr 2014 | B2 |
8768889 | Martin | Jul 2014 | B1 |
8768927 | Yoon et al. | Jul 2014 | B2 |
8856103 | Barber et al. | Oct 2014 | B2 |
8886614 | Morris | Nov 2014 | B2 |
9098201 | Benjamin et al. | Aug 2015 | B2 |
9355060 | Barber et al. | May 2016 | B1 |
9430390 | Mukherjee et al. | Aug 2016 | B2 |
9454560 | Cha et al. | Sep 2016 | B2 |
9684682 | Mukherjee et al. | Jun 2017 | B2 |
9792318 | Schreter et al. | Oct 2017 | B2 |
20010039609 | Houldsworth et al. | Nov 2001 | A1 |
20020016820 | DuVal et al. | Feb 2002 | A1 |
20040260684 | Agrawal et al. | Dec 2004 | A1 |
20050018683 | Zaho et al. | Jan 2005 | A1 |
20050033741 | Dombroski et al. | Feb 2005 | A1 |
20060218176 | Sun Hsu et al. | Sep 2006 | A1 |
20070136317 | Przywara | Jun 2007 | A1 |
20070244850 | Hoppe et al. | Oct 2007 | A1 |
20070245119 | Hoppe | Oct 2007 | A1 |
20080126706 | Newport et al. | May 2008 | A1 |
20080133583 | Artan et al. | Jun 2008 | A1 |
20080162402 | Holmes et al. | Jul 2008 | A1 |
20090006399 | Raman et al. | Jan 2009 | A1 |
20090024568 | Al-Omari et al. | Jan 2009 | A1 |
20090100223 | Murayama et al. | Apr 2009 | A1 |
20090187586 | Olson | Jul 2009 | A1 |
20090210445 | Draese et al. | Aug 2009 | A1 |
20090222659 | Miyabayashi et al. | Sep 2009 | A1 |
20100088309 | Petculescu et al. | Apr 2010 | A1 |
20100114668 | Beavin et al. | May 2010 | A1 |
20100131540 | Xu et al. | May 2010 | A1 |
20100199066 | Artan et al. | Aug 2010 | A1 |
20100223253 | Gopal et al. | Sep 2010 | A1 |
20110060876 | Liu | Mar 2011 | A1 |
20110066593 | Ahluwalia et al. | Mar 2011 | A1 |
20110078134 | Bendel et al. | Mar 2011 | A1 |
20110107021 | Muthukumarasamy et al. | May 2011 | A1 |
20110283082 | McKenney et al. | Nov 2011 | A1 |
20110307471 | Sheinin | Dec 2011 | A1 |
20120011133 | Faerber et al. | Jan 2012 | A1 |
20120011144 | Transier et al. | Jan 2012 | A1 |
20120036134 | Malakhov | Feb 2012 | A1 |
20120117055 | Al-omari et al. | May 2012 | A1 |
20120136846 | Song et al. | May 2012 | A1 |
20120136889 | Jagannathan et al. | May 2012 | A1 |
20120143877 | Kumar et al. | Jun 2012 | A1 |
20120158729 | Mital et al. | Jun 2012 | A1 |
20120166400 | Sinclair et al. | Jun 2012 | A1 |
20120173517 | Lang et al. | Jul 2012 | A1 |
20120260349 | Nagai et al. | Oct 2012 | A1 |
20120303633 | He et al. | Nov 2012 | A1 |
20120310917 | Sheinin et al. | Dec 2012 | A1 |
20130046949 | Colgrove et al. | Feb 2013 | A1 |
20130138628 | Bensberg et al. | May 2013 | A1 |
20130218934 | Lin et al. | Aug 2013 | A1 |
20130325900 | Barber et al. | Dec 2013 | A1 |
20140006379 | Arndt et al. | Jan 2014 | A1 |
20140025648 | Corbett et al. | Jan 2014 | A1 |
20140074819 | Idicula | Mar 2014 | A1 |
20140108489 | Glines et al. | Apr 2014 | A1 |
20140129568 | Kim et al. | May 2014 | A1 |
20140214794 | Attaluri et al. | Jul 2014 | A1 |
20140214795 | Attaluri et al. | Jul 2014 | A1 |
20140214855 | Attaluri | Jul 2014 | A1 |
20140215019 | Ahrens | Jul 2014 | A1 |
20140337375 | Yue | Nov 2014 | A1 |
20140372388 | Attaluri et al. | Dec 2014 | A1 |
20140372392 | Liu | Dec 2014 | A1 |
20150088813 | Lahiri et al. | Mar 2015 | A1 |
20150301743 | Nagao et al. | Oct 2015 | A1 |
20160147457 | Legler et al. | May 2016 | A1 |
20160173445 | Mosko | Jun 2016 | A1 |
20160232169 | Archak | Aug 2016 | A1 |
Number | Date | Country |
---|---|---|
1522409 | Aug 2004 | CN |
1968212 | May 2007 | CN |
101013427 | Aug 2007 | CN |
101067822 | Nov 2007 | CN |
1003672239 | Feb 2008 | CN |
101231657 | Jul 2008 | CN |
101388042 | Mar 2009 | CN |
102893265 | Jan 2013 | CN |
103635902 | Mar 2014 | CN |
104021205 | Sep 2014 | CN |
104317966 | Jan 2015 | CN |
0457707 | Nov 1991 | EP |
2015078136 | Jun 2015 | WO |
Entry |
---|
Anonymous, “System and Method for Usage Aware Row Storage in Database Systems”, Jul. 23, 2010, pp. 1-4, IP.com, United States. |
Anonymous, “High Performance Technique Using Join Collocation in a Massively Parallel Processing Relational Database Implementation”, Jun. 14, 2012, pp. 1-5, IP.com, United States. |
Anonymous, “CashMap: Processor Cache-Aware Implementation of Hash Tables”, Jul. 5, 2013, pp. 1-7, IP.com, United States. |
Anonymous, “Efficient Grouping Over Joins of Compressed Tables”, Apr. 6, 2010, pp. 1-6, IP.com, United States. |
Korn, D., et al., “The VCDIFF Generic Differencing and Compression Data Format (RFC3284)”, Jul. 1, 2002, pp. 1-31, Network Working Group, IP.com, United States. |
Hu, K. et al. “Rapid multi-dimension hierarchical algorithm in data warehouse system”, Computer Integrated Manufacturing Systems, Jan. 2007, pp. 196-201, vol. 13, No. 1, China [English-language translation: Abstract Only]. |
Spyros, B., et al., “Design and Evaluation of Main Memory Hash Join Algorithms for Multi-core CPUs”, SIGMOD Int'l Conference on Management of Data, Ju. |
Raman, V., et al., “DB2 with BLU Acceleration: So Much More than Just a Column Store”, Proceedings of the VLDB Endowment, Aug. 2013, pp. 1-12, vol. 6, No. 11, ACM, United States. |
Mell, P., et al., “The NIST Definition of Cloud Computing”, National Institute of Standards and Technology Special Publication 800-145, Sep. 2011, pp. 1-7, U.S. Department of Commerce, United States. |
Yan, W.P. et al., “Performing Group-By Before Join”, Proceedings 10th Int'l Conference on Data Engineering, 1994, pp. 1-30, IEEE, United States. |
Cleary, J.G., “Compact Hash Tables Using Bidirectional Linear Probing”, IEEE Transactions on Computers, Sep. 1994, pp. 828-834, vol. C-33, No. 9, United States. |
Nan Hua. H., et al., “Rank-Indexed Hashing: A Compact Construction of Bloom Filters and Variants”, IEEE, 2008, pp. 73-82, United States. |
Xu, Y., “A Multi-Dirnesional Progressive Perfect Hashing for High-Speed String Matching”, Seventh ACM/ IEEE Symposium on Architectures for Networking and Communications Systems, 2011, pp. 167-177, IEEE Computer Society, United States. |
Chang, S., “Recent Advances of Compact Hashing for Large-Scale Visual Search”, Columbia University, Oct. 2012, pp. 1-44, United States. |
Wang, W., et al.; “Investigating Memory Optimization of Hash-index for Next Generation Sequencing on Multi-core Architecture”, IPDPSW IEEE 26th Inter. Conf., May 21-25, 2012, pp. 665-674, IEEE Computer Society, United States. |
Cutt, B., et al.; “Improving Join Performance for Skewed Databases”, IEEE, 2008, pp. 1-5, United States. |
Li, Q., et al.; “Adaptively Reordering Joins during Query Execution”, IEEE, 2007, pp. 26-35, United States. |
Internet Society, et al., “The VCDIFF Generic Differencing and Compression Data Format (RFC3284)”, Jul. 1, 2002, pp. 1-31, Network Working Group, IP.com, United States. |
Hu, K. et al. “Rapid multi-dimension hierarchical algorithm in data warehouse system”, Computer Integrated Manufacturing Systems, Jan. 2007, p. 196-201, vol. 13, No. 1, China. |
Lehman, T.J. “Design and Performance Evaluation of a Main Memory Relational Database System.” 1986, PhD Dissertation, 334 pages, [Abstract Only], University of Washington, Madison, WI. |
Leis, V., et al., “The Adaptive Radix Tree: ARTful Indexing for Main-Memory Databases”, IEEE 29th International Conference on Data Engineering (ICDE), Apr. 8, 2013, pp. 38-49, IEEE, United States. |
Levandoski, J., et al., “The Bw-Tree: A B-tree for New Hardware Platforms”, IEEE 29th International Conference on Data Engineering (ICDE), Apr. 8, 2013, pp. 1-12, IEEE, United States. |
Prokopec, A. et al., “Lock-Free Resizeable Concurrent Tries”, Languages and Compilers for Parallel Computing, 2013, pp. 156-170, vol. 7146, Springer Berlin Heidelberg, Germany. |
Areias, M. et al., “A Simple and Efficient Lock-Free Hash Trie Design for Concurrent Tabling”, Theory and Practice of Logic Programming, May 14, 2014, pp. 1-10, Arxiv.org, Cornell University Library, United States. |
Pandis, I. et al., “PLP: Page Latch-free Shared-everything OLTP”, Proceedings of the 37th International Conference on Very Large Data Bases (VLDB Endowment), Aug. 29, 2011-Sep. 3, 2011, pp. 610-621, vol. 4, No. 10, United States. |
Sewall, J. et al., “PALM: Parallel Architecture-Friendly Latch-Free Modifications to B+ Trees on Many-Core Processors”, Proceedings of the 37th International Conference on Very Large Data Bases (VLDB Endowment), Aug. 29, 2011-Sep. 3, 2011, pp. 795-806, vol. 4, No. 11, United States. |
Gao, H. et al., “Lock-free dynamic hash tables with open addressing”, Journal of Distributed Computing, Jul. 2005, pp. 21-42, vol. 18, Issue 1, United Kingdom. |
Xu, Y., “A Multi-Dimensional Progressive Perfect Hashing for High-Speed String Matching”, Seventh ACM/ IEEE Symposium on Architectures for Networking and Communications Systems, 2011, pp. 167-177, IEEE Computer Society, United States. |
Marek, R., et al., “TID Hash Joins,” CIKIM, 1994, pp. 42-49, Gaithersburg, MD, United States. |
List of IBM Patents or Patent Applications Treated as Related. |
Chinese Office Action dated Dec. 5, 2018 for Chinese Patent Application No. 201610177904.0 from ChinesePatent Office, pp. 1-15, Beijing, China. |
Chinese Office Action dated Feb. 12, 2019 for Chinese Patent Application No. 201610177904.0 from ChinesePatent Office, pp. 1-8, Beijing, China. |
List of IBM Patents or Patent Applications Treated as Related: Attaluri, G.K., et al., U.S. Appl. No. 16/427,190, filed May 30, 2019. |
Number | Date | Country | |
---|---|---|---|
20180137123 A1 | May 2018 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14664710 | Mar 2015 | US |
Child | 15850363 | US |