Method and Apparatus for Performing Similarity Searching

Information

  • Patent Application
  • 20170124255
  • Publication Number
    20170124255
  • Date Filed
    January 11, 2017
    7 years ago
  • Date Published
    May 04, 2017
    7 years ago
Abstract
A system and method for performing similarity searching is disclosed wherein programmable logic devices such as field programmable gate arrays (FPGAs) can be used to implement Bloom filters for identifying possible matches between a query and data. The Bloom filters can be implemented in a parallel architecture where the different parallel Bloom filters share access to the same memory units. Further, a hash table may be generated to map a set of strings to keys. In other examples, the hash table may be used to map a set of substrings to a position in a larger string.
Description
THE NAMES OF PARTIES TO A JOINT RESEARCH AGREEMENT

Washington University, a corporation of the State of Missouri, and Data Search Systems, Inc., a corporation of the State of Delaware (now Exegy Incorporated, a corporation of the State of Delaware), are parties to a Joint Research Agreement.


FIELD OF THE INVENTION

The present invention relates to the field of biosequence similarity searching. In particular, the present invention relates to the field of searching large databases of biological sequences for strings that match a query sequence.


BACKGROUND OF THE INVENTION

The amount of biosequence data being produced each year is growing exponentially. Extracting useful information from this massive amount of data efficiently is becoming an increasingly difficult task. The databases of genomic DNA and protein sequences are an essential resource for modern molecular biology. This is where a computational search of these databases can show that a DNA sequence acquired in the lab is similar to other sequences of a known biological function, revealing both its role in the cell and its history over evolutionary time. A decade of improvement in DNA sequencing technology has driven exponential growth of biosequence databases such as NCBI GenBank, which has doubled in size every twelve (12) to sixteen (16) months for the last decade and now stands at over forty-five (45) billion characters. These technological gains have also generated more novel sequences, including entire mammalian genomes, which keep search engines very busy.


Examples of this type of searching can be found in a paper entitled Biosequence Similarity Search On The Mercury System, P. Krishnamurthy, J. Buhler, R. D. Chamberlain, M. A. Franklin, K. Gyang, and J. Lancaster, In Proceedings of the 15th IEEE International Conference on Application-Specific Systems, Architectures, and Processors (ASAP04), September 2004, Pages 365-375 (2004). Another example is a paper entitled: NCBI BLASTN STAGE 1 IN RECONFIGURABLE HARDWARE, by Kwame Gyang, Department of Computer Science and Engineering, Washington University, August 2004, Technical Report WUCSE-2005-30. Another example is a paper entitled “BLASTN Redundancy Filter in Reprogrammable Hardware”, by C. Behrens, J. Lancaster, and B. Wun, Department of Computer Science and Engineering, Washington University, Final Project Submission, Fall 2003. These papers are each incorporated by reference, in their entirety.


There is a growing need for a fast and cost effective mechanism for extracting useful information from biosequence data.


SUMMARY OF INVENTION

In one aspect of this invention, a system for generating a hash table for use in mapping a set of strings to keys is disclosed that comprises: (1) a processor configured to provide near perfect hashing on the set of strings to (i) identify a location in a hash table, utilizing an H3 family of hash functions, corresponding to each string, and (ii) generate the hash table, the hash table being configured to store, at each identified location therein, a key, and (2) a memory in communication with the processor for storing the hash table, wherein the processor is configured to provide near perfect hashing via hash functions, utilizing the H3 family of hash functions, that are guaranteed to be of full rank.


In another aspect of this invention, a system for generating a hash table for use in mapping a set of substrings to a position in a larger string is disclosed that comprises: (1) a processor configured to (i) provide near perfect hashing on the set of substrings to identify a position in a hash table corresponding to each substring, and (ii) generate the hash table, the hash table being configured to store, at each identified location therein, a pointer to a position of each substring in the larger string, and (2) a memory in communication with the processor for storing the hash table.


In still another aspect of this invention, a method for generating a hash table for use in mapping a set of strings is disclosed that comprises: (1) hashing the set of strings with a processor to identify a position in the hash table corresponding to each string, (2) generating the hash table, the hash table being configured to store, at each identified location therein, a pointer to a position of each string, and (3) storing the hash table within a memory, the memory in communication with the processor.


These are merely some of the innumerable aspects of the present invention and should not be deemed an all-inclusive listing of the innumerable aspects associated with the present invention.





BRIEF DESCRIPTION OF DRAWINGS

For a better understanding of the present invention, reference may be made to the accompanying drawings in which:



FIG. 1 provides a block diagram overview of a three (3) stage pipeline in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention broken down into three (3) main stages;



FIG. 2 provides an exemplary block diagram of a Stage 1 pipeline, e.g., BLASTN, in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 3 is a general schematic of a hardware platform for a biological data pipeline in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 4 depicts a sample portion of a biological sequence data stream;



FIG. 5 depicts a query sequence can be broken down into multiple w-mers in connection with the database sequence shown in FIG. 3;



FIG. 6(a) depicts an exemplary Bloom filter stream in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 6(b) shows an arrangement of Bloom filters configured to share the same dual port block random access memories (BRAMs) in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 6(c) depicts a Bloom filter stage that comprises sixteen (16) Bloom filters formed from four (4) Bloom filter arrangements, as shown in FIG. 6(b), in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 7 is a schematic and flow chart of control logic that is either within or associated with a Bloom filter stage and readily configured to identify the starting position within the database sequence of the database w-mer that is a possible match with a query w-mer;



FIG. 8 illustrates an exemplary hashing stage and hash table in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 9 depicts an exemplary data window associated with the analysis of ungapped extension with analysis in two directions within a predefined window of data;



FIG. 10 depicts an exemplary data window associated with the analysis of ungapped extension with analysis in a single direction within a predefined data window in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 11 is a block diagram overview of the three (3) stage pipeline of FIG. 1, wherein the first and second stages are each broken down into two (2) substages;



FIG. 11A is a block diagram overview of the three (3) stage pipeline of an alternative embodiment of the present invention shown in FIG. 1, wherein only the first stage is broken down into two (2) substages;



FIG. 12 is an illustrative, but nonlimiting, listing of pseudocode utilized in an ungapped extension algorithm in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 13 is an illustrative, but nonlimiting, block diagram of a prefilter stage for an ungapped extension in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 13A is an illustrative, but nonlimiting, block diagram of a window lookup module of the prefilter stage for an ungapped extension, as shown in FIG. 13;



FIG. 14 is an illustrative, but nonlimiting, block diagram of a base comparator, as shown in FIG. 13, in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 15 is an illustrative, but nonlimiting, block diagram of an array of scoring stages, as shown in FIG. 13, in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 16 is an illustrative, but nonlimiting, block diagram of a single two step scoring stage, as shown in FIG. 15, in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 17 is a graphical representation of throughput of an overall pipeline as a function of ungapped extension throughput for queries of sizes 20 kbases and 25 kbases;



FIG. 18 is a graphical representation of speedup of an overall pipeline as a function of ungapped extension throughput for queries of sizes 20 kbases and 25 kbases with prefiltering in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention;



FIG. 19 is a table of sensitivity results for an illustrative, but nonlimiting, prefilter for the ungapped extension in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention; and



FIG. 20 is a flow chart of the algorithm shown in FIG. 12 utilized in an ungapped extension in accordance with a preferred, nonlimiting, illustrative embodiment of the present invention.





DETAILED DESCRIPTION OF THE INVENTION

In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the invention. However, it will be understood by those skilled in the art that the present invention may be practiced without these specific details. In other instances, well-known methods, procedures, and components have not been described in detail so as to obscure the present invention.


In an effort to improve upon the speed of biological similarity searching, the inventors herein have developed a novel and unique method and system whereby Stage 1 of the software pipeline known in the art as BLAST (Basic Local Alignment Search Tool), and more particularly the BLAST application BLASTN, is implemented via programmable logic device, e.g., reconfigurable hardware, such as, but not limited to, FPGAs (field programmable gate arrays). In addition, this present invention decides in Stage 2 of the software pipeline, ungapped extension, whether each w-mer emitted from word matching is worth inspecting by the more computationally intensive, gapped extension by shifting through pattern matches between query and database. It is important to identify and discard spurious w-mers as early as possible in the BLAST pipeline because later stages are increasingly more complex.


BLAST compares a query sequence to a biosequence database to find other sequences that differ from it by a small number of edits (single-character insertions, deletions, or substitutions). Because direct measurement of the edit distance between sequences is computationally expensive, BLAST uses a variety of heuristics to identify small portions of a large database that are worth comparing carefully to the biosequence query.


BLAST includes a pipeline of computations that filter a stream of characters from a database to identify meaningful matches to a query. To keep pace with growing databases and queries, this stream must be filtered at increasingly higher rates. One path to higher performance is to develop a specialized processor that offloads part of BLAST's computation from a general-purpose processor, e.g., CPU. Illustrative, but nonlimiting, examples of processors that are known to accelerate or replace BLAST include the ASIC-based Paracel GeneMatcher™ and the FPGA-based TimeLogic DecypherBLAST™ engine.


There is also a recently developed new design for an accelerator, i.e., the FPGA-based MERCURY BLAST engine. This MERCURY BLAST engine utilizes fine-grained parallelism in BLAST's algorithms and the high input/output (I/O) bandwidth of current commodity computing systems to deliver a speedup that is at least one (1) to two (2) orders of magnitude over the software BLAST that can be deployed on a standard processor located in a laboratory.


The MERCURY BLAST engine is a multistage pipeline and this present invention involves word matching, i.e., a first stage (Stage 1), and an ungapped extension stage, i.e., a second stage (Stage 2). The present invention includes a prefilter stage between the word matching stage and the ungapped extension stage that sifts through pattern matches between a query sequence and a biological database sequence and decides whether to perform a more accurate, but computationally expensive, comparison between them. The present invention also provides a more fruitful acceleration of variable-length string matching that is robust to character substitutions. The illustrative, but nonlimiting, preferred implementation is compact, runs at high clock rates, and can process one pattern match for every clock cycle.



FIG. 1 shows a block diagram of a high level overview of a three (3) stage pipeline that is generally indicated by numeral 50. The database sequences 52 are shown entering the pipeline 50. The first stage (Stage 1) 54 provides a word matching function that detects substrings of fixed length “w” in the stream that perfectly match a substring of the query, e.g., w=11, for DNA. These short matches are also known as “w-mers.” Each matching w-mer is forwarded to a second stage (Stage 2), which is an ungapped extension 56, which extends the w-mer on either side to identify a longer pair of sequences around the w-mer that match with at most a small number of mismatched characters. These longer matches are identified as high-scoring segment pairs (“HSP”s) or ungapped alignments.


Finally, every high-scoring segment pair (HSP) that has both enough matches and sufficiently few mismatches is passed to the third stage (Stage 3), which is gapped extension 58, which preferably uses the Smith-Waterman dynamic programming algorithm to extend the high-scoring segment pair (HSP) into a gapped alignment. Gapped alignment is a pair of similar regions that may differ by only a few arbitrary edits. The software, e.g., BLAST, reports only final gapped alignments with many matches and few edits 60.


Research has determined that with the standard NCBI BLASTN software program, 83.9% of the computational time was spent in the first stage (Stage 1) 54 and 15.9% of the time was spent in the second stage (Stage 2) 56. Therefore, to achieve a significant speedup, e.g., greater than six (6) times, of BLASTN would require acceleration of both the first stage (Stage 1) 54 and the second stage (Stage 2) 56. NCBI is the National Center for Biotechnology Information, which is a multi-disciplinary research group, composed of computer scientists, molecular biologists, mathematicians, biochemists, research physicians, and structural biologists concentrating on basic and applied research in computational molecular biology utilizing a variety of software databases and tools.


A block diagram overview of the first stage (Stage 1) 54, referenced in FIG. 1, is shown in FIG. 2, which is a pipeline 100, e.g., BLASTN Stage 1, in accordance with a preferred embodiment of the present invention. The input processing unit 104, the Bloom filter stage 106, the string buffer unit 108, the hashing stage 110, the hash (look-up) table 112, and the redundancy filter 114 are preferably implemented on a programmable logic device. An illustrative, but nonlimiting, example of a programmable logic device 102 can include a field programmable gate array (FPGA). A hash table 112 is preferably implemented on a memory unit, e.g., static random access memory (SRAM) 118, which can be off-chip from the programmable logic device 102.


The placement of the hash table 112 as being off-chip is optional and is only a result of the memory constraints for field programmable gate arrays (FPGAs), wherein readily available FPGAs do not have sufficient capacity to store the hash table and provide the functionality of the various pipeline stages. However, if capacity for a field programmable gate array (FPGA) sufficiently increases in the future, the hash table 112 can also be implemented on an FPGA.


A preferred hardware platform for pipeline 100 is that disclosed in FIG. 3 and is generally indicated by numeral 800. Additional details about this hardware platform are disclosed in: (1) U.S. Pat. No. 6,711,558 entitled “Associative Database Scanning and Information Retrieval”; (2) pending U.S. patent application Ser. No. 10/153,151, filed May 21, 2002 and entitled “Associative Database Scanning and Information Retrieval Using FPGA Devices” (published as U.S. patent application publication 2003/0018630); (3) pending PCT application PCT/US04/16398 filed May 21, 2004 and entitled “Intelligent Storage and Processing Using FPGA Devices”; and (4) a paper on “Achieving Real Data Throughput for an FPGA Co-Processor on Commodity Server Platforms” by Chamberlain et al., published in Proc. of 1st Workshop on Building Block Engine Architectures for Computers and Networks, Boston, Mass., October 2004; and a paper on “Acceleration of Ungapped Extension in Mercury BLAST” by Lancaster et al., was published on Nov. 12, 2005 for the Seventh (7th) Workshop on Media and Streaming Processors held in conjunction with the Thirty-Eighth (38th) International Symposium on Microarchitecture (MICRO-38) in Barcelona, Spain, the entire disclosures of all of which are incorporated herein by reference.


The input processing unit 104 is preferably configured to receive a biological sequence data stream 120 as an input from a sequence database, which is shown in FIG. 2. This database stream 120 represents a database sequence that can be broken down into a plurality of substrings, each substring having a base length “w.” Such substrings are referred to herein as database w-mers. The input processing unit 104 preferably outputs a parallel stream of q database w-mers (or sequence w-mers) 122i, wherein q represents the total number of database w-mers that can be simultaneously processed by the Bloom filter stage 106.


A depiction of a sample portion of a biological sequence data stream is indicated by numeral 120, which is shown in FIG. 4. Stream 120 comprises a sequence of bases (in this example, A, C, G, or T, which correspond to the bases of a DNA sequence). In this example, each base can be represented by 2 or 4 bits, as is known in the art. If w equals 11 and q equals 8, this stream 120 can be converted by the input processing unit 104 into 8 parallel w-mers as shown in FIG. 5.


The Bloom filter stage 106, as shown in FIG. 2, operates to process each database w-mer to determine whether it is a possible match with a query w-mer. A query w-mer being a substring of base length “w” derived from a query sequence. By definition, Bloom filters will not produce any false negatives between the database w-mers and the query w-mers, but Bloom filters are highly likely to produce some number of false positive matches between the database w-mers and the query w-mers. Because of the presence of these false positives, the “yes” responses of the Bloom filter to database w-mers are best referred to as possible matches.


Before processing database w-mers, the Bloom filters within Bloom filter stage 106 will need to be programmed/keyed with query w-mers from a query sequence (or query string). The query sequence is some string of bases for which a person wants to compare with the biological database sequence to find matches. The query sequence can be broken down into multiple w-mers as shown in FIG. 5 in connection with the database sequence. For a given search, the value of w is the same for both the query w-mers and database w-mers. Each Bloom filter within Bloom filter stage 106 is preferably programmed with all of the query w-mers developed from the query sequence. In an illustrative, but nonlimiting, preferred embodiment, the Bloom filter stage 106 comprises a plurality of parallel Bloom filters. Also, it is preferred, but not necessary, that the parallel Bloom filters are copies of one another. A Bloom filter is said to be in parallel with another Bloom filter when both of those Bloom filters are configured to simultaneously process w-mers. Programming of the Bloom filter stage 106 can be achieved via appropriate control commands asserted by the input processing unit 104 (in response to control input to the pipeline 100).


A preferred Bloom filter architecture within the Bloom filter stage 106 is shown in FIGS. 6(a)-6(c). FIG. 6(a) depicts an exemplary Bloom filter 400. Within the Bloom filter 400, a database w-mer is processed by “k” hash functions 402, each hash function “i” operating to map the database w-mer to a bit in the l×m dual port block RAM (BRAM) unit 404i that corresponds to that hash function i. If all of the bits to which that database w-mer maps within BRAMs 404l through 404k are set (e.g., equal to 1), then logic unit 406, e.g., AND logic unit, will receive all 1's on its input lines, thereby producing a “yes” response as to whether that database w-mer is a possible match with a query w-mer.


The bits within the l×m dual port BRAMs 404 are set if during the Bloom filter programming process, a query w-mer maps to that bit location. Because the Bloom filter is programmed with all of the query w-mers from the query sequence, the output from logic unit 406, e.g., AND logic unit, will never produce a false negative, but may produce a false positive.


The hash functions 402 are preferably the same hash functions selected from the H3 family of hash functions as used in the hashing stage 110. In the illustrative, but nonlimiting, example of FIG. 6(a), the l×m memory units 404 to which the hash functions 402 map can include dual-ported block random access memories (BRAMs). However, by virtue of being dual ported, multiple Bloom filters can share access to the same BRAM 404, thereby providing a substantial savings in the resources consumed on the programmable logic device 102, e.g., FPGA, when implementing the Bloom filter stage 106.


A description of the operation of Bloom filters can be found in U.S. patent application Ser. No. 10/640,513, which was published as U.S. Patent Application No. 2005/0086520, and was filed on Aug. 14, 2003, which is incorporated herein by reference in its entirety.


As an additional realization of consumed resources, the dual port BRAMs 404 are preferably double clocked to allow four Bloom filters 106 to share the same dual port BRAM 404. Thus, during a given clock cycle, 4 accesses can be made to each dual port BRAM 404i; thereby allowing each dual port BRAM to support four Bloom filters 106 each clock cycle.


An arrangement 410 of four (4) Bloom filters 106 configured to share the same k BRAMs 404l through 404k is shown in FIG. 6(b). Further still, if it is feasible to triple clock the BRAMs 404, it is worth noting that each dual port BRAM 404 could support six (6) Bloom filters.


A Bloom filter stage 106 that includes at least sixteen (16) Bloom filters is shown in FIG. 6(c) formed from four (4) Bloom filter arrangements 410 that are shown in FIG. 6(b). Such a Bloom filter stage 106 is capable of simultaneously processing a different database w-mer through each parallel Bloom filter. As such, the Bloom filter stage 106 can determine whether possible matches exist for multiple (sixteen (16) in a preferred, but nonlimiting, embodiment) database w-mers simultaneously. While the Bloom filter stage 106 comprises sixteen (16) parallel Bloom filters in a preferred embodiment, it should be noted that more or less parallel Bloom filters can be used in the practice of the present invention, as would be understood by a person having ordinary skill in the art following the teachings presented herein.


The control logic that is either within Bloom filter stage 106 or associated with the Bloom filter stage 106 can be readily configured to identify the starting position within the database sequence of the database w-mer that is a possible match with a query w-mer. An illustrative, but nonlimiting, example of how this control logic could be configured with an illustrative flow chart diagram is shown in FIG. 7 and is generally indicated by numeral 900. The first step is for a counter to be updated each time a new clock cycle is received to reflect the new starting point within the database sequence of w-mer for that clock cycle, as indicated by numeral 902. Depending on which logic unit 406i (or units), e.g., AND logic unit, finds a possible match 904, an appropriate starting position for a possibly matching w-mer can be recorded. If a possible match is found, then i can be recorded 906 and for each recorded value of i, return a starting point (SP) plus the value of i as the pointer for a new starting position for a possible match in the database sequence 908. The process then determines if there is a new clock cycle 910. If the answer is negative, then this determination will be repeated until there is a positive determination and the starting point (SP) can be increased by q indicated by process step 912, wherein q represents the total number of database w-mers that can be simultaneously processed by the Bloom filter stage 106. The process then returns to step 904 to determine if a possible match can be found.


If no match is found in process step 904, the process then determines if there is a new clock cycle 910. If the answer is negative, then this determination will be repeated until there is a positive determination and the starting point (SP) can be increased by q indicated by process step 912, wherein q represents the total number of database w-mers that can be simultaneously processed by the Bloom filter stage 106. The process then returns to step 904 to determine if a possible match can be found.


Each of these starting positions for a possible match is preferably then passed on to the string buffer unit 108, as shown in FIG. 2, as an output from the Bloom filter stage 106. Optionally, an output 124i from the Bloom filter stage can comprise the actual database w-mer that triggered the possible match, either in addition to or instead of the starting position pointer.


A preferred range of values for w is 5 to 15, with 11 being the most preferred value. However, it should be noted that values outside this range can also be used if desired by a practitioner of the invention. A preferred value for q is 16. A preferred value for k is 6. Preferred values for m are 32 kbits or 64 kbits. However, other values for q, k, and m can be used if desired by a practitioner of the invention. Additional details about the preferred Bloom filter stage 106 as well as alternative embodiments are described in publications that were previously incorporated by reference.


The string buffer unit 108 operates to receive the parallel outputs 124l through 124q that comprise pointers that identify the starting position(s) in the database sequence for each database w-mer that is a possible match with a query w-mer and/or the possibly matching database w-mers themselves. String buffer unit 108 produces as an output 126 a serialized stream of pointers to these starting positions and/or possibly matching database w-mers. Additional details about the preferred string buffer unit 108 are described in publications that were previously incorporated by reference.


For each received database w-mer pointer or database w-mer within stream 126, hashing stage 110 operates to confirm that a possible match exists between a database w-mer and a query w-mer and determine a pointer to the query w-mer that may possibly match the database w-mer and is shown in FIG. 2. If stream 126 includes only pointers and not the database w-mers themselves, then it is preferred that hashing stage 110 also receive stream 120 so that the hashing stage can extract the appropriate possibly matching database w-mers from stream 120 using the starting position pointers in stream 126. Hash functions within the hash stage operate to map each possibly matching database w-mer to a position in hash table 112.


An exemplary hashing stage 110 and hash table 112 in accordance with a preferred embodiment of the present invention is shown in FIG. 8. The hash table 112 preferably comprises a primary table 602, a secondary table 604, and a duplicate table 606. Stored at each address in primary table 602 is preferably an entry that includes a pointer to the starting position in the query sequence for the query w-mer that was mapped to that address by the hash functions for hashing stage 110. For any query w-mers that are duplicated at multiple locations within the query sequence, a duplicate bit in that query w-mer's corresponding entry in primary table 602 is preferably set.


To build the hash table 112, hash functions are used to map each query w-mer to an address in the hashing stage 110. The creation of these hash functions is preferably performed by software prior to streaming the database w-mers through the pipeline 100. Once the hash functions are created by software, these hash functions are preferably loaded into hashing stage 110 during the programming process prior to streaming the database w-mers through the pipeline. As new query biological sequences are used to program the hashing stage 110, there is a displacement table utilized by the hashing functions that is preferably updated to reflect the new query w-mers. The hashing functions themselves can remain the same. Thereafter, the hash table 112 can be programmed by processing the query w-mers through the hashing stage 110 while an appropriate control command is asserted. If desired, ordinary hash functions can be used to build the hash table 112 and process database w-mers in the hashing stage 110. However, because ordinary hashing typically produces multiple collisions, and because the hash table 112 can be stored off the programmable logic device 102, e.g., FPGA, in memory 118, e.g., SRAM, it is desirable use hashing techniques other than ordinary hashing to minimize the number of memory look-ups, and thereby avoid a potential pipeline bottleneck. In the hashing process, a collision occurs if when building a hash table 112, two different query w-mers map to the same address in the primary table 602. With ordinary hashing, such collisions are not uncommon. To resolve such collisions, a secondary table 604 is preferably used that provides different addresses for the different w-mers that mapped to the same address in the primary table 602, as is well-known. A collision bit in the address of the primary table 602 where the collision occurred is then set to identify the need to access the secondary table to properly map that w-mer.


A secondary set of hash functions are preferably used by the hashing stage 110 to map database w-mers to the secondary table 604. This secondary set of hash functions can be either ordinary hash functions, perfect hash functions, or near perfect hash functions, depending upon the preferences of a practitioner of the present invention. In a preferred hashing operation, the hash functions that map database w-mers to the primary table 602 and the hash functions that map database w-mers to the secondary table 604 simultaneously process all received database w-mers. If the address in the primary table 602 to which the database w-mer maps has its collision bit set, then the hashing stage logic uses the address in the secondary table 604 to which that database w-mer mapped, thereby avoiding the need for the hashing stage to re-hash a database w-mer if that database w-mer maps to an address where a collision occurred.


In a first alternative embodiment that seeks to minimize the number of hash table look-ups, perfect hashing functions are used to build the hash table 112 and perform the mapping of hashing stage 110. If perfect hashing is used to build hash table 112, then the secondary table 604 is not needed. Perfect hashing functions are explained in greater detail in some of the publications incorporated by reference.


In a second alternative embodiment, there is a need to reduce the number of hash table look-ups so that near perfect hashing functions are used to build the hash table 112 and perform the mapping of hashing stage 110. Near perfect hashing can be defined as hashing that operates to provide a decreased likelihood of collisions relative to the likelihood of collisions provided by ordinary hashing functions, and wherein the hashing approaches a perfect hash the longer that it executes. Near perfect hashing functions are explained in greater detail in some of the publications incorporated by reference. The preferred near perfect hashing functions are those in the family H3 chosen to be of full rank. The family of hashing functions that is designated as “H3” as well as the concept of “full rank” are fully described and disclosed in numerous publications in this area. More generally, the near-perfect hashing algorithm described herein may be used to generate a mapping of a set of binary strings of common length equal to n, referred to as keys, to indices into a hash table, so that few or no pairs of keys map to the same location in the table. The hashing logic of hashing stage 110 is preferably implemented on the programmable logic device 102, e.g., FPGA, to map possibly matching database w-mers to their respective positions in the query sequence.


If the hashing stage 110 maps a possibly matching database w-mer to an empty entry in the hash table 112, then that database w-mer can be discarded as a false positive. If the hashing stage 110 maps a possibly matching database w-mer to an entry in the hash table 112 that contains a pointer to a position in the query sequence, then neither the duplicate bit nor the collision bit will be set. Then that pointer points to the only starting position in the query sequence for the query w-mer that is a possible match to that database w-mer.


If the hashing stage 110 maps a possibly matching database w-mer to an entry in the hash table 112 for which the collision bit is set, then the hashing stage 110 looks to the address in the secondary table 604 where the secondary hash function mapped that possibly matching database w-mer. Once the proper entry in the secondary table 604 for the database w-mer is identified, then the secondary table entry is processed in the same manner as entries in the primary table (whose collision bits are not set).


If the hashing stage 110 maps a possibly matching database w-mer to an entry in the hash table 112 that contains a pointer to a position in the query sequence, and the duplicate bit is set, then that pointer points to one of a plurality of starting positions in the query sequence for the query w-mer that is a possible match to that database w-mer. To obtain the starting position(s) for each duplicate query w-mer, the duplicate table 606 is then preferably accessed. Also included in each primary table entry whose duplicate bit is set is a pointer to an address in the duplicate table 606 for that query w-mer. At that address, the duplicate table address preferably includes an entry that identifies the total number of duplicate query w-mers that exist in the duplicate table 606 for a given query w-mer. This total number is preferably followed by a contiguous array comprising, for each w-mer that is a duplicate of the query w-mer at issue, a pointer to the starting position in the query sequence for that duplicate query w-mer. Alternatively, linked list techniques can be used to chain the duplicate w-mer pointers together. Further still, the system can be configured to cancel a query sequence search if, during programming of the hash table 110, the number of duplicate w-mers in the query sequence is sufficiently large to exceed the capacity of the duplicate table 606.


Therefore, the hashing stage 110 operates to identify for each database w-mer that is a possible match to a query w-mer, the starting position in the query sequence for each query w-mer that is a possible match thereto. The preferred output 128 from the hashing stage 110 is a stream of ordered pairs, where each ordered pair comprises a pointer to the starting position in the database sequence of the database w-mer that is a possible match with a query w-mer and a pointer to a starting position in the query sequence of the query w-mer that possibly matches the database w-mer. If desired, comparison logic (not shown) can be optionally implemented on the programmable logic device 102, e.g., FPGA, following the hashing stage that operates to compare the database w-mer and query w-mer for each ordered pair to eliminate false positives.


It is worth noting that the role of the duplicate table 606 can be removed from the hash table 112 and placed either in separate memory, e.g., SRAM, or on the programmable logic device 102, e.g., FPGA, as its own pipeline stage downstream from the hashing stage 110 if desired by a practitioner of the invention.


The redundancy filter stage 114 is preferably configured to remove ordered pairs from stream 128 that are deemed redundant with other ordered pairs within stream 128. The database-to-query w-mer matches that are deemed redundant and non-redundant on the basis of a user-specified redundancy threshold N as well as the preferred design of a redundancy filter stage 114 is described in documents that were previously incorporated by reference. The output 130 from redundancy filter stage 114 is preferably a stream of ordered pairs as with stream 128, with the exception that the ordered pairs for matches that are deemed redundant with other ordered pairs are not present therein.


Also, it is worth noting that the implementation on an FPGA of the FPGA pipeline stages for the programmable logic device 102 described herein is within the skill of a person having ordinary skill in the art following the teachings herein and the teachings of (1) U.S. Pat. No. 6,711,558 entitled “Associative Database Scanning and Information Retrieval”; (2) pending U.S. patent application Ser. No. 10/153,151, filed May 21, 2002 and entitled “Associative Database Scanning and Information Retrieval Using FPGA Devices” (published as U.S. patent application publication 2003/0018630); and (3) pending PCT application PCT/US04/16398 filed May 21, 2004 and entitled “Intelligent Storage and Processing Using FPGA Devices”; the entire disclosures of all of which have been incorporated herein by reference. For example, see the description for FIG. 8 in PCT application PCT/US04/16398.


After the first stage (Stage 1) for word matching 54, being accelerated, as previously described and shown in FIG. 1, the performance of the second stage (Stage 2) that relates to the ungapped extension 56 directly determines the performance of the overall pipelined application.


The purpose of extending a w-mer is to determine, as quickly and accurately as possible, whether the w-mer arose by chance alone, or whether it may indicate a significant match. Ungapped extension in the second stage (Stage 2) 56 must decide whether each w-mer emitted from word matching is worth inspecting by the more computationally intensive, gapped extension. It is important to identify and discard spurious w-mers as early as possible in the BLAST pipeline because later stages are increasingly more complex. There exists a delicate balance between the stringency of the filter and its sensitivity, i.e. the number of biologically significant alignments that are found. A highly stringent filter is needed to minimize time spent in fruitless gapped extension, but the filter must not discard w-mers that legitimately identify long query-database matches with few differences. In the illustrative, but nonlimiting, embodiment of an implementation of a programmable logic device 102, e.g., FPGA, this filtering computation must also be parallelizable and simple enough to fit in a limited area.


The implementation of the second stage (Stage 2) 56 builds on the concepts utilized in the implementation of the first stage (Stage 1) 54 described above and disclosed in P. Krishnamurthy, J. Buhler, R. D. Chamberlain, M. A. Franklin, K. Gyang, and J. Lancaster, Biosequence Similarity Search On The Mercury System, In Proceedings of the 15th IEEE International Conference on Application-Specific Systems, Architectures, and Processors (ASAP04), Pages 365-375 (2004), which is incorporated herein by reference in its entirety.


The second stage (Stage 2) 56 deploys an ungapped extension stage utilizing a programmable logic device, e.g., FPGA 74, which operates as a prefilter stage. This design utilizes the speed of the implementation of the programmable logic device, e.g., FPGA 74, to greatly reduce the number of w-mers passed to software while retaining the flexibility of the software implementation on those w-mers that pass in the pipeline 70.


Preferably, but not necessarily, a w-mer must pass the second stage (Stage 2) 56 ungapped extension, which may utilize both hardware and software, before being released to the third stage (Stage 3) gapped extension 58. By focusing on data streaming nature of the application with overall throughput and not latency, the addition of another processing stage is very useful.


Utilizing a first illustrative software program, e.g., NCBI BLASTN, the second stage (Stage 2) 56 is now described. The ungapped extension of a w-mer into an HSP runs in two steps utilizing NCBI BLASTN. The w-mer is extended back toward the beginnings of the two sequences, then forward toward their ends. As the HSP extends over each character pair, that pair receives a reward +α if the characters match or a penalty −β if they mismatch. An HSP's score is the sum of these rewards and penalties over all its pairs. The end of the HSP in each direction is chosen to maximize the total score of that direction's extension. If the final HSP scores above a user-defined threshold, it is passed on to the third stage (Stage 3), gapped extension 58.


For long biological sequences, it is useful to terminate ungapped extension before reaching the ends of the sequences, especially if no high scoring HSP is likely to be found. In the illustrative software, BLASTN, implements early termination by an X-drop mechanism. The algorithm tracks the highest score achieved by any extension of the w-mer thus far; if the current extension scores at least X below this maximum, further extension in that direction is terminated. Ungapped extension with X-dropping allows the BLASTN software to recover HSPs of arbitrary length while limiting the average search space for a given w-mer. However, because the regions of extension can in principle be quite long, this heuristic is not as suitable for fast implementation in a typical programmable logic device 74, e.g., FPGA. Note that even though extension in both directions can be done in parallel, this was not sufficient to achieve a desired speedup on processing.


The prior processing of the second stage (Stage 2) 56 utilizing ungapped extension is shown in FIG. 9 utilizing NCBI BLAST. The parameters used here are w=5, α=1, β=−3, and X-drop=10. W is the length of a substring, α is a positive reward for character match, β is a negative penalty for character mismatch and X-drop is the reduction below the maximum level of current extension scores where further extension is terminated. The use of X-dropping allows the software to recover HSPs of an arbitrary length while limiting the average search space for a given w-mer but it is still not very fast or efficient.


The ungapped extension analysis with a preferred software program, i.e., NCBI BLAST, is generally indicated by numeral 62 and begins at the end of the w-mer and extends left. The extension stops when the running score drops 10 below the maximum score (as indicated by the arrows). This is also indicated as the user modifiable “X-drop” parameter. The same computation is then performed in the other direction. The final substring is the concatenation of the best substrings from the left and right extensions.


In the present invention, a prefilter is utilized to improve processing efficiency. The extension algorithm for a given w-mer is performed in a single forward pass over a fixed size window. This extension begins by calculating the limits of a fixed window of length Lw, which is centered on the w-mer in both the query and database stream. Afterward, the appropriate substrings of the query and the database stream are then fetched into buffers. Once the substrings are buffered, then the extension algorithm begins. The extension algorithm in pseudo code is shown on FIG. 12 and generally indicated by numeral 66 and the associated flow chart is shown on FIG. 20 and is generally indicated by numeral 500. This algorithm 66 also lends itself to hardware implementation despite the sequential expression of the computation.


This prefilter extension algorithm implements a dynamic programming recurrence that simultaneously computes the start and end of the best high-scoring segment pair (HSP) in a predetermined window. The first step 502 is to calculate predetermined window boundaries. The current position within a window is indicated by a variable “i′”. There is then a reset of the score “γ” of the best high-scoring segment pair (HSP) that terminates at position i of the window and the score “Γ” of the best high-scoring segment pair (HSP) ending at or before i to zero (0). This is followed by a reset of the variable “B” as well as the two endpoints Bmax and Emax of the best high scoring segment pair (HSP) ending at or before i to the value of zero (0).


First, for every position i in a window up to the length of the window Lw, a determination is made as to whether characters in the window match 504 and the score contribution of each character pair in the window is computed, using a reward +α, indicated by numeral 506 and a penalty −β, indicated by numeral 508 as the implementation for providing rewards for pattern matches and penalties for pattern mismatches, respectively. These contributions can be calculated independently in parallel for each pair. If there is a reward +α, it is added to the score γ, i.e., γ=γ+α. and if there is a penalty −β, it is subtracted from the score, i.e., γ=Γ−β.


Then, the algorithm determines if the score y of the best high-scoring segment pair (HSP) is greater than zero (0) 510. If this determination is positive, then there is a determination if the score “γ” of the best high-scoring segment pair (HSP) that terminates at position i of the window is greater than the score “Γ” of the best high-scoring segment pair (HSP) ending at or before i and if i (position in the window) is greater than the end of a short word match w-mer (WmerEnd) 511.


If this determination in step 510 is negative, then there is a determination of whether i is less than the start of the short word match w-mer (WmerStart) 512. If i is less than the start of the short word match w-mer, then the variable B is set to the current position within a window i incremented by one (1) and the score “γ” of the best high-scoring segment pair (HSP) is set to (0) zero 516. The next step is then a determination if i is equal to the length of the window Lw 519. If the determination in step 511 is positive, then the score “Γ” of the best high-scoring segment pair (HSP) ending at or before i is set to the score “γ” of the best high-scoring segment pair (HSP) that terminates at i, the endpoint Bmax of the best high-scoring segment pair (HSP) “Γ” ending at or before i is set to the variable B and the endpoint Emax of the best high-scoring segment pair (HSP) is set to i (position in the window) 514. The next step is then a determination if i is equal to the length of the window Lw 519. Moreover, if the determination in step 511 is negative, then there is also a determination if i is equal to the length of the window Lw 519.


If the determination in step 519 is negative, then the variable i is incremented by one (1) in step 517 and the process is returned to step 504. If this determination in step 519 is positive, then the software program proceeds to the next program step 518, which is a determination whether the score of the best high-scoring segment pair (HSP) “Γ” is greater than the variable T in step 518. If this determination is positive, then a value of “true” is returned 520 and the software program ends 527.


If the determination in step 518 is negative, this is followed by a determination of whether the position of the endpoint Bmax of the best high-scoring segment pair (HSP) is equal to zero 522. If this determination is positive, then a value of “true” is returned 526 and the software program ends 527.


Finally, if the determination in step 522 is negative, there is a determination of whether the endpoint Emax of the best high-scoring segment pair (HSP) is equal to the length of the window Lw 524. If this determination is positive, then a value of “true” is returned 526 and the algorithm ends 527 and if this determination is negative, then a value of “false” is returned 528 and the algorithm will again end 527.


In summary, the algorithm tracks Γ, which is the score of the highest scoring HSP ending before i, along with its endpoints Bmax and Emax. Therefore, if ΓLw is greater than a user-defined threshold score, the w-mer passes the prefilter and is forwarded to (Stage 2b) 84 for the ungapped extension.


There are two aspects of this prefilter algorithm. The first aspect is that the recurrence requires that the HSP found by the algorithm pass through its original matching w-mer and that a higher-scoring HSP in the window that does not contain this w-mer is ignored. This constraint ensures that, if two distinct biological features appear in a single window, the w-mers generated from each have a chance to generate two independent HSPs. Otherwise, both w-mers might identify only the feature with the higher-scoring HSP, causing the other feature to be ignored.


The second aspect is if the highest scoring HSP intersects the bounds of the window, it is passed on to software regardless of the score. This heuristic ensures that HSPs that might extend well beyond the window boundaries are properly found by software, which has no fixed size window limits, rather than being prematurely eliminated.


An illustrative example of the ungapped extension prefilter utilizing the illustrative Mercury BLAST is shown in FIG. 10. Using the same parameters of FIG. 9 with the addition of a new parameter Lw, which is the length of a predetermined window, then Lw=19, w=5, α=1, β=−3, the present invention, e.g., Mercury BLAST, ungapped extension, as generally indicated by numeral 65, begins at the leftmost base of the window (indicated by brackets) and moves right, calculating the best-scoring substring in the window. In this example, the algorithms gave the same result in both FIGS. 9 and 10. However, this is not necessarily the situation in general.


An illustrative system of the present invention is shown in schematic form as the application pipeline and is generally indicated by numeral 70 is shown in FIG. 11. This system is preferably an infrastructure designed to accelerate disk-based computations and exploits the high input/output (I/O) bandwidth available from modern disks by streaming data 52 directly from the disk medium to programmable logic devices such as, but not limited to, FPGAs (field programmable gate arrays) 74.


The streaming data is delivered to the hardware word matching prefilter 76, which is designated as Stage 1a. After the data stream is prefiltered, it passes into a word matching module 80, as previously described above and is designated as Stage 1b. The word matching module 80 utilizes memory 78, e.g., SRAM (static random access memory that retains data bits in its memory as long as power is being supplied and does not have to be periodically refreshed). The combination of Stage 1a and 1b, 76 and 80 form the previously described first stage (Stage 1) 54, that was previously shown in FIG. 1.


The data then passes into an ungapped extension prefilter 82 which includes dedicated hardware and/or firmware, which is designated as Stage 2a. The output of the prefilter then passes into the remainder of Stage 2, which is the ungapped extension 84, which is designated as Stage 2b and includes software. The combination of Stage 2a and 2b, 82 and 84 form the previously described second stage (Stage 2) 56, previously shown in FIG. 1.


In the alternative, if there are enough hardware/firmware resources available, the software ungapped extension 84, which is designated as Stage 2b, can be eliminated with the hardware/firmware ungapped extension prefilter 82 functioning as the sole ungapped extension, which is designated as Stage 2 and shown in FIG. 11A. This methodology can provide higher throughput since the software portion of the ungapped extension can be eliminated in the utilization of the biological sequence similarity search computation, e.g., BLAST, search.


The data stream finally passes into the gapped extension 58, which is designated as Stage 3. The ungapped extension and gapped extension can be operated by a wide variety of computing mechanisms, including, but not limited to, a processor 88. The prefilter (Stage 2a) 82 lends itself to hardware implementation despite the sequential expression of the computation in FIG. 12.


The ungapped extension prefilter design 82 is fully pipelined internally and accepts one match per clock. The prefilter is parameterizable with the parameters chosen based on a number of design-specific constraints. The commands are supported to configure parameters such as a match score, a mismatch score, and a cutoff threshold. Therefore, there is a trade-off between sensitivity and throughput that is left to the discretion of the user.


Since the w-mer matching stage generates more output than input, two independent data paths are used for input into the ungapped extension, i.e., Stage 2a, which is generally indicated by numeral 82 and is shown in FIG. 13. The w-mers/commands and the data are parsed with the w-mers/commands are received on one path 92, and the data from the database is received on the other path 93. The commands are supported by an active control valid signal and the commands are also transmitted in a predetermined format. Also, the commands are designed so that the programmable logic device 74, e.g., FPGA, shown in FIG. 11, does not need to be reconfigured.


The module is organized into three (3) pipelined stages. This includes an extension controller 94, a window lookup module 95 and a scoring module 98. There is an extension controller 94 that parses the input to demultiplex the shared w-mers command 92 and database stream 93. All w-mer matches and the database stream flow through the extension controller 94 into the window lookup module 95. The window lookup module 95 is the module that is responsible for fetching the appropriate substrings of the database stream and the query to form an alignment window.


The window lookup module 95 is illustrated in additional structural detail in FIG. 13A. Preferably, but not necessarily, a query is stored on-chip utilizing dual-ported random access memory (RAM) on the programmable logic device 74, e.g., FPGA, shown in FIG. 11. The query is streamed 202 at the beginning of each biological sequence similarity search computation, e.g., BLAST search, and is fixed until the end of the database is reached. The size of the query stream 202 is limited to the amount of memory, e.g., block random access memories (RAMs) that are provided in the query buffer 204. The database is also streamed 206 in a similar manner as the query stream 202 and provided to a database buffer 208. The database buffer 208 is preferably, but not necessarily, a circular buffer. The use of a circular buffer is helpful to retain a necessary section of the database stream 206 that any windows that are formed from the arriving w-mers might require. Since the w-mer generation is performed in the first stage 54, as shown in FIG. 1, only a small portion of the database stream 206 needs to be buffered in the database buffer 208 in order to accommodate each extension request.


The database buffer 208 is preferably, but not necessarily, built to allow a fixed distance of w-mers that are out-of-order to be processed correctly. For the previously referenced BLASTP implementation, this is important since the w-mers in the word matching stage 80, as shown in FIG. 11, will possibly be out-of-order.


The window lookup module 95 preferably, but not necessarily, is organized as a six (6) stage pipeline that is shown in FIG. 13A. This includes a first stage 210, a second stage 212, a third stage 214, a fourth stage 216, a fifth stage 218 and a sixth stage 220. Extensive pipelining may be necessary to keep up with requests from the previous stage, i.e., extension controller 94 shown in FIG. 13, which may come once per clock pulse.


The first stage of the pipeline 210 calculates the beginning of the query and database windows based on the incoming w-mer and a configurable window size. The offset of the query 222 and the offset of the database 224 is provided to the first stage 210. This first stage of the pipeline 210 calculates the beginning of the query window and database windows based on the incoming w-mer and a configurable window size. The offset of the beginning of the window is then passed to the buffer modules, i.e., the second stage 212, the third stage 214, the fourth stage 216 as well as the fifth stage 218, which begins the task of reading a superset of the window from memory, e.g., block RAMs. An extra word of data must be retrieved from the memory, e.g., block RAMs, because there is no guarantee that the window boundaries will fall on a word boundary. Therefore, an extra word is fetched on each lookup so that the exact window can be constructed from a temporary buffer holding a window size worth of data plus the extra word. The next four (4) stages, i.e., the second stage 212, the third stage 214, the fourth stage 216 and the fifth stage 218, move the input data lock-step with the buffer lookup process and ensure that pipeline stalls are handled in a correct manner. Finally, the superset of the query and database windows arrive to the final stage or sixth stage for format output 220. The output includes a seed out 226, a query window 228 and a database window 230 so that the correct window of the buffers is extracted and registered as the output.


In an illustrative, but nonlimiting, embodiment, the query is buffered on-chip using the dual-ported block random access memory (BRAM) on to programmable logic devices such as, but not limited to, FPGAs (field programmable gate arrays) 74, as shown in FIG. 11. Preferably, but not necessarily, the database stream is preferably buffered in a circular buffer, which can be created from BRAMs. As previously mentioned, the w-mer generation is done in the first stage for the extension controller 94. Only a small amount of the database stream 93 needs to be buffered to accommodate each input w-mer because they arrive from Stage 1a, 76, and Stage 1b, 80, in order with respect to the database stream 93.


Since the illustrative, but nonlimiting, dual-ported block random access memories (BRAMs) are a highly-utilized resource in Stage 1a, 76, and Stage 1b, 80, the memories, e.g., dual-ported block random access memories BRAMs, are time-multiplexed to create a quad ported BRAM structure. After the window is fetched, it is passed into the scoring module 98 and stored in registers. The scoring module 98 implements the recurrence of the extension algorithm 66 on FIG. 12. Since the computation is too complex to be done in a single cycle, the scorer is extensively pipelined.


The first stage of the scoring pipeline 98 is shown in FIG. 13. The base comparator 96 receives every base pair in parallel registers as shown by numerals 302 and 304, respectively as shown in FIGS. 13 and 14. These pairs are fed into a plurality of comparators 306 that assign a comparison score to each base pair in the window. In the illustrative but preferred embodiment, for DNA sequences, e.g., BLASTN, the base comparator 96 assigns a reward +α, generally indicated by numeral 310, to each matching base pair and a penalty −β, generally indicated by numeral 312 to each mismatching pair. The score computation is the same for protein sequence analysis, e.g., BLASTP, except there are many more choices for the score. In BLASTP, the reward +α and the penalty −β are replaced with a value from a lookup table that is indexed by the concatenation of the two symbols 302, 304. The calculation of all of the comparison scores is performed in a single cycle using Lw comparators. After the scores are calculated, the calculated scores are delivered to later scoring stages 309.


The scoring module 98 is preferably, but not necessarily, arranged as a classic systolic array. The data from the previous stage are read on each clock pulse and results are output to the following stage on the next clock pulse. Storage for comparison scores in successive pipeline stages 97 decrease in every successive stage and is shown in FIG. 13. This decrease is possible because the comparison score for window position “i” is consumed in the ith pipeline stage and may then be discarded, since later stages inspect only window positions that are greater than i.


This structure of data movement is shown in more detail in FIG. 15, which is generally indicated by numeral 97 as a systolic array of scoring stages. The darkened registers 326 hold the necessary comparison scores for the w-mer being processed in each pipeline stage. Although FIG. 15 shows a single comparison score being dropped for each scorer stage for ease of explanation, the preferred embodiment includes dropping two comparison scores per stage. There is also a plurality of pipeline scoring stages 330. The data flows from left to right on FIG. 15. The light registers 328 are pipeline calculation registers and are used to transfer the state of the recurrence from a previous scoring stage to the next with each column of registers containing a different w-mer in the pipeline 70. Initialization 322 is provided to the light registers 328, dark registers 326 and scoring stages 330.


The interface of an individual scoring stage is generally indicated by numeral 330 and is shown in FIG. 16. The values shown entering the top of the scoring stage 330 are the state of dynamic programming recurrence propagated from the previous scoring stage. These values are read as input to a first register 342 and the results are stored in the output registers 364 shown on the right. Each scoring stage 362 in the pipeline 97 contains combinational logic to implement the dynamic programming recurrence shown in Lines 12-19 of the algorithm described in FIG. 12. This includes a value for γ indicated by numeral 344, a value for B indicated by numeral 346, a value for Γ indicated by numeral 348, a maximum value of B, i.e., Bmax, indicated by numeral 350 and a maximum value of E, i.e., Emax, indicated by numeral 352.


The data entering from the left of the second register 360 are the comparison scores 358, i.e., α/β, and the database offset 354 and query offset 356 for a given w-mer, which are independent of the state of the recurrence. In order to sustain a high clock frequency design, each scoring stage computes only two iterations of the loop per clock cycle, resulting in Lw/2 scoring stages for a complete calculation. This provides step 1, indicated by numeral 361 and step 2, indicated by numeral 363, which implements two iterations of Lines 12 through 19 of the algorithm 66 shown in FIG. 12. This is merely an illustrative, but nonlimiting, number of loop iterations within a single stage since any number of loop iterations can be utilized. This tradeoff is preferred when contrasting area and speed utilizing the preferred and illustrative, but nonlimiting, hardware. Therefore, there are Lw/2 independent w-mers being processed simultaneously in the scoring stages 362 of the processor when the pipeline 70 is full.


The pipeline calculation registers output 364 that are provided to subsequent scoring stages, e.g., i+1. This includes a value for y indicated by numeral 366, a value for B indicated by numeral 368, a value for Γ indicated by numeral 370, a maximum value of B, i.e., Bmax, indicated by numeral 372 and a maximum value of E, i.e., Emax, indicated by numeral 374.


The final pipeline stage of the scoring module is the threshold comparator 99 that is shown in FIG. 13. The comparator takes the fully-scored segment and makes a decision to discard or keep the segment. This decision is based on the score of the alignment relative to a user-defined threshold T, as well as the position of the highest-scoring substring. If the maximum score is above the threshold, the segment is passed on. Additionally, if the maximal scoring substring intersects either boundary of the window, the segment is also passed on, regardless of the score. If neither condition holds, the substring of a predetermined length, i.e., segment, is discarded. The segments that are passed on are indicated by numeral 100 on FIG. 13.


In an optional and illustrative implementation, the ungapped extension prefilter 82, as shown in FIG. 13, utilizes approximately 38% of the logic cells and 27 BRAMs on a reconfigurable hardware such as, but not limited to, FPGAs (field programmable gate arrays) 74. An illustrative, but nonlimiting, example includes a Xilinx® Virtex-II 6000® series FPGA. Xilinx, Inc., is a Delaware Corporation, having a place of business at 2100 Logic Drive, San Jose, Calif. 95124-3400. This includes the infrastructure for moving data in and out of the reconfigurable hardware, e.g., FPGA, itself. The illustrative, but nonlimiting, current design runs at 96 MHz, processing one w-mer per clock. The full Mercury BLASTN design utilizes approximately 65% of the logic cells and 134 BRAMs.


There are many aspects to the performance of an ungapped extension prefilter 82, as shown in FIG. 13. First is the individual stage throughput. The ungapped extension stage (Stage 2a) 82 must run fast enough to not be a bottleneck in the overall pipeline 70. Second, the ungapped extension stage (Stage 2a) 82 must effectively filter as many w-mers as possible, since downstream stages are even more computationally expensive. Finally, the above must be achieved without inadvertently dropping a large percentage of the significant alignments (i.e., the false negative rate must be limited).


First, throughput performance for the ungapped extension stage alone is a function of data input rate. In an illustrative, but nonlimiting, example, the ungapped extension stage accepts one w-mer per clock and runs at 96 MHz. Hence the maximum throughput of the filter is one (1) input match/cycle times 96 MHz=96 Mmatches/second. This provides a speedup of twenty-five (25) over the software ungapped extension 82 executed on the previously described baseline system, as shown in FIG. 1. This analysis is based solely on modeling, based on assumed parameters and constraints, and actual results can vary.



FIG. 17 shows the application throughput (quantified by the ingest rate of the complete pipeline 70, in million bases per second) as a function of the performance attainable in the second stage (Stage 2) 56, as shown in FIG. 1, which is quantified by the ingest rate of the second stage (Stage 2) 56 alone, in million matches per second). FIG. 17 includes results for both 25,000-base double stranded queries and 20,000-base double stranded queries.



FIG. 18 plots the resulting speedup versus throughput (for both query sizes) for the preferred embodiment of the present invention with a prefiltering stage (Stage 2a) 82 as shown on FIG. 11 relative to the software implementation. Once again, this analysis is based solely on modeling, based on assumed parameters and constraints, and actual results can vary. The software profiling shows, for 25,000-base queries, an average execution time for the second stage (Stage 2) 56 alone of 0.265 μs/match. This corresponds to a throughput of 3.8 Mmatches/s, plotted towards the left in FIGS. 17 and 18. As the performance of the second stage (Stage 2) 56 is increased, the overall pipeline 70 performance increases proportionately until the second stage (Stage 2) 56 is no longer the bottleneck stage.


To explore the impact this performance in the second stage (Stage 2) 56 has on the overall system, we return to the graphs of FIGS. 17 and 18. Above approximately 35 Mmatches/s, the overall system throughput is at its maximum rate of 1400 Mbases/s, with a speedup of forty-eight (48 times over that of software, e.g., NCBI BLASTN, for a 25,000-base query) and a speedup of thirty-eight (38 times over that of software, e.g., NCBI BLASTN, for a 20,000-base query).


There are two (2) parameters of ungapped extension that affect its sensitivity. The first parameter is the score threshold used, which affects the number of HSPs that are produced. The second parameter is the length of the window, which can affect the number of false negatives that are produced. The HSPs that have enough mismatches before the window boundary to be below the score threshold but have many matches immediately outside the boundary will be incorrectly rejected.


To evaluate the functional sensitivity of hardware ungapped extension 82, measurements were performed using an instrumented version of software, e.g., NCBI BLASTN. A software emulator of the new ungapped extension algorithm was placed in front of the standard ungapped extension stage, e.g., NCBI. Then, the statistics were gathered which show how many w-mers arrived at the ungapped extension prefilter stage 82, and how many passed.


These statistics were collected both for ungapped extension, the second stage (Stage 2) 56, FIG. 1, alone, e.g., NCBI BLASTN, and with the hardware emulator in place. The dataset was generated from the human and mouse genomes. The queries were statistically significant samples of various sizes (e.g., 10 kbase, 100 kbase, and 1 Mbase). The database stream was the mouse genome with low-complexity and repetitive sequences removed.



FIG. 19 is a table, generally indicated by numeral 380, which summarizes the results for a window size of 64 bases, which is the window size used in the illustrative, but nonlimiting hardware implementation. A score threshold of twenty (20) corresponds to the default value in ungapped extension or second stage (Stage 2) 82, FIG. 11, e.g., NCBI BLASTN. The reject fraction is the measured ratio of output HSPs over input w-mers. This value quantifies the effectiveness of the overall second stage (Stage 2) 82, FIG. 11, at filtering w-mers so they need not be processed in (Stage 2b) 84. The percent found is the percentage of gapped alignments present in the MERCURY BLAST's output relative to NCBI BLASTN output, as shown in FIG. 19.


Using a window length of 64 bases, the ungapped extension prefilter 82 is able to filter out between 99.5% and 99.99% of all its input. For instance, a threshold of seventeen (17) missed HSPs provides a good tradeoff between a high reject fraction while keeping the vast majority (99.95%) of the significant HSPs. This translates into five (5) missed HSPs out of 10,334 HSPs found by software system, e.g., NCBI BLAST.


Therefore, biosequence similarity searching can be accelerated practically by a pipeline 70 designed to filter high-speed streams of character data utilizing a performance-critical ungapped extension stage 56. The present invention is a highly parallel and pipelined implementation yields results comparable to those obtained from software, such as BLASTN, while running twenty-five times faster and enabling the entire accelerator to run approximately forty to fifty times faster.


For example, the techniques of the present invention can be used to implement word matching for seeded alignment searching techniques other than biological sequence similarity searching. Furthermore, the techniques of the present invention can also be applied to BLAST versions other than BLASTN, such as BLASTP, BLASTX, TBLASTN, and TBLASTX. Further still, the preferred embodiment has been described wherein the starting position of each matching w-mer within either the database sequence or the query sequence is determined by the various pipeline stages. However, it should be noted that these positions need not necessarily be the starting position. For example, the ending position or some intermediate position can also be used to identify the positions of the matching w-mers within the database and query sequences. These changes and modifications should be considered as alternative embodiments for the present invention, and the invention should be considered as limited only by the scope of the claims appended hereto and their legal equivalents.


Thus, there has been shown and described several embodiments of a novel invention. As is evident from the foregoing description, certain aspects of the present invention are not limited by the particular details of the examples illustrated herein, and it is therefore contemplated that other modifications and applications, or equivalents thereof, will occur to those skilled in the art. The terms “have,” “having,” “includes” and “including” and similar terms as used in the foregoing specification are used in the sense of “optional” or “may include” and not as “required.” Many changes, modifications, variations and other uses and applications of the present construction will, however, become apparent to those skilled in the art after considering the specification and the accompanying drawings. All such changes, modifications, variations and other uses and applications which do not depart from the spirit and scope of the invention are deemed to be covered by the invention which is limited only by the claims that follow.

Claims
  • 1. A system for generating a hash table for use in mapping a set of strings to keys, the system comprising: a processor configured to provide near perfect hashing on the set of strings to (1) identify a location in a hash table, utilizing an H3 family of hash functions, corresponding to each string, and (2) generate the hash table, the hash table being configured to store, at each identified location therein, a key; anda memory in communication with the processor for storing the hash table, wherein the processor is configured to provide near perfect hashing via hash functions, utilizing the H3 family of hash functions, that are guaranteed to be of full rank.
  • 2. The system of claim 1, wherein the near perfect hashing stage is configured with functions A and B selected from the H3 family of hash functions, wherein A and B are guaranteed to be of full rank.
  • 3. The system of claim 1, wherein the mapping of the set of strings to keys is performed by the processor configured to provide perfect hashing on the set of strings.
  • 4. A system for generating a hash table for use in mapping a set of substrings to a position in a larger string, the system comprising: a processor configured to (1) provide near perfect hashing on the set of substrings to identify a position in a hash table corresponding to each substring, and (2) generate the hash table, the hash table being configured to store, at each identified location therein, a pointer to a position of each substring in the larger string; anda memory in communication with the processor for storing the hash table.
  • 5. The system of claim 4, wherein the near perfect hashing stage is configured with functions A and B selected from the H3 family of hash functions, wherein A and B are guaranteed to be of full rank.
  • 6. The system of claim 4, wherein the mapping of the set of strings to keys is performed by the processor configured to provide perfect hashing on the set of strings.
  • 7. A method for generating a hash table for use in mapping a set of strings, the method comprising: hashing the set of strings with a processor to identify a position in the hash table corresponding to each string;generating the hash table, the hash table being configured to store, at each identified location therein, a pointer to a position of each string; andstoring the hash table within a memory, the memory in communication with the processor.
  • 8. The method of claim 7, wherein a pipeline includes a Bloom filter located upstream of the hashing done on the set of strings to identify the position in the hash table corresponding to each string.
  • 9. The method of claim 8, wherein a possible match from the Bloom filter is mapped to the position in the hash table corresponding to the string that caused the possible match.
  • 10. The method of claim 9, wherein the memory configured to store the hash table stores at least one pointer to the position in the string for the possible match from the Bloom filter.
  • 11. The method of claim 7, wherein the hash table used for mapping the set of strings is created via perfect hashing.
  • 12. The method of claim 7, wherein the hash table used for mapping the set of strings is created via near perfect hashing.
  • 13. The system of claim 12, wherein the near perfect hashing is configured with functions A and B selected from an H3 family of hash functions, wherein A and B are guaranteed to be of full rank.
CROSS-REFERENCE AND PRIORITY CLAIM TO RELATED APPLICATIONS

This application is a divisional of U.S. patent application Ser. No. 13/969,935 filed Aug. 19, 2013, now U.S. Pat. No. ______, which is a divisional of U.S. patent application Ser. No. 13/046,395, filed Mar. 11, 2011, now U.S. Pat. No. 8,515,682, which is a divisional of U.S. patent application Ser. No. 11/359,285, filed Feb. 22, 2006, now U.S. Pat. No. 7,917,299, the entire disclosures of which are incorporated herein by reference, which claims the benefit of U.S. Provisional Application No. 60/658,418, filed on Mar. 3, 2005 and also claims the benefit of U.S. Provisional Application No. 60/736,081, filed on Nov. 11, 2005.

STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT

This invention was made with government support under grant/contract numbers 0217334, 0237902, 0313203, 0427794, and R42 HG003225 awarded by the National Institutes of Health (NIH). The government has certain rights in the invention.

Provisional Applications (2)
Number Date Country
60658418 Mar 2005 US
60736081 Nov 2005 US
Divisions (3)
Number Date Country
Parent 13969935 Aug 2013 US
Child 15403687 US
Parent 13046395 Mar 2011 US
Child 13969935 US
Parent 11359285 Feb 2006 US
Child 13046395 US