Method and system for high performance pattern indexing

Information

  • Patent Grant
  • 9323794
  • Patent Number
    9,323,794
  • Date Filed
    Tuesday, November 27, 2012
    11 years ago
  • Date Issued
    Tuesday, April 26, 2016
    8 years ago
Abstract
Disclosed herein is a method and system for accelerating the generation of pattern indexes. In exemplary embodiments, regular expression pattern matching can be performed at high speeds on data to determine whether a pattern is present in the data. Pattern indexes can then be built based on the results of such regular expression pattern matching. Reconfigurable logic such a field programmable gate arrays (FPGAs) can be used to hardware accelerate these operations.
Description

This patent application is related to U.S. patent application Ser. No. 11/938,709, filed Nov. 12, 2007, now U.S. Pat. No. 7,660,793, the entire disclosure of which is incorporated herein by reference.


This patent application is also related to U.S. patent application Ser. No. 12/640,891, filed Dec. 17, 2009, now U.S. Pat. No. 8,156,101, and U.S. patent application Ser. No. 13/442,442, filed Apr. 9, 2012, published as U.S. Pat. App. Pub. 2013/0007000.


FIELD OF THE INVENTION

The present invention is generally directed toward the field of metadata generation and data indexing, particularly the indexing of high speed data streams.


TERMINOLOGY

The following paragraphs provide several definitions for various terms used herein. These paragraphs also provide background information relating to these terms.

  • GPP: As used herein, the term “general-purpose processor” (or GPP) refers to a hardware device having a fixed form and whose functionality is variable, wherein this variable functionality is defined by fetching instructions and executing those instructions (for example, an Intel Xeon processor or an AMD Opteron processor), of which a conventional central processing unit (CPU) is a common example.
  • Reconfigurable Logic: As used herein, the term “reconfigurable logic” refers to any logic technology whose form and function can be significantly altered (i.e., reconfigured) in the field post-manufacture. This is to be contrasted with a GPP, whose function can change post-manufacture, but whose form is fixed at manufacture.
  • Software: As used herein, the term “software” refers to data processing functionality that is deployed on a GPP or other processing devices, wherein software cannot be used to change or define the form of the device on which it is loaded.
  • Firmware: As used herein, the term “firmware” refers to data processing functionality that is deployed on reconfigurable logic or other processing devices, wherein firmware may be used to change or define the form of the device on which it is loaded.
  • Coprocessor: As used herein, the term “coprocessor” refers to a computational engine designed to operate in conjunction with other components in a computational system having a main processor (wherein the main processor itself may comprise multiple processors such as in a multi-core processor architecture). Typically, a coprocessor is optimized to perform a specific set of tasks and is used to offload tasks from a main processor (which is typically a GPP) in order to optimize system performance. The scope of tasks performed by a coprocessor may be fixed or variable, depending on the architecture of coprocessor. Examples of fixed coprocessor architectures include Graphics Processor Units which perform a broad spectrum of tasks and floating point numeric coprocessors which perform a relatively narrow set of tasks. Examples of reconfigurable coprocessor architectures include reconfigurable logic devices such as Field Programmable Gate Arrays (FPGAs) which may be reconfigured to implement a wide variety of fixed or programmable computational engines. The functionality of a coprocessor may be defined via software and/or firmware.
  • Hardware Acceleration: As used herein, the term “hardware acceleration” refers to the use of software and/or firmware implemented on a coprocessor for offloading one or more processing tasks from a main processor to decrease processing latency for those tasks relative to the main processor.
  • Enterprise: As used herein, the term “enterprise” refers to any business organization or governmental entity that stores and/or processes data (referred to as “enterprise data”) as part of its ongoing operations.
  • Database: As used herein, the term “database” refers to a persistent data store with indexing capabilities to expedite query processing. Various database management system (DBMS) implementations might be categorized as relational (RDBMS), object-oriented (OODBMS), hierarchical, etc.; however, the dominant architecture in today's industry is a relational, row-column, structured query language (SQL)-capable database. An ANSI-standard SQL database engine is a mature software architecture that can retrieve structured data in response to a query, usually in an efficient manner.
  • Structured Data: As used herein, the term “structured data” refers to data that has been normalized and persisted to a relational database. Normalization is the data design process of putting data into a tabular, row-column format and abstracting duplicate data into separate tables. Structured data in relational columns is capable of being indexed with B-tree indexes, significantly speeding access to the data in these columns. In SQL terms, structured columns have size limits. These columns may have constraints and referential integrity applied to them in order to ensure consistent data quality. Examples of common structured SQL datatypes are: INT(eger), NUMBER, CHAR(acter), VARCHAR, DATE, TIMESTAMP.
  • Unstructured Data: As used herein, the term “unstructured data” refers to data that falls outside the scope of the definition above for structured data. Thus, the term unstructured data encompasses files, documents or objects with free form text or embedded values included therein. This data includes the complete set of bytes, often including binary-format data, that was used by the application that generated it. Examples of unstructured data include word processing documents (e.g., Microsoft Word documents in their native format), Adobe Acrobat documents, emails, image files, video files, audio files, and other files in their native formats relative to the software application that created them. In SQL terms, unstructured columns have very large, if not unlimited size. Common examples of unstructured SQL datatypes are: BLOB, TEXT, XML, RAW, and IMAGE. Unstructured objects may also be stored outside the database, for example in operating system files. Access to these external objects from within the database engine uses links in the metadata in the database table to the storage location.


There are a number of reasons why XML is will not normally be categorized as “structured” as that term is used herein:

    • XML may have large or unlimited sized values
    • XML often does not have strongly enforced datatyping
    • XML has a flexible schema
    • XML values in elements and attributes is often not as rigidly conformed and carefully cleansed as traditional “structured” database columns


      Although the concept of “semi-structured” data with flexible schemas is emerging, particularly for XML, for present purposes everything that is not has not been normalized and persisted to a relational database will be considered unstructured data. As such, a column that is of the XML datatype would thus fall under this present definition of “unstructured data”.
  • Metadata: As used herein, the term “metadata” in the context of data objects and documents refers to data that describes or characterizes a data object or document. Examples of object and document metadata include but are not limited to file type, byte size, date created, date last modified, author, title, information about the document's/object's data source (optionally including the name and version number of the program used to generate the document), information about whether the data matches other data, subject coverage, classification information (e.g. information about its concepts, person/place/entity names that are found within the document/data object, word counts, etc.), position information relating to data within a document/object, or other content-derived information about documents/objects.
  • Bus: As used herein, the term “bus” refers to a logical bus which encompasses any physical interconnect for which devices and locations are accessed by an address. Examples of buses that could be used in the practice of the present invention include, but are not limited to the PCI family of buses (e.g., PCI-X and PCI-Express) and HyperTransport buses.
  • Pipelining: As used herein, the terms “pipeline”, “pipelined sequence”, or “chain” refer to an arrangement of application modules wherein the output of one application module is connected to the input of the next application module in the sequence. This pipelining arrangement allows each application module to independently operate on any data it receives during a given clock cycle and then pass its output to the next downstream application module in the sequence during another clock cycle.
  • Full-Text Search: As used herein, the term “full-text search” refers to scanning through the entire body of a document or object, considering every word or byte. This processing may allow for approximations, tokenization based on tagging for a flexible schema, wild-carding, or complex matching.
  • Text Analytics and Text Mining: As used herein, the terms “text analytics” and “text mining” refer to algorithms that operate on document objects using complex language concepts such as semantics. Examples of text analytic/text mining processing include: named entity recognition, content extraction, document classification, document summarization, natural language processing, statistical pattern learning, and relevance ranking.


BACKGROUND OF THE INVENTION

Enterprises such as corporations, institutions, agencies, and other entities have massive amounts of data that they need to manage. While some of an enterprise's critical data are normalized, structured, and stored in relational databases, most enterprises' data (generally thought to be around 80% of enterprise data) is unstructured. With conventional computing systems, effective management and efficient access to such unstructured data is problematic.


Indexing is a well-known technique that is used to increase the efficiency by which data can be searched. An index is a list of terms and pointers associated with a collection of data. An example of such an index 100 is shown in FIG. 1. Index 100 comprises a plurality of index entries 102, with each index entry 102 comprising a term 104 (see the “term” column in the table) and one or more pointers 106 (see the “pointer(s)” column in the table). The terms 104 in an index can be words, phrases, or other information associated with the data. In many situations, these terms are user-specified. Each pointer 106 in an index corresponds to the term 104 for that entry 102 and identifies where that term can be found in the data. With unstructured data, the data collection often comprises a plurality of documents. Examples of documents include items such a word processing files, spreadsheet files, emails, images, Adobe Acrobat files, web pages, books, pages of books, etc.


However, the inventors note their belief that conventional indexing techniques require a tremendous amount of time to generate an effective index. Even relatively small data sets can take days to effectively index with conventional indexing techniques deployed in software on central processors such as GPPs because of indexing's computationally-intensive nature. Because of the sheer volume of data that enterprises encounter on a daily basis, it is simply not practical for enterprises to index all of the data in its possession (and to which it has access) using these conventional indexing techniques. Instead, enterprises are forced to make a priori decisions as to which data will be subjected to indexing; this is particularly true for unstructured data which comprises the bulk of most enterprises' data. In doing so, enterprises are left without an effective means for efficiently managing and searching much of its data.


SUMMARY OF THE INVENTION

In an effort to address this need in the art, the inventors herein disclose a technique for hardware-accelerating the generation of metadata for data. This data may comprise both structured and/or unstructured data. From this metadata, rich indexes can be generated to enable robust and high performance data searching and analysis. With embodiments of the present invention, data is streamed into a coprocessor, and metadata for the indexes is generated at bus bandwidth rates, thereby leading to dramatic improvements in indexing latency. Because of such improvements in latency, much larger amounts (if not all) of an enterprise's data can be efficiently and effectively indexed.


In doing so, the present invention preferably harnesses the underlying hardware-accelerated technology disclosed in the following patents and patent applications: U.S. Pat. No. 6,711,558 entitled “Associated Database Scanning and Information Retrieval”, U.S. Pat. No. 7,139,743 entitled “Associative Database Scanning and Information Retrieval using FPGA Devices”, U.S. Patent Application Publication 2006/0294059 entitled “Intelligent Data Storage and Processing Using FPGA Devices”, U.S. Patent Application Publication 2007/0067108 entitled “Method and Apparatus for Performing Biosequence Similarity Searching”, U.S. Patent Application Publication 2008/0086274 entitled “Method and Apparatus for Protein Sequence Alignment Using FPGA Devices” (published from U.S. application Ser. No. 11/836,947, filed Aug. 10, 2007), U.S. Patent Application Publication 2007/0130140 entitled “Method and Device for High Performance Regular Expression Pattern Matching”, U.S. Patent Application Publication 2007/0260602 entitled “Method and Apparatus for Approximate Pattern Matching” (published from U.S. application Ser. No. 11/381,214, filed May 2, 2006), U.S. Patent Application Publication 2007/0174841 entitled “Firmware Socket Module for FPGA-Based Pipeline Processing”, and U.S. Patent Application Publication 2007/0237327 entitled “Method and System for High Throughput Blockwise Independent Encryption/Decryption”), the entire disclosures of each of which are incorporated herein by reference.


As documents are streamed into a coprocessor, any of a number of metadata generation operations can be flexibly performed on the document stream. For example, exact matching and approximate matching operations can be performed on the words within a document stream to find which words within the document stream are “hits” against the words in any of a number of dictionaries. From these “hits”, rich indexes can quickly be generated.


For ease of reference, the term “documents” will be often be used to describe the unstructured data that is indexed in accordance with various embodiments of the present invention. However, it should be noted that the use of the term “documents” is meant to be only exemplary as other forms unstructured data can also be indexed using the techniques described herein. Also, the term “words” is used to describe bits within a data stream that are grouped together for the purposes of metadata generation operations. Words preferably correspond to the bytes that form discrete character strings, wherein the characters may be any combination of letters, numbers, or other characters.


In accordance with an extremely powerful aspect of an embodiment of the invention, the coprocessor can perform regular expression pattern matching on a received document stream to detect whether any of a plurality of pre-defined patterns exist within the document stream. Examples of patterns that can detected using regular expression pattern matching include social security numbers, credit card numbers, telephone numbers, email address, etc. By developing indexes based on these patterns, an enterprise can effectively answer an indefinite pattern-based question such as “How many of our documents contain a credit card number?” without having to re-analyze all of the enterprise's documents. Instead, the generated index pertaining to the credit card pattern can be accessed to lookup all documents that have been identified as containing a credit card number.


Furthermore, classification and clustering operations can be performed on documents using a coprocessor to further analyze and index those documents.


The high performance stream indexing that is enabled by the present invention has a myriad of useful applications. For example, the email traffic within and/or incoming to an enterprise can be efficiently indexed “on the fly” before or as that email traffic reaches its addressees.


Similarly, the web traffic (e.g., the web pages downloaded by user computers within an enterprise) can also be indexed “on the fly” using various embodiments of the present invention.


Web search engines could employ the embodiments of the present invention to stream web page content through an appliance configured to perform the metadata generation and index generation techniques described herein. The web search engine can then apply search queries against the generated index(es) when formulating responses to those search queries. It is expected that the present invention can be employed to dramatically reduce the lag time that it takes for new web pages to become “findable” through Internet search engines.


Also, the embodiments of the present invention can be employed to operate on data being saved to a disk by any computers within an enterprise.


Another application pertains to analyzing and indexing data streams such as incoming sales transactions, incoming news wire information, and log files.


Yet another application pertains to retroactively indexing existing enterprise data, including indexing previously-unindexed enterprise data and re-indexing previously-indexed enterprise data.


These and other features and advantages of the present invention will be apparent to those having ordinary skill in the art upon review of the following description and drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 depicts an exemplary index;



FIG. 2(a) illustrates an exemplary metadata generating appliance in accordance with an embodiment of the present invention;



FIG. 2(b) illustrates an exemplary metadata generating appliance in accordance with another embodiment of the present invention;



FIG. 2(c) illustrates an exemplary metadata generating appliance in accordance with yet another embodiment of the present invention;



FIGS. 3(a) and (b) illustrate exemplary printed circuit boards for use in the appliances of FIGS. 2(a)-(c);



FIG. 4 illustrates an example of how a firmware pipeline can be deployed across multiple reconfigurable logic devices;



FIG. 5(a) is a high level block diagram view of how a coprocessor can be used to generate metadata for data that streams therethrough;



FIG. 5(b) depicts an exemplary index that can be created from the metadata produced by the coprocessor of FIG. 5(a);



FIG. 6 depicts an exemplary firmware pipeline that can be deployed within reconfigurable logic to perform a desired metadata generation operation;



FIG. 7(a) depicts an exemplary firmware pipeline that can be used to perform the stream pre-processing operation shown in FIG. 6;



FIG. 7(b) depicts an exemplary output stream of words and feature vectors that can be produced by the firmware pipeline of FIG. 7(a);



FIG. 8(a) depicts an exemplary specialized index generation operation that seeks to exactly match the words within the incoming data stream against the words in a dictionary to thereby generate a specialized index;



FIG. 8(b) depicts an exemplary specialized index generation operation that seeks to approximately match the words within the incoming data stream against the words in a dictionary to thereby generate a specialized index;



FIG. 9(a) depicts an exemplary specialized index generation operation that seeks to exactly match the words within the incoming data stream against the words in a plurality of different dictionaries to thereby generate a plurality of specialized indexes;



FIG. 9(b) depicts an exemplary specialized index generation operation that seeks to match the words within the incoming data stream against a plurality of different regular expression patterns to thereby generate a plurality of specialized indexes;



FIG. 9(c) depicts an exemplary specialized index generation operation that seeks to match the words within the incoming data stream against the words in a plurality of different dictionaries and against a plurality of different regular expression patterns to thereby generate a plurality of specialized indexes;



FIG. 10 depicts an exemplary technique for inserting the generated indexes into operational indexes maintained by a relational database;



FIG. 11(a) depicts a stream pre-processing firmware pipeline that includes a word counting module;



FIG. 11(b) depicts an exemplary index that includes metadata relating to the word counting function of FIG. 11(a);



FIG. 11(c) depicts an exemplary firmware pipeline wherein the generated metadata corresponds to classification data about the incoming documents;



FIG. 12(a) depicts an exemplary classification operation that produces class scores for incoming documents to assess their relevance to a plurality of different classifications;



FIG. 12(b) depicts an exemplary classification operation that determines a classification for each incoming document;



FIGS. 13(a) and (b) depict exemplary firmware pipelines that employ histogram generation to enable document clustering;



FIGS. 14(a) and (b) depict exemplary firmware pipelines that perform classification, specialized index generation, and histogram generation operations on incoming documents;



FIG. 15 depicts an exemplary environment in which the appliance of FIGS. 2(a)-(c) can be employed;



FIG. 16 illustrates an exemplary overview of a document ingest pre-processing operation that can employ the metadata generation techniques described herein;



FIG. 17 illustrates an exemplary logical view of a document ingest pre-processing operation in accordance with the FIG. 2(c) embodiment; and



FIG. 18 illustrates an exemplary data flow within the appliance of FIG. 2(c) for a document ingest pre-processing operation that can employ the metadata generation techniques described herein.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS


FIG. 2(a) depicts an exemplary embodiment for a metadata generating appliance 200. While an embodiment of appliance 200 can be referred to as a metadata generating appliance, it should be noted that functionalities in addition to metadata generation can be supported by appliance 200, as explained in the above-referenced and incorporated U.S. Pat. No. 7,660,793.


Preferably, appliance 200 employs a hardware-accelerated data processing capability through coprocessor 450 to generate the desired metadata. Within appliance 200, a coprocessor 450 is positioned to receive data that streams into the appliance 200 from a network 420 (via network interface 410). Network 420 preferably comprises an enterprise network (whether LAN or WAN), in which various disparate data sources are located. Such incoming data may comprise both structured and unstructured data as appliance 200 can provide beneficial metadata generation operations for both data types.


The computer system defined by processor 412 and RAM 408 can be any commodity computer system as would be understood by those having ordinary skill in the art. For example, the computer system may be an Intel Xeon system or an AMD Opteron system. Thus, processor 412, which serves as the central or main processor for appliance 200, preferably comprises a GPP.


In a preferred embodiment, the coprocessor 450 comprises a reconfigurable logic device 402. Preferably, data streams into the reconfigurable logic device 402 by way of system bus 406, although other design architectures are possible (see FIG. 3(b)). Preferably, the reconfigurable logic device 402 is a field programmable gate array (FPGA), although this need not be the case. System bus 406 can also interconnect the reconfigurable logic device 402 with the appliance's processor 412 as well as the appliance's RAM 408. In a preferred embodiment, system bus 406 may be a PCI-X bus or a PCI-Express bus, although this need not be the case.


The reconfigurable logic device 402 has firmware modules deployed thereon that define its functionality. The firmware socket module 404 handles the data movement requirements (both command data and target data) into and out of the reconfigurable logic device, thereby providing a consistent application interface to the firmware application module (FAM) chain 350 that is also deployed on the reconfigurable logic device. The FAMs 350i of the FAM chain 350 are configured to perform specified data processing operations on any data that streams through the chain 350 from the firmware socket module 404. Preferred examples of FAMs that can be deployed on reconfigurable logic in accordance with a preferred embodiment of the present invention are described below.


The specific data processing operation that is performed by a FAM is controlled/parameterized by the command data that FAM receives from the firmware socket module 404. This command data can be FAM-specific, and upon receipt of the command, the FAM will arrange itself to carry out the data processing operation controlled by the received command. For example, within a FAM that is configured to perform an exact match operation, the FAM's exact match operation can be parameterized to define the key(s) that the exact match operation will be run against. In this way, a FAM that is configured to perform an exact match operation can be readily re-arranged to perform a different exact match operation by simply loading new parameters for one or more different keys in that FAM.


Once a FAM has been arranged to perform the data processing operation specified by a received command, that FAM is ready to carry out its specified data processing operation on the data stream that it receives from the firmware socket module. Thus, a FAM can be arranged through an appropriate command to process a specified stream of data in a specified manner. Once the FAM has completed its data processing operation, another command can be sent to that FAM that will cause the FAM to re-arrange itself to alter the nature of the data processing operation performed thereby. Not only will the FAM operate at hardware speeds (thereby providing a high throughput of data through the FAM), but the FAMs can also be flexibly reprogrammed to change the parameters of their data processing operations.


The FAM chain 350 preferably comprises a plurality of firmware application modules (FAMs) 350a, 350b, . . . that are arranged in a pipelined sequence. However, it should be noted that within the firmware pipeline, one or more parallel paths of FAMs 350i can be employed. For example, the firmware chain may comprise three FAMs arranged in a first pipelined path (e.g., FAMs 350a, 350b, 350c) and four FAMs arranged in a second pipelined path (e.g., FAMs 350d, 350e, 350f, and 350g), wherein the first and second pipelined paths are parallel with each other. Furthermore, the firmware pipeline can have one or more paths branch off from an existing pipeline path. A practitioner of the present invention can design an appropriate arrangement of FAMs for FAM chain 350 based on the processing needs of a given application.


A communication path 430 connects the firmware socket module 404 with the input of the first one of the pipelined FAMs 350a. The input of the first FAM 350a serves as the entry point into the FAM chain 350. A communication path 432 connects the output of the final one of the pipelined FAMs 350m with the firmware socket module 404. The output of the final FAM 350m serves as the exit point from the FAM chain 350. Both communication path 430 and communication path 432 are preferably multi-bit paths.


The nature of the software and hardware/software interfaces used by appliance 200, particularly in connection with data flow into and out of the firmware socket module are described in greater detail in the above-referenced and incorporated U.S. Patent Application Publication 2007/0174841.



FIG. 2(b) depicts another exemplary embodiment for appliance 200. In the example of FIG. 2(b), appliance 200 includes a relational database management system 304 that is in communication with bus 406 via disk controller 414. Thus, the data that is streamed through the coprocessor 450 may also emanate from RDBMS 304. Furthermore, the metadata generated by coprocessor 450 can be stored as structured data within RDBMS 304 from which it can be used as an index for various data searching/analysis operations. Such indexes can take the form of B-tree indexes. An example of such usage for the appliance 200 of FIG. 2(b) is described in the above-referenced and incorporated U.S. Pat. No. 7,660,793.



FIG. 2(c) depicts another exemplary embodiment for appliance 200. In the example of FIG. 2(c), appliance 200 also includes a data store 306 of unstructured data that is in communication with bus 406 via disk controller 416. Thus, the data that is streamed through the coprocessor 450 may also emanate from data store 306. Furthermore, the unstructured data that is streamed through coprocessor 450 to generate metadata therefor can optionally be stored within data store 306. As described in the above-referenced and incorporated U.S. Pat. No. 7,660,793, the metadata tables within RDBMS 304 can be used to render data search/analysis operations on the unstructured data stored within data store 306 more efficient.



FIG. 3(a) depicts a printed circuit board or card 700 that can be connected to the PCI-X or PCI-e bus 406 of a commodity computer system for use as a coprocessor 450 in appliance 200 for any of the embodiments of FIGS. 2(a)-(c). In the example of FIG. 3(a), the printed circuit board includes an FPGA 402 (such as a Xilinx Virtex II FPGA) that is in communication with a memory device 702 and a PCI-X bus connector 704. A preferred memory device 702 comprises SRAM and DRAM memory. A preferred PCI-X or PCI-e bus connector 704 is a standard card edge connector.



FIG. 3(b) depicts an alternate configuration for a printed circuit board/card 700. In the example of FIG. 3(b), a bus 706 (such as a PCI-X or PCI-e bus), one or more disk controllers 708, and a disk connector 710 are also installed on the printed circuit board 700. Any commodity disk interface technology can be supported, as is understood in the art. In this configuration, the firmware socket 404 also serves as a PCI-X to PCI-X bridge to provide the processor 412 with normal access to any disk(s) connected via the private PCI-X bus 706. It should be noted that a network interface can be used in addition to or in place of the disk controller and disk connector shown in FIG. 3(b).


It is worth noting that in either the configuration of FIG. 3(a) or 3(b), the firmware socket 404 can make memory 702 accessible to the bus 406, which thereby makes memory 702 available for use by an OS kernel as the buffers for transfers to the FAMs from a data source with access to bus. It is also worth noting that while a single FPGA 402 is shown on the printed circuit boards of FIGS. 3(a) and (b), it should be understood that multiple FPGAs can be supported by either including more than one FPGA on the printed circuit board 700 or by installing more than one printed circuit board 700 in the appliance 200. FIG. 4 depicts an example where numerous FAMs in a single pipeline are deployed across multiple FPGAs.



FIG. 5(a) depicts at a high level a coprocessor 450 that receives an incoming data stream, generates metadata for the data stream, and outputs the data stream together with the generated metadata. Examples of different hardware-accelerated metadata generation operations will be described in greater detail hereinafter. Any of a number of text mining and text analytic operations can be employed to generate metadata. Some examples of metadata generation operations include but are not limited to part-of-speech tagging, information and entity extraction, document classification, document clustering, and text summarization. In these examples, the data stream under consideration will be a stream of documents. As noted, this usage in only exemplary as data streams other than documents can be readily processed using coprocessor 450 and the techniques described herein. These data streams may comprise unstructured and/or structured data.



FIG. 5(b) depicts an exemplary index 100 that can built from the metadata generated by coprocessor 450. As described in connection with FIG. 1, each table entry 102 comprises a term 104 and its associated pointer 106. In the example of FIG. 5(b), each pointer 106 comprises a document identifier Di and one or more position identifiers pi. Each document identifier identifies a document in which the term 104 corresponding to that pointer 106 is present. Each position identifier in the pointer identifies a position within the identified document where that term 104 is located. Preferably, the position identifier is represented by a byte offset relative to the start of a document. Thus, the pointers for the term “Azkaban” in FIG. 5(b) establish that “Azkaban” can be found in Document D12 at position p1 and in Document D9919 at positions p1 and p2. Any of a number of well-known techniques can be used to map pointers such as the ones shown in FIG. 5(b) to the documents stored within a file system of an enterprise.



FIG. 6 depicts exemplary firmware 350 that can be employed in reconfigurable logic 402 to perform a metadata generation operation 604 on an incoming stream of documents 600. An example of a metadata generation operation 604 includes a stream pre-processing operation 602. The stream pre-processing preferably comprises various position and feature vector extraction operations. These operations may employ techniques known in the art as cryptographic hashing to provide the term descriptor and tables to enable the values and lists described herein to be generated and exploited.



FIG. 7(a) depicts an exemplary sequence of processing modules for the stream pre-processing stage 602. A word parsing module 750 preferably operates to identify white-space and/or punctuation delimited terms in the documents and record their positions. These recorded positions can then be used to help determine word positions within documents (e.g., the pi values in the pointers of FIG. 5(b)). A stop list filter module 752 preferably operates to remove words from the document stream that appear on a stop list. This may be desirable to remove common words (e.g. “the”, “a”, “an”) for which indexing may not be particularly helpful from the document stream. The stop list is preferably adjustable to control which words will be removed from the document stream. A stemming module 754 preferably operates to stem words in the document stream to their roots. Examples of stemming operations include removing plurals, removing prefixes, removing suffixes, removing gerunds, etc. It should also be noted that other processing modules can be included in stream pre-processing module 602 if desired. For example, a spell-checking module can be used to correct possible spelling errors that exist within a word stream.



FIG. 7(b) depicts an exemplary output data stream from the stream pre-processing operation 602, wherein the stream includes the parsed words 760 such that each parsed word is accompanied by a position identifier 758 for that word. Furthermore, a document identifier 756 is present that identifies when a new document is present within the stream. While the example of FIG. 7(b) shows a serial stream of such data, it should be noted that the document identifiers 756 and position identifiers 758 can flow in bit streams within firmware 350 that are parallel to a bit stream of words 760.


Returning to FIG. 7(a), as words stream through the pre-processing module 602, a memory containing an index for the documents can be generated. This index 100 can serve as a general index that is populated with entries for each different word encountered by the stream pre-processing module 602 together with its associated position vector information (e.g., document identifier and position identifier as shown in FIG. 5(b)). If a word that has already been added to general index 100 is encountered again within the stream, then the pointer information in index 100 for that word can be updated to reflect the latest position at which the word has been found.


It should be noted that the memory used to store the index 100 as words stream through the pre-processing module 602 can be located in any of a number of places. For example, this memory can be located in the available on-chip memory of the FPGA 402. This memory could also be memory device 702 shown in FIGS. 3(a) and (b). Further still, this memory could be RAM 408. This memory could even be a location such as RDBMS 304. Preferably, coprocessor 450 performs the action of updating index 100. Techniques such as hashing can be used to insert new data into the index 100.


It should be noted that each of the modules shown in FIG. 7(a) as well as the stream pre-processing operation 602 itself is an optional feature. If desired, the metadata generation operation 604 can be performed on documents 600 that have not been pre-processed.



FIG. 8(a) depicts an exemplary embodiment for the metadata generation operation 604 wherein a specialized index generation operation 850 is also performed. In FIG. 8(a), words and position vectors stream into an exact matching module 800. This exact matching module 800 is loaded with the words 804 from a dictionary 802. The exact matching modules will use words 804 as keys against which the streaming words are compared to determine whether any exact matches exist therebetween. As matches are found by module 800, a specialized index 860 maintained in the memory is updated with the matching word information and the corresponding pointers for those matching words (see FIG. 5(b)). The position vectors found by the stream pre-processing operation 602 (e.g., document identifiers 756 and position identifiers 758) can be used as the pointer information. Thus, in addition to the general index 100 described in connection with FIG. 7(a), a specialized index 860 (that is specific as to dictionary 802) can also be efficiently generated via the data processing operations of coprocessor 450.


Any of a number of exact matching techniques can be used to perform the exact matching operation. For example, the hardware-accelerated matching techniques disclosed in the above-referenced and incorporated U.S. Pat. Nos. 6,711,558 and 7,139,743 and U.S. Patent Application Publications 2006/0294059, 2007/0130140, and 2007/0260602.


Any collection of words can be used as dictionary 802. Examples of dictionaries 802 that can be used include large word collections such as full English language dictionaries, full foreign language dictionaries, scientific/technical dictionaries, etc. Smaller word collections can also be used.



FIG. 8(b) depicts an alternate embodiment for index generation wherein an approximate matching module 810 is used rather than an exact matching module. It may be desirable to employ an approximate matching module 810 for the index generation operation 850 to reduce the likelihood that spelling errors and the like do not result in document text being mis-indexed. Examples of hardware-accelerated approximate matching techniques suitable for use as approximate matching module 810 are also disclosed in the above-referenced and incorporated U.S. Pat. Nos. 6,711,558 and 7,139,743 and U.S. Patent Application Publications 2006/0294059, 2007/0130140, and 2007/0260602. It should be noted that if a spell-checking module as described above is used in the stream pre-processing stage, it may be unnecessary to employ an approximate matching module to accommodate mis-spellings and the like.


It should also be noted that a plurality of exact matching/approximate matching modules 800/810 can be deployed in parallel to distribute the workload of matching streaming words to the word keys 804 across multiple matching modules 800/810. However, if desired, a single serial pipelined sequence of compare registers can be used by an matching module 800/810 to sequentially compare the streaming words with word keys 804 if desired.


It should be noted that, like the memory used to store general index 100, the memory used to store the specialized indexes 860 can be located in any of a number of places (e.g., the on-chip memory of the FPGA 402, memory device 702, etc.).


A powerful aspect of various embodiments of the present invention is the ability to concurrently generate multiple specialized indexes. An example of this is shown in FIG. 9(a). With this embodiment, the matching module 800/810 matches the incoming words against multiple dictionaries 802i, each dictionary 802i having its own collection of words which will serve as the keys against which the incoming words are judged. As shown, the words within dictionaries such as an English dictionary, a French dictionary, a medical dictionary, and a technical dictionary can be compared against the incoming word stream. As matches are found between the incoming words and the words within any of these dictionaries, a specialized index associated with each of these dictionaries can be updated with pointer information. FIG. 9(a) depicts an English dictionary index 902 and a technical dictionary index 904 as examples of such specialized indexes. Preferably, these specialized indexes are pre-populated in the term column with all of the words within its associated dictionary. As that term is found within the word stream by the matching module 800/810, the pointer information for that term can be updated as appropriate.


The general index 100 in such an embodiment can be updated when words streaming through the exact matching module do not find a match in any of the dictionaries 802. Furthermore, the entries in the other specialized indexes can be merged into the general index if desired to thereby make general index 100 applicable to all of the words within document stream 600. This merger can be performed offline or concurrently with the updating of the other indexes. Otherwise, the coprocessor 450 can be configured to update the general index 100 as words stream out of the stream pre-processing module 602 and update the specialized indexes as words stream out of the exact/approximate matching module 800/810.


Another powerful aspect of various embodiments of the present invention is the ability to perform regular expression pattern matching on incoming words. An example of this is shown in FIG. 9(b). Regular expressions, as is well-known in the art, can be used to detect patterns of varying degrees of complexity within a word stream. An exemplary pattern for which regular expressions are useful means of detection is a social security number (SSN). SSNs exhibit the pattern xxx-xx-xxxx, wherein x can be any number between 0 and 9. Items such as phone numbers, credit card numbers, primary account numbers (PANs), e-mails, URLs, and others also exhibit patterns that are detectable through regular expressions.


As shown in FIG. 9(b), a regular expression matching module 910 can be deployed on coprocessor 450 (preferably in firmware 350 on a reconfigurable logic device 402). The regular expression matching module 910 can be configured to detect one or more user-defined patterns. For example, the regular expression matching module 910 in FIG. 9(b) is configured to detect words (including word portions) that exhibit an e-mail pattern 912, a phone number pattern 914, a credit card number pattern 916, and an SSN pattern 918. As these patterns are found within the word stream by the regular expression matching module 910, a specialized index associated with each of these patterns can be updated with the content of the matching pattern and with the pointer information for that matching pattern. FIG. 9(b) depicts an e-mail index 920 and a SSN index 922 as examples of such indexes. Thus, if 123-45-6789 is encountered in the word stream, then the regular expression matching module 910 will identify this pattern as a SSN pattern, and the regular expression matching module 910 will add an entry to the SSN index 922 comprising the term “123-45-6789” and a pointer to that term in the word stream. Preferably, as described in connection with FIG. 9(a), a general index 100 is also updated as words stream through the regular expression matching module 910, and the regular expression matching module 910 does not find any pattern matches.


An example of a hardware-accelerated regular expression matching technique suitable for use as regular expression matching module 910 is disclosed in the above-referenced and incorporated U.S. Patent Application Publication 2007/0130140.


Also, as noted above in connection with the matching modules 800/810, various combinations of sequential and parallel regular expression matching modules 910 can be employed within coprocessor 450 to distribute the pattern matching workload across multiple resources.


The power of such pattern indexes shown by FIG. 9(b) cannot be understated. It is typically extremely difficult for a person within an enterprise to efficiently get an answer to a question such as “Which documents within our enterprise include a SSN?” or a command such as “Show me all documents within our enterprise that include a credit card number”. Similarly, it is difficult for a web search engine to provide relevant results to an Internet search query seeking to find web pages which contain credit card numbers or SSNs without being specific as to individual numbers. However, through the indexing techniques disclosed in connection with FIG. 9(b), the answers to questions such as these are readily available.



FIG. 9(c) illustrates an exemplary embodiment wherein the hardware-accelerated index generation operation 850 includes one or more exact/approximate matching modules 800/810 and one or more regular expression matching modules 910 to generate corresponding specialized indexes (e.g., indexes 902, 904, 920, and 922) and a general index 100. Once again the power of such efficiently-generated indexes cannot be understated. These indexes can serve to answer complicated query commands such as “Show me all documents which include a credit card number and contain the word ‘stolen’” without re-analyzing the entire set of documents 600.


It should be noted that coprocessor 450 is preferably configured to change the dictionaries and/or patterns used by the matching modules 800/810/910. Preferably, such changes are made in response to user commands received as command data by the firmware socket module 404. These changes may comprise adding/removing dictionaries/patterns and/or adding/removing words to/from the various dictionaries.


It should also be noted that the coprocessor 450 can deploy these modules 800/810 and 910 in a sequential, parallel, or hybrid sequential-parallel arrangement as desired.



FIG. 10 depicts an embodiment wherein periodic batch transfers of the generated indexes (e.g., indexes 100, 902, 904, 920, and 922) are performed to merge them into their corresponding operational counterparts within an RDBMS (e.g., RDBMS 304 or some other RDBMS within enterprise network 420). Thus, batch insert/merge operations can be performed to insert the content of general index 100 into operational general index 1000. Similarly, the content of indexes 902, 904, 920, and 922 can be inserted into operational indexes 1002, 1004, 1020, and 1022 respectively. These transfers can be performed periodically at scheduled intervals or can be performed on an as needed basis as the available memory space for indexes 100, 902, 904, 920, and 922 runs low.


After the operational indexes have been updated, the newly indexed data can then be accessed via standardized queries such as SQL commands that are applied to the operational indexes stored by SQL databases. The above-referenced and incorporated U.S. Pat. No. 7,660,793 discloses a particularly efficient query handling technique for queries directed toward unstructured data for which an RDBMS maintains a metadata index.


Another metadata generation operation that can be performed by the stream pre-processing module 602 is a word counting operation 1150, as shown in FIG. 11(a). In order to ascertain the relevance of the documents 600 during a query handling operation, the indexing engine preferably obtains statistical information about the words in the documents. This statistical information may include a count of the number of times a particular word appears in a document and a count of the number of times a particular word appears in the entire document collection (should the documents 600 be grouped into one or more collections). Another parameter that can be determined and used for relevance assessments is the size (in words or bytes) of each document. By knowing these parameters, the relevance of a document retrieved during querying can be calculated using well-known probabilistic relevance models such as the Okapi BM25 model. See Karen Spärck Jones, Steve Walker, and Stephen E. Robertson. “A Probabilistic Model of Information Retrieval: Development and Comparative Experiments (parts 1 and 2).”, Information Processing and Management, 36 (6):779-840. 2000, the entire disclosure of which is incorporated herein by reference.


Through a word counting module 1150, word lists for documents and document collections are generated as the words stream therethrough. Associated with each word on the list will be a count value that is incremented each time the associated word is encountered in the stream. FIG. 11(b) depicts an exemplary index 1154 wherein the pointers 106 include a word count parameter we that identifies how many times each term appears in the referenced document. These pointers may also include parameters that identify the size of each document (although this information can be maintained in a separate index table). Table 1154 may also include an additional column 1160 that includes a “collection count” parameter that identifies how many times each term appears in a document collection. Appropriate flags can be added to the bits in the word and position vector stream to identify collection separations. Because the output from the word counting module preferably includes this count information, FIG. 11(a) depicts the output from the word counting module 1150 as including count vectors.


Another metadata generation operation 604 that can be performed by coprocessor 450 is a classification operation 1100. An exemplary pipeline for such actions is shown in FIG. 11(c). An example of a document classification operation that can be performed on documents includes language classification. With language classification, the document can be applied to a statistical n-gram algorithm that is configured to identify the language that the text within the document most closely matches. See William B. Cavnar and John M. Trenkle, “N-Gram-Based Text Categorization”, Proceedings of SDAIR-94, 3rd Annual Symposium on Document Analysis and Information Retrieval, Las Vegas, pages 161-175, 1994, the entire disclosure of which is incorporated herein by reference. Other document classification operations may employ hidden Markov models (HMMs) to learn some type of classification for the document. See Ludovic Denoyer, Hugo Zaragoza and Patrick Gallinari, “HMM-based Passage Models for Document Classification and Ranking”, Proceedings of ECIR-01, 23rd European Colloquium Information Retrieval Research, Darmstatd, DE, pages 126-135, 2001, the entire disclosure of which is incorporated herein by reference.



FIG. 12(a) depicts an exemplary classification operation wherein a list of words is maintained for a plurality of different classes. For example, a first class 1202 is associated with words 1204 and so on such that class n 1206 is associated with words 1208. These words can be used by one or more matching modules 800/810 as keys that are compared against incoming words. As the key words find a match to a word within a document, a scoring memory 1210 that tracks a count of matches for each class can be updated as appropriate. Thus, when the matching module 800/810 finds a match between a given word within Document X and a word 1204 within the first class 1202, then the “count” field for class 1 in the scoring memory 1210 associated with that document can be updated. These counts can serve as class scores for each document to thereby indicate how relevant a given document is to a particular class. Preferably, the scoring memory 1210 is refreshed upon encountering a new document in the word stream. Scoring memory 1210 is preferably maintained in the available on-chip memory of the FPGA 402 or in memory device 702 (see FIGS. 3(a) and (b)).


The different classes can pertain to classifications such as “news”, “sports”, “legal”, “medicine”, etc. The words within each class can be words that are generally expected to be found in a document that pertains to the subject defined by the class. For example, the word “baseball” may be present in the word list for a “sports” classification, and the word “estop” may be present in the word list for a “legal” classification. Another exemplary document classification can be language classification. In such an application, each class shown in FIG. 12(a) can be associated with a different language, wherein the words within each class would comprise a full (or redacted) dictionary for that language. With reference to the embodiment of FIG. 9(a), it can be understood that such language classification operations can function as an adjunct to specialized indexing, wherein the matching operations shown in FIG. 9(a) can also operate to update a scoring memory 1210 as appropriate. Yet another exemplary document classification can be “reading comprehension level”. Each class can be associated with a different comprehension level (e.g., “5th grade reading level”, “high school freshman reading level”, “undergraduate reading level”, etc.), and the words within each class can be words expected to be encountered in documents belonging to the corresponding comprehension level. It should be noted that the count vector information produced by the word counting module of FIG. 11(a) may be helpful in aiding classifications such as “comprehension level”.



FIG. 12(b) depicts an exemplary classification embodiment wherein decision logic 1212 is employed to assign one or more classes to a document based on the class counts for that document as recorded by scoring memory 1210. For example, each class can have an associated threshold value. If the class count meets or exceeds the threshold value for its class, then the document can be tagged as belonging to that class. As such, it may be possible for the same document to belong to multiple classes should its counts in those classes meet or exceed the classes' assigned threshold values.


Furthermore, it should be noted that one or more regular expression matching modules 910 can be advantageously employed in connection with classification operation 1100 (possibly in combination with one or more matching modules 800/810). For example, if a document is found to contain a credit card number, then an inference could possibly be drawn that the document should be classified as a “sales record”, “personal record”, or the like. Regular expression matching can also be useful to generate counts of how many words have particular character lengths. For example, the regular expression “xxxx” where x can be any character can be used to update a count of 4 letter words, while the regular expression “xxxxx” can be used to update a count of 5 letter words, and so on. Such word length counts can be useful for assessing classifications such as “comprehension level” discussed above.


The classification module 1100 may be used to enrich the word stream emanating from the module 1100 by adding bit strings for each document to the stream, wherein each bit string tags each document as belonging to a particular class. For example, if the classification module is configured to determine which of 20 classifications a document belongs to, a 20 bit string can be added to the output stream for each document, wherein each bit is associated with a particular classification and flags whether the document belongs to that classification. As such, bit X in this bit string can be set aside to flag whether the document contains a credit card number. Thus, if a regular expression matching module 910 detects the presence of a credit card number pattern within a document, the bit string for that document can be updated to set bit X to high. Similarly, bit Y in the bit string can be set aside to flag whether the document is a legal document. If the classification operation results in a determination that the document should be considered a legal document, then bit Y can be set to high.


It should also be noted that classification operation 1100 can be considered as also belonging to the specialized index generation category of metadata generation operations because the classification information produced by classification operation 1100 can also be used to generate specialized indexes of documents by class. Such indexes can serve as powerful searching aids in that they can be used to answer questions such as “How many Spanish language documents are within the enterprise?” and conduct research such as “What relationships exist between sports and medicine?” and “Find all legal documents in which John Smith is mentioned”.


Also, it should be noted that the classification operation can be performed in software (e.g., software executed by processor 412) using the general and specialized indexes generated from embodiments such as those shown in FIGS. 9(a)-(c). These different indexes can be cross-correlated with pre-defined word lists for the different classes to determine the classification information for each document.


With reference to FIG. 13(a), and following from FIG. 11(a), another metadata generation operation 604 that can be performed by coprocessor 450 is a histogram generation operation 1300 to thereby aid document clustering. With document clustering, it is desired to know the relevance of different documents to each other. To aid such clustering determinations, the coprocessor 450 can be configured to perform the histogram generation operation 1300 on documents 600 (wherein this operation can be similar if not identical to the word counting operation described above). These histograms can identify the frequency of words within a document. From the generated histograms, software can be used to perform feature vector expansion and identify document clusters (operation 1302) that are likely related to each other with respect to common subject matter, etc. Feature vector expansion is a standard technique wherein the feature vectors of individual documents are expanded to include words from all the documents in the collection. Clustering can be performed using any of the well-known techniques such as K-means clustering, fuzzy C-means clustering, hierarchical clustering, etc. See M. Steinbach, G. Karypis, and V. Kumar. “A comparison of document clustering techniques.”, KDD Workshop on Text Mining, 2000, the entire disclosure of which is incorporated herein by reference.


Should the stream pre-processing module 602 already employ word counting, then the histogram generation stage 1300 can be omitted, as shown in FIG. 13(b).


It should be noted that the clustering operations shown by FIGS. 13(a) and (b) can be considered as also belonging to the specialized index generation category of metadata generation operations because the clustering information can also be used to generate specialized indexes of documents by clusters. Such indexes can serve as powerful searching aids in that they can be used to answer questions such as “Which documents within an enterprise are similar to Document X?” and “Which books available for sale from this website are similar to Book X?”.



FIGS. 14(a) and (b) depict exemplary pipelines wherein the hardware-accelerated metadata generation operation 604 includes classification 1100, specialized index generation 850, and histogram generation 1300 (or word counting 1150).



FIG. 15 depicts an exemplary environment in which appliance 200 can be effectively employed. FIG. 15 depicts an enterprise computing system 1504 that receives a large volume of incoming data from the Internet 1500. After this incoming data passes enterprise firewall 1502, it can be streamed through appliance 200 which taps into the firewall output. Appliance 200 can thus be used to generate metadata and indexes for the incoming data as that data reaches the enterprise and before it lands in data storage somewhere within enterprise network 420. The data processed by appliance 200 can also include data originating from within the enterprise computing system 1504. Furthermore, appliance 200 can optionally be configured to output its generated metadata for delivery to other repositories within enterprise network 420. Further still, queries can optionally be applied to appliance 200 to conduct full-text searching or other data analysis operations on data indexed by appliance 200. As noted, an example of such a use for appliance 200 is disclosed in the above-referenced and incorporated U.S. Pat. No. 7,660,793.



FIG. 16 depicts an overview of document ingest pre-processing in accordance with an embodiment where appliance 200 is used to index documents within an enterprise. Preferably, through some form of document ingest GUI 1600 displayed on a user computer within enterprise network 420, the user is able to specify which document(s) should be ingested into data store 306 of FIG. 2(c). Optionally, the user can also key in various forms of metadata about the document(s) to be ingested; however this need not be the case as the coprocessor 450 will be arranged to perform the desired metadata generation operations automatically, as described above. In response to an appropriate command 1612 delivered to appliance 200 from GUI 1600, one or more documents 600 stored in a data store 308 accessible through the enterprise network 420, but external to the appliance 200, is delivered to the appliance 200. Various adapters can be employed in document retrieval functionality 1652 employed by the appliance 200 to provide access to documents stored on common file systems such as NTFS, FAT, CIFS, various flavors of Unix file systems, as well as Web access over HTTP.


Should the coprocessor employ a reconfigurable logic device 402 with firmware 350 deployed thereon, FAMs resident in firmware pipeline 350 will be arranged to perform a document metadata generation operation on the documents it receives. Examples of these metadata generation operations are described above.


The document metadata 1650 that is produced by the operation of firmware 350 can then be stored in RDBMS 304, wherein the RDBMS engine operates to generate and maintain an index of this document metadata that can later be queried using standardized database queries to identify which documents within data store 306 should be processed through the coprocessor 450 at hardware speeds during query processing. After the received document 600 has been processed by firmware 350, document 600 can then be ingested into the appliance by storing it in the data store 306 of unstructured data. The acts of metadata generation and document ingestion preferably operate virtually concurrently in near-real time. It should be noted that the document metadata 1650 can optionally be stored in a structured database external to the appliance 200.



FIG. 17 depicts this document ingest pre-processing 1700 as a logical flow. At step 1, the user interacts with GUI 1600 to identify a new document 600 for ingestion into appliance 200. GUI 1600 may optionally be configured to allow the user to specify what metadata is to be generated from document 600. Next, at step 2, the document 600 is retrieved from its original source location (either an enterprise document store 308, the Internet or some other network that is accessible to the enterprise network 420). Then, firmware 350 performs its document metadata generation operation 604 on document 600 to produce document metadata 1650. At step 3, the document 600 is then stored in the file system of data store 306, and the document metadata (including its location in the file system of data store 306) is saved into relational tables of the RDBMS 304. FIG. 18 illustrates this data flow superimposed over the appliance 200 of FIG. 2(c).


In this manner, appliance 200 can now use the document metadata 1650 indexed by RDBMS 304 to aid decisions as to which documents should be subjected to a query-specified data processing operation (e.g., a full-text searching operation) via coprocessor 450. Furthermore, because standardized RDBMS technology has been leveraged within appliance 200, standardized database queries that are well-known to a number of users can be used to decide which documents are to be subjected to the coprocessor-based data processing operation when processing a given query. This type of query processing is disclosed in the above-referenced and incorporated U.S. Pat. No. 7,660,793.


While in the preferred embodiment disclosed herein the coprocessor 450 comprises a reconfigurable logic device 402 such as an FPGA, it should be noted that the coprocessor 450 can be realized using other processing devices. For example, the coprocessor 450 may comprise graphics processor units (GPUs), general purpose graphics processors, chip multi-processors (CMPs), dedicated memory devices, complex programmable logic devices, application specific integrated circuits (ASICs), and other I/O processing components. Moreover, it should be noted that appliance 200 may employ a plurality of coprocessors 450 in either or both of a sequential and a parallel multi-coprocessor architecture.


While the present invention has been described above in relation to its preferred embodiments, various modifications may be made thereto that still fall within the invention's scope. Such modifications to the invention will be recognizable upon review of the teachings herein. Accordingly, the full scope of the present invention is to be defined solely by the appended claims and their legal equivalents.

Claims
  • 1. An apparatus comprising: a field programmable gate array (FPGA); and a bus through which streaming data is delivered to the FPGA at a bus bandwidth rate;the FPGA having firmware logic deployed thereon, the firmware logic configured to(1) receive the streaming data,(2) perform regular expression pattern matching on the streaming data with respect to a plurality of patterns to detect whether any pattern matches exist within the streaming data,and (3) concurrently build, at the bus bandwidth rate, a plurality of pattern indexes for the streaming data based on detected pattern matches, wherein each pattern index corresponds to a pattern against which regular expression pattern matching was performed and comprises location data for detected pattern matches.
  • 2. The apparatus of claim 1 wherein the firmware logic is further configured to perform the regular expression pattern matching on the streaming data at hardware speeds.
  • 3. The apparatus of claim 1 wherein the bus is a PCI-Express bus.
  • 4. The apparatus of claim 1 further comprising a network interface, the network interface configured to receive the streaming data from a network and deliver the streaming data to the FPGA through the bus at the bus bandwidth rate.
  • 5. The apparatus of claim 4 further comprising a processor, the processor configured to manage a flow of the streaming data to the FPGA via the network interface and the bus.
  • 6. The apparatus of claim 1 further comprising a disk controller, the disk configured to receive the streaming data from a disk and deliver the streaming data to the FPGA through the bus at the bus bandwidth rate.
  • 7. The apparatus of claim 6 further comprising a processor, the processor configured to manage a flow of the streaming data to the FPGA via the network interface and the bus.
  • 8. The apparatus of claim 1 wherein the firmware logic comprises a plurality of regular expression pattern matching modules in parallel, each regular expression pattern matching module being keyed with a pattern; each regular expression pattern matching module being configured to (1) receive the streaming data, (2) perform a regular expression pattern matching operation on the streaming data to detect any pattern matches that exist within the streaming data with respect to its keyed pattern, and (3) operate simultaneously in parallel with the other regular expression pattern matching modules.
  • 9. The apparatus of claim 8 wherein each regular expression pattern matching module is keyed with a different pattern.
  • 10. The apparatus of claim 8 wherein the streaming data comprises a plurality of unstructured data objects; wherein the pattern indexes comprise a plurality of terms and a plurality of pointers associated with the terms, the terms corresponding to portions of the data stream for which a pattern match was detected, wherein each pointer identifies where at least one data object related to the term associated with that pointer can be located in a computer system;wherein the firmware logic is further configured to (1) pre-process the streaming data by (i) parsing the data objects into a plurality of words, (ii) generating a data object identifier that corresponds to each data object, wherein each data object identifier is indicative of its corresponding data object's position within the streaming data, and (iii) generating a position identifier that corresponds to each parsed word, wherein each position identifier is indicative of its corresponding word's position within the streaming data, and (2) build the pattern indexes by generating the pointers for the pattern indexes based at least in part on the generated data object identifiers and the generated position identifiers for the detected pattern matches.
  • 11. The apparatus of claim 10 wherein the firmware logic is further configured to build a general index from the pre-processed streaming data using the generated data object identifiers and the generated position identifiers such that (1) the general index's terms comprise the words within the pre-processed streaming data and (2) the terms' associated pointers comprise the data object identifiers and position identifiers corresponding to those words.
  • 12. The apparatus of claim 8 wherein the firmware logic further comprises an exact matching module in parallel with the parallel regular expression pattern matching modules, the exact matching module being keyed with a dictionary, the dictionary comprising a plurality of words; wherein the exact matching module is configured to perform an exact matching operation on the streaming data to thereby detect any exact matches that exist within the streaming data with respect to any of the dictionary words; andwherein the firmware logic is further configured to build a dictionary index for the streaming data based on the detected exact matches.
  • 13. The apparatus of claim 12 wherein the exact matching module and the parallel regular expression pattern matching modules are configured to operate on the streaming data at hardware speeds.
  • 14. The apparatus of claim 8 wherein the firmware logic further comprises an approximate matching module in parallel with the parallel regular expression pattern matching modules, the approximate matching module being keyed with a dictionary, the dictionary comprising a plurality of words; wherein the approximate matching module is configured to perform an approximate matching operation on the streaming data to thereby detect any approximate matches that exist within the streaming data with respect to any of the dictionary words; andwherein the firmware logic is further configured to build a dictionary index for the streaming data based on the detected approximate matches.
  • 15. The apparatus of claim 14 wherein the approximate matching module and the parallel regular expression pattern matching modules are configured to operate on the streaming data at hardware speeds.
  • 16. The apparatus of claim 8 wherein the streaming data comprises a plurality of data objects, and wherein the firmware logic is further configured to, in response to a detection from by a regular expression pattern matching operation that data within a data object is a pattern match with respect to a pattern, tag that data object as belonging to a class.
  • 17. The apparatus of claim 16 wherein the firmware logic is further configured to build a class index based on the tagged data objects, the class index being associated with the class and configured to identify the data objects that belong to the associated class.
  • 18. The apparatus of claim 1 wherein at least one of the patterns comprises a credit card number pattern.
  • 19. The apparatus of claim 1 wherein at least one of the patterns comprises a social security number pattern.
  • 20. The apparatus of claim 1 wherein at least one of the patterns comprises an email address pattern.
  • 21. The apparatus of claim 1 wherein at least one of the patterns comprises a telephone number pattern.
  • 22. The apparatus of claim 1 wherein at least one of the patterns comprises an Internet uniform resource locator (URL) pattern.
  • 23. The apparatus of claim 1 wherein the streaming data comprises a plurality of unstructured data objects; wherein the pattern indexes comprise a plurality of terms and a plurality of pointers associated with the terms, the terms corresponding to portions of the streaming data for which a pattern match was detected, wherein each pointer identifies where at least one data object related to the term associated with that pointer can be located in a computer system;wherein the firmware logic is further configured to (1) pre-process the streaming data by (i) parsing the data objects into a plurality of words, (ii) generating a data object identifier that corresponds to each data object, wherein each data object identifier is indicative of its corresponding data object's position within the streaming data, and (iii) generating a position identifier that corresponds to each parsed word, wherein each position identifier is indicative of its corresponding word's position within the streaming data, and (2) build the pattern indexes by generating the pointers for the pattern indexes based at least in part on the generated data object identifiers and the generated position identifiers for the detected pattern matches.
  • 24. A method comprising: delivering streaming data to a field programmable gate array (FPGA) via a bus at a bus bandwidth rate;the FPGA receiving the streaming data, the FPGA having firmware logic deployed thereon;the firmware logic performing regular expression pattern matching on the streaming data with respect to a plurality of patterns to detect whether any pattern matches exist within the streaming data; andthe firmware logic concurrently building, at the bus bandwidth rate as the streaming data continues to be delivered to the FPGA, a plurality of pattern indexes for the streaming data based on detected pattern matches, wherein each pattern index corresponds to a pattern against which regular expression pattern matching was performed and comprises location data for detected pattern matches.
  • 25. The method of claim 24 wherein the performing step comprises the firmware logic performing the regular expression pattern matching on the streaming data at hardware speeds as the streaming data continues to be received by the FPGA.
  • 26. The method of claim 24 wherein the bus is a PCI-Express bus.
  • 27. The method of claim 26 wherein the delivering step comprises delivering the streaming data to the FPGA via the PCI-Express bus from a network source.
  • 28. The method of claim 26 wherein the delivering step comprises delivering the streaming data to the FPGA via the PCI-Express bus from a disk source.
  • 29. The method of claim 24 wherein at least one of the patterns comprises a credit card number pattern.
  • 30. The method of claim 24 wherein at least one of the patterns comprises a social security number pattern.
  • 31. The method of claim 24 wherein at least one of the patterns comprises an email address pattern.
  • 32. The method of claim 24 wherein at least one of the patterns comprises a telephone number pattern.
  • 33. The method of claim 24 wherein at least one of the patterns comprises an Internet uniform resource locator (URL) pattern.
  • 34. The method of claim 24 wherein the streaming data comprises a plurality of web pages, and wherein each pattern index comprises a plurality of pointers to the web pages which contain the pattern corresponding to that pattern index.
  • 35. The method of claim 34 further comprising: an Internet search engine receiving a pattern query; andthe Internet search engine looking up the web pages that are responsive to the pattern query based on at least one of the pattern indexes; andthe Internet search engine providing search results for the pattern query in response to the looking up step.
  • 36. The method of claim 35 further comprising merging the pattern indexes into a plurality of operational pattern indexes, and wherein the looking up step is performed on the operational indexes.
  • 37. The method of claim 35 wherein the performing step comprises the firmware logic performing the regular expression pattern matching on the streaming data at hardware speeds as the streaming data continues to be received by the FPGA.
  • 38. The method of claim 24 wherein the firmware logic comprises a plurality of regular expression pattern matching modules in parallel, each regular expression pattern matching module being keyed with a pattern, and wherein the performing step comprises: each regular expression pattern matching module (1) receiving the streaming data, (2) performing a regular expression pattern matching operation on the streaming data to detect any pattern matches that exist within the streaming data with respect to its keyed pattern, and (3) operating simultaneously in parallel with the other regular expression pattern matching modules.
  • 39. The method of claim 38 wherein each regular expression pattern matching module is keyed with a different pattern.
  • 40. The method of claim 38 wherein the streaming data comprises a plurality of unstructured data objects, wherein the pattern indexes comprise a plurality of terms and a plurality of pointers associated with the terms, the terms corresponding to portions of the data stream for which a pattern match was detected, wherein each pointer identifies where at least one data object related to the term associated with that pointer can be located in a computer system, the method further comprising: the firmware logic pre-processing the streaming data by (i) parsing the data objects into a plurality of words, (ii) generating a data object identifier that corresponds to each data object, wherein each data object identifier is indicative of its corresponding data object's position within the streaming data, and (iii) generating a position identifier that corresponds to each parsed word, wherein each position identifier is indicative of its corresponding word's position within the streaming data; andwherein the building step comprises the firmware logic building the pattern indexes by generating the pointers for the pattern indexes based at least in part on the generated data object identifiers and the generated position identifiers for the detected pattern matches.
  • 41. The method of claim 40 further comprising the firmware logic building a general index from the pre-processed streaming data using the generated data object identifiers and the generated position identifiers such that (1) the general index's terms comprise the words within the pre-processed streaming data and (2) the terms' associated pointers comprise the data object identifiers and position identifiers corresponding to those words.
  • 42. The method of claim 38 wherein the firmware logic further comprises an exact matching module in parallel with the parallel regular expression pattern matching modules, the exact matching module being keyed with a dictionary, the dictionary comprising a plurality of words, the method further comprising: the exact matching module performing an exact matching operation on the streaming data to thereby detect any exact matches that exist within the streaming data with respect to any of the dictionary words; andthe firmware logic building a dictionary index for the streaming data based on the detected exact matches.
  • 43. The method of claim 42 wherein the exact matching module and the parallel regular expression pattern matching modules are operating at hardware speeds on the streaming data as the streaming data continues to be received by the FPGA.
  • 44. The method of claim 38 wherein the firmware logic further comprises an approximate matching module in parallel with the parallel regular expression pattern matching modules, the approximate matching module being keyed with a dictionary, the dictionary comprising a plurality of words, the method further comprising: the approximate matching module performing an approximate matching operation on the streaming data to thereby detect any approximate matches that exist within the streaming data with respect to any of the dictionary words; andthe firmware logic is further building a dictionary index for the streaming data based on the detected approximate matches.
  • 45. The method of claim 44 wherein the approximate matching module and the parallel regular expression pattern matching modules are operating at hardware speeds on the streaming data as the streaming data continues to be received by the FPGA.
  • 46. The method of claim 38 wherein the streaming data comprises a plurality of data objects, the method further comprising: the firmware logic, in response to the regular expression pattern matching operation detecting that data within a data object is a pattern match with respect to a pattern, tagging that data object as belonging to a class.
  • 47. The method of claim 46 further comprising: the firmware logic building a class index based on the tagged data objects, the class index being associated with the class and configured to identify the data objects that belong to the associated class.
  • 48. The method of claim 24 wherein the streaming data comprises a plurality of unstructured data objects, wherein the pattern indexes comprise a plurality of terms and a plurality of pointers associated with the terms, the terms corresponding to portions of the streaming data for which a pattern match was detected, wherein each pointer identifies where at least one data object related to the term associated with that pointer can be located in a computer system, the method further comprising: the firmware logic pre-processing the streaming data by (i) parsing the data objects into a plurality of words, (ii) generating a data object identifier that corresponds to each data object, wherein each data object identifier is indicative of its corresponding data object's position within the streaming data, and (iii) generating a position identifier that corresponds to each parsed word, wherein each position identifier is indicative of its corresponding word's position within the streaming data; andwherein the building step comprises the firmware logic building the pattern indexes by generating the pointers for the pattern indexes based at least in part on the generated data object identifiers and the generated position identifiers for the detected pattern matches.
  • 49. A method comprising: delivering streaming data to a field programmable gate array (FPGA) via a bus at a bus bandwidth rate, wherein the streaming data comprises a plurality of web pages;the FPGA receiving the streaming data, the FPGA having firmware logic deployed thereon;the firmware logic performing regular expression pattern matching on the streaming data with respect to a plurality of patterns to detect whether any pattern matches exist within the streaming data;the firmware logic concurrently building, at the bus bandwidth rate as the streaming data continues to be delivered to the FPGA, a plurality of pattern indexes for the streaming data based on detected pattern matches, wherein each pattern index corresponds to a pattern against which regular expression pattern matching was performed and comprises a plurality of pointers to the web pages which contain the pattern corresponding to that pattern index;an Internet search engine receiving a pattern query;the Internet search engine looking UP the web pages that are responsive to the pattern query based on at least one of the pattern indexes; andthe Internet search engine providing search results for the pattern query in response to the looking UP step.
  • 50. The method of claim 49 wherein the bus is a PCI-Express bus.
  • 51. An apparatus comprising: a processor;a coprocessor; anda bus over which streaming data flows to the coprocessor at a bus bandwidth rate;wherein the processor is configured to manage a flow of the streaming data to the coprocessor;wherein the coprocessor comprises a plurality of regular expression pattern matching modules in parallel, each regular expression pattern matching module being keyed with a pattern;each regular expression pattern matching module being configured to(1) receive the streaming data,(2) perform a regular expression pattern matching operation on the streaming data to detect any pattern matches that exist within the streaming data with respect to its keyed pattern, and(3) operate simultaneously in parallel with the other regular expression pattern matching modules; andwherein the coprocessor is further configured to concurrently build, at the bus bandwidth rate, a plurality of pattern indexes for the streaming data based on detected pattern matches, wherein each pattern index corresponds to a pattern against which regular expression pattern matching was performed and comprises location data for detected pattern matches.
  • 52. The apparatus of claim 51 wherein the coprocessor comprises a reconfigurable logic device, the regular expression pattern matching modules being implemented by the reconfigurable logic device.
  • 53. The apparatus of claim 51 wherein the coprocessor comprises a graphics processor unit (GPU), the regular expression pattern matching modules being implemented by the GPU.
  • 54. The apparatus of claim 51 wherein the coprocessor comprises a chip multi-processor (CMP), the regular expression pattern matching modules being implemented by the CMP.
  • 55. The apparatus of claim 51 wherein the streaming data comprises a plurality of unstructured data objects; wherein the pattern indexes comprise a plurality of terms and a plurality of pointers associated with the terms, the terms corresponding to portions of the streaming data for which a pattern match was detected, wherein each pointer identifies where at least one data object related to the term associated with that pointer can be located in a computer system;wherein the coprocessor is further configured to (1) pre-process the streaming data by (i) parsing the data objects into a plurality of words, (ii) generating a data object identifier that corresponds to each data object, wherein each data object identifier is indicative of its corresponding data object's position within the streaming data, and (iii) generating a position identifier that corresponds to each parsed word, wherein each position identifier is indicative of its corresponding word's position within the streaming data, and (2) build the pattern indexes by generating the pointers for the pattern indexes based at least in part on the generated data object identifiers and the generated position identifiers for the detected pattern matches.
CROSS-REFERENCE AND PRIORITY CLAIM TO RELATED PATENT APPLICATIONS

This patent application is a continuation of U.S. patent application Ser. No. 11/938,732, filed Nov. 12, 2007, now U.S. Pat. No. 8,326,819, which claims priority to U.S. provisional patent application Ser. No. 60/865,629, filed Nov. 13, 2006, the entire disclosures of each of which are incorporated herein by reference.

US Referenced Citations (523)
Number Name Date Kind
3296597 Scantlin et al. Jan 1967 A
3573747 Adams et al. Apr 1971 A
3581072 Nymeyer May 1971 A
3601808 Vlack Aug 1971 A
3611314 Pritchard, Jr. et al. Oct 1971 A
3729712 Glassman Apr 1973 A
3824375 Gross et al. Jul 1974 A
3848235 Lewis et al. Nov 1974 A
3906455 Houston et al. Sep 1975 A
4044334 Bachman et al. Aug 1977 A
4081607 Vitols et al. Mar 1978 A
4298898 Cardot Nov 1981 A
4314356 Scarbrough Feb 1982 A
4385393 Chaure et al. May 1983 A
4412287 Braddock, III Oct 1983 A
4464718 Dixon et al. Aug 1984 A
4550436 Freeman et al. Oct 1985 A
4674044 Kalmus et al. Jun 1987 A
4823306 Barbic et al. Apr 1989 A
4903201 Wagner Feb 1990 A
4941178 Chuang Jul 1990 A
5023910 Thomson Jun 1991 A
5038284 Kramer Aug 1991 A
5050075 Herman et al. Sep 1991 A
5063507 Lindsey et al. Nov 1991 A
5077665 Silverman et al. Dec 1991 A
5101353 Lupien et al. Mar 1992 A
5101424 Clayton et al. Mar 1992 A
5140644 Kawaguchi et al. Aug 1992 A
5140692 Morita Aug 1992 A
5161103 Kosaka et al. Nov 1992 A
5163131 Row et al. Nov 1992 A
5179626 Thomson Jan 1993 A
5226165 Martin Jul 1993 A
5243655 Wang Sep 1993 A
5249292 Chiappa Sep 1993 A
5255136 Machado et al. Oct 1993 A
5258908 Hartheimer et al. Nov 1993 A
5265065 Turtle Nov 1993 A
5267148 Kosaka et al. Nov 1993 A
5270922 Higgins Dec 1993 A
5297032 Trojan et al. Mar 1994 A
5313560 Maruoka et al. May 1994 A
5315634 Tanaka et al. May 1994 A
5319776 Hile et al. Jun 1994 A
5327521 Savic et al. Jul 1994 A
5339411 Heaton, Jr. Aug 1994 A
5347634 Herrell et al. Sep 1994 A
5371794 Diffie et al. Dec 1994 A
5375055 Togher et al. Dec 1994 A
5388259 Fleischman et al. Feb 1995 A
5396253 Chia Mar 1995 A
5404411 Banton et al. Apr 1995 A
5404488 Kerrigan et al. Apr 1995 A
5418951 Damashek May 1995 A
5421028 Swanson May 1995 A
5432822 Kaewell, Jr. Jul 1995 A
5440723 Arnold et al. Aug 1995 A
5461712 Chelstowski et al. Oct 1995 A
5463701 Kantner, Jr. et al. Oct 1995 A
5465353 Hull et al. Nov 1995 A
5481735 Mortensen et al. Jan 1996 A
5488725 Turtle et al. Jan 1996 A
5497317 Hawkins et al. Mar 1996 A
5497488 Akizawa et al. Mar 1996 A
5500793 Deming, Jr. et al. Mar 1996 A
5517642 Bezek et al. May 1996 A
5535384 Kasahara Jul 1996 A
5544352 Egger Aug 1996 A
5546578 Takada et al. Aug 1996 A
5651125 Witt et al. Jul 1997 A
5684980 Casselman Nov 1997 A
5687297 Coonan et al. Nov 1997 A
5701464 Aucsmith Dec 1997 A
5704060 Del Monte Dec 1997 A
5721898 Beardsley et al. Feb 1998 A
5740244 Indeck et al. Apr 1998 A
5740466 Geldman et al. Apr 1998 A
5774835 Ozawa et al. Jun 1998 A
5774839 Shlomot Jun 1998 A
5781772 Wilkinson, III et al. Jul 1998 A
5781921 Nichols Jul 1998 A
5802290 Casselman Sep 1998 A
5805832 Brown et al. Sep 1998 A
5809483 Broka et al. Sep 1998 A
5813000 Furlani Sep 1998 A
5819273 Vora et al. Oct 1998 A
5819290 Fujita et al. Oct 1998 A
5826075 Bealkowski et al. Oct 1998 A
5832212 Cragun et al. Nov 1998 A
5845266 Lupien et al. Dec 1998 A
5857176 Ginsberg Jan 1999 A
5864738 Kessler et al. Jan 1999 A
5870730 Furuya et al. Feb 1999 A
5873071 Ferstenberg et al. Feb 1999 A
5886701 Chauvin et al. Mar 1999 A
5905974 Fraser et al. May 1999 A
5913211 Nitta Jun 1999 A
5930753 Potamianos et al. Jul 1999 A
5943421 Grabon Aug 1999 A
5943429 Händel Aug 1999 A
5950196 Pyreddy et al. Sep 1999 A
5963923 Garber Oct 1999 A
5978801 Yuasa Nov 1999 A
5987432 Zusman et al. Nov 1999 A
5991881 Conklin et al. Nov 1999 A
5995963 Nanba et al. Nov 1999 A
6016483 Rickard et al. Jan 2000 A
6023755 Casselman Feb 2000 A
6023760 Karttunen Feb 2000 A
6028939 Yin Feb 2000 A
6044375 Shmueli et al. Mar 2000 A
6044407 Jones et al. Mar 2000 A
6058391 Gardner May 2000 A
6064739 Davis May 2000 A
6067569 Khaki et al. May 2000 A
6070172 Lowe May 2000 A
6073160 Grantham et al. Jun 2000 A
6105067 Batra Aug 2000 A
6134551 Aucsmith Oct 2000 A
6138176 McDonald et al. Oct 2000 A
RE36946 Diffie et al. Nov 2000 E
6147890 Kawana et al. Nov 2000 A
6147976 Shand et al. Nov 2000 A
6169969 Cohen Jan 2001 B1
6173270 Cristofich et al. Jan 2001 B1
6175874 Imai et al. Jan 2001 B1
6178494 Casselman Jan 2001 B1
6185531 Schwartz et al. Feb 2001 B1
6226676 Crump et al. May 2001 B1
6236980 Reese May 2001 B1
6272616 Fernando et al. Aug 2001 B1
6279113 Vaidya Aug 2001 B1
6289440 Casselman Sep 2001 B1
6304858 Mosler et al. Oct 2001 B1
6317728 Kane Nov 2001 B1
6317795 Malkin et al. Nov 2001 B1
6336150 Ellis et al. Jan 2002 B1
6339819 Huppenthal et al. Jan 2002 B1
6370645 Lee et al. Apr 2002 B1
6377942 Hinsley et al. Apr 2002 B1
6381242 Maher, III et al. Apr 2002 B1
6389532 Gupta et al. May 2002 B1
6397259 Lincke et al. May 2002 B1
6397335 Franczek et al. May 2002 B1
6412000 Riddle et al. Jun 2002 B1
6415269 Dinwoodie Jul 2002 B1
6430272 Maruyama et al. Aug 2002 B1
6442545 Feldman et al. Aug 2002 B1
6456632 Baum et al. Sep 2002 B1
6463474 Fuh et al. Oct 2002 B1
6484209 Momirov Nov 2002 B1
6499107 Gleichauf et al. Dec 2002 B1
6535868 Galeazzi et al. Mar 2003 B1
6564263 Bergman et al. May 2003 B1
6578147 Shanklin et al. Jun 2003 B1
6594643 Freeny, Jr. Jul 2003 B1
6601104 Fallon Jul 2003 B1
6604158 Fallon Aug 2003 B1
6625150 Yu Sep 2003 B1
6633868 Min et al. Oct 2003 B1
6658377 Anward et al. Dec 2003 B1
6658423 Pugh et al. Dec 2003 B1
6691301 Bowen Feb 2004 B2
6704816 Burke Mar 2004 B1
6710702 Averbuch et al. Mar 2004 B1
6711558 Indeck et al. Mar 2004 B1
6765918 Dixon et al. Jul 2004 B1
6766304 Kemp, II et al. Jul 2004 B2
6772132 Kemp, II et al. Aug 2004 B1
6772170 Pennock et al. Aug 2004 B2
6772345 Shetty Aug 2004 B1
6782394 Landeck et al. Aug 2004 B1
6785677 Fritchman Aug 2004 B1
6804667 Martin Oct 2004 B1
6807156 Veres et al. Oct 2004 B1
6839686 Galant Jan 2005 B1
6850906 Chadha et al. Feb 2005 B1
6870837 Ho et al. Mar 2005 B2
6877044 Lo et al. Apr 2005 B2
6882747 Thawonmas et al. Apr 2005 B2
6886103 Brustoloni et al. Apr 2005 B1
6901461 Bennett May 2005 B2
6931408 Adams et al. Aug 2005 B2
6944168 Paatela et al. Sep 2005 B2
6971017 Stringer et al. Nov 2005 B2
6978223 Milliken Dec 2005 B2
6980976 Alpha et al. Dec 2005 B2
6981054 Krishna Dec 2005 B1
7016910 Egilsson et al. Mar 2006 B2
7016914 Nayak Mar 2006 B2
7019674 Cadambi et al. Mar 2006 B2
7024408 Dehlinger et al. Apr 2006 B2
7028250 Ukrainczyk et al. Apr 2006 B2
7046848 Olcott May 2006 B1
7051037 Thomas et al. May 2006 B1
7054854 Hattori et al. May 2006 B1
7058735 Spencer Jun 2006 B2
7065475 Brundobler Jun 2006 B1
7082427 Seibel et al. Jul 2006 B1
7089188 Logan et al. Aug 2006 B2
7092956 Ruediger Aug 2006 B2
7093023 Lockwood et al. Aug 2006 B2
7096179 Zhu et al. Aug 2006 B2
7106905 Simske Sep 2006 B2
7113954 Vogel Sep 2006 B2
7117437 Chen et al. Oct 2006 B2
7120079 McCollum et al. Oct 2006 B2
7120699 Stork et al. Oct 2006 B2
7124140 Barton Oct 2006 B2
7127424 Kemp, II et al. Oct 2006 B2
7127510 Yoda et al. Oct 2006 B2
7130913 Fallon Oct 2006 B2
7139743 Indeck et al. Nov 2006 B2
7149715 Browne et al. Dec 2006 B2
7167980 Chiu Jan 2007 B2
7177833 Marynowski et al. Feb 2007 B1
7181437 Indeck et al. Feb 2007 B2
7181608 Fallon et al. Feb 2007 B2
7222114 Chan et al. May 2007 B1
7224185 Campbell et al. May 2007 B2
7225188 Gai et al. May 2007 B1
7228289 Brumfield et al. Jun 2007 B2
7249118 Sandler et al. Jul 2007 B2
7251629 Marynowski et al. Jul 2007 B1
7277887 Burrows et al. Oct 2007 B1
7287037 An et al. Oct 2007 B2
7305383 Kubesh et al. Dec 2007 B1
7305391 Wyschogrod et al. Dec 2007 B2
7321937 Fallon Jan 2008 B2
7356498 Kaminsky et al. Apr 2008 B2
7363277 Dutta et al. Apr 2008 B1
7372875 Hadzic et al. May 2008 B2
7376755 Pandya May 2008 B2
7386564 Abdo et al. Jun 2008 B2
7406444 Eng et al. Jul 2008 B2
7408932 Kounavis et al. Aug 2008 B2
7411957 Stacy et al. Aug 2008 B2
7415723 Pandya Aug 2008 B2
7420931 Nanda et al. Sep 2008 B2
7444515 Dharmapurikar et al. Oct 2008 B2
7454418 Wang Nov 2008 B1
7457834 Jung et al. Nov 2008 B2
7461064 Fontoura et al. Dec 2008 B2
7467155 McCool et al. Dec 2008 B2
7478431 Nachenberg Jan 2009 B1
7480253 Allan Jan 2009 B1
7487264 Pandya Feb 2009 B2
7487327 Chang et al. Feb 2009 B1
7496108 Biran et al. Feb 2009 B2
7536462 Pandya May 2009 B2
7558925 Bouchard et al. Jul 2009 B2
7565525 Vorbach et al. Jul 2009 B2
7580719 Karmarkar Aug 2009 B2
7606968 Branscome et al. Oct 2009 B2
7627693 Pandya Dec 2009 B2
7631107 Pandya Dec 2009 B2
7636703 Taylor Dec 2009 B2
7660793 Indeck et al. Feb 2010 B2
7668849 Narancic et al. Feb 2010 B1
7685121 Brown et al. Mar 2010 B2
7685254 Pandya Mar 2010 B2
7701945 Roesch et al. Apr 2010 B2
7761459 Zhang et al. Jul 2010 B1
7783862 Cameron Aug 2010 B2
7788293 Pasztor et al. Aug 2010 B2
7805392 Steele et al. Sep 2010 B1
7827190 Pandya Nov 2010 B2
7831606 Pandya Nov 2010 B2
7831607 Pandya Nov 2010 B2
7840482 Singla et al. Nov 2010 B2
7856545 Casselman Dec 2010 B2
7856546 Casselman et al. Dec 2010 B2
7870217 Pandya Jan 2011 B2
7890692 Pandya Feb 2011 B2
7899976 Pandya Mar 2011 B2
7899977 Pandya Mar 2011 B2
7899978 Pandya Mar 2011 B2
7908259 Branscome et al. Mar 2011 B2
7912808 Pandya Mar 2011 B2
7917299 Buhler et al. Mar 2011 B2
7921046 Parsons et al. Apr 2011 B2
7930314 Gupta Apr 2011 B2
7944920 Pandya May 2011 B2
7945528 Cytron et al. May 2011 B2
7949650 Indeck et al. May 2011 B2
7953743 Indeck et al. May 2011 B2
7954114 Chamberlain et al. May 2011 B2
7996348 Pandya Aug 2011 B2
8005966 Pandya Aug 2011 B2
8027893 Burrows et al. Sep 2011 B1
8030888 Pandya et al. Oct 2011 B2
8051022 Pandya Nov 2011 B2
8055601 Pandya Nov 2011 B2
8069102 Indeck et al. Nov 2011 B2
8095508 Chamberlain et al. Jan 2012 B2
8131697 Indeck et al. Mar 2012 B2
8156101 Indeck et al. Apr 2012 B2
8181239 Pandya May 2012 B2
8200599 Pandya Jun 2012 B2
8224800 Branscome et al. Jul 2012 B2
8229918 Branscome et al. Jul 2012 B2
8234267 Branscome et al. Jul 2012 B2
8244718 Chamdani et al. Aug 2012 B2
8260754 Tatebe et al. Sep 2012 B2
8274682 Parkinson Sep 2012 B2
8326819 Indeck et al. Dec 2012 B2
8402022 Frank et al. Mar 2013 B2
8407122 Parsons et al. Mar 2013 B2
8458081 Parsons et al. Jun 2013 B2
8478680 Parsons et al. Jul 2013 B2
8515682 Buhler et al. Aug 2013 B2
8549024 Indeck et al. Oct 2013 B2
8595104 Parsons et al. Nov 2013 B2
8600856 Parsons et al. Dec 2013 B2
8601086 Pandya Dec 2013 B2
8620881 Chamberlain et al. Dec 2013 B2
8626624 Parsons et al. Jan 2014 B2
8655764 Parsons et al. Feb 2014 B2
8737606 Taylor et al. May 2014 B2
8751452 Chamberlain et al. Jun 2014 B2
8762249 Taylor et al. Jun 2014 B2
8768805 Taylor et al. Jul 2014 B2
8768888 Chamberlain et al. Jul 2014 B2
8880501 Indeck et al. Nov 2014 B2
20010014093 Yoda et al. Aug 2001 A1
20010047473 Fallon Nov 2001 A1
20010056547 Dixon Dec 2001 A1
20020010825 Wilson Jan 2002 A1
20020031125 Sato Mar 2002 A1
20020049841 Johnson et al. Apr 2002 A1
20020069370 Mack Jun 2002 A1
20020069375 Bowen Jun 2002 A1
20020072893 Wilson Jun 2002 A1
20020091826 Comeau et al. Jul 2002 A1
20020095512 Rana et al. Jul 2002 A1
20020100029 Bowen Jul 2002 A1
20020101425 Hamid Aug 2002 A1
20020103663 Bankier et al. Aug 2002 A1
20020105911 Pruthi et al. Aug 2002 A1
20020129140 Peled et al. Sep 2002 A1
20020138376 Hinkle Sep 2002 A1
20020143521 Call Oct 2002 A1
20020156998 Casselman Oct 2002 A1
20020162025 Sutton et al. Oct 2002 A1
20020166063 Lachman et al. Nov 2002 A1
20020180742 Hamid Dec 2002 A1
20020199173 Bowen Dec 2002 A1
20030009693 Brock et al. Jan 2003 A1
20030014521 Elson et al. Jan 2003 A1
20030014662 Gupta et al. Jan 2003 A1
20030018630 Indeck et al. Jan 2003 A1
20030023653 Dunlop et al. Jan 2003 A1
20030023876 Bardsley et al. Jan 2003 A1
20030028408 RuDusky Feb 2003 A1
20030028690 Appleby-Alis et al. Feb 2003 A1
20030028864 Bowen Feb 2003 A1
20030033234 RuDusky Feb 2003 A1
20030033450 Appleby-Alis Feb 2003 A1
20030033514 Appleby-Allis et al. Feb 2003 A1
20030033588 Alexander Feb 2003 A1
20030033594 Bowen Feb 2003 A1
20030035547 Newton Feb 2003 A1
20030037037 Adams et al. Feb 2003 A1
20030037321 Bowen Feb 2003 A1
20030041129 Applcby-Allis Feb 2003 A1
20030043805 Graham et al. Mar 2003 A1
20030046668 Bowen Mar 2003 A1
20030051043 Wyschogrod et al. Mar 2003 A1
20030055769 RuDusky Mar 2003 A1
20030061409 RuDusky Mar 2003 A1
20030065607 Satchwell Apr 2003 A1
20030065943 Geis et al. Apr 2003 A1
20030074177 Bowen Apr 2003 A1
20030074582 Patel et al. Apr 2003 A1
20030078865 Lee Apr 2003 A1
20030079060 Dunlop Apr 2003 A1
20030099254 Richter May 2003 A1
20030105620 Bowen Jun 2003 A1
20030105721 Ginter et al. Jun 2003 A1
20030110229 Kulig et al. Jun 2003 A1
20030115485 Milliken Jun 2003 A1
20030117971 Aubury Jun 2003 A1
20030120460 Aubury Jun 2003 A1
20030121010 Aubury Jun 2003 A1
20030140337 Aubury Jul 2003 A1
20030154368 Stevens et al. Aug 2003 A1
20030163715 Wong Aug 2003 A1
20030169877 Liu et al. Sep 2003 A1
20030177253 Schuehler et al. Sep 2003 A1
20030184593 Dunlop Oct 2003 A1
20030187662 Wilson Oct 2003 A1
20030208430 Gershon Nov 2003 A1
20030221013 Lockwood et al. Nov 2003 A1
20040000928 Cheng et al. Jan 2004 A1
20040010612 Pandya Jan 2004 A1
20040015502 Alexander et al. Jan 2004 A1
20040019703 Burton Jan 2004 A1
20040028047 Hou et al. Feb 2004 A1
20040049596 Schuehler et al. Mar 2004 A1
20040054924 Chuah et al. Mar 2004 A1
20040064737 Milliken et al. Apr 2004 A1
20040100977 Suzuki et al. May 2004 A1
20040111632 Halperin Jun 2004 A1
20040117645 Okuda et al. Jun 2004 A1
20040162826 Wyschogrod et al. Aug 2004 A1
20040170070 Rapp et al. Sep 2004 A1
20040177340 Hsu et al. Sep 2004 A1
20040196905 Yamane et al. Oct 2004 A1
20040205149 Dillon et al. Oct 2004 A1
20050005145 Teixeira Jan 2005 A1
20050044344 Stevens Feb 2005 A1
20050086520 Dharmapurikar et al. Apr 2005 A1
20050108518 Pandya May 2005 A1
20050171943 Ichino et al. Aug 2005 A1
20050175010 Wilson et al. Aug 2005 A1
20050187974 Gong Aug 2005 A1
20050195832 Dharmapurikar et al. Sep 2005 A1
20050229254 Singh et al. Oct 2005 A1
20050283743 Mulholland et al. Dec 2005 A1
20060020715 Jungck Jan 2006 A1
20060031263 Arrouye et al. Feb 2006 A1
20060036693 Hulten et al. Feb 2006 A1
20060047636 Mohania et al. Mar 2006 A1
20060053295 Madhusudan et al. Mar 2006 A1
20060059213 Evoy Mar 2006 A1
20060101005 Yang et al. May 2006 A1
20060129745 Thiel et al. Jun 2006 A1
20060136570 Pandya Jun 2006 A1
20060198375 Baik et al. Sep 2006 A1
20060215691 Kobayashi et al. Sep 2006 A1
20060242123 Williams, Jr. Oct 2006 A1
20060259417 Marynowski et al. Nov 2006 A1
20060269148 Farber et al. Nov 2006 A1
20060294059 Chamberlain et al. Dec 2006 A1
20070011175 Langseth et al. Jan 2007 A1
20070011183 Langseth et al. Jan 2007 A1
20070011317 Brandyburg et al. Jan 2007 A1
20070011687 Ilik et al. Jan 2007 A1
20070061594 Ginter et al. Mar 2007 A1
20070067108 Buhler et al. Mar 2007 A1
20070067481 Sharma et al. Mar 2007 A1
20070078837 Indeck et al. Apr 2007 A1
20070094199 Deshpande et al. Apr 2007 A1
20070112748 Angell et al. May 2007 A1
20070112837 Houh et al. May 2007 A1
20070118500 Indeck et al. May 2007 A1
20070130140 Cytron et al. Jun 2007 A1
20070156574 Marynowski et al. Jul 2007 A1
20070156669 Marchisio et al. Jul 2007 A1
20070174841 Chamberlain et al. Jul 2007 A1
20070179935 Lee et al. Aug 2007 A1
20070209068 Ansari et al. Sep 2007 A1
20070237327 Taylor et al. Oct 2007 A1
20070244859 Trippe et al. Oct 2007 A1
20070260602 Taylor Nov 2007 A1
20070260814 Branscome et al. Nov 2007 A1
20070277036 Chamberlain et al. Nov 2007 A1
20070294157 Singla et al. Dec 2007 A1
20080005062 Gupta et al. Jan 2008 A1
20080021874 Dahl et al. Jan 2008 A1
20080077582 Reed Mar 2008 A1
20080082502 Gupta Apr 2008 A1
20080084573 Horowitz et al. Apr 2008 A1
20080086274 Chamberlain et al. Apr 2008 A1
20080104542 Cohen et al. May 2008 A1
20080109413 Indeck et al. May 2008 A1
20080114724 Indeck et al. May 2008 A1
20080114725 Indeck et al. May 2008 A1
20080114760 Indeck et al. May 2008 A1
20080126320 Indeck et al. May 2008 A1
20080133453 Indeck et al. Jun 2008 A1
20080133519 Indeck et al. Jun 2008 A1
20080175239 Sistanizadeh et al. Jul 2008 A1
20080183688 Chamdani et al. Jul 2008 A1
20080189251 Branscome et al. Aug 2008 A1
20080189252 Branscome et al. Aug 2008 A1
20080243675 Parsons et al. Oct 2008 A1
20080253395 Pandya Oct 2008 A1
20090262741 Jungck et al. Oct 2009 A1
20090287628 Indeck et al. Nov 2009 A1
20100082895 Branscome et al. Apr 2010 A1
20100094858 Indeck et al. Apr 2010 A1
20100174770 Pandya Jul 2010 A1
20100198850 Cytron et al. Aug 2010 A1
20100198920 Wong et al. Aug 2010 A1
20110040701 Singla et al. Feb 2011 A1
20110066832 Casselman et al. Mar 2011 A1
20110125960 Casselman May 2011 A1
20110167083 Branscome et al. Jul 2011 A1
20110178911 Parsons et al. Jul 2011 A1
20110178912 Parsons et al. Jul 2011 A1
20110178917 Parsons et al. Jul 2011 A1
20110178918 Parsons et al. Jul 2011 A1
20110178919 Parsons et al. Jul 2011 A1
20110178957 Parsons et al. Jul 2011 A1
20110179050 Parsons et al. Jul 2011 A1
20110184844 Parsons et al. Jul 2011 A1
20110218987 Branscome et al. Sep 2011 A1
20110231446 Buhler et al. Sep 2011 A1
20110252008 Chamberlain et al. Oct 2011 A1
20110291615 Pandya et al. Dec 2011 A1
20120109849 Chamberlain et al. May 2012 A1
20120110316 Chamberlain et al. May 2012 A1
20120116998 Indeck et al. May 2012 A1
20120117610 Pandya May 2012 A1
20120130922 Indeck et al. May 2012 A1
20120215801 Indeck et al. Aug 2012 A1
20130007000 Indeck et al. Jan 2013 A1
20130018835 Pandya Jan 2013 A1
20130148802 Taylor et al. Jun 2013 A1
20130159449 Taylor et al. Jun 2013 A1
20130262287 Parsons et al. Oct 2013 A1
20130290163 Parsons et al. Oct 2013 A1
20140025656 Indeck et al. Jan 2014 A1
20140040109 Parsons et al. Feb 2014 A1
20140067830 Buhler et al. Mar 2014 A1
20140089163 Parsons et al. Mar 2014 A1
20140164215 Parsons et al. Jun 2014 A1
20140180903 Parsons et al. Jun 2014 A1
20140180904 Parsons et al. Jun 2014 A1
20140180905 Parsons et al. Jun 2014 A1
20140181133 Parsons et al. Jun 2014 A1
Foreign Referenced Citations (70)
Number Date Country
0573991 Dec 1993 EP
0880088 Nov 1996 EP
0851358 Jul 1998 EP
0887723 Dec 1998 EP
0911738 Apr 1999 EP
0989754 Mar 2000 EP
57-137978 Aug 1982 JP
58-102378 Jun 1983 JP
61-107596 May 1986 JP
63-143668 Jun 1988 JP
02136900 May 1990 JP
03014075 Jan 1991 JP
H03156677 Jul 1991 JP
04-205174 Jul 1992 JP
5101102 Apr 1993 JP
05-217289 Aug 1993 JP
09-054797 Feb 1997 JP
09145544 Jun 1997 JP
9-269930 Oct 1997 JP
2000-076270 Mar 2000 JP
2000285134 Oct 2000 JP
2000286715 Oct 2000 JP
2001357048 Dec 2001 JP
2002101089 Apr 2002 JP
2005242997 Sep 2005 JP
2005527041 Sep 2005 JP
2008063973 May 2008 NO
2005081855 Sep 2005 WF
9010910 Sep 1990 WO
9409443 Apr 1994 WO
9737735 Oct 1997 WO
9852130 Nov 1998 WO
9905814 Feb 1999 WO
9955052 Oct 1999 WO
0041136 Jul 2000 WO
0122425 Mar 2001 WO
0135216 May 2001 WO
0139577 Jun 2001 WO
0161913 Aug 2001 WO
0180082 Oct 2001 WO
0180082 Oct 2001 WO
0180558 Oct 2001 WO
0190890 Nov 2001 WO
02061525 Aug 2002 WO
02082271 Oct 2002 WO
03100650 Apr 2003 WO
03036845 May 2003 WO
03100662 Dec 2003 WO
03104943 Dec 2003 WO
2004014065 Feb 2004 WO
2004017604 Feb 2004 WO
2004042560 May 2004 WO
2004042561 May 2004 WO
2004042562 May 2004 WO
2004042574 May 2004 WO
2005017708 Feb 2005 WO
2005026925 Mar 2005 WO
2005048134 May 2005 WO
2005114339 Dec 2005 WO
2006023948 Mar 2006 WO
2004042569 Apr 2006 WO
2006060571 Jun 2006 WO
2006096324 Sep 2006 WO
2007064685 Jun 2007 WO
2007079095 Jul 2007 WO
2007087507 Aug 2007 WO
2008022036 Feb 2008 WO
2008063974 May 2008 WO
2008073824 Jun 2008 WO
2010077829 Jul 2010 WO
Non-Patent Literature Citations (212)
Entry
Lockwood, “Building Networks with Reprogrammable Hardware”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Evolvable Internet Hardware Platforms”, NASA/DoD Workshop on Evolvable Hardware (EHW'01), Long Beach, CA, Jul. 12-14, 2001, pp. 271-279.
Lockwood, “Hardware Laboratory Configuration”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Introduction”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Platform and Methodology for Teaching Design of Hardware Modules in Internet Routers and Firewalls”, IEEE Computer Society International Conference on Microelectronic Systems Education (MSE'2001), Las Vegas, NV, Jun. 17-18, 2001, pp. 56-57.
Lockwood, “Protocol Processing on the FPX”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Simulation and Synthesis”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Simulation of the Hello World Application for the Field-Programmable Port Extender (FPX)”, Washington University, Applied Research Lab, Spring 2001 Gigabits Kits Workshop.
Madhusudan, “Design of a System for Real-Time Worm Detection”, Hot Interconnects, pp. 77-83, Stanford, CA, Aug. 2004, found at http://www.hoti.org/hoti12/program/papers/2004/paper4.2.pdf.
Madhusudan, “Design of a System for Real-Time Worm Detection”, Power Point Presentation in Support of Master's Thesis, Washington Univ., Dept. of Computer Science and Engineering, St. Louis, MO, Aug. 2004.
Mao et al., “Cluster-based Online Monitoring System of Web Traffic”, Dept. of Computer Science and Technology, Tsinghua Univ., Bejing, 100084 P.R. China.
Mosanya et al., “A FPGA-Based Hardware Implementation of Generalized Profile Search Using Online Arithmetic”, ACM/Sigda International Symposium on Field Programmable Gate Arrays (FPGA '99), Feb. 21-23, 1999, pp. 101-111, Monterey, CA, USA.
Moscola et al., “FPGrep and FPSed: Regular Expression Search and Substitution for Packet Streaming in Field Programmable Hardware”, Dept. of Computer Science, Applied Research Lab, Washington University, Jan. 8, 2002, unpublished, pp. 1-19, St. Louis, MO.
Moscola et al., “FPSed: A Streaming Content Search-and-Replace Module for an Internet Firewall”, Proc. of Hot Interconnects, 11th Symposium on High Performance Interconnects, pp. 122-129, Aug. 20, 2003.
Moscola, “FPGrep and FPSed: Packet Payload Processors for Managing the Flow of Digital Content on Local Area Networks and the Internet”, Master's Thesis, Sever Institute of Technology, Washington University, St. Louis, MO, Aug. 2003.
Motwani et al., “Randomized Algorithms”, 1995, pp. 215-216, Cambridge University Press.
Mueller, “Upgrading and Repairing PCs, 15th Anniversary Edition”, 2004, pp. 63-66, 188, Que.
Navarro, “A Guided Tour to Approximate String Matching”, ACM Computing Surveys, vol. 33, No. 1, Mar. 2001, pp. 31-88.
Necker et al., “TCP-Stream Reassembly and State Tracking in Hardware”, School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA.
Niewczas et al., “A Pattern Matching Algorithm for Verification and Analysis of Very Large IC Layouts”, ACM, Apr. 1998, pp. 129-134.
Notice of Allowance for U.S. Appl. No. 11/938,709 dated Aug. 27, 2009.
Notice of Allowance for U.S. Appl. No. 12/640,891 dated Nov. 9, 2011.
Nunez et al., “The X-MatchLITE FPGA-Based Data Compressor”, Euromicro Conference 1999, Proceedings, Italy, Sep. 8-10, 1999, pp. 126-132, Los Alamitos, CA.
Nwodoh et al., “A Processing System for Real-Time Holographic Video Computation”, Reconfigurable Technology: FPGAs for Computing and Application, Proceedings for the SPIE, Sep. 1999, Boston, pp. 129-140, vol. 3844.
Office Action for JP Application 2009-536535 dated Aug. 21, 2012.
Office Action for JP Application 2009-536536 dated Jul. 17, 2012.
Office Action for U.S. Appl. No. 10/550,323 dated Jan. 3, 2011.
Office Action for U.S. Appl. No. 11/938,709 dated Jan. 13, 2009.
Office Action for U.S. Appl. No. 11/938,709 dated Jul. 14, 2008.
Partial International Search Report for PCT/US03/15638 dated Feb. 3, 2004.
Prakash et al., “OC-3072 Packet Classification Using BDDs and Pipelined SRAMs”, Department of Electrical and Computer Engineering, The University of Texas at Austin.
Pramanik et al., “A Hardware Pattern Matching Algorithm on a Dataflow”; Computer Journal; Jul. 1, 1985; pp. 264-269; vol. 28, No. 3; Oxford University Press, Surrey, Great Britain.
Prosecution History for U.S. Appl. No. 11/938,709, now U.S. Pat. No. 7,660,793, filed Nov. 12, 2007.
Prosecution History for U.S. Appl. No. 12/640,891, now U.S. Pat. No. 8,156,101, filed Dec. 17, 2009.
Ramakrishna et al., “A Performance Study of Hashing Functions for Hardware Applications”, Int. Conf. on Computing and Information, May 1994, pp. 1621-1636, vol. 1, No. 1.
Ramakrishna et al., “Efficient Hardware Hashing Functions for High Performance Computers”, IEEE Transactions on Computers, Dec. 1997, vol. 46, No. 12.
Ramesh et al., “Automatic Selection of Tuning Parameters for Feature Extraction Sequences”, IEEE, Jun. 21-23, 1994, pp. 672-677.
Ratha et al., “Convolution on Splash 2”, Proceedings of IEEE Symposium on FPGAS for Custom Computing Machines, Apr. 19, 1995, pp. 204-213, Los Alamitos, California.
Ratha et al., “FPGA-based coprocessor for text string extraction”, IEEE, Sep. 11-13, 2000, pp. 217-221.
Roberts, “Internet Still Growing Dramatically Says Internet Founder”, Press Release, Caspian Networks, Inc.—Virtual Pressroom.
Roesch, “Snort—Lightweight Intrusion Detection for Networks”, Proceedings of LISA '99: 13th Systems Administration Conference; Nov. 7-12, 1999; pp. 229-238; USENIX Association, Seattle, WA USA.
Roy, “A bounded search algorithm for segmented channel routing for FPGA's and associated channel architecture issues”, IEEE, Nov. 11, 1993, pp. 1695-1705, vol. 12.
Sachin Tandon, “A Programmable Architecture for Real-Time Derivative Trading”, Master's Thesis, University of Edinburgh, 2003.
Schmerken, “With Hyperfeed Litigation Pending, Exegy Launches Low-Latency Ticker Plant”, in Wall Street & Technology Blog, Mar. 20, 2007, pp. 1-2.
Schmit, “Incremental Reconfiguration for Pipelined Applications”, FPGAs for Custom Computing Machines, Proceedings, The 5th Annual IEEE Symposium, Dept. of ECE, Carnegie Mellon University, Apr. 16-18, 1997, pp. 47-55, Pittsburgh, PA.
Schuehler et al., “Architecture for a Hardware Based, TCP/IP Content Scanning System”, IEEE Micro, 24(1):62-69, Jan.-Feb. 2004, USA.
Schuehler et al., “TCP-Splitter: A TCP/IP Flow Monitor in Reconfigurable Hardware”, Hot Interconnects 10 (Hotl-10), Stanford, CA, Aug. 21-23, 2002, pp. 127-131.
Seki et al., “High Speed Computation of Shogi With FPGA”, Proceedings of 58th Convention Architecture, Software Science, Engineering, Mar. 9, 1999, pp. 1-133-1-134.
Shah, “Understanding Network Processors”, Version 1.0, University of California-Berkeley, Sep. 4, 2001.
Shalunov et al., “Bulk TCP Use and Performance on Internet 2”, ACM SIGCOMM Internet Measurement Workshop, 2001.
Shasha et al., “Database Tuning”, 2003, pp. 280-284, Morgan Kaufmann Publishers.
Shirazi et al., “Quantitative Analysis of Fpga-based Database Searching”, Journal of VLSI Signal Processing Systems for Signal, Image, and Video Technology, May 2001, pp. 85-96, vol. 28, No. 1/2, Kluwer Academic Publishers, Dordrecht, NL.
Sidhu et al., “Fast Regular Expression Matching Using FPGAs”, IEEE Symposium on Field Programmable Custom Computing Machines (FCCM 2001), Apr. 2001.
Sidhu et al., “String Matching on Multicontext FPGAs Using Self-Reconfiguration”, FPGA '99: Proceedings of the 1999 ACM/SIGDA 7th International Symposium on Field Programmable Gate Arrays, Feb. 1999, pp. 217-226.
Singh et al., “The EarlyBird System for Real-Time Detection on Unknown Worms”, Technical report CS2003-0761, Aug. 2003.
Skiena et al., “Programming Challenges: The Programming Contest Training Manual”, 2003, pp. 30-31, Springer.
Sourdis and Pnevmatikatos, “Fast, Large-Scale String Match for a 10Gbps FPGA-based Network Intrusion Detection System”, 13th International Conference on Field Programmable Logic and Applications, 2003.
Steinbach et al., “A Comparison of Document Clustering Techniques”, KDD Workshop on Text Mining, 2000.
Tan et al., “A High Throughput String Matching Architecture for Intrusion Detection and Prevention”, ISCA 2005: 32nd Annual International Symposium on Computer Architecture, pp. 112-122, 2005.
Tau et al., “Transit Note #114: A First Generation DPGA Implementation”, Jan. 1995, 9 pages.
Taylor et al., “Dynamic Hardware Plugins (DHP): Exploiting Reconfigurable Hardware for High-Performance Programmable Routers”, Computer Networks, 38(3): 295-310 (16), Feb. 21, 2002, and online at http://www.cc.gatech.edu/classes/AY2007/cs8803hpc—fall/papers/phplugins.pdf.
Taylor et al., “Generalized Rad Module Interface Specification of the Field Programmable Port Extender (FPX) Version 2”, Washington University, Department of Computer Science, Technical Report, Jul. 5, 2001, pp. 1-10.
Taylor et al., “Modular Design Techniques for the FPX”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Taylor et al., “Scalable Packet Classification using Distributed Crossproducting of Field Labels”, Proceedings of IEEE Infocom, Mar. 2005, pp. 1-12, vol. 20, No. 1.
Taylor, “Models, Algorithms, and Architectures for Scalable Packet Classification”, doctoral thesis, Department of Computer Science and Engineering, Washington University, St. Louis, MO, Aug. 2004, pp. 1-201.
Uluski et al., “Characterizing Antivirus Workload Execution”, Sigarch Comput. Archit. News, vol. 33, No. 1, pp. 90-98, Mar. 2005.
U.S. Appl. No. 61/421,545, filed Dec. 9, 2010 (Taylor et al.).
Villasenor et al, “Configurable Computing Solutions for Automatic Target Recognition”, FPGAS for Custom Computing Machines, 1996, Proceedings, IEEE Symposium on Napa Valley, CA, Apr. 17-19, 1996, pp. 70-79, 1996 IEEE, Napa Valley, CA, Los Alamitos, CA, USA.
Waldvogel et al., “Scalable High-Speed Prefix Matching”, ACM Transactions on Computer Systems, Nov. 2001, pp. 440-482, vol. 19, No. 4.
Ward et al., “Dynamically Reconfigurable Computing: A Novel Computation Technology with Potential to Improve National Security Capabilities”, May 15, 2003, A White Paper Prepared by Star Bridge Systems, Inc. [retrieved Dec. 12, 2006]. Retrieved from the Internet: <URL: http://www.starbridgesystems.com/resources/whitepapers/Dynamically%20Reconfigurable%20Computing.pdf.
Weaver et al., “Very Fast Containment of Scanning Worms”, Proc. USENIX Security Symposium 2004, San Diego, CA, Aug. 2004, located at http://www.icsi.berkely.edu/˜nweaver/containment/containment.pdf.
Web-Pop (Professional Options Package) (www.pmpublishing.com).
West et al., “An FPGA-Based Search Engine for Unstructured Database”, Proc. of 2nd Workshop on Application Specific Processors, Dec. 2003, San Diego, CA.
Wooster et al., “HTTPDUMP Network HTTP Packet Snooper”, Apr. 25, 1996.
Worboys, “GIS: A Computing Perspective”, 1995, pp. 245-247, 287, Taylor & Francis Ltd.
Yamaguchi et al., “An Approach for Homology Search with Reconfigurable Hardware”, Google, 2001, pp. 374-375.
Yamaguchi et al., “High Speed Homology Search with FPGAs”, Proceedings Pacific Symposium on Biocomputing, Jan. 3-7, 2002, pp. 271-282, vol. 7, Online, Lihue, Hawaii, USA.
Yan et al., “Enhancing Collaborative Spam Detection with Bloom Filters”, 2006, IEEE, pp. 414-425.
Yoshitani et al., “Performance Evaluation of Parallel vol. Rendering Machine Re Volver/C40”, Study Report of Information Processing Society, Mar. 5, 1999, pp. 79-84, vol. 99, No. 21.
“A Reconfigurable Computing Model for Biological Research Application of Smith-Waterman Analysis to Bacterial Genomes” A White Paper Prepared by Star Bridge Systems, Inc. [retrieved Dec. 12, 2006]. Retrieved from the Internet: <URL: http://www.starbridgesystems.com/resources/whitepapers/Smith%20 Waterman%20Whitepaper.pdf.
“Lucent Technologies Delivers “PayloadPlus” Network Processors for Programmable, MultiProtocol, OC-48c Processing”, Lucent Technologies Press Release, downloaded from http://www.lucent.com/press/1000/0010320.meb.html on Mar. 21, 2002.
“Overview, Field Programmable Port Extender”, Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002, pp. 1-4.
“Payload Plus™ Agere System Interface”, Agere Systems Product Brief, Jun. 2001, downloaded from Internet, Jan. 2002, pp. 1-6.
“RFC793: Transmission Control Protocol, Darpa Internet Program, Protocol Specification”, Sep. 1981.
“Technology Overview”, Data Search Systems Incorporated, downloaded from the http://www.datasearchsystems.com/tech.htm on Apr. 19, 2004.
“The Field-Programmable Port Extender (FPX)”, downloaded from http://www.arl.wustl.edu/arl/ in Mar. 2002.
Aldwairi et al., “Configurable String Matching Hardware for Speeding up Intrusion Detection”, SIRARCH Comput. Archit. News, vol. 33, No. 1, pp. 99-107, Mar. 2005.
Amanuma et al., “A FPGA Architecture for High Speed Computation”, Proceedings of 60th Convention Architecture, Software Science, Engineering, Mar. 14, 2000, pp. 1-163-1-164, Information Processing Society, Japan.
Amer-Yahia et al., “XQuery 1.0 and XPath 2.0 Full-Text 1.0”, W3C Working Draft, http://www.w3.org/TR/query-full-text/, May 18, 2007—parts 1-4.
Anerousis et al., “Using the At&T Labs PacketScope for Internet Measurement, Design, and Performance Analysis”, Network and Distributed Systems Research Laboratory, AT&T Labs-Research, Florham, Park, NJ, Oct. 1997.
Anonymous, “Method for Allocating Computer Disk Space to a File of Known Size”, IBM Technical Disclosure Bulletin, vol. 27, No. 10B, Mar. 1, 1985, New York.
Arnold et al., “The Splash 2 Processor and Applications”, Proceedings 1993 IEEE International Conference on Computer Design: VLSI in Computers and Processors (ICCD '93), Oct. 3, 1993, pp. 482-485, IEEE Computer Society, Cambridge, MA USA.
Artan et al., “Multi-packet Signature Detection using Prefix Bloom Filters”, 2005, IEEE, pp. 1811-1816.
Asami et al., “Improvement of DES Key Search on FPGA-Based Parallel Machine “RASH””, Proceedings of Information Processing Society, Aug. 15, 2000, pp. 50-57, vol. 41, No. SIG5 (HPS1), Japan.
Baboescu et al., “Scalable Packet Classification,” SIGCOMM'01, Aug. 27-31, 2001, pp. 199-210, San Diego, California, USA; http://www.ecse.rpi.edu/homepages/shivkuma/teaching/sp2001/readings/baboescu-pkt-classification.pdf.
Baer, “Computer Systems Architecture”, 1980, pp. 262-265; Computer Science Press, Potomac, Maryland.
Baeza-Yates et al., “New and Faster Filters for Multiple Approximate String Matching”, Random Structures and Algorithms (RSA), Jan. 2002, pp. 23-49, vol. 20, No. 1.
Baker et al., “High-throughput Linked-Pattern Matching for Intrusion Detection Systems”, ANCS 2005: Proceedings of the 2005 Symposium on Architecture for Networking and Communications Systems, pp. 193-202, ACM Press, 2005.
Barone-Adesi et al., “Efficient Analytic Approximation of American Option Values”, Journal of Finance, vol. 42, No. 2 (Jun. 1987), pp. 301-320.
Behrens et al., “BLASTN Redundancy Filter in Reprogrammable Hardware,” Final Project Submission, Fall 2003, Department of Computer Science and Engineering, Washington University.
Berk “JLex. A lexical analyzer generator for Java™”, downloaded from http://www.cs.princeton.edu/˜appel/modern/java/Jlex/ in Jan. 2002, pp. 1-18.
Bloom, “Space/Time Trade-offs in Hash Coding With Allowable Errors”, Communications of the ACM, Jul. 1970, pp. 422-426, vol. 13, No. 7, Computer Usage Company, Newton Upper Falls, Massachusetts, USA.
Braun et al., “Layered Protocol Wrappers for Internet Packet Processing in Reconfigurable Hardware”, Proceedings of Hot Interconnects 9 (Hotl-9) Stanford, CA, Aug. 22-24, 2001, pp. 93-98.
Braun et al., “Protocol Wrappers for Layered Network Packet Processing in Reconfigurable Hardware”, IEEE Micro, Jan.-Feb. 2002, pp. 66-74.
Brodie et al., “Dynamic Reconfigurable Computing”, in Proc. of 9th Military and Aerospace Programmable Logic Devices International Conference, Sep. 2006.
Cavnar et al., “N-Gram-Based Text Categorization”, Proceedings of SDAIR-94, 3rd Annual Symposium on Document Analysis and Information Retrieval, Las Vegas, pp. 161-175, 1994.
Celko, “Joe Celko's Data & Databases: Concepts in Practice”, 1999, pp. 72-74, Morgan Kaufmann Publishers.
Chamberlain et al., “Achieving Real Data Throughput for an FPGA Co-Processor on Commodity Server Platforms”, Proc. of 1st Workshop on Building Block Engine Architectures for Computers and Networks, Oct. 2004, Boston, MA.
Chamberlain et al., “The Mercury System: Embedding Computation Into Disk Drives”, 7th High Performance Embedded Computing Workshop, Sep. 2003, Boston, MA.
Chamberlain et al., “The Mercury System: Exploiting Truly Fast Hardware for Data Search”, Proc. of Workshop on Storage Network Architecture and Parallel I/Os, Sep. 2003, New Orleans, LA.
Chaney et al., “Design of a Gigabit ATM Switch”, Washington University, St. Louis.
Cho et al., “Deep Packet Filter with Dedicated Logic and Read Only Memories”, 12th Annual IEEE Symposium on Field-Programmable Custom Computing Machines, Apr. 2004.
Cho, “A Fast Regular Expression Indexing Engine”, Proc. of 18th Int'l Conv. on Data Engineering, 2001, pp. 1-12.
Choi et al., “Design of a Flexible Open Platform for High Performance Active Networks”, Allerton Conference, 1999, Champaign, IL.
Clark et al., “Scalable Pattern Matching for High Speed Networks”, Proceedings of the 12th Annual IEEE Symposium on Field-Programmable Custom Computing Machines, 2004; FCCM 2004, 20-23 Apr. 2004; pp. 249-257; IEEE Computer Society; Cambridge, MA USA.
Cloutier et al., “VIP: An Fpga-Based Processor for Image Processing and Neural Networks”, Proceedings of Fifth International Conference on Microelectronics for Neural Networks, Feb. 12, 1996, pp. 330-336, Los Alamitos, California.
Compton et al., “Configurable Computing: a Survey of Systems and Software”, Technical Report, Northwestern University, Dept. of ECE, 1999.
Compton et al., “Reconfigurable Computing: A Survey of Systems and Software”, Technical Report, Northwestern University, Dept. of ECE, 1999, presented by Yi-Gang Tai.
Cong et al., “An Optional Technology Mapping Algorithm for Delay Optimization in Lookup-Table Based FPGA Designs”, IEEE, 1992, pp. 48-53.
Cuppu and Jacob, “Organizational Design Trade-Offs at the Dram, Memory Bus and Memory Controller Level: Initial Results,” Technical Report UMB-SCA-1999-2, Univ. of Maryland Systems & Computer Architecture Group, Nov. 1999, pp. 1-10.
Denoyer et al., “HMM-based Passage Models for Document Classification and Ranking”, Proceedings of ECIR-01, 23rd European Colloquim Information Retrieval Research, Darmstatd, DE, pp. 126-135, 2001.
Dharmapurikar et al., “Deep Packet Inspection Using Parallel Bloom Filters,” IEEE Micro, Jan.-Feb. 2004, vol. 24, Issue: 1, pp. 52-61.
Dharmapurikar et al., “Deep Packet Inspection Using Parallel Bloom Filters,” Symposium on High Performance Interconnects (Hotl), Stanford, California, 2003, pp. 44-51.
Dharmapurikar et al., “Design and Implementation of a String Matching System for Network Intrusion Detection using FPGA-based Bloom Filters”, Proc. of 12th Annual IEEE Symposium on Field Programmable Custom Computing Machines, 2004, pp. 1-10.
Dharmapurikar et al., “Longest Prefix Matching Using Bloom Filters,” SIGCOMM, 2003, pp. 201-212.
Dharmapurikar et al., “Robust TCP Stream Reassembly in the Presence of Adversaries”, Proc. of the 14th Conference on USENIX Security Symposium—vol. 14, 16 pages, Baltimore, MD, 2005; http://www.icir.org/vern/papers/TcpReassembly/TCPReassembly.pdf.
Dharmapurikar, “Fast and Scalable Pattern Matching for Content Filtering”, ACM, ANCS 05, 2005, pp. 183-192.
Ebeling et al., “RaPiD—Reconfigurable Pipelined Datapath”, University of Washington, Dept. of Computer Science and Engineering, Sep. 23, 1996, Seattle, WA.
Feldmann, “BLT: Bi-Layer Tracing of HTTP and TCP/IP”, AT&T Labs-Research, Florham Park, NJ, USA.
Fernandez, “Template Matching of Binary Targets in Grey-Scale Images: A Nonparametric Approach”, Pattern Recognition, 1997, pp. 1175-1182, vol. 30, No. 7.
Forgy, “RETE: A Fast Algorithm for the Many Pattern/Many Object Pattern Matching Problem”, Artificial Intelligence, 1982, pp. 17-37, vol. 19.
Franklin et al., “An Architecture for Fast Processing of Large Unstructured Data Sets.” Proc. of 22nd Int'l Conf. on Computer Design, Oct. 2004, pp. 280-287.
Franklin et al., “Assisting Network Intrusion Detection with Reconfigurable Hardware”, Symposium on Field-Programmable Custom Computing Machines (FCCM 2002), Apr. 2002, Napa, California.
Fu et al., “The FPX KCPSM Module: An Embedded, Reconfigurable Active Processing Module for the Field Programmable Port Extender (FPX)”, Washington University, Department of Computer Science, Technical Report WUCS-01-14, Jul. 2001.
Gavrila et al., “Multi-feature Hierarchical Template Matching Using Distance Transforms”, IEEE, Aug. 16-20, 1998, vol. 1, pp. 439-444.
Gokhale et al., “Reconfigurable Computing: Accelerating Computation With Field-Programmable Gate Arrays”, 2005, pp. 1-3, 7, 11-15, 39, 92-93, Springer.
Google Search Results Page for “field programmable gate array financial calculation stock market” over dates of Jan. 1, 1990-May 21, 2002, 1 page.
Guerdoux-Jamet et al., “Systolic Filter for Fast DNA Similarity Search”, IEEE, 1995, pp. 145-156.
Gunther et al., “Assessing Document Relevance with Run-Time Reconfigurable Machines”, IEEE Symposium on FPGAs for Custom Computing Machines, 1996, pp. 10-17, Proceedings, Napa Valley, CA.
Gupta et al, “High-Speed Implementations of Rule-Based Systems,” ACM Transactions on Computer Systems, May 1989, pp. 119-146, vol. 7, Issue 2.
Gupta et al., “Packet Classification on Multiple Fields”, Computer Systems Laboratory, Stanford University, Stanford, CA.
Gupta et al., “PMM: A Parallel Architecture for Production Systems,” Proceedings of the IEEE, Apr. 1992, pp. 693-696, vol. 2.
Gurtov, “Effect of Delays on TCP Performance”, Cellular Systems Development, Sonera Corporation, online at http://cs.helsinki.fi/u/gurtov/papers/pwc01.pdf.
Gyang, “NCBI BLASTN Stage 1 in Reconfigurable Hardware,” Technical Report WUCSE-2005-30, Aug. 2004, Department of Computer Science and Engineering, Washington University, St. Louis, MO.
Halaas et al., “A Recursive MISD Architecture for Pattern Matching”, IEEE Transactions on Very Large Scale Integration, vol. 12, No. 7, pp. 727-734, Jul. 2004.
Hauck et al., “Software Technologies for Reconfigurable Systems”, Northwestern University, Dept. of ECE, Technical Report, 1996.
Hayes, “Computer Architecture and Organization”, Second Edition, 1988, pp. 448-459, McGraw-Hill, Inc.
Herbordt et al., “Single Pass, Blast-Like, Approximate String Matching on FPGAs”, 14th Annual IEEE Symposium on Field-Programmable Custom Computing Machines (FCCM'06), Apr. 2006, pp. 1-10, IEEE.
Hezel et al., “FPGA-Based Template Matching Using Distance Transforms”, Proceedings of the 10th Annual IEEE Symposium on Field-Programmable Custom Computing Machines, Apr. 22, 2002, pp. 89-97, IEEE Computer Society, USA.
Hollaar, “Hardware Systems for Text Information Retrieval”, Proceedings of the Sixth Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Jun. 6-8, 1983, pp. 3-9, Baltimore, Maryland, USA.
Hutchings et al., “Assisting Network Intrusion Detection with Reconfigurable Hardware”, FCCM 2002: 10th Annual IEEE Symposium on Field-Programmable Custom Computing Machines, 2002.
Ibrahim et al., “Lecture Notes in Computer Science: Database and Expert Systems Applications”, 2000, p. 769, vol. 1873, Springer.
International Search Report for PCT/US2001/011255 dated Jul. 10, 2003.
International Search Report for PCT/US2002/033286 dated Jan. 22, 2003.
International Search Report for PCT/US2003/015638 dated May 6, 2004.
International Search Report for PCT/US2004/016021 dated Aug. 18, 2005.
International Search Report for PCT/US2004/016398 dated Apr. 12, 2005.
International Search Report for PCT/US2005/030046; Sep. 25, 2006.
International Search Report for PCT/US2006/006105 dated Oct. 31, 2006.
International Search Report for PCT/US2006/045653 dated Jul. 8, 2008.
International Search Report for PCT/US2007/060835 dated Jul. 9, 2007.
International Search Report for PCT/US2007/084464 dated Oct. 2, 2008.
International Search Report for PCT/US2007/084466 dated Jul. 23, 2008.
Jacobson et al., “RFC 1072: TCP Extensions for Long-Delay Paths”, Oct. 1988.
Jacobson et al., “tcpdump—dump traffic on a network”, Jun. 30, 1997, online at www.cse.cuhk.edu.hk/˜cslui/CEG4430/tcpdump.ps.gz.
Johnson et al., “Pattern Matching in Reconfigurable Logic for Packet Classification”, College of Computing, Georgia Institute of Technology, Atlanta, GA.
Jones et al., “A Probabilistic Model of Information Retrieval: Development and Status”, Information Processing and Management, Aug. 1998, 76 pages.
Keutzer et al., “A Survey of Programmable Platforms—Network Proc”, University of California-Berkeley, pp. 1-29.
Koloniari et al., “Content-Based Routing of Path Queries in Peer-to-Peer Systems”, pp. 1-19, E. Bertino et al. (Eds.): EDBT 2004, LNCS 2992, pp. 29-47, 2004, copyright by Springer-Verlag, Germany.
Krishnamurthy et al., “Biosequence Similarity Search on the Mercury System”, Proceedings of the 15th IEEE International Conference on Application-Specific Systems, Architectures, and Processors (ASAP04), Sep. 2004, pp. 365-375.
Kulig et al., “System and Method for Controlling Transmission of Data Packets Over an Information Network”, pending U.S. Patent Application.
Lancaster et al., “Acceleration of Ungapped Extension in Mercury BLAST”, Seventh (7th) Workshop on Media and Streaming Processors, Nov. 12, 2005, Thirty-Eighth (38th) International Symposium on Microarchitecture (MICRO-38), Barcelona, Spain.
Li et al., “Large-Scale IP Traceback in High-Speed Internet: Practical Techniques and Theoretical Foundation”, Proceedings of the 2004 IEEE Symposium on Security and Privacy, 2004, pp. 1-15.
Lin et al., “Real-Time Image Template Matching Based on Systolic Array Processor”, International Journal of Electronics; Dec. 1, 1992; pp. 1165-1176; vol. 73, No. 6; London, Great Britain.
Lockwood et al., “Field Programmable Port Extender (FPX) for Distributed Routing and Queuing”, ACM International Symposium on Field Programmable Gate Arrays (FPGA 2000), Monterey, CA, Feb. 2000, pp. 137-144.
Lockwood et al., “Hello, World: A Simple Application for the Field Programmable Port Extender (FPX)”, Washington University, Department of Computer Science, Technical Report WUCS-00-12, Jul. 11, 2000.
Lockwood et al., “Parallel FPGA Programming over Backplane Chassis”, Washington University, Department of Computer Science, Technical Report WUCS-00-11, Jun. 12, 2000.
Lockwood et al., “Reprogrammable Network Packet Processing on the Field Programmable Port Extender (FPX)”, ACM International Symposium on Field Programmable Gate Arrays (FPGA 2001), Monterey, CA, Feb. 2001, pp. 87-93.
Lockwood, “An Open Platform for Development of Network Processing Modules in Reprogrammable Hardware”, IEC DesignCon 2001, Santa Clara, CA, Jan. 2001, Paper WB-19.
Altschul et al., “Basic Local Alignment Search Tool”, J. Mol. Biol., Oct. 5, 1990, 215, pp. 403-410.
Baker et al., “Time and Area Efficient Pattern Matching on FPGAs”, ACM, Feb. 22-24, 2004, pp. 223-232.
Batory, “Modeling the Storage Architectures of Commercial Database Systems”, ACM Transactions on Database Systems, Dec. 1985, pp. 463-528, vol. 10, issue 4.
Bianchi et al., “Improved Queueing Analysis of Shared Buffer Switching Networks”, ACM, Aug. 1993, pp. 482-490.
Chenna et al., “Multiple Sequence Alignment with the Clustal Series of Programs”, Nucleic Acids Research, 2003, vol. 31, No. 13, pp. 3497-3500.
Cholleti, “Storage Allocation in Bounded Time”, MS Thesis, Dept. of Computer Science and Engineering, Washington Univeristy, St. Louis, MO (Dec. 2002). Available as Washington University Technical Report WUCSE-2003-2.
Dehon, “DPGA-coupled Microprocessors Commodity ICs for the Early 21st Century”, Institute of Electrical and Electronics Engineers, 1994, pp. 31-39.
Edgar, “Muscle: Multiple Sequence Alignment with High Accuracy and High Throughput”, Nucleic Acids Research, 2004, vol. 32, No. 5, pp. 1792-1797.
Eick et al., “Hardware Accelerated Algorithms for Semantic Processing of Document Streams”, IEEE Aerospace Conference, 2006, pp. 1-14, Piscataway, NJ, USA.
Extended European Search Report for EP Application 07845064.0 dated May 6, 2013.
Extended European Search Report for EP Application 07845066.5 dated May 21, 2013.
FIPS 197, “Advanced Encryption Standard”, National Institute of Standards and Technology (2001).
FIPS Pub. 46-3. Data Encryption Standard (DES). Revised version of 46-2. Reaffirmed Oct. 25, 1999.
Gokhale et al., “Reconfigurable Computing: Accelerating Computation with Field-Programmable Gate Arrays”Springer, 2005, pp. 1-36.
Gokhale et al., “Reconfigurable Computing: Accelerating Computation with Field-Programmable Gate Arrays”Springer, 2005, pp. 1-54, 92-96.
Hoinville, et al. “Spatial Noise Phenomena of Longitudinal Magnetic Recording Media”, IEEE Transactions on Magnetics, vol. 28, No. 6, Nov. 1992.
Howe, Data Analysis for Database Design Third Edition, 2001, 335 pages, Butterworth-Heinemann.
International Search Report for PCT/US2007/075723 dated Jul. 25, 2008.
International Search Report for PCT/US2008/065955 dated Aug. 22, 2008.
International Search Report for PCT/US2008/066929 dated Aug. 29, 2008.
Invitation to Pay Additional Fees and Annex to Form PCT/ISA/206 Communication Relating to the Results of the Partial International Search for International Application PCT/US2003/015638 dated Feb. 3, 2004.
Jeanmougin et al., “Multiple Sequence Alignment with Clustal X”, TIBS, 1998, vol. 23, pp. 403-405.
Lancaster, “Design and Evaluation of a BLAST Ungapped Extension Accelerator, Master's Thesis, May 2006”, Thesis (http://cse.seas.wustedu/Research/File Download.asp?489), 2006, pp. Cover pp. 1-4; iii-x; pp. 1-65, Washington University in St. Louis.
Office Action for JP Application 2009-536535 dated Apr. 23, 2013.
Office Action for JP Application 2009-536535 dated Nov. 12, 2013.
Office Action for JP Application 2009-536536 dated Apr. 2, 2013.
Office Action for JP Application 2013-160166 dated Jul. 15, 2014.
Office Action for JP Application 2013-160166 dated Mar. 31, 2015.
Prosecution History for U.S. Appl. No. 11/938,732, now U.S. Pat. No. 8,326,819, filed Nov. 12, 2007.
Sugawara et al., “Over 10Gbps String Matching Mechanism for Multi-Stream Packet Scanning Systems”, Field Programmable Logic and Application Lecture Notes in Computer Science, 2004, pp. 484-493, vol. 3203.
Tau et al., “A First Generation DPGA Implementation”, Third Canadian Workshop of Field-Programmable Devices,May 1995, pp. 138-143.
Thompson et al., “The CLUSTAL—X Windows Interface: Flexible Strategies for Multiple Sequence Alignment Aided by Quality Analysis Tools”, Nucleic Acids Research, 1997, vol. 25, No. 24, pp. 4876-4882.
Ziv et al., “A Universal Algorithm for Sequential Data Compression”, IEEE Trans. Inform. Theory, IT-23(3): 337-343 (1997).
Related Publications (1)
Number Date Country
20130086096 A1 Apr 2013 US
Provisional Applications (1)
Number Date Country
60865629 Nov 2006 US
Continuations (1)
Number Date Country
Parent 11938732 Nov 2007 US
Child 13686338 US