Method and apparatus for high-speed processing of financial market depth data

Information

  • Patent Grant
  • 10929930
  • Patent Number
    10,929,930
  • Date Filed
    Friday, August 24, 2018
    6 years ago
  • Date Issued
    Tuesday, February 23, 2021
    3 years ago
Abstract
A variety of embodiments for hardware-accelerating the processing of financial market depth data are disclosed. A coprocessor, which may be resident in a ticker plant, can be configured to update order books based on financial market depth data at extremely low latency. Such a coprocessor can also be configured to enrich a stream of limit order events pertaining to financial instruments with data from a plurality of updated order books.
Description
TERMINOLOGY

The following paragraphs provide several definitions for various terms used herein. These paragraphs also provide background information relating to these terms.


Financial Instrument: As used herein, a “financial instrument” refers to a contract representing an equity ownership, debt, or credit, typically in relation to a corporate or governmental entity, wherein the contract is saleable. Examples of financial instruments include stocks, bonds, options, commodities, currency traded on currency markets, etc. but would not include cash or checks in the sense of how those items are used outside the financial trading markets (i.e., the purchase of groceries at a grocery store using cash or check would not be covered by the term “financial instrument” as used herein; similarly, the withdrawal of $100 in cash from an Automatic Teller Machine using a debit card would not be covered by the term “financial instrument” as used herein).


Financial Market Data: As used herein, the term “financial market data” refers to data contained in or derived from a series of messages that individually represent a new offer to buy or sell a financial instrument, an indication of a completed sale of a financial instrument, notifications of corrections to previously-reported sales of a financial instrument, administrative messages related to such transactions, and the like. Feeds of messages which contain financial market data are available from a number of sources and exist in a variety of feed types—for example, Level 1 feeds and Level 2 feeds as discussed herein.


Basket: As used herein, the term “basket” refers to a collection comprising a plurality of elements, each element having one or more values. The collection may be assigned one or more Net Values (NVs), wherein a NV is derived from the values of the plurality of elements in the collection. For example, a basket may be a collection of data points from various scientific experiments. Each data point may have associated values such as size, mass, etc. One may derive a size NV by computing a weighted sum of the sizes, a mass NV by computing a weighted sum of the masses, etc. Another example of a basket would be a collection of financial instruments, as explained below.


Financial Instrument Basket: As used herein, the term “financial instrument basket” refers to a basket whose elements comprise financial instruments. The financial instrument basket may be assigned one or more Net Asset Values (NAVs), wherein a NAV is derived from the values of the elements in the basket. Examples of financial instruments that may be included in baskets are securities (stocks), bonds, options, mutual funds, exchange-traded funds, etc. Financial instrument baskets may represent standard indexes, exchange-traded funds (ETFs), mutual funds, personal portfolios, etc. One may derive a last sale NAV by computing a weighted sum of the last sale prices for each of the financial instruments in the basket, a bid NAV by computing a weighted sum of the current best bid prices for each of the financial instruments in the basket, etc.


GPP: As used herein, the term “general-purpose processor” (or GPP) refers to a hardware device having a fixed form and whose functionality is variable, wherein this variable functionality is defined by fetching instructions and executing those instructions, of which a conventional central processing unit (CPU) is a common example. Exemplary embodiments of GPPs include an Intel Xeon processor and an AMD Opteron processor.


Reconfigurable Logic: As used herein, the term “reconfigurable logic” refers to any logic technology whose form and function can be significantly altered (i.e., reconfigured) in the field post-manufacture. This is to be contrasted with a GPP, whose function can change post-manufacture, but whose form is fixed at manufacture.


Software: As used herein, the term “software” refers to data processing functionality that is deployed on a GPP or other processing devices, wherein software cannot be used to change or define the form of the device on which it is loaded.


Firmware: As used herein, the term “firmware” refers to data processing functionality that is deployed on reconfigurable logic or other processing devices, wherein firmware may be used to change or define the form of the device on which it is loaded.


Coprocessor: As used herein, the term “coprocessor” refers to a computational engine designed to operate in conjunction with other components in a computational system having a main processor (wherein the main processor itself may comprise multiple processors such as in a multi-core processor architecture). Typically, a coprocessor is optimized to perform a specific set of tasks and is used to offload tasks from a main processor (which is typically a GPP) in order to optimize system performance. The scope of tasks performed by a coprocessor may be fixed or variable, depending on the architecture of the coprocessor. Examples of fixed coprocessor architectures include Graphics Processor Units which perform a broad spectrum of tasks and floating point numeric coprocessors which perform a relatively narrow set of tasks. Examples of reconfigurable coprocessor architectures include reconfigurable logic devices such as Field Programmable Gate Arrays (FPGAs) which may be reconfigured to implement a wide variety of fixed or programmable computational engines. The functionality of a coprocessor may be defined via software and/or firmware.


Hardware Acceleration: As used herein, the term “hardware acceleration” refers to the use of software and/or firmware implemented on a coprocessor for offloading one or more processing tasks from a main processor to decrease processing latency for those tasks relative to the main processor.


Bus: As used herein, the term “bus” refers to a logical bus which encompasses any physical interconnect for which devices and locations are accessed by an address. Examples of buses that could be used in the practice of the present invention include, but are not limited to the PCI family of buses (e.g., PCI-X and PCI-Express) and HyperTransport buses.


Pipelining: As used herein, the terms “pipeline”, “pipelined sequence”, or “chain” refer to an arrangement of application modules wherein the output of one application module is connected to the input of the next application module in the sequence. This pipelining arrangement allows each application module to independently operate on any data it receives during a given clock cycle and then pass its output to the next downstream application module in the sequence during another clock cycle.


BACKGROUND AND SUMMARY OF THE INVENTION

The process of trading financial instruments may be viewed broadly as proceeding through a cycle as shown in FIG. 1. At the top of the cycle is the exchange which is responsible for matching up offers to buy and sell financial instruments. Exchanges disseminate market information, such as the appearance of new buy/sell offers and trade transactions, as streams of events known as market data feeds. Trading firms receive market data from the various exchanges upon which they trade. Note that many traders manage diverse portfolios of instruments requiring them to monitor the state of multiple exchanges. Utilizing the data received from the exchange feeds, trading systems make trading decisions and issue buy/sell orders to the financial exchanges. Orders flow into the exchange where they are inserted into a sorted “book” of orders, triggering the publication of one or more events on the market data feeds.


Exchanges keep a sorted listing of limit orders for each financial instrument, known as an order book. As used herein, a “limit order” refers to an offer to buy or sell a specified number of shares of a given financial instrument at a specified price. Limit orders can be sorted based on price, size, and time according to exchange-specific rules. Many exchanges publish market data feeds that disseminate order book updates as order add, modify, and delete events. These feeds belong to a class of feeds known as level 2 data feeds. It should be understood that each exchange may be a little different as to when data is published on the feed and how much normalization the exchange performs when publishing events on the feed, although it is fair to expect that the amount of normalization in the level 2 feed is minimal relative to a level 1 feed. These feeds typically utilize one of two standard data models: full order depth or price aggregated depth. As shown in FIG. 2(a), full order depth feeds contain events that allow recipients to construct order books that mirror the order books used by the exchange matching engines. This is useful for trading strategies that require knowledge of the state of specific orders in the market.


As shown in FIG. 2(b), price aggregated depth feeds contain events that allow recipients to construct order books that report the distribution of liquidity (available shares) over prices for a given financial instrument.


Order book feeds are valuable to electronic trading as they provide what is generally considered the fastest and deepest insight into market dynamics. The current set of order book feeds includes feeds for order books of equities, equity options, and commodities. Several exchanges have announced plans to provide new order book feeds for derivative instruments such as equity options. Given its explosive growth over the past several years, derivative instrument trading is responsible for the lion's share of current market data traffic. The Options Price Reporting Authority (OPRA) feed is the most significant source of derivatives market data, and it belongs to the class of feeds known as “level 1” feeds. Level 1 feeds report quotes, trades, trade cancels and corrections, and a variety of summary events. For a given financial instrument, the highest buy price and lowest sell price comprise the “best bid and offer” (BBO) that are advertised as the quote. As an exchange's sorted order book listing changes due to order executions, modifications, or cancellations, the exchange publishes new quotes. When the best bid and offer prices match in the exchange's order book, the exchange executes a trade and advertises the trade transaction on its level 1 market data feed. Note that some amount of processing is required prior to publishing a quote or trade event because of the latency incurred by the publisher's computer system when processing limit orders to build order books and identify whether trades or quotes should be generated. Thus, level 1 data feeds from exchanges or other providers possess inherent latency relative to viewing “raw” order events on order book feeds. A feed of raw limit order data belongs to a class of feeds known as “level 2” feeds.


In order to minimize total system latency, many electronic trading firms ingest market data feeds, including market data feeds of limit orders, directly into their own computer systems from the financial exchanges. While some loose standards are in place, most exchanges define unique protocols for disseminating their market data. This allows the exchanges to modify the protocols as needed to adjust to changes in market dynamics, regulatory controls, and the introduction of new asset classes. The ticker plant resides at the head of the platform and is responsible for the normalization, caching, filtering, and publishing of market data messages. A ticker plant typically provides a subscribe interface to a set of downstream trading applications. By normalizing data from disparate exchanges and asset classes, the ticker plant provides a consistent data model for trading applications. The subscribe interface allows each trading application to construct a custom normalized data feed containing only the information it requires. This is accomplished by performing subscription-based filtering at the ticker plant.


In traditional market data platforms known to the inventors, the ticker plant may perform some normalization tasks on order book feeds, but the task of constructing sorted and/or price-aggregated views of order books is typically pushed to downstream components in the market data platform. The inventors believe that such a trading platform architecture increases processing latency and the number of discrete systems required to process order book feeds. As an improvement over such an arrangement, an embodiment of the invention disclosed herein enables a ticker plant to perform order feed processing (e.g., normalization, price-aggregation, sorting) in an accelerated and integrated fashion, thereby increasing system throughput and decreasing processing latency. In an exemplary embodiment, the ticker plant employs a coprocessor that serves as an offload engine to accelerate the building of order books. Financial market data received on a feed into the ticket plant can be transferred on a streaming basis to the coprocessor for high speed processing.


Thus, in accordance with an exemplary embodiment of the invention, the inventors disclose a method for generating an order book view from financial market depth data, the method comprising: (1) maintaining a data structure representative of a plurality of order books for a plurality of financial instruments, and (2) hardware-accelerating a processing of a plurality of financial market depth data messages to update the order books within the data structure. Preferably the hardware-accelerating step is performed by a coprocessor within a ticker plant. The inventors also disclose a system for generating an order book view from financial market depth data, the system comprising: (1) a memory for storing a data structure representative of a plurality of order books for a plurality of financial instruments, and (2) a coprocessor configured to process of a plurality of financial market depth data messages to update the order books within the data structure.


Using these order books, the method and system can also produce views of those order books for ultimate delivery to interested subscribers. The inventors define two general classes of book views that can be produced in accordance with various exemplary embodiments: stream views (unsorted, non-cached) and summary views (sorted, cached). Stream views provide client applications with a normalized stream of updates for limit orders or aggregated price-points for the specified regional symbol, composite symbol, or feed source (exchange). Summary views provide client applications with multiple sorted views of the book, including composite views (a.k.a. “virtual order books”) that span multiple markets.


In an exemplary embodiment, stream views comprise a normalized stream of updates for limit orders or aggregated price-points for the specified regional symbol, composite symbol, or feed source (exchange). Following the creation of a stream subscription, a ticker plant can be configured to provide a client application with a stream of normalized events containing limit order or price point updates. As stream subscriptions do not provide sorting, it is expected that stream view data would be employed by client applications that construct their own book views or journals from the normalized event stream from one or more specified exchanges.


An example of a stream view that can be generated by various embodiments is an order stream view. An order stream view comprises a stream of normalized limit order update events for one or more specified regional symbols. The normalized events comprise fields such as the type of update (add, modify, delete), the order price, order size, exchange timestamp, and order identifier (if provided by the exchange). Another example of an order stream view is an order exchange stream view that comprises a stream of normalized limit order update events for one or more specified exchanges or clusters of instruments within an exchange. The normalized events comprise fields such as the type of update (add, modify, delete), the order price, order size, exchange timestamp, and order identifier (if provided by the exchange).


Another example of a stream view that can be generated by various embodiments is a price stream view. A price stream view comprises a stream of normalized price level update events for one or more specified regional symbols. The normalized events comprise fields such as the type of update (add, modify, delete), the aggregated price, order volume at the aggregated price, and the order count at the aggregated price. Another example of a price stream view is a price exchange stream view. A price exchange stream view comprises a stream of normalized price level update events for one or more specified exchanges or clusters of instruments within an exchange. The normalized events comprise fields such as the type of update (add, modify, delete), the aggregated price, order volume at the aggregated price, and order count at the aggregated price.


Another example of a stream view that can be generated by various embodiments is an aggregate stream view. An aggregate stream view comprises a stream of normalized price level update events for one or more specified composite symbols. The normalized events comprise fields such as the type of update (add, modify, delete), the (virtual) aggregated price, (virtual) order volume at the aggregated price, and (virtual) order count at the aggregated price.


As explained in the above-referenced and incorporated U.S. Patent Application Publication 2008/0243675, a regional symbol serves to identify a financial instrument traded on a particular exchange while a composite symbol serves to identify a financial instrument in the aggregate on all of the exchanges upon which it trades. It should be understood that embodiments of the invention disclosed herein may be configured to store both regional and composite records for the same financial instrument in situations where the financial instrument is traded on multiple exchanges.


Summary views provide liquidity insight, and the inventors believe it is highly desirable to obtain such liquidity insight with ultra low latency. In accordance with an embodiment disclosed herein, by offloading a significant amount of data processing from client applications to a ticker plant, the ticker plant frees up client processing resources, thereby enabling those client resources to implement more sophisticated trading applications that retain first mover advantage.


An example of a summary view that can be generated by various embodiments is an order summary view. An order summary view represents a first-order liquidity view of the raw limit order data disseminated by a single feed source. The inventors define an order summary view to be a sorted listing comprising a plurality of individual limit orders for a given financial instrument on a given exchange. The sort order is preferably by price and then by time (or then by size for some exchanges). An example of an order summary view is shown in FIG. 3.


Another example of a summary view that can be generated by various embodiments is a price summary view. A price summary view represents a second-order liquidity view of the raw limit order data disseminated by a single feed source. The inventors define a price summary view to be a sorted listing comprising a plurality of price levels for a given financial instrument on a given exchange, wherein each price level represents an aggregation of same-priced orders from that exchange. The price level timestamp in the summary view preferably reports the timestamp of the most recent event at that price level from that exchange. An example of a price summary view is shown in FIG. 4(a). Note that a price summary view produced by an embodiment disclosed herein may be limited to a user-specified number of price points starting from the top of the book.


Another example of a summary view that can be generated by various embodiments is a spliced price summary view. A spliced price summary view represents a second-order, pan-market liquidity view of the raw limit order data disseminated by multiple feed sources. The inventors define a spliced price summary view to be a sorted listing comprising a plurality of price levels for a given financial instrument across all contributing exchanges where each price level represents an aggregation of same-priced orders from a unique contributing exchange. The price level timestamp in the spliced price summary view preferably reports the timestamp of the most recent event at that price level for the specified exchange. An example of a spliced price summary view is shown in FIG. 4(b). Note that a spliced price summary view produced by an embodiment disclosed herein may be limited to a user-specified number of price points starting from the top of the book.


Another example of a summary view that can be generated by various embodiments is an aggregate price summary view. An aggregate price summary view represents a third-order, pan-market liquidity view of the raw limit order data disseminated by multiple feed sources. The inventors define an aggregate price summary view to be a sorted listing comprising a plurality of price levels for a given financial instrument where each price level represents an aggregation of same-priced orders from all contributing exchanges. The price level timestamp in the aggregate price summary view preferably reports the timestamp of the most recent event at that price level from any contributing exchange. An example of an aggregate price summary view is shown in FIG. 4(c). Note that an aggregate price summary view produced by an embodiment disclosed herein may be limited to a user-specified number of price points starting from the top of the book.


The inventors further note that financial exchanges have continued to innovate in order to compete and to provide more efficient markets. One example of such innovation is the introduction of ephemeral regional orders in several equity markets (e.g., FLASH orders on NASDAQ, BOLT orders on BATS) that provide regional market participants the opportunity to view specific orders prior to public advertisement. Another example of such innovation is implied liquidity in several commodity markets (e.g. CME, ICE) that allow market participants to trade against synthetic orders whose price is derived from other derivative instruments. In order to capture and distinguish this type of order or price level in an order book, the inventors define the concept of attributes and apply this concept to the data structures employed by various embodiments disclosed herein. Each entry in an order book or price book may have one or more attributes. Conceptually, attributes are a vector of flags that may be associated with each order book or price book entry. By default, every order or aggregated price level is “explicit” and represents a limit order to buy or sell the associated financial instrument entered by a market participant. In some equity markets, an order or price level may be flagged using various embodiments disclosed herein with an attribute to indicate whether the order or price level relates to an ephemeral regional order (ERO). Similarly, in some commodity markets, an order or price level may be flagged using various embodiments disclosed herein to indicate whether the order or price level relates to an implied liquidity.


By capturing such attributes in the data structures employed by exemplary embodiments, the inventors note that these attributes thus provide another dimension to the types of book views that various embodiments disclosed herein generate. For example, one commodity trading application may wish to view a price aggregated book that omits implied liquidity, another commodity trading application may wish to view a price aggregated book with the explicit and implied price levels shown independently (spliced view), while another commodity trading application may wish to view a price aggregated book with explicit and implied entries aggregated by price. These three examples of attribute-based book views are shown in FIGS. 5, 6 and 7, respectively.


Thus, in accordance with an exemplary embodiment, the inventors disclose the use of attribute filtering and price level merging to capture the range of options in producing book views for books that contain entries with attributes. Attribute filtering allows applications to specify which entries should be included and/or excluded from the book view. Price level merging allows applications to specify whether or not entries that share the same price but differing attributes should be aggregated into a single price level.


The inventors also disclose several embodiments wherein a coprocessor can be used to enrich a stream of limit order events pertaining to financial instruments with order book data, both stream view order book data and summary view order book data, as disclosed herein.


These and other features and advantages of the present invention will be described hereinafter to those having ordinary skill in the art.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 depicts an exemplary process cycle for trading financial instruments;



FIG. 2(a) depicts an exemplary limit order event and its relation to a full order depth book;



FIG. 2(b) depicts an exemplary limit order event and its relation to a price aggregated depth order book;



FIG. 3 depicts exemplary bid and ask order summary views;



FIG. 4(a) depicts exemplary bid and ask price summary views;



FIG. 4(b) depicts exemplary bid and ask spliced price summary views;



FIG. 4(c) depicts exemplary bid and ask aggregate price summary views;



FIG. 5 depicts an exemplary price book view, with implied attributes filtered out;



FIG. 6 depicts an exemplary price book view, including a spliced attribute view;



FIG. 7 depicts an exemplary price book view, including a price merged attribute view;



FIGS. 8(a) and (b) depict examples of suitable platforms for processing market depth data;



FIGS. 9(a) and (b) depict exemplary printed circuit boards for use as a coprocessor;



FIG. 10 depicts an example of how a firmware pipeline can be deployed across multiple reconfigurable logic devices;



FIGS. 11(a)-(c) depicts various embodiments of a processing module for processing limit order data;



FIGS. 12(a)-(c) depict various embodiments of a pipeline for generating stream views of order books;



FIG. 13 depicts an exemplary embodiment of a compression function used to generate a hash key for symbol mapping;



FIG. 14 depicts an exemplary embodiment of a hash function for symbol mapping;



FIG. 15 depicts an exemplary embodiment for generating a global exchange identifier (GEID) for symbol mapping;



FIG. 16 depicts an exemplary embodiment of a module configured to enrich limit order events with normalization and price aggregation data;



FIG. 17 depicts an exemplary embodiment for storing and accessing limit order records and price point records, wherein such records are stored in a shared memory;



FIG. 18 depicts another exemplary embodiment for storing and accessing limit order records and price point records, wherein such records are partitioned across multiple physical memories;



FIG. 19 depicts an exemplary limit order event;



FIG. 20 depicts an exemplary limit order record;



FIG. 21(a) depicts an exemplary regional price point record;



FIG. 21(b) depicts an exemplary composite price point record;



FIG. 22 depicts an exemplary enriched limit order event;



FIG. 23 depicts an exemplary architecture for an order normalization and price aggregation (ONPA) module;



FIG. 24 depicts an exemplary embodiment where an API in a client application is configured to produce a sorted view of an order book from a stream view of an order book provided by a ticker plant;



FIGS. 25(a)-(d) depict various embodiments of a pipeline for generating summary views of order books;



FIG. 26 depicts an example of a B+ tree;



FIG. 27 depicts an example of how a module can be configured to access sorted order book data within a data structure;



FIG. 28 depicts an exemplary architecture for a sorted view update (SVU) module;



FIG. 29 depicts an exemplary pipeline that includes a value cache update (VCU) module operating on synthetic quote events created by an SVU module;



FIG. 30 depicts an exemplary pipeline that includes a basket calculation engine (BCE) module driven in part by synthetic quote events created by an SVU module; and



FIG. 31 depicts an exemplary ticker plant architecture in which a pipeline configured to process financial market depth data can be employed.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

Examples of suitable platforms for implementing exemplary embodiments of the invention are shown in FIGS. 8(a) and (b). FIG. 8(a) depicts a system 800 employs a hardware-accelerated data processing capability through coprocessor 840 to process financial market depth data. Within system 800, a coprocessor 840 is positioned to receive data that streams into the system 800 from a network 820 (via network interface 810). In a preferred embodiment, system 800 is employed to receive financial market limit order data and process financial market depth data. Network 820 thus preferably comprises a network through which system 800 can access a source for Level 2 financial data such as the exchanges themselves (e.g., NYSE, NASDAQ, etc.) or a third party provider (e.g., extranet providers such as Savvis or BT Radians). Such incoming data preferably comprises a series of financial market data messages, the messages representing events such as limit orders relating to financial instruments. These messages can exist in any of a number of formats, as is known in the art.


The computer system defined by processor 812 and RAM 808 can be any commodity computer system as would be understood by those having ordinary skill in the art. For example, the computer system may be an Intel Xeon system or an AMD Opteron system. Thus, processor 812, which serves as the central or main processor for system 800, preferably comprises a GPP.


In a preferred embodiment, the coprocessor 840 comprises a reconfigurable logic device 802. Preferably, data streams into the reconfigurable logic device 802 by way of system bus 806, although other design architectures are possible (see FIG. 9(b)). Preferably, the reconfigurable logic device 802 is a field programmable gate array (FPGA), although this need not be the case. System bus 806 can also interconnect the reconfigurable logic device 802 with the processor 812 as well as RAM 808. In a preferred embodiment, system bus 806 may be a PCI-X bus or a PCI-Express bus, although this need not be the case.


The reconfigurable logic device 802 has firmware modules deployed thereon that define its functionality. The firmware socket module 804 handles the data movement requirements (both command data and target data) into and out of the reconfigurable logic device, thereby providing a consistent application interface to the firmware application module (FAM) chain 850 that is also deployed on the reconfigurable logic device. The FAMs 850i of the FAM chain 850 are configured to perform specified data processing operations on any data that streams through the chain 850 from the firmware socket module 804. Examples of FAMs that can be deployed on reconfigurable logic in accordance with a preferred embodiments of the present invention are described below.


The specific data processing operation that is performed by a FAM is controlled/parameterized by the command data that FAM receives from the firmware socket module 804. This command data can be FAM-specific, and upon receipt of the command, the FAM will arrange itself to carry out the data processing operation controlled by the received command. For example, within a FAM that is configured to perform an exact match operation between data and a key, the FAM's exact match operation can be parameterized to define the key(s) that the exact match operation will be run against. In this way, a FAM that is configured to perform an exact match operation can be readily re-arranged to perform a different exact match operation by simply loading new parameters for one or more different keys in that FAM. As another example pertaining to baskets, a command can be issued to the one or more FAMs that make up a basket calculation engine to add/delete one or more financial instruments to/from the basket.


Once a FAM has been arranged to perform the data processing operation specified by a received command, that FAM is ready to carry out its specified data processing operation on the data stream that it receives from the firmware socket module. Thus, a FAM can be arranged through an appropriate command to process a specified stream of data in a specified manner. Once the FAM has completed its data processing operation, another command can be sent to that FAM that will cause the FAM to re-arrange itself to alter the nature of the data processing operation performed thereby. Not only will the FAM operate at hardware speeds (thereby providing a high throughput of data through the FAM), but the FAMs can also be flexibly reprogrammed to change the parameters of their data processing operations.


The FAM chain 850 preferably comprises a plurality of firmware application modules (FAMs) 850a, 850b, . . . that are arranged in a pipelined sequence. However, it should be noted that within the firmware pipeline, one or more parallel paths of FAMs 850i can be employed. For example, the firmware chain may comprise three FAMs arranged in a first pipelined path (e.g., FAMs 850a, 850b, 850c) and four FAMs arranged in a second pipelined path (e.g., FAMs 850d, 850e, 850f, and 850g), wherein the first and second pipelined paths are parallel with each other. Furthermore, the firmware pipeline can have one or more paths branch off from an existing pipeline path. A practitioner of the present invention can design an appropriate arrangement of FAMs for FAM chain 850 based on the processing needs of a given application.


A communication path 830 connects the firmware socket module 804 with the input of the first one of the pipelined FAMs 850a. The input of the first FAM 850a serves as the entry point into the FAM chain 850. A communication path 832 connects the output of the final one of the pipelined FAMs 850m with the firmware socket module 804. The output of the final FAM 850m serves as the exit point from the FAM chain 850. Both communication path 830 and communication path 832 are preferably multi-bit paths.


The nature of the software and hardware/software interfaces used by system 800, particularly in connection with data flow into and out of the firmware socket module are described in greater detail in U.S. Patent Application Publication 2007/0174841, the entire disclosure of which is incorporated herein by reference.



FIG. 8(b) depicts another exemplary embodiment for system 800. In the example of FIG. 8(b), system 800 includes a data store 842 that is in communication with bus 806 via disk controller 814. Thus, the data that is streamed through the coprocessor 840 may also emanate from data store 842. Data store 842 can be any data storage device/system, but it is preferably some form of mass storage medium. For example, data store 842 can be a magnetic storage device such as an array of Seagate disks.



FIG. 9(a) depicts a printed circuit board or card 900 that can be connected to the PCI-X or PCI-e bus 806 of a commodity computer system for use as a coprocessor 840 in system 800 for any of the embodiments of FIGS. 8(a)-(b). In the example of FIG. 9(a), the printed circuit board includes an FPGA 802 (such as a Xilinx Virtex 5 FPGA) that is in communication with a memory device 902 and a PCI-X bus connector 904. A preferred memory device 902 comprises SRAM and DRAM memory. A preferred PCI-X or PCI-e bus connector 904 is a standard card edge connector.



FIG. 9(b) depicts an alternate configuration for a printed circuit board/card 900. In the example of FIG. 9(b), a bus 906 (such as a PCI-X or PCI-e bus), one or more disk controllers 908, and a disk connector 910 are also installed on the printed circuit board 900. Any commodity disk interface technology can be supported, as is understood in the art. In this configuration, the firmware socket 804 also serves as a PCI-X to PCI-X bridge to provide the processor 812 with normal access to any disk(s) connected via the private PCI-X bus 906. It should be noted that a network interface can be used in addition to or in place of the disk controller and disk connector shown in FIG. 9(b).


It is worth noting that in either the configuration of FIG. 9(a) or 9(b), the firmware socket 804 can make memory 902 accessible to the bus 806, which thereby makes memory 902 available for use by an OS kernel as the buffers for transfers to the FAMs from a data source with access to bus. It is also worth noting that while a single FPGA 802 is shown on the printed circuit boards of FIGS. 9(a) and (b), it should be understood that multiple FPGAs can be supported by either including more than one FPGA on the printed circuit board 900 or by installing more than one printed circuit board 900 in the system 800. FIG. 10 depicts an example where numerous FAMs in a single pipeline are deployed across multiple FPGAs.



FIGS. 11(a)-(c) depict examples of processing modules 1100 that can be employed within coprocessor 840 to process limit order events. The processing module 1100 of FIG. 11(a) is configured to generate a stream view of processed limit order data. The processing module 1100 of FIG. 11(b) is configured to generate a summary view of processed limit order data, and the processing module 1100 of FIG. 11(c) is configured to generate both a stream view and a summary view of processed limit order data.


In the exemplary embodiments of FIG. 12(a)-(c), a data processing module 1100 for generating a stream view of processed limit order data can be realized via a pipeline 1200. In the example of FIG. 12(a), the pipeline comprises a message parsing (MP) module 1204 that receives raw messages 1202. These messages 1202 comprise a stream of financial market data events, of which at least a plurality comprise limit order events. Downstream from the MP module 1204 is a symbol mapping (SM) module 1206, and downstream from the SM module 1206 is an Order Normalization and Price Aggregation (ONPA) module 1208. The ONPA module 1208, as explained below, is configured to generate the stream view of the limit order data contained in limit order events.


The MP module 1204 is configured to parse the incoming stream of raw messages 1202 into a plurality of parsed messages having data fields that can be understood by downstream modules. Exemplary embodiments for such an MP module are described in the above-referenced and incorporated U.S. Patent Application Publication 2008/0243675. Thus, the MP modules is configured to process incoming raw messages 1202 to create limit order events that can be understood by downstream modules.


The SM module 1206 resolves a unique symbol identifier for the base financial instrument and the associated market center for a received event. Input events may contain a symbol field that uniquely identifies the base financial instrument. In this case, the symbol mapping stage performs a one-to-one translation from the input symbol field to the symbol identifier, which is preferably a minimally-sized binary tag that provides for efficient lookup of associated state information for the financial instrument. Thus, the SM module 1206 operates to map the known symbol for a financial instrument (or set of financial instruments) as defined in the parsed message to a symbology that is internal to the platform (e.g., mapping the symbol for IBM stock to an internal symbol “12345”). Preferably, the internal platform symbol identifier (ID) is an integer in the range 0 to N−1, where N is the number of entries in a symbol index memory. Also, the symbol ID may formatted as a binary value of size M=log2(N) bits. The format of financial instrument symbols in input exchange messages varies for different message feeds and financial instrument types. Typically, the symbol is a variable-length ASCII character string. A symbology ID is an internal control field that uniquely identifies the format of the symbol string in the message. As shown in FIG. 13, a symbology ID is preferably assigned by a feed handler and present in all incoming messages, as the symbol string format is typically shared by all messages on a given input feed.


An exemplary embodiment of the SM module 1206 maps each unique symbol character string to a unique binary number of size M bits. In such an exemplary embodiment, the symbol mapping FAM performs a format-specific compression of the symbol to generate a hash key of size K bits, where K is the size of the entries in a symbol index memory. The symbology ID may be used to lookup a Key Code that identifies the symbol compression technique that should be used for the input symbol. Preferably, the symbol mapping FAM compresses the symbol using format-specific compression engines and selects the correct compressed symbol output using the key code. Also, the key code can be concatenated with the compressed symbol to form the hash key. In doing so, each compression technique is allocated a subset of the range of possible hash keys. This ensures that hash keys will be unique, regardless of the compression technique used to compress the symbol. An example is shown in FIG. 13 wherein the ASCII symbol for a financial instrument is compressed in parallel by a plurality of different compression operations (e.g., alpha-numeric ticker compression, ISIN compression, and commodity compression). Compression techniques for different symbologies can be selected and/or devised on an ad hoc basis as desired by a practitioner. A practitioner is free to select a different compression operation as may be appropriate for a given symbology. Based on the value of the key code, the SM module will pass one of the concatenations of the key code and compression results as the output from the multiplexer for use as the hash key.


Alternatively, the format-specific compression engines may be implemented in a programmable processor. The key code may then be used to fetch a sequence of instructions that specify how the symbol should be compressed.


Once the hash key is generated, the SM module 1206 maps the hash key to a unique address in a symbol index memory in the range 0 to N−1. The symbol index memory may be implemented in a memory “on-chip” (e.g., within the reconfigurable logic device) or in “off-chip” high speed memory devices such as SRAM and SDRAM that are accessible to the reconfigurable logic device. Preferably, this mapping is performed by a hash function. A hash function attempts to minimize the number of probes, or table lookups, to find the input hash key. In many applications, additional meta-data is associated with the hash key. In an exemplary embodiment, the location of the hash key in the symbol index memory is used as the unique internal Symbol ID for the financial instrument.



FIG. 14 shows an exemplary embodiment of a hash function to perform this mapping that represents a novel combination of known hashing methods. The hash function of FIG. 14 uses near-perfect hashing to compute a primary hash function, then uses open-addressing to resolve collisions. The hash function H(x) is described as follows:

H(x)=(h1(x)+(i*h2(x)))mod N
h1(x)=A(x)⊕d(x)
d(x)=T(B(x))
h2(x)=C(x)


The operand x is the hash key generated by the previously described compression stage. The function h1(x) is the primary hash function. The value i is the iteration count. The iteration count i is initialized to zero and incremented for each hash probe that results in a collision. For the first hash probe, hash function H(x)=h1(x), thus the primary hash function determines the first hash probe. The preferred hash function disclosed herein attempts to maximize the probability that the hash key is located on the first hash probe. If the hash probe results in a collision, the hash key stored in the hash slot does not match hash key x, the iteration count is incremented and combined with the secondary hash function h2(x) to generate an offset from the first hash probe location. The modulo N operation ensures that the final result is within the range 0 to N−1, where N is the size of the symbol index memory. The secondary hash function h2(x) is designed so that its outputs are prime relative to N. The process of incrementing i and recomputing H(x) continues until the input hash key is located in the table or an empty table slot is encountered. This technique of resolving collisions is known as open-addressing.


The primary hash function, h1(x), is computed as follows. Compute hash function B(x) where the result is in the range 0 to Q−1. Use the result of the B(x) function to lookup a displacement vector d(x) in table T containing Q displacement vectors. Preferably the size of the displacement vector d(x) in bits is equal to M. Compute hash function A(x) where the result is M bits in size. Compute the bitwise exclusive OR, ⊕, of A(x) and d(x). This is one example of near-perfect hashing where the displacement vector is used to resolve collisions among the set of hash keys that are known prior to the beginning of the query stream. Typically, this fits well with streaming financial data where the majority of the symbols for the instruments trading in a given day is known. Methods for computing displacement table entries are known in the art.


The secondary hash function, h2(x), is computed by computing a single hash function C(x) where the result is always prime relative to N. Hash functions A(x), B(x), and C(x) may be selected from the body of known hash functions with favorable randomization properties. Preferably, hash functions A(x), B(x), and C(x) are efficiently implemented in hardware. The set of H3 hash functions are good candidates. (See Krishnamurthy et al., “Biosequence Similarity Search on the Mercury System”, Proc. of the IEEE 15th Int'l Conf. on Application-Specific Systems, Architectures and Processors, September 2004, pp. 365-375, the entire disclosure of which is incorporated herein by reference).


Once the hash function H(x) produces an address whose entry is equal to the input hash key, the address is passed on as the new Symbol ID to be used internally by the ticker plant to reference the financial instrument. As shown in FIG. 14, the result of the hash key compare function may be used as a valid signal for the symbol ID output.


Hash keys are inserted in the table when an exchange message contains a symbol that was unknown at system initialization. Hash keys are removed from the table when a financial instrument is no longer traded. Alternatively, the symbol for the financial instrument may be removed from the set of known symbols and the hash table may be cleared, recomputed, and initialized. By doing so, the displacement table used for the near-perfect hash function of the primary hash may be optimized. Typically, financial markets have established trading hours that allow for after-hours or overnight processing. The general procedures for inserting and deleting hash keys from a hash table where open-addressing is used to resolve collisions is well-known in the art.


In an exemplary embodiment, the SM module 1210 can also be configured to compute a global exchange identifier (GEID) that maps the exchange code and country code fields in the exchange message to an integer in the range 0 to G−1, as shown in FIG. 15. Similar to the symbol field for financial instruments, the exchange code and country code fields uniquely identify the source of the exchange message. Thus, the global exchange identifier (GEID) preferably comprises a binary tag that uniquely identifies a particular exchange for this the message is relevant. The value of G should be selected such that it is larger than the total number of sources (financial exchanges) that will be generating input messages for a given instance of the system. Hashing could be used to map the country codes and exchange codes to the GEID. Alternatively, a “direct addressing” approach can be used to map country and exchange codes to GEIDs. For example, the exchange code and country codes can each be represented by two character codes, where the characters are 8-bit upper-case ASCII alpha characters. These codes can then be truncated to 5-bit characters in embodiment where only 26 unique values of these codes are needed. For each code, these truncated values are concatenated to generate a 10-bit address that is used to lookup a compressed intermediate value in a stage 1 table. Then the compressed intermediate values for the exchange and country code can be concatenated to generate an address for a stage 2 lookup. The result of the stage 2 lookup is the GEID. The size of the intermediate values and the stage 2 address will depend on the number of unique countries and the max number of exchanges in any one country, which can be adjusted as new exchanges open in different countries.


The ONPA module 1208 then receives a stream of incoming limit order events 1600, as shown in FIG. 16. The ONPA module accesses a memory 1602 that stores data structures which comprise the various order books tracked by the system to determine (1) how to update the data structures in view of the received limit order event, and (2) how to enrich the limit order event in view of its relation to the tracked order books. The output from the ONPA module 1208 is an outgoing stream of enriched limit order events 1604.



FIG. 19 depicts an exemplary limit order event 1600. As exemplary data fields, the limit order event comprises a symbol field 1902 and GEID field 1904 (as mapped by the SM module 1206). The event may also include a reference number field 1906, typically assigned by the publisher of the event to identify a particular limit order. Additional data fields for the event 1600 comprise a flag 1908 to identify whether the limit order pertains to a bid or ask, a field 1910 to identify a price for the limit order, a field 1912 to identify the size for the limit order (e.g., share count), and a timestamp field 1914. Furthermore, the event 1600 preferably comprises one or more flags 1916 that identify whether one or more attributes are applicable to the event. For example, as discussed above, the value of the attribute flag field 1916 can identify whether the limit order pertains to an ephemeral regional order (e.g., a FLASH order on NASDAQ or a BOLT order on BATS) and whether the limit order is implied. Lastly, the limit order event 1600 may comprise one or more flags 1918 for identifying whether the limit order event is an add event, modify event or delete event. Thus the add, modify, delete (AMD) flags field 1918 will enable an ONPA module 1208 to decide whether a received limit order event represents a new limit order (the add flag), a modification to a pre-existing limit order (the modify flag), or a deletion of an pre-existing limit order (the delete flag).


It should be understood that many limit order events 1600 will not have the same fields shown in the example of FIG. 19, whether there be differences in the number of fields and/or the types of fields. For example, many limit order events will have fields that vary based on the value of the AMD flag field 1918. As another example, some limit order events 1600 will not include a symbol field 1902. In such an instance, the symbol mapping task would be performed by the ONPA module 1208 rather than the SM module 1206. As an example of such a circumstance, in order to conserve bandwidth on data transmission links, several market centers minimize the size of event messages by sending “static” information regarding outstanding limit orders only once. Typically, the first message advertising the addition of a new limit order includes all associated information (such as the symbol, source identifier, price and size of the order, etc.) as well as a reference number for the order. Subsequent messages reporting modification or deletion of the order may only include a reference number that uniquely identifies the order (thus omitting a symbol field from the event). In a preferred embodiment, one of the roles for the ONPA module 1208 is to pass complete order information to downstream consumers of the event messages. For limit order event messages, the ONPA module 1208 normalizes the messages by ensuring that all desired fields are present in a regular format and the message consistently communicates the market event. For example, a market center may choose to advertise all events using order modification events. In such a scenario, with a preferred embodiment, it is the responsibility of the ONPA module 1208 to determine if the event results in the addition of a new order, modification of an existing order, or deletion of an existing order. In practice, market centers have disparate data models for communicating limit order events; the ONPA module 1208 ensures that a consistent data model is presented to downstream processing blocks and trading applications. All output events contain a consistent set of fields where each field has a well-defined type. For example, some exchanges may only send the order reference number in a delete event. The ONPA module 1208 fills in missing fields such as the price and size of the deleted order.


In order to resolve the symbol identifier for input events lacking a symbol field, the ONPA module 1208 can use another identifying field (such as an order reference number). In this case, the ONPA module 1208 performs a many-to-one translation to resolve the symbol identifier, as there may be many outstanding orders to buy or sell a given financial instrument. It is important to note that this many-to-one mapping requires maintaining a dynamic set of items that map to a given symbol identifier, items may be added, modified, or removed from the set at any time as orders enter and execute at market centers.


While there are several viable approaches to solve the order normalization problem, the preferred method is to maintain a record for every outstanding limit order advertised by the set of input market data feeds. An example of such a limit order record 2000 is shown in FIG. 20. An event reporting the creation of a new limit order must contain a symbol field, thus when the event arrives at the ONPA module 1208 it will contain the symbol identifier resolved by the symbol mapping stage. If the event is from a market center that uses order reference numbers to minimize subsequent message sizes, the event will also contain an order reference number. The ONPA module 1208 maintains a map to the order record, where the record 2000 may contain the symbol identifier, order reference number, price, size, and other fields provided by the market center. Preferably, the ONPA module 1208 also assigns a unique internal order identifier 2002 that may be used to directly address the order record by other system components.


In the example of FIG. 20, the limit order record 2000 comprises a plurality of fields, such as:

    • A unique internal identifier field 2002 as noted herein.
    • A symbol field 2004 as noted herein.
    • A GEID field 2006 as noted herein.
    • A reference number field 2008 as noted herein.
    • A bid/ask flag field 2010 as noted herein.
    • A price field 2012 as noted herein.
    • A size field 2014 as noted herein. As explained below, the value of this field in the record 2000 may be updated over time as order modify events are received.
    • A timestamp field 2016 as noted herein
    • A flag field 2018 for a first attribute (A0) (e.g., to identify whether the order pertains to an ephemeral regional order)
    • A flag field for a second attribute (A1) (e.g., to identify whether the order is an implied order. Together the A0 and A1 flags can be characterized as an order attribute vector 2030 within the limit order record 2000.
    • An interest vector field 2022 that serves to identify downstream subscribers that have an interest in the subject limit order. Optionally, this vector can be configured to not only identify which subscribers are interested in which limit orders but also what fields in each limit order record each subscriber has an interest in.


Once again, however, it should be noted that limit order records 2000 can be configured to have more or fewer and/or different data fields.


Preferably, the mapping of a received limit order event 1600 to a limit order record 2000 is performed using hashing in order to achieve constant time access performance on average. The hash key may be constructed from the order reference number, symbol identifier, or other uniquely identifying fields. The type of hash key is determined by the type of market center data feed. Upstream feed handlers that perform pre-normalization of the events set flags in the event that notify the ONPA module as to what type of protocol the exchange uses and what fields are available for constructing unique hash keys. For example, this information may be encoded in the GEID field 2006 or in some other field of the limit order event 1600. There are a variety of hash functions that could be used by the ONPA module. In the preferred embodiment, the ONPA employs H3 hash functions as discussed above and in the above-referenced and incorporated U.S. Patent Application Publication 2008/0243675 due to their efficiency and amenability to parallel hardware implementation. Hash collisions may be resolved in a number of ways. In the preferred embodiment, collisions are resolved via chaining, creating a linked list of entries that map to the same hash slot. A linked list is a simple data structure that allows memory to be dynamically allocated as the number of entries in the set changes.


Once the record is located, the ONPA module updates fields in the record and copies fields from the record to the message, filling in missing fields as necessary to normalize the output message. It is during this step that the ONPA module may modify the type of the message to be consistent with the result of the market event. For example, if the input message is a modify event that specifies that 100 shares should be subtracted from the order (due to a partial execution at the market center) and the outstanding order is for 100 shares, then the ONPA will change the type of the message to a delete event, subject to market center rules. Note that market centers may dictate rules such as whether or not zero size orders may remain on an order book. In another scenario, if the outstanding order was for 150 shares, the ONPA module would update the size field 2014 of the limit order record to replace the 150 value with 50 reflect the removal of 100 shares from the order. In general, the ONPA module attempts to present the most descriptive and consistent view of the market data events. Hardware logic within the ONPA module can be configured to provide these updating and normalization tasks.


In addition to normalizing order messages, the ONPA module may additionally perform price aggregation in order to support price aggregated views of the order book. Preferably, the ONPA module maintains an independent set of price point records. In this data structure, a record is maintained for each unique price point in an order book. At minimum, the set of price point records preferably contain the price, volume (sum of order sizes at that price, which can be referred to as the price volume), and order count (total number of orders at that price). Order add events increase the volume and order count fields, order delete events decrease the volume and order count fields, etc. Price point records are created when an order event adds a new price point to the book. Likewise, price point records are deleted when on order event removes the only record with a given price point from the book. Preferably, the ONPA module updates AMD flags in an enriched limit order event 1604 that specify if the event resulted in the addition, modification, or deletion of a price entry in the book (see the price AMD field 2226 in FIG. 22). This information may be used to optimize downstream sorting engines. Preferably, the ONPA module also assigns a unique internal price identifier to each price record that may be used to directly address the price record by other system components.


Note that mapping a limit order event 1600 to a price point record is also a many-to-one mapping problem. Preferably, the set of price point records is maintained using a hash mapping, similar to the order records. In order to locate the price point record associated with an order event, the hash key is constructed from fields such as the symbol identifier, global exchange identifier, and price. Preferably, hash collisions are resolved using chaining as with the order record data structure. Other data structures may be suitable for maintaining the sets of order and price point records, but hash maps have the favorable property of constant time accesses (on average).


In order to support efficient attribute filtering and price level merging in downstream book views, the ONPA module preferably maintains a price attribute vector as part of the price point records, wherein the price attribute vectors also comprise a vector of volumes and price counts in each price point record. For example, the price point record may include the following fields: price, volume (total shares or lots at this price), order count (total orders at this price), attribute flags, attribute volume 0 (total shares or lots at this price with attribute 0), order count 0 (total orders at this price with attribute 0), attribute volume 1, attribute order count 1, etc. Examples of such price point records are shown in FIGS. 21(a) and (b). In general, the number of unique attributes for a given financial instrument is expected to be small. Preferably, the ONPA module is configurable to allow the number of possible attributes to be defined dynamically for a given financial instrument.


The ONPA module may append the volume, order count, and price attribute to events when creating enriched limit order events 1604. Preferably, the ONPA module maintains price interest vectors that specify if any downstream applications or components require the price aggregated and/or attribute information. Furthermore, the ONPA module preferably updates flags in the event that specify if the event resulted in the addition, modification, or deletion of a price entry in the book as defined by attribute (see the price AMD field 2226 in FIG. 22).


The data structure used to store price point records preferably separately maintains regional price point records 2100 and composite price point records 2150 for limit orders. A regional price point record 2100 stores price point information for limit orders pertaining to a financial instrument traded on a specific regional exchange. A composite price point record 2150 stores price point information for limits order pertaining to a financial instrument traded across multiple exchanges. An example of a regional price point record is shown in FIG. 21(a), and an example of a composite price point record is shown in FIG. 21(b).


An exemplary regional price point record 2100 comprises a plurality of fields, such as:

    • A unique internal (UI) regional price identifier field 2102 for providing an internal identifier with respect to the subject price point record.
    • A symbol field 2104 as noted herein.
    • A GEID field 2106 as noted herein.
    • A bid/ask flag field 2108 as noted herein.
    • A price field 2110 as noted herein.
    • A volume field 2112, which identifies the volume of shares across all limit orders for the financial instrument (see the symbol field 2104) on the regional exchange (see the GEID field 2106) at the price identified in the price field 2110.
    • A count field 2114, which comprises a count of how many limit orders make up the volume 2112.
    • A timestamp field 2116 as noted above (which preferably is representative of the timestamp 1914 for the most recent limit order event 1600 that caused an update to the subject price point record.
    • A regional price attribute vector 2140, which as noted above, preferably not only flags whether any attributes are applicable to at least a portion of the volume making up the subject price point record, but also provides a breakdown of a volume and count for each attribute. For example, a flag 2118 to indicate whether attribute A0 is applicable, together with a volume 2120 and count 2122 for attribute A0, and a flag 2124 to indicate whether attribute A1 is applicable, together with a volume 2126 and count 2128 for attribute A1.
    • An interest vector field 2130 that serves to identify downstream subscribers that have an interest in the subject price point record. Optionally, this vector can be configured to not only identify which subscribers are interested in the regional price point record but also what fields in each regional price point record should each subscriber has an interest in.


An exemplary composite price point record 2150 comprises a plurality of fields, such as:

    • A unique internal (UI) composite price identifier field 2152 for providing an internal identifier with respect to the subject price point record.
    • A symbol field 2154 as noted herein.
    • A bid/ask flag field 2156 as noted herein.
    • A price field 2158 as noted herein.
    • A volume field 2160, which essentially comprises the sum of the volume fields 2112 for all regional price point records which are aggregated together in the composite price point record.
    • A count field 2162, which essentially comprises the sum of the count fields 2114 for all regional price point records which are aggregated together in the composite price point record.
    • A timestamp field 2164 as noted above (which preferably is representative of the timestamp 1914 for the most recent limit order event 1600 that caused an update to the subject price point record.
    • A composite price attribute vector 2180, which as noted above, preferably not only flags whether any attributes are applicable to at least a portion of the volume making up the subject price point record, but also provides a breakdown of a volume and count for each attribute. For example, a flag 2166 to indicate whether attribute A0 is applicable, together with a volume 2168 and count 2170 for attribute A0, and a flag 2172 to indicate whether attribute A1 is applicable, together with a volume 2174 and count 2176 for attribute A1.
    • An interest vector field 2178 that serves to identify downstream subscribers that have an interest in the subject price point record. Optionally, this vector can be configured to not only identify which subscribers are interested in the composite price point record but also what fields in each composite price point record should each subscriber has an interest in.


Once again, however, it should be noted that regional and composite price point records 2100 and 2150 can be configured to have more or fewer and/or different data fields.


In the preferred embodiment, parallel engines update and maintain the order and price aggregation data structures in parallel. In one embodiment, the data structures are maintained in the same physical memory. In this case, the one or more order engines and one or more price engines interleave their accesses to memory, masking the memory access latency of the memory technology and maximizing throughput of the system. There are a variety of well-known techniques for memory interleaving. In one embodiment, an engine controller block utilizes a time-slot approach where each engine is granted access to memory at regular intervals. In another embodiment, a memory arbitration block schedules outstanding memory requests on the shared interface and notifies engines when their requests are fulfilled. Preferably, the memory technology is a high-speed dynamic memory such as DDR3 SDRAM. In another embodiment, the order and price data structures are maintained in separate physical memories. As in the single memory architecture, multiple engines may interleave their accesses to memory in order to mask memory access latency and maximize throughput.



FIG. 17 shows an example of how a single shared memory may be partitioned to support order normalization, regional price aggregation, and composite price aggregation. In this example, each hash table is allocated a portion of the memory space. Hash collisions are resolved by chaining, creating a linked list of entries that map to the same hash slot. Linked list entries for all three hash tables are dynamically allocated from a common memory space. FIG. 18 shows an example of how the ONPA module data structures may be partitioned across multiple physical memories. In this particular example, the normalization data structure is stored in one physical memory, while the regional and composite price aggregation data structures are stored in a second physical memory. This architecture allows memory accesses to be performed in parallel.


The ONPA module 1208, upon receipt of a limit order event 1600, thus (1) processes data in the limit order event 1600 to access memory 1602 (which may be multiple physical memories) and retrieve a limit order record 2000, regional price point record 2100 and composite price point record 2150 as appropriate, (2) processes data in the limit order event and retrieved records to update the records as appropriate, and (3) enriches the limit order event 1600 with new information to create an enriched limit order event 1604. An example of such an enriched limit order event 1604 is shown in FIG. 22. Preferably, the ONPA module appends a number of fields onto the limit order event 1600 that provide downstream subscribers with valuable information about market depth. In the example of FIG. 22, the enriched limit order event 1604 comprises fields such as:

    • A field 2202 for the unique internal (UI) identifiers (such as UI ID 2002, UI regional price ID 2102 and UI composite price ID 2152).
    • A symbol field 2204 as noted herein.
    • A GEID field 2206 as noted herein.
    • A reference number field 2208 as noted herein.
    • A bid/ask flag field 2210 as noted herein.
    • A price field 2212 as noted herein.
    • A size field 2214 as noted herein.
    • A timestamp field 2216 as noted herein.
    • A regional volume field 2218 and a regional count field 2220. It should be noted that the ONPA module 1208 is preferably configured to append these fields onto the limit order event based on the updated volume and count values for the retrieved regional price point record pertinent to that limit order event.
    • A composite volume field 2222 and a composite count field 2224. It should be noted that the ONPA module 1208 is preferably configured to append these fields onto the limit order event based on the updated volume and count values for the retrieved composite price point record pertinent to that limit order event.
    • A field 2226 for identifying whether the limit order pertains to an add, modify or delete (AMD) (a field that the ONPA module may update based on the content of the limit order event relative to its pertinent limit order record).
    • A price AMD field 2228 for identifying whether the limit order event caused an addition, modification, or deletion from a regional price point record and/or a composite price point record. The ONPA module can append this field onto the limit order event based on how the regional and composite price point records were processed in response to the content of the received limit order record 1600.
    • An enriched attribute vector 2250, which the ONPA module can append to the limit order event as a consolidation of the updated attribute vectors 2140 and 2180 for the regional and composite price point records that are pertinent to the limit order event. This enriched attribute vector can also include the order attribute vector 2030 from the pertinent limit order record. Thus, the enriched attributed vector 2250 may comprise an order attribute vector field 2230, a regional price attribute vector field 2232 and a composite price attribute vector field 2234.
    • An interest vector field 2236 that serves to identify downstream subscribers that have an interest in data found in the enriched limit order event. The ONPA module can append the interest vector as a consolidation of the interest vectors 2022, 2130 and 2178 for the limit order, regional price point and composite price point records that are pertinent to the limit order event.


Once again, however, it should be noted the ONPA module can be configured to enrich limit order events with more and fewer and/or different data fields.


The outgoing enriched limit order events 1604 thus serve as the stream view of processed limit order data 1210 that can be produced by the ONPA module 1208 at extremely low latency.


A block diagram of an exemplary embodiment for the ONPA module 1208 is shown in FIG. 23.


The Extractor module is responsible for extracting from an input market event the fields needed by the rest of the modules within the ONPA module and presenting those fields in parallel to downstream modules. The Extractor also forwards the event to the Blender module for message reconstruction.


The Price Normalizer module converts variably typed prices into normalized fixed-sized prices (e.g. 64-bit values). In the preferred embodiment, the new prices are in units of either billionths or 256ths. The power of 2 price conversions may be performed by simple shifts. The power of 10 price conversions take place in a pipeline of shifts and adds.


In the preferred embodiment, the Hash modules are responsible for doing the following:

    • Order Hash Module—Hashing the symbol, exchange identifier, and order reference number to create an address offset into the static order region of the memory which contains the first entry in the linked list that contains or will contain the desired order.
    • Price Hash Module—Hashing the symbol, exchange identifier, and price to create an address offset into the static price region of the memory (for both regional price records and composite price records) which contains the first entry in the linked list that contains or will contain the desired price level.
    • Header Hash Module—Hashing the symbol and the exchange identifier to create an address offset into the static header region that contains pointers to the first entries in both the order refresh list and the price refresh list.


With respect to such refresh lists, the inventors note that a refresh event can be used to initialize the book view provided to client applications. Thus, one of the responsibilities of the Level 2 processing pipeline can be to generate book snapshots for client application initialization. At subscription time, a refresh event provides a snapshot of the book at a particular instant in time. It is at this point that the appropriate bits in the interest vector are set in the appropriate data structures. Following the refresh event, incremental update events are delivered to the client application in order to update the client application's view of the book. Refresh events may be processed in-line with incremental update events in the FAM pipeline. In order to minimize the overhead of generating the book snapshot, refresh events may be processed asynchronously. So long as the snapshot of the book is an atomic event that records the event sequence number of the most recent update, the snapshot need not be processed synchronous to all incremental update traffic. Synchronizing buffers in the client API may be used to buffer incremental updates received prior to receipt of the refresh event. When the refresh event is received, incremental updates in the synchronization buffer are processed. Updates with sequence numbers less than or equal to the sequence number noted in the refresh event are discarded.


The Order Engine module is responsible for the following:

    • Traversing the hash linked list for limit order records.
    • Traversing the refresh linked lists for refreshes.
    • Performing adds, modifies, and deletes on the limit order records.
    • Performing necessary maintenance to the linked lists.


The Price Engine module is responsible for the following:

    • Traversing the hash linked list for price levels in the regional and composite price point records.
    • Traversing the refresh linked lists for refreshes.
    • Performing aggregation tasks for order adds, modifies, and deletes at a given price level for the regional and composite price point records.
    • Adding and deleting price levels as appropriate from the data structure.
    • Performing necessary maintenance to the linked lists.


The Cache module optimizes performance by maintaining the most recently accessed records in a fast, on-chip memory. The Cache module is responsible for the following:

    • Storing limit order records, price point records, and header nodes for fast and non-stale access.
    • Keeping track of which SDRAM addresses are cached.
    • Fetching un-cached data from SDRAM when requested by an order or price engine.
    • Providing the local on-chip memory address for direct access when an SDRAM address is queried.


The Operation FIFO module is responsible for the following:

    • Storing operation data during pipeline hibernation.
    • Monitoring next pointer information from the order and price engines during deletes for automatic address forwarding.


The Refresh Queue module is configured to store refreshes that are received while another refresh is currently being processed. The Blender module may only able to check one order refresh and one price refresh at a time, which limits the number of concurrent refreshes. The SDRAM arbiter module arbitrates accesses from the order and price engines to the two SDRAM interfaces. The Blender module also constructs an outgoing enriched and normalized event using the original event and various fields created by the order and price engines (see FIG. 22). It also coalesces micro-events from refreshes to create outgoing requested refresh images. The Blender module is also preferably configured to normalize conflicts between the refresh and the event stream.


If desired by a practitioner, the ONPA module's stream view output 1210, comprising the enriched limit order events, may be transmitted directly to clients with appropriate stream view subscriptions. An Interest and Entitlement Filtering (IEF) module 1210 can be located downstream from the ONPA module as shown in FIGS. 12 (b) and (c). The IEF module 1210 can be configured to utilize the mapped symbol index to resolve the set of interested and entitled clients for the given enriched limit order event, as described in the above-referenced and incorporated U.S. Patent Application Publication 2008/0243675. Also, for enriched limit order events 1604 that include an interest vector field 2234, the IEF module 1210 can also utilize such an interest vector to identify interested and entitled clients. The filtering aspect of the IEF module may be extended to filter specific fields from enriched limit order events 1604 based on the type of book view specified by client applications. For example, a sorted order view does not require price aggregated volume and order count fields, and the IEF module 1210 can be configured to remove those fields from enriched limit order events for clients who want a sorted order view. Event transmission latency can be reduced and downstream network bandwidth can be conserved by omitting these fields from update events.


As shown in FIG. 12(c), a message formatting (MF) module 1212 can be deployed downstream from the IEF module 1210 to format outgoing enriched limit order events destined for interested and entitled subscribers to a message format expected by those subscribers. An exemplary embodiment for the MF module 1212 is described in the above-referenced and incorporated U.S. Patent Application Publication 2008/0243675.


As noted above, some clients may prefer to receive a stream view comprising enriched limit order events because they will build their own sorted data structures of the order book data. Thus, in one embodiment of the invention, the output of the pipeline 1200 shown in FIGS. 12(a)-(c) can be transmitted to a consuming system (client machine) where it is processed by an Application Programming Interface (API) associated with the ticker plant, as shown in connection with FIG. 24. The API performs the sorting and presents the client application with a summary view of the data. The sorting task may be performed using a variety of techniques known in the art.


However, other clients may prefer to receive the summary view from the ticker plant itself. For additional embodiments of the invention, the inventors disclose sorting techniques that can be deployed in pipeline 1200 implemented within the ticker plant coprocessor to create summary views of the order data. However, it should be noted that these sorting techniques could also be performed in an API as shown in FIG. 24 if desired by a practitioner.



FIGS. 25(a)-(d) depict exemplary embodiments of the pipeline 1200 where a Sorted View Update (SVU) module 2500 is included to create a summary view 2502 of the order books from the enriched limit order events. With the examples of FIGS. 25(b)-(d), it should be noted that the pipeline 1200 is configured to provide both stream views 1210 to interest clients and summary views 2402 to interest clients.


While the SVU module 2500 can be configured to provide sorting functionality via any of a number of techniques, with a preferred embodiment, the SVU module employs sorting engines to independently maintain each side (bid and ask) of each order book. Since input order and price events only affect one side of the order book, accessing each side of the book independently reduces the potential number of entries that must be accessed and provides for parallel access to the sides of the book. Each side of the book is maintained in sorted order in physical memory. The book is “anchored” at the bottom of the memory allocation for the book, i.e. the last entry is preferably always stored at the last address in the memory allocation. As a consequence the location of the “top” of the book (the first entry) varies as the composition of the order book changes. In order to locate the top of the book, the SVU module 2500 maintains a record that contains pointers to the top of the bid and ask side of the book, as well as other meta-data that may describe the book. The record may be located directly by using the symbol map index. Note that inserting an entry into the book moves entries above the insertion location up one memory location. Deleting an entry in the book moves entries above the insertion location down one memory location. While these operations may result in large numbers of memory copies, performance is typically good as the vast majority of order book transactions affect the top positions in the order book. Since the price AMD field 2226 in the enriched limit order event 1604 specifies whether or not a price entry has been inserted or deleted, the sorting engine within the SVU module 2500 can make use of this information to make a single pass through the sorted memory array. Furthermore, since the price aggregation engine within the ONPA module maintains all volume, order count, and attribute information for each price entry in the price point records, the entries in the SVU data structure only need to store the values required for sorting.


For regional order summary views, the SVU module preferably maintains a pair of bid and ask books for each symbol on each exchange upon which it trades. The entries in each book are from one exchange. Since the order engine within the ONPA module maintains all information associated with an order, the SVU data structure only needs to maintain the fields necessary for sorting and the unique order identifier assigned by the ONPA module. In some cases, only the price and timestamp are required for sorting.


For price summary views, the SVU module preferably maintains a pair of spliced bid and ask books for each symbol. The entries in each book are from every exchange upon which the symbol trades. Since the price aggregation engine within the ONPA module maintains all information associated with a price level, the SVU data structure only needs to maintain the fields necessary for sorting (i.e. price) and the unique price identifier assigned by the ONPA module. Composite, spliced, and regional views of the price book may be synthesized from this single spliced book. Attribute filtered and price merged views of the price book may be synthesized in the same way. A price book sorting engine in the SVU module computes the desired views by aggregating multiple regional entries to produce composite entries and positions, and filters unwanted regional price entries to produce regional entries and positions. These computations are performed as the content of each book is streamed from memory through the engine. In order to minimize the memory bandwidth consumed for each update event, the engine requests chunks of memory that are typically smaller in size than the entire order book. Typically, a default memory chunk size is specified at system configuration time. Engines request the default chunk size in order to fetch the top of the book. If additional book entries must be accessed, the engines request the next memory chunk, picking up at the next address relative to the end of the previous chunk. In order to mask the latency of reading chunks of memory, processing, and requesting the next chunk of memory, multiple engines interleave their accesses to memory. As within the ONPA module, interleaving is accomplished by using a memory arbitration block that schedules memory transactions for multiple engines. Note that a time-slot memory controller may also be used. Engines may operate on unique symbols in parallel without affecting the correctness of the data.


In another embodiment of the Sorted View Update module, each side of the book is organized as a hierarchical multi-way tree. The depth of a multi-way tree is dictated by the number of child branches leaving each node. A B+ tree is an example of a multi-way tree where all entries in the tree are stored at the same level, i.e. the leaf level. Typically, the height of a multi-way tree is minimized in order to minimize the number of nodes that must be visited in order to reach a leaf node in the tree, i.e. the desired entry. An example of a B+ tree is shown in FIG. 26. Note that the leaf nodes may be organized as a linked list such that the entire contents of the tree may be accessed by navigating to the leftmost child and following the pointers to the next child node. This feature can be exploited in the SVU module to quickly produce a refresh event, a snapshot of the entire contents of the order book in order to initialize newly subscribed clients. Furthermore, a direct pointer to the leftmost child may be stored along with the root node in order to provide fast access to the top of the sorted book.



FIG. 26 shows a simple example of a set of positive integers sorted using a B+ tree. Note that all of the stored integers are stored in the leaf nodes. As is well-known in the art, B+ tree nodes vary in size between a minimum and maximum size threshold that is dictated by the branching factor of the tree. Note that “next pointers” are stored in each leaf node to create a linked list of leaf nodes that may be quickly traversed to retrieve the entire sorted order. Internal tree nodes contain integers that dictate ranges stored the child nodes. For example, all integers less than or equal to 7 are stored in the left sub-tree of the root node.


The SVU module can be configured to utilize hierarchical B+ trees to minimize the number of memory accesses required to update the various book views. As shown in FIG. 27, the SVU module can be configured to maintain a header record that contains pointers to the root of the composite and regional book trees. The root tree is the price aggregated tree. Each leaf of the price aggregated tree contains a pointer to a B+ tree that maintains a sort of the orders at the given price point. With such an embodiment, the SVU module preferably maintains a hierarchical B+ tree for the composite views and each regional view of the book. Note that each side of the book (bid and ask) corresponds to a hierarchical B+ tree. Note that hierarchical B+ trees may also be used in the embodiment where sorting is performed by the API on the client machine. Furthermore, note that a portion of the sorting may be offloaded to the API on the client machine. For example, construction of the spliced view of the book may be offloaded to the API by subscribing to the summary view of all regional books.


Similar to the insertion sorting engines in the previous embodiment, a parallel set of tree traversal engines can operate in parallel and interleave their accesses to memory. Furthermore, the SVU module may optionally cache recently accessed tree nodes in on-chip memory in order to further reduce memory read latency.



FIG. 28 depicts an exemplary embodiment for the SVU module 2500. Like the ONPA module, the SVU module 2500 of FIG. 28 utilizes a functional pipeline to achieve parallelism among the dispatching and sorting engines. It also uses data parallelism to instantiate multiple sorting engines in order to make full use of the available memory bandwidth to maximize message throughput performance.


The Extractor module provides the same service of extracting necessary fields for processing as described in connection with the ONPA module, and the Extractor module further propagates those fields to the Dispatcher. The Extractor module also preferably propagates the full event to the Blender module.


The Dispatcher module is responsible for fetching the header record that contains pointers to the composite and regional book trees. A cache positioned between the Dispatcher module and the SDRAM Arbiter module provides quick access to recently accessed header records. The operation FIFO module stores event fields and operational state while the Dispatcher module is waiting for memory operations to complete. This allows the Dispatcher module to operate on multiple events in parallel.


When the book pointers have been received from memory, the Dispatcher module passes the event fields and the book pointers to one of several parallel sorting engine modules. All events for a given symbol are preferably processed by the same sorting engine module, but events for different symbols may be processed in parallel by different sorting engine modules. The Dispatcher module may balance the workload across the parallel sorting engines using a variety of techniques well-known in the art. For example, the inventors have found that a random distribution of symbols across sorting engines provides an approximately even load balance on average. Note that a Dispatch Buffer module resides between the Dispatcher module and the sorting engines. This buffer maintains separate queues of pending events for each sorting engine. It reduces the probability of head-of-line blocking when a single sorting engine is backlogged. Pending events for that engine are buffered, while events scheduled for other sorting engines may be processed when the associated sorting engine is ready. The sorting engine may utilize the modified insertion sort or B+ tree sorting data structures described above. In the preferred embodiment, the B+ tree sorting data structure is used. The sorting engine is responsible for:

    • Inserting and removing price levels from the sorted data structure
    • Inserting and removing orders from sorted price levels in the sorted listing
    • Identifying the relative position of the price level
    • Identifying the relative position of the order


The sorting engines include the relative price and order position in outgoing events. For example, the sorting engines can be configured to append data fields onto the order events it processes that identify the sort position for the order and/or price within the various books maintained by the data structure. Thus, the SVU module can create the summary views by further enriching the limit order events that it receives with sort position information for one or more books maintained by the pipeline. This positional information may be in the form of a scalar position (e.g. 3rd price level) or a pointer to the previous entry (e.g. pointer to the previous price level). Each sorting engine has an associated cache, operation FIFO, and refresh FIFO. The cache provides fast access to recently accessed memory blocks, which optimizes latency performance for back-to-back operations to the top of the same order book. Each sorting engine may operate on multiple events in parallel by storing in-process fields and state in its operation FIFO. Note that the sorting engines ensure correctness and data structure coherency by monitoring for accesses to the same data structure nodes. Similarly, the sorting engines incrementally process refresh events to service new client subscriptions by employing the refresh queue, similar to the ONPA module.


The output of each sorting engine is passed to the Blender module which constructs the normalized output event by blending the positional information from the sorting engine with the event fields passed by the Extractor module. Note that the Blender maintains a queue for each sorting engine that stores the pending event fields.


Level 2 updates from the ticker plant may be delivered to client applications in the form of discrete events that update the state of one or more book views presented by the client API. For summary views, the API preferably maintains a mirror data structure to the SVU module. For example, if the SVU module employs B+ trees the API preferably maintains a B+ tree. This allows the SVU module to include parent/child pointers in the update event. Specifically, the SVU module assigns a locally unique identifier to each node in the B+ tree for the given book view. The SVU module enriches the update events with these identifiers to specify the maintenance operations on affected nodes in the data structure. For example, the update event may specify that the given price identifier be added to a given node. This allows the API to perform constant time updates to its data structure using direct addressing and prevents the need for a tree search operation.


Level 2 updates from the ticker plant may also be delivered to client applications in the form of snapshots of the top N levels of a book view, where N is typically on the order of 10 or less. N may be specified by the ticker plant or the subscribing application. In the case that the book view is natively maintained by the SVU module, the snapshot is readily produced by simply reading the first N entries in the sorted data structure. When B+ trees are used, nodes may be sized such that snapshots may be produced in a single memory read. In the case that the SVU module synthesizes the book view (such as composite or attribute filtered views), the SVU preferably reads a sufficient number of entries from the underlying sorted view to produce N entries in the synthesized view. Snapshot delivery from the ticker plant significantly reduces the amount of processing required on client systems by the API to produce the specified book views. The API simply copies the snapshot into a memory array and presents the view to the client application.


In accordance with another embodiment, the pipeline 1200 can leverage its likely ability to generate quote messages before that quote appears in a Level 1 feed published by an exchange. As shown in FIG. 29, the pipeline 1200 can be augmented to pass synthetic quote events from the SVU module to a Value Cache Update (VCU) module 2900 that handles Level 1 traffic. With the embodiment of FIG. 29, the SVU module is configured with logic for detecting when an order or price book update event modifies the top of the book (best bid or offer) (e.g., new price, size, or timestamp). In response to detecting such an event, the SVU module generates a synthetic quote event that contains the new bid or ask price and aggregate order size at that price for the given regional and/or composite view of the book for the associated symbol. The VCU module 2900 receives this synthetic quote event, updates a Last Value Cache (LVC) data structure that contains the associated Level 1 record for the symbol and transmits a Level 1 quote message to interested and entitled subscribers. Records in the LVC data structure represent a current state of a financial instrument in the market place. As indicated above, because of the extreme low latency provided by the ONPA and SVU modules, it is expected that the SVU module will be able to recognize when order events affect the top of a book before that situation is reflected in a conventional level 1 feed from an exchange. Thus, by recognizing and delivering synthetic quote events to the VCU module 2900, the inventors believe that the pipeline 1200 of FIG. 29 is configured to generate level 1 events from a feed of level 2 events with favorably low latency.


It should be noted that an exemplary embodiment for the VCU module 2900 is described in the above-referenced and incorporated U.S. Patent Application Publication 2008/0243675 (e.g., see FIGS. 15(a) and (b) therein).


In yet another exemplary embodiment, the pipeline 1200 of FIG. 29 can be further configured to include a basket calculation engine (BCE) module 3000 downstream from the VCU module 2900. An exemplary embodiment for the BCE module 3000 is described in the above-referenced and incorporated U.S. Patent Application Publication 2009/0182683. The BCE module 3000 can be configured to operate on level 1 event resulting from the synthetic quote event to effectively drive Net Asset Value (NAV) computations for baskets of financial instruments from low latency Level 2 data. The inventors believe that this chaining of synthetic quote generation and basket calculation is capable of providing a considerable speed advantage for a number of trading strategies including index arbitrage.


The aforementioned embodiments of the pipeline 1200 may be implemented in a variety of parallel processing technologies including: Field Programmable Gate Arrays (FPGAs), Chip Multi-Processors (CMPs), Application Specific Integrated Circuits (ASICs), Graphics Processing Units (GPUs), and multi-core superscalar processors. Furthermore, such pipelines 1200 may be deployed on coprocessor 840 of a ticker plant platform 3100 as shown in FIG. 31. It should also be noted that the pipelines disclosed herein can be implemented not only in the ticker plants resident in trading firms but also in the ticker plants resident within the exchanges themselves to accelerate the exchanges' abilities to create and maintain their own order books. Additional details regarding such a ticker plant platform 3100 can be found in the above-referenced and incorporated U.S. Patent Application Publications 2009/0182683 and 2008/0243675. In summary, the financial market data from the exchanges is received at the O/S supplied protocol stack 3102 executing in the kernel space on processor 812 (see FIGS. 8(a)-(b)). An upject driver 3104 delivers this exchange data to multi-threaded feed pre-processing software 3112 executing in the user-space on processor 812. These threads may then communicate data destined for the coprocessor 840 to the hardware interface driver software 3106 running in the kernel space.


Instructions from client applications may also be communicated to the hardware interface driver 3106 for ultimate delivery to coprocessor 840 to appropriately configure pipeline 1200 that is instantiated on coprocessor 840. Such instructions arrive at an O/S supplied protocol stack 3110 from which they are delivered to a request processing software module 3116. A background and maintenance processing software module 3114 thereafter determines whether the client application has the appropriate entitlement to issue such instructions. If so entitled, the background and maintenance processing block 3114 communicates a command instruction to the hardware interface driver 3106 for delivery to the coprocessor to appropriately update the pipeline 1200 to reflect any appropriate instructions.


The hardware interface driver 3106 then can deliver an interleaved stream of financial market data and commands to the coprocessor 840 for consumption thereby. Details regarding this stream transfer are described in the above-referenced and incorporated U.S. Patent Application Publication 2007/0174841. Outgoing data from the coprocessor 840 returns to the hardware interface driver 3106, from which it can be supplied to MDC driver 3108 for delivery to the client connections (via protocol stack 3110) and/or delivery to the background and maintenance processing block 3114.


While the present invention has been described above in relation to its preferred embodiments, various modifications may be made thereto that still fall within the invention's scope. Such modifications to the invention will be recognizable upon review of the teachings herein. Accordingly, the full scope of the present invention is to be defined solely by the appended claims and their legal equivalents.

Claims
  • 1. A system for applying specific computer technology to reduce latency and increase throughput with respect to updating a plurality of order books from financial market depth data, the system comprising: a memory for storing a plurality of data structures that represent a plurality of independent sets of order records for a plurality of financial instruments, the order records comprising (1) a plurality of limit order records representative of a plurality of limit orders for a plurality of financial instruments and (2) a plurality of price point records for a plurality of the limit orders; anda member of the group consisting of a reconfigurable logic device, a graphics processing unit (GPU), and a chip multi-processor (CMP), the member configured to (1) receive a plurality of limit order events, each limit order event pertaining to a limit order for a financial instrument on an exchange and comprising a plurality of data fields and (2) process a plurality of the limit order events to build and update the limit order records and the price point records independently within the data structures based on the processed limit order events;wherein the member comprises a first processing engine and a second processing engine;wherein the first processing engine is configured to generate a plurality of updates for the limit order records in the memory based on the limit order events;wherein the second processing engine is configured to generate a plurality of updates for the price point records in the memory based on the limit order events; andwherein the first and second processing engines are configured to perform their respective operations in parallel with each other.
  • 2. The system of claim 1 wherein the first and second processing engines are configured to interleave accesses to the memory to update the limit order records and the price point records.
  • 3. The system of claim 1 wherein the member is resident within a ticker plant, the ticker plant also including a processor, and wherein the member is configured to serve as an offload engine relative to the processor.
  • 4. The system of claim 3 further comprising: a trading application configured to (1) access a plurality of the updated order records and a plurality of the updated price point records via the ticker plant and (2) generate a plurality of financial instrument trade orders based on the accessed order records.
  • 5. The system of claim 1 wherein the first processing engine comprises an order engine, wherein the second processing engine comprises a price engine; wherein the order engine is configured to identify a plurality of adds, modifies, and deletes with respect to a plurality of limit order records based on the limit order events for updating the limit order records in the memory;wherein the price engine is configured to aggregate a plurality of adds, modifies, and deletes at a plurality of price levels with respect to a plurality of price point records based on the limit order events for updating the price point records in the memory; andwherein the order engine and the price engine are configured to perform their respective operations in parallel with each other.
  • 6. The system of claim 5 wherein the memory comprises a single physical memory in which the limit order records and the price point records are stored.
  • 7. The system of claim 6 wherein the order engine and the price engine are further configured to interleave their accesses to the single physical memory.
  • 8. The system of claim 5 wherein the memory comprises a first physical memory in which the limit order records are stored and a second physical memory in which the price point records are stored.
  • 9. The system of claim 8 wherein the order engine and the price engine are further configured to interleave their accesses to the first and second physical memories.
  • 10. The system of claim 5 wherein the price point records comprise a plurality of regional price point records and a plurality of composite price point records, each regional price point record being representative of an aggregation of limit orders for a financial instrument at a particular price on a given exchange, each composite price point record being representative of an aggregation of limit orders for a financial instrument at a particular price across a plurality of different exchanges.
  • 11. The system of claim 5 wherein a plurality of the limit order records are stored in the memory as a first hash table, wherein a plurality of the price point records are stored in the memory as a second hash table, and wherein the member further comprises an order hash component upstream from the order engine and a price hash component upstream from the price engine; wherein the order hash component is configured to generate a first hash key for a limit order event from an order reference number field or a symbol identifier field of the limit order event;wherein the order engine is further configured to locate a limit order record in the first hash table based on the generated first hash key;wherein the price hash component is configured to generate a second hash key for a limit order event from a symbol identifier field, a global exchange identifier field, and a price field of the limit order event; andwherein the price engine is further configured to locate a price point record in the second hash table based on the generated second hash key.
  • 12. The system of claim 11 wherein the price point records comprise a plurality of regional price point records and a plurality of composite price point records, each regional price point record being representative of an aggregation of limit orders for a financial instrument at a particular price on a given exchange, each composite price point record being representative of an aggregation of limit orders for a financial instrument at a particular price across a plurality of different exchanges.
  • 13. The system of claim 5 wherein the memory is external to the member; wherein the member is further configured to (1) cache a plurality of accessed limit order records and a plurality of accessed price point records in an on-chip memory cache and (2) track which limit order records and which price point records are cached in the on-chip memory cache;wherein the order engine is further configured to (1) receive limit order event data pertaining to a cached limit order record and (2) based on the tracking of which limit order records are cached in the on-chip memory cache, access the limit order records pertaining to a cached limit order record via the on-chip memory cache rather than the external memory; andwherein the price engine is further configured to (1) receive limit order event data pertaining to a cached price point record and (2) based on the tracking of which price point records are cached in the on-chip memory cache, access the price point records pertaining to a cached price point record via the on-chip memory cache rather than the external memory.
  • 14. The system of claim 13 wherein the on-chip memory cache comprises a first on-chip memory cache and a second on-chip memory cache; wherein the member is further configured to (1) cache a plurality of the accessed limit order records in the first on-chip memory cache, (2) cache a plurality of the accessed price point records in the second on-chip memory cache, (3) track which limit order records are cached in the first on-chip memory cache, and (4) track which price point records are cached in the second on-chip memory cache;wherein the order engine is further configured to, based on the tracking of which limit order records are cached in the first on-chip memory cache, access the limit order records pertaining to a cached limit order record via the first on-chip memory cache rather than the external memory; andwherein the price engine is further configured to, based on the tracking of which price point records are cached in the second on-chip memory cache, access the price point records pertaining to a cached price point record via the second on-chip memory cache rather than the external memory.
  • 15. The system of claim 1 wherein the member comprises a reconfigurable logic device.
  • 16. The system of claim 15 wherein the reconfigurable logic device comprises a field programmable gate array (FPGA).
  • 17. The system of claim 1 wherein the member comprises a GPU.
  • 18. The system of claim 1 wherein the member comprises a CMP.
  • 19. The system of claim 1 wherein the member is further configured to, for each of a plurality of the received limit order events, (1) map a subject limit order event to a limit order record in the memory using a first hash key that is derived from a first plurality of fields of the subject limit order event, (2) retrieve the mapped limit order record from the memory for updating by the first processing engine, (3) map the subject limit order event to a price point record in the memory using a second hash key that is derived from a second plurality of fields of the subject limit order event, and (4) retrieve the mapped price point record from the memory for updating by the second processing engine.
  • 20. The system of claim 19 wherein the first plurality of fields comprises (1) a field that holds data indicative of a symbol for the subject limit order event, (2) a field that holds data indicative of an exchange for the subject limit order event, and (3) a field that holds data indicative of an order reference number for the subject limit order event; and wherein the second plurality of fields comprises (1) the field that holds data indicative of the symbol for the subject limit order event, (2) the field that holds data indicative of the exchange for the subject limit order event, and (3) a field that holds data indicative of a price for the subject limit order event.
  • 21. The system of claim 19 wherein each of a plurality of the price point records comprises (1) a field that holds data indicative of a financial instrument identifier for a subject price point record, (2) a field that holds data indicative of a price for the subject price point record, (3) a field that holds data indicative of a volume for the subject price point record, and (4) a field that holds data indicative of an order count for the subject price point record.
  • 22. The system of claim 21 wherein each of a plurality of the price point records further comprises a field that holds data indicative of a record identifier for the subject price point record.
  • 23. The system of claim 19 wherein the price point records comprise a plurality of regional price point records and a plurality of composite price point records, wherein each regional price point record comprises aggregated price information about limit orders for a subject financial instrument on a specific regional exchange, and wherein each composite price point record comprises aggregated price information about limit orders for the subject financial instrument across a plurality of exchanges on which the subject financial instrument trades.
  • 24. The system of claim 1 wherein each of a plurality of the price point records comprises (1) a field that holds data indicative of a financial instrument identifier for a subject price point record, (2) a field that holds data indicative of a price for the subject price point record, (3) a field that holds data indicative of a volume for the subject price point record, and (4) a field that holds data indicative of an order count for the subject price point record.
  • 25. The system of claim 24 wherein each of a plurality of the price point records further comprises a field that holds data indicative of a record identifier for the subject price point record.
  • 26. The system of claim 24 wherein each of a plurality of the limit order records comprises (1) a field that holds data indicative of a financial instrument identifier for a subject limit order record, (2) a field that holds data indicative of an exchange for the subject limit order record, (3) a field that holds data indicative of an order reference number for the subject limit order record, (4) a field that holds data indicative of a price for the subject limit order record, and (5) a field that holds data indicative of a size for the subject limit order record.
  • 27. The system of claim 1 wherein the price point records comprise a plurality of regional price point records and a plurality of composite price point records, wherein each regional price point record comprises aggregated price information about limit orders for a subject financial instrument on a specific regional exchange, and wherein each composite price point record comprises aggregated price information about limit orders for the subject financial instrument across a plurality of exchanges on which the subject financial instrument trades.
  • 28. The system of claim 27 wherein the second processing engine is further configured to generate updates for the regional price point records and the composite price point records based on the limit order events.
  • 29. The system of claim 1 wherein the member is further configured to generate order book data for delivery to one or more downstream consumers, wherein the order book data includes an order summary view based on the limit order records.
  • 30. The system of claim 1 wherein the member is further configured to generate order book data for delivery to one or more downstream consumers, wherein the order book data includes a price summary view based on the price point records.
  • 31. The system of claim 1 wherein the member is further configured to generate order book data for delivery to one or more downstream consumers, wherein the order book data includes a spliced price summary view based on the price point records.
  • 32. The system of claim 1 wherein the member is further configured to generate and communicate order book data for delivery to one or more downstream consumers, wherein the order book data includes an aggregate price summary view based on the price point records.
  • 33. A method for applying specific computer technology to reduce latency and increase throughput with respect to updating a plurality of order books from financial market depth data, the method comprising: maintaining a plurality of data structures in memory that represent a plurality of independent sets of order records for a plurality of financial instruments, the order records comprising (1) a plurality of limit order records representative of a plurality of limit orders for a plurality of financial instruments and (2) a plurality of price point records for a plurality of the limit orders;receiving a plurality of limit order events, each limit order event pertaining to a limit order for a financial instrument on an exchange and comprising a plurality of data fields; andprocessing of a plurality of the limit order events to build and update the limit order records and the price point records independently within the data structures based on the processed limit order events, wherein the processing step comprises (1) a first processing engine generating a plurality of updates for the limit order records in the memory based on the limit order events and (2) a second processing engine generating a plurality of updates for the price point records in the memory based on the limit order events, wherein the first processing engine and the second processing engine perform their respective generating steps in parallel with each other; andwherein the processing step is performed by a member of the group consisting of a reconfigurable logic device, a graphics processing unit (GPU), and a chip multi-processor (CMP), the first and second processing engines being deployed on the member.
  • 34. The method of claim 33 wherein the processing step comprises the first and second processing engines interleaving accesses to the memory to update the limit order records and the price point records.
  • 35. The method of claim 33 further comprising, for each of a plurality of the received limit order events, the member (1) mapping a subject limit order event to a limit order record in the memory using a first hash key that is derived from a first plurality of fields of the subject limit order event, (2) retrieving the mapped limit order record from the memory for updating by the first processing engine, (3) mapping the subject limit order event to a price point record in the memory using a second hash key that is derived from a second plurality of fields of the subject limit order event, and (4) retrieving the mapped price point record from the memory for updating by the second processing engine.
  • 36. The method of claim 35 wherein the first plurality of fields comprises (1) a field that holds data indicative of a symbol for the subject limit order event, (2) a field that holds data indicative of an exchange for the subject limit order event, and (3) a field that holds data indicative of an order reference number for the subject limit order event; and wherein the second plurality of fields comprises (1) the field that holds data indicative of the symbol for the subject limit order event, (2) the field that holds data indicative of the exchange for the subject limit order event, and (3) a field that holds data indicative of a price for the subject limit order event.
  • 37. The method of claim 35 wherein each of a plurality of the price point records comprises (1) a field that holds data indicative of a financial instrument identifier for a subject price point record, (2) a field that holds data indicative of a price for the subject price point record, (3) a field that holds data indicative of a volume for the subject price point record, and (4) a field that holds data indicative of an order count for the subject price point record.
  • 38. The method of claim 37 wherein each of a plurality of the price point records further comprises a field that holds data indicative of a record identifier for the subject price point record.
  • 39. The method of claim 35 wherein the price point records comprise a plurality of regional price point records and a plurality of composite price point records, wherein each regional price point record comprises aggregated price information about limit orders for a subject financial instrument on a specific regional exchange, and wherein each composite price point record comprises aggregated price information about limit orders for the subject financial instrument across a plurality of exchanges on which the subject financial instrument trades.
  • 40. The method of claim 35 wherein each of a plurality of the price point records comprises (1) a field that holds data indicative of a financial instrument identifier for a subject price point record, (2) a field that holds data indicative of a price for the subject price point record, (3) a field that holds data indicative of a volume for the subject price point record, and (4) a field that holds data indicative of an order count for the subject price point record.
  • 41. The method of claim 40 wherein each of a plurality of the price point records further comprises a field that holds data indicative of a record identifier for the subject price point record.
  • 42. The method of claim 40 wherein each of a plurality of the limit order records comprises (1) a field that holds data indicative of a financial instrument identifier for a subject limit order record, (2) a field that holds data indicative of an exchange for the subject limit order record, (3) a field that holds data indicative of an order reference number for the subject limit order record, (4) a field that holds data indicative of a price for the subject limit order record, and (5) a field that holds data indicative of a size for the subject limit order record.
  • 43. The method of claim 33 wherein the price point records comprise a plurality of regional price point records and a plurality of composite price point records, wherein each regional price point record comprises aggregated price information about limit orders for a subject financial instrument on a specific regional exchange, and wherein each composite price point record comprises aggregated price information about limit orders for the subject financial instrument across a plurality of exchanges on which the subject financial instrument trades.
  • 44. The method of claim 43 wherein the second processing engine generating step comprises the second processing engine generating updates for the regional price point records and the composite price point records based on the limit order events.
  • 45. The method of claim 33 further comprising: the member generating order book data for delivery to one or more downstream consumers, wherein the order book data includes an order summary view based on the limit order records.
  • 46. The method of claim 33 further comprising: the member generating order book data for delivery to one or more downstream consumers, wherein the order book data includes a price summary view based on the price point records.
  • 47. The method of claim 33 further comprising: the member generating order book data for delivery to one or more downstream consumers, wherein the order book data includes a spliced price summary view based on the price point records.
  • 48. The method of claim 33 further comprising: the member generating order book data for delivery to one or more downstream consumers, wherein the order book data includes an aggregate price summary view based on the price point records.
CROSS-REFERENCE AND PRIORITY CLAIM TO RELATED PATENT APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 14/315,571, filed Jun. 26, 2014, now U.S. Pat. No. 10,062,115, which is a divisional of U.S. patent application Ser. No. 13/154,804, filed Jun. 7, 2011, now U.S. Pat. No. 8,768,805, which is a divisional of International Patent Application No. PCT/US2009/067935, filed Dec. 14, 2009, and designating the U.S., which claims priority to U.S. patent application 61/122,673, filed Dec. 15, 2008, entitled “Method and Apparatus for High-Speed Processing of Financial Market Depth Data”, the entire disclosures of each of which are incorporated herein by reference. This application is related to (1) U.S. patent application Ser. No. 13/132,408, which is a national phase entry of International Patent Application No. PCT/US2009/067935 and claims priority to U.S. patent application 61/122,673, and (2) U.S. patent application Ser. No. 13/154,824, filed Jun. 7, 2011, now U.S. Pat. No. 8,762,249, which is a divisional of International Patent Application No. PCT/US2009/067935 and claims priority to U.S. patent application 61/122,673. This application is also related to (1) U.S. patent application Ser. No. 12/013,302, filed Jan. 11, 2008, entitled “Method and System for Low Latency Basket Calculation”, published as U.S. Patent Application Publication 2009/0182683, (2) U.S. patent application Ser. No. 11/765,306, filed Jun. 19, 2007, entitled “High Speed Processing of Financial Information Using FPGA Devices”, published as U.S. Patent Application Publication 2008/0243675, (3) U.S. patent application Ser. No. 11/760,211, filed Jun. 8, 2007, entitled “Method and System for High Speed Options Pricing”, published as U.S. Patent Application Publication 2007/0294157, and (4) U.S. patent application Ser. No. 11/561,615, filed Nov. 20, 2006, entitled “Method and Apparatus for Processing Financial Information at Hardware Speeds Using FPGA Devices”, published as U.S. Patent Application Publication 2007/0078837, the entire disclosures of each of which are incorporated herein by reference.

US Referenced Citations (611)
Number Name Date Kind
2046381 Hicks et al. Jul 1936 A
3082402 Scantlin Mar 1963 A
3296597 Scantlin et al. Jan 1967 A
3573747 Adams et al. Apr 1971 A
3581072 Nymeyer May 1971 A
3601808 Mack Aug 1971 A
3611314 Pritchard, Jr. et al. Oct 1971 A
3729712 Glassman Apr 1973 A
3824375 Gross et al. Jul 1974 A
3848235 Lewis et al. Nov 1974 A
3906455 Houston et al. Sep 1975 A
4044334 Bachman et al. Aug 1977 A
4081607 Vitols et al. Mar 1978 A
4298898 Cardot Nov 1981 A
4300193 Bradley et al. Nov 1981 A
4314356 Scarbrough Feb 1982 A
4385393 Chaure et al. May 1983 A
4412287 Braddock, III Oct 1983 A
4464718 Dixon et al. Aug 1984 A
4550436 Freeman et al. Oct 1985 A
4674044 Kalmus et al. Jun 1987 A
4811210 McAulay Mar 1989 A
4823306 Barbic et al. Apr 1989 A
4868866 Williams, Jr. Sep 1989 A
4903201 Wagner Feb 1990 A
4941178 Chuang Jul 1990 A
5023910 Thomson Jun 1991 A
5038284 Kramer Aug 1991 A
5050075 Herman et al. Sep 1991 A
5063507 Lindsey et al. Nov 1991 A
5077665 Silverman et al. Dec 1991 A
5101353 Lupien et al. Mar 1992 A
5101424 Clayton et al. Mar 1992 A
5126936 Champion et al. Jun 1992 A
5140692 Morita Aug 1992 A
5161103 Kosaka et al. Nov 1992 A
5163131 Row et al. Nov 1992 A
5179626 Thomson Jan 1993 A
5226165 Martin Jul 1993 A
5243655 Wang Sep 1993 A
5249292 Chiappa Sep 1993 A
5255136 Machado et al. Oct 1993 A
5258908 Hartheimer et al. Nov 1993 A
5265065 Turtle Nov 1993 A
5267148 Kosaka et al. Nov 1993 A
5270922 Higgins Dec 1993 A
5297032 Trojan et al. Mar 1994 A
5313560 Maruoka et al. May 1994 A
5315634 Tanaka et al. May 1994 A
5319776 Hile et al. Jun 1994 A
5327521 Savic et al. Jul 1994 A
5339411 Heaton, Jr. Aug 1994 A
5371794 Diffie et al. Dec 1994 A
5375055 Togher et al. Dec 1994 A
5388259 Fleischman et al. Feb 1995 A
5396253 Chia Mar 1995 A
5404488 Kerrigan et al. Apr 1995 A
5418951 Damashek May 1995 A
5432822 Kaewell, Jr. Jul 1995 A
5461712 Chelstowski et al. Oct 1995 A
5465353 Hull et al. Nov 1995 A
5481735 Mortensen et al. Jan 1996 A
5488725 Turtle et al. Jan 1996 A
5497317 Hawkins et al. Mar 1996 A
5497488 Akizawa et al. Mar 1996 A
5500793 Deming, Jr. et al. Mar 1996 A
5517642 Bezek et al. May 1996 A
5535384 Kasahara Jul 1996 A
5544352 Egger Aug 1996 A
5546578 Takada et al. Aug 1996 A
5596569 Madonna et al. Jan 1997 A
5619574 Johnson et al. Apr 1997 A
5633865 Short May 1997 A
5651125 Witt et al. Jul 1997 A
5684980 Casselman Nov 1997 A
5701464 Aucsmith Dec 1997 A
5712942 Jennings et al. Jan 1998 A
5721898 Beardsley et al. Feb 1998 A
5740244 Indeck et al. Apr 1998 A
5740466 Geldman et al. Apr 1998 A
5774835 Ozawa et al. Jun 1998 A
5774839 Shlomot Jun 1998 A
5781772 Wilkinson, III et al. Jul 1998 A
5781921 Nichols Jul 1998 A
5802290 Casselman Sep 1998 A
5805832 Brown et al. Sep 1998 A
5809483 Broka et al. Sep 1998 A
5813000 Furlani Sep 1998 A
5819273 Vora et al. Oct 1998 A
5819290 Fujita et al. Oct 1998 A
5826075 Bealkowski et al. Oct 1998 A
5845266 Lupien et al. Dec 1998 A
5857176 Ginsberg Jan 1999 A
5864738 Kessler et al. Jan 1999 A
5870730 Furuya et al. Feb 1999 A
5873071 Ferstenberg et al. Feb 1999 A
5884286 Daughtery, III Mar 1999 A
5905974 Fraser et al. May 1999 A
5913211 Nitta Jun 1999 A
5926775 Brumley et al. Jul 1999 A
5930753 Potamianos et al. Jul 1999 A
5943421 Grabon Aug 1999 A
5943429 Händel Aug 1999 A
5950196 Pyreddy et al. Sep 1999 A
5963923 Garber Oct 1999 A
5978801 Yuasa Nov 1999 A
5987432 Zusman et al. Nov 1999 A
5991881 Conklin et al. Nov 1999 A
5995963 Nanba et al. Nov 1999 A
6006264 Colby et al. Dec 1999 A
6016483 Rickard et al. Jan 2000 A
6023755 Casselman Feb 2000 A
6023760 Karttunen Feb 2000 A
6028939 Yin Feb 2000 A
6044407 Jones et al. Mar 2000 A
6058391 Gardner May 2000 A
6061662 Makivic May 2000 A
6064739 Davis May 2000 A
6067569 Khaki et al. May 2000 A
6070172 Lowe May 2000 A
6073160 Grantham et al. Jun 2000 A
6084584 Nahi et al. Jul 2000 A
6105067 Batra Aug 2000 A
6134551 Aucsmith Oct 2000 A
6138176 McDonald et al. Oct 2000 A
6147976 Shand et al. Nov 2000 A
6169969 Cohen Jan 2001 B1
6173270 Cristofich et al. Jan 2001 B1
6173276 Kant et al. Jan 2001 B1
6178494 Casselman Jan 2001 B1
6195024 Fallon Feb 2001 B1
6226676 Crump et al. May 2001 B1
6236980 Reese May 2001 B1
6243753 Machin et al. Jun 2001 B1
6247060 Boucher et al. Jun 2001 B1
6263321 Daughtery, III Jul 2001 B1
6272616 Fernando et al. Aug 2001 B1
6278982 Korhammer et al. Aug 2001 B1
6279113 Vaidya Aug 2001 B1
6279140 Slane Aug 2001 B1
6289440 Casselman Sep 2001 B1
6295530 Ritchie et al. Sep 2001 B1
6304858 Mosler et al. Oct 2001 B1
6309424 Fallon Oct 2001 B1
6317728 Kane Nov 2001 B1
6317795 Malkin et al. Nov 2001 B1
6321258 Stollfus et al. Nov 2001 B1
6336150 Ellis et al. Jan 2002 B1
6339819 Huppenthal et al. Jan 2002 B1
6370592 Kumpf Apr 2002 B1
6370645 Lee et al. Apr 2002 B1
6377942 Hinsley et al. Apr 2002 B1
6397259 Lincke et al. May 2002 B1
6397335 Franczek et al. May 2002 B1
6412000 Riddle et al. Jun 2002 B1
6415269 Dinwoodie Jul 2002 B1
6418419 Nieboer et al. Jul 2002 B1
6430272 Maruyama et al. Aug 2002 B1
6456982 Pilipovic Sep 2002 B1
6463474 Fuh et al. Oct 2002 B1
6493682 Horrigan et al. Dec 2002 B1
6499107 Gleichauf et al. Dec 2002 B1
6535868 Galeazzi et al. Mar 2003 B1
6546375 Pang et al. Apr 2003 B1
6578147 Shanklin et al. Jun 2003 B1
6581098 Kumpf Jun 2003 B1
6591302 Boucher et al. Jul 2003 B2
6594643 Freeny, Jr. Jul 2003 B1
6597812 Fallon et al. Jul 2003 B1
6601094 Mentze et al. Jul 2003 B1
6601104 Fallon Jul 2003 B1
6604158 Fallon Aug 2003 B1
6624761 Fallon Sep 2003 B2
6625150 Yu Sep 2003 B1
6691301 Bowen Feb 2004 B2
6704816 Burke Mar 2004 B1
6711558 Indeck et al. Mar 2004 B1
6765918 Dixon et al. Jul 2004 B1
6766304 Kemp, II et al. Jul 2004 B2
6772132 Kemp, II et al. Aug 2004 B1
6772136 Kant et al. Aug 2004 B2
6772345 Shetty Aug 2004 B1
6778968 Gulati Aug 2004 B1
6785677 Fritchman Aug 2004 B1
6801938 Bookman et al. Oct 2004 B1
6804667 Martin Oct 2004 B1
6807156 Veres et al. Oct 2004 B1
6820129 Courey, Jr. Nov 2004 B1
6839686 Galant Jan 2005 B1
6847645 Potter et al. Jan 2005 B1
6850906 Chadha et al. Feb 2005 B1
6877044 Lo et al. Apr 2005 B2
6886103 Brustoloni et al. Apr 2005 B1
6901461 Bennett May 2005 B2
6931408 Adams et al. Aug 2005 B2
6944168 Paatela et al. Sep 2005 B2
6978223 Milliken Dec 2005 B2
6981054 Krishna Dec 2005 B1
7003488 Dunne et al. Feb 2006 B2
7024384 Daughtery, III Apr 2006 B2
7046848 Olcott May 2006 B1
7058735 Spencer Jun 2006 B2
7065475 Brundobler Jun 2006 B1
7089206 Martin Aug 2006 B2
7089326 Boucher et al. Aug 2006 B2
7093023 Lockwood et al. Aug 2006 B2
7099838 Gastineau et al. Aug 2006 B1
7103569 Groveman et al. Sep 2006 B1
7117280 Vasudevan Oct 2006 B2
7124106 Stallaert et al. Oct 2006 B1
7127424 Kemp, II et al. Oct 2006 B2
7130913 Fallon Oct 2006 B2
7139743 Indeck et al. Nov 2006 B2
7149715 Browne et al. Dec 2006 B2
7161506 Fallon Jan 2007 B2
7167980 Chiu Jan 2007 B2
7177833 Marynowski et al. Feb 2007 B1
7181437 Indeck et al. Feb 2007 B2
7181608 Fallon et al. Feb 2007 B2
7212998 Muller et al. May 2007 B1
7222114 Chan et al. May 2007 B1
7224185 Campbell et al. May 2007 B2
7225188 Gai et al. May 2007 B1
7228289 Brumfield et al. Jun 2007 B2
7249118 Sandler et al. Jul 2007 B2
7251629 Marynowski et al. Jul 2007 B1
7257842 Barton et al. Aug 2007 B2
7277887 Burrows et al. Oct 2007 B1
7287037 An et al. Oct 2007 B2
7305383 Kubesh et al. Dec 2007 B1
7305391 Wyschogrod et al. Dec 2007 B2
7321937 Fallon Jan 2008 B2
7356498 Kaminsky et al. Apr 2008 B2
7363277 Dutta et al. Apr 2008 B1
7376755 Pandya May 2008 B2
7378992 Fallon May 2008 B2
7386046 Fallon et al. Jun 2008 B2
7406444 Eng et al. Jul 2008 B2
7411957 Stacy et al. Aug 2008 B2
7414975 Olderdissen Aug 2008 B2
7415723 Pandya Aug 2008 B2
7417568 Fallon et al. Aug 2008 B2
7454418 Wang Nov 2008 B1
7457834 Jung et al. Nov 2008 B2
7461064 Fontoura et al. Dec 2008 B2
7478431 Nachenberg Jan 2009 B1
7487264 Pandya Feb 2009 B2
7487327 Chang et al. Feb 2009 B1
7496108 Biran et al. Feb 2009 B2
7499966 Elnozahy et al. Mar 2009 B2
7536462 Pandya May 2009 B2
7539845 Wentzlaff et al. May 2009 B1
7558753 Neubert et al. Jul 2009 B2
7558925 Bouchard et al. Jul 2009 B2
7565315 West et al. Jul 2009 B2
7565525 Vorbach et al. Jul 2009 B2
7580719 Karmarkar Aug 2009 B2
7587476 Sato Sep 2009 B2
7598958 Kelleher Oct 2009 B1
7603303 Kraus et al. Oct 2009 B1
7617291 Fan et al. Nov 2009 B2
7627693 Pandya Dec 2009 B2
7631107 Pandya Dec 2009 B2
7636703 Taylor Dec 2009 B2
7660761 Zhou et al. Feb 2010 B2
7668849 Narancic et al. Feb 2010 B1
7685121 Brown et al. Mar 2010 B2
7701945 Roesch et al. Apr 2010 B2
7714747 Fallon May 2010 B2
7761459 Zhang et al. Jul 2010 B1
7788293 Pasztor et al. Aug 2010 B2
7827190 Pandya Nov 2010 B2
7831606 Pandya Nov 2010 B2
7831607 Pandya Nov 2010 B2
7840482 Singla et al. Nov 2010 B2
7856545 Casselman Dec 2010 B2
7856546 Casselman et al. Dec 2010 B2
7870217 Pandya Jan 2011 B2
7890692 Pandya Feb 2011 B2
7899976 Pandya Mar 2011 B2
7899977 Pandya Mar 2011 B2
7899978 Pandya Mar 2011 B2
7908213 Monroe et al. Mar 2011 B2
7908259 Branscome et al. Mar 2011 B2
7912808 Pandya Mar 2011 B2
7917299 Buhler et al. Mar 2011 B2
7921046 Parsons et al. Apr 2011 B2
7944920 Pandya May 2011 B2
7945528 Cytron et al. May 2011 B2
7949650 Indeck et al. May 2011 B2
7953743 Indeck et al. May 2011 B2
7954114 Chamberlain et al. May 2011 B2
7991667 Kraus et al. Aug 2011 B2
7996348 Pandya Aug 2011 B2
8005966 Pandya Aug 2011 B2
8015099 Reid Sep 2011 B2
8024253 Peterffy et al. Sep 2011 B2
8027893 Burrows et al. Sep 2011 B1
8030888 Pandya et al. Oct 2011 B2
8032440 Hait Oct 2011 B1
8046283 Burns et al. Oct 2011 B2
8051022 Pandya Nov 2011 B2
8055601 Pandya Nov 2011 B2
8069102 Indeck et al. Nov 2011 B2
8073763 Merrin et al. Dec 2011 B1
8095508 Chamberlain et al. Jan 2012 B2
8131697 Indeck et al. Mar 2012 B2
8140416 Borkovec et al. Mar 2012 B2
8156101 Indeck et al. Apr 2012 B2
8175946 Hamati et al. May 2012 B2
8181239 Pandya May 2012 B2
8200599 Pandya Jun 2012 B2
8213518 Wang et al. Jul 2012 B1
8281026 Lankford et al. Oct 2012 B2
8332829 Mattson Dec 2012 B2
8374986 Indeck et al. Feb 2013 B2
8407122 Parsons et al. Mar 2013 B2
8458081 Parsons et al. Jun 2013 B2
8478680 Parsons et al. Jul 2013 B2
8515682 Buhler et al. Aug 2013 B2
8549024 Indeck et al. Oct 2013 B2
8595104 Parsons et al. Nov 2013 B2
8600856 Parsons et al. Dec 2013 B2
8601086 Pandya Dec 2013 B2
8620881 Chamberlain et al. Dec 2013 B2
8626624 Parsons et al. Jan 2014 B2
8762249 Taylor et al. Jun 2014 B2
8768805 Taylor et al. Jul 2014 B2
8768888 Chamberlain et al. Jul 2014 B2
8843408 Singla et al. Sep 2014 B2
8879727 Taylor et al. Nov 2014 B2
8880501 Indeck et al. Nov 2014 B2
8983063 Taylor et al. Mar 2015 B1
9020928 Indeck et al. Apr 2015 B2
9047243 Taylor et al. Jun 2015 B2
9176775 Chamberlain et al. Nov 2015 B2
9582831 Parsons et al. Feb 2017 B2
9813976 Velusamy et al. Nov 2017 B2
10062115 Taylor et al. Aug 2018 B2
10360632 Parsons et al. Jul 2019 B2
10504184 Parsons et al. Dec 2019 B2
10817945 Parsons et al. Oct 2020 B2
20010003193 Woodring et al. Jun 2001 A1
20010013048 Imbert de Tremiolles et al. Aug 2001 A1
20010041012 Hsieh et al. Nov 2001 A1
20010042040 Keith Nov 2001 A1
20010047473 Fallon Nov 2001 A1
20010056547 Dixon Dec 2001 A1
20020010825 Wilson Jan 2002 A1
20020019812 Board et al. Feb 2002 A1
20020023010 Rittmaster et al. Feb 2002 A1
20020038276 Buhannic et al. Mar 2002 A1
20020038279 Samuelson et al. Mar 2002 A1
20020069375 Bowen Jun 2002 A1
20020072893 Wilson Jun 2002 A1
20020080871 Fallon et al. Jun 2002 A1
20020082967 Kaminsky et al. Jun 2002 A1
20020095519 Philbrick et al. Jul 2002 A1
20020100029 Bowen Jul 2002 A1
20020101425 Hamid Aug 2002 A1
20020105911 Pruthi et al. Aug 2002 A1
20020129140 Peled et al. Sep 2002 A1
20020138376 Hinkle Sep 2002 A1
20020143521 Call Oct 2002 A1
20020150248 Kovacevic Oct 2002 A1
20020156998 Casselman Oct 2002 A1
20020162025 Sutton et al. Oct 2002 A1
20020166063 Lachman et al. Nov 2002 A1
20020169873 Zodnik Nov 2002 A1
20020180742 Hamid Dec 2002 A1
20020198813 Patterson et al. Dec 2002 A1
20020199173 Bowen Dec 2002 A1
20030009411 Ram et al. Jan 2003 A1
20030009693 Brock et al. Jan 2003 A1
20030014521 Elson et al. Jan 2003 A1
20030014662 Gupta et al. Jan 2003 A1
20030018630 Indeck et al. Jan 2003 A1
20030023653 Dunlop et al. Jan 2003 A1
20030023876 Bardsley et al. Jan 2003 A1
20030028408 RuDusky Feb 2003 A1
20030028690 Appleby-Alis et al. Feb 2003 A1
20030028864 Bowen Feb 2003 A1
20030033234 RuDusky Feb 2003 A1
20030033240 Balson et al. Feb 2003 A1
20030033450 Appleby-Alis Feb 2003 A1
20030033514 Appleby-Allis et al. Feb 2003 A1
20030033588 Alexander Feb 2003 A1
20030033594 Bowen Feb 2003 A1
20030035547 Newton Feb 2003 A1
20030037037 Adams et al. Feb 2003 A1
20030037321 Bowen Feb 2003 A1
20030041129 Appleby-Allis Feb 2003 A1
20030043805 Graham et al. Mar 2003 A1
20030046668 Bowen Mar 2003 A1
20030051043 Wyschogrod et al. Mar 2003 A1
20030055658 RuDusky Mar 2003 A1
20030055769 RuDusky Mar 2003 A1
20030055770 RuDusky Mar 2003 A1
20030055771 RuDusky Mar 2003 A1
20030055777 Ginsberg Mar 2003 A1
20030061409 RuDusky Mar 2003 A1
20030065607 Satchwell Apr 2003 A1
20030065943 Geis et al. Apr 2003 A1
20030074177 Bowen Apr 2003 A1
20030074582 Patel et al. Apr 2003 A1
20030078865 Lee Apr 2003 A1
20030079060 Dunlop Apr 2003 A1
20030086300 Noyes et al. May 2003 A1
20030093343 Huttenlocher et al. May 2003 A1
20030093347 Gray May 2003 A1
20030097481 Richter May 2003 A1
20030099254 Richter May 2003 A1
20030105620 Bowen Jun 2003 A1
20030105721 Ginter et al. Jun 2003 A1
20030110229 Kulig et al. Jun 2003 A1
20030115485 Milliken Jun 2003 A1
20030117971 Aubury Jun 2003 A1
20030120460 Aubury Jun 2003 A1
20030121010 Aubury Jun 2003 A1
20030126065 Eng et al. Jul 2003 A1
20030140337 Aubury Jul 2003 A1
20030154368 Stevens et al. Aug 2003 A1
20030163715 Wong Aug 2003 A1
20030177253 Schuehler et al. Sep 2003 A1
20030184593 Dunlop Oct 2003 A1
20030187662 Wilson Oct 2003 A1
20030191876 Fallon Oct 2003 A1
20030208430 Gershon Nov 2003 A1
20030217306 Harthcock et al. Nov 2003 A1
20030221013 Lockwood et al. Nov 2003 A1
20030233302 Weber et al. Dec 2003 A1
20040000928 Cheng et al. Jan 2004 A1
20040010612 Pandya Jan 2004 A1
20040015502 Alexander et al. Jan 2004 A1
20040015633 Smith Jan 2004 A1
20040019703 Burton Jan 2004 A1
20040028047 Hou et al. Feb 2004 A1
20040034587 Amberson et al. Feb 2004 A1
20040049596 Schuehler et al. Mar 2004 A1
20040059666 Waelbroeck et al. Mar 2004 A1
20040062245 Sharp et al. Apr 2004 A1
20040064737 Milliken et al. Apr 2004 A1
20040073703 Boucher et al. Apr 2004 A1
20040111632 Halperin Jun 2004 A1
20040162826 Wyschogrod et al. Aug 2004 A1
20040170070 Rapp et al. Sep 2004 A1
20040177340 Hsu et al. Sep 2004 A1
20040186804 Chakraborty et al. Sep 2004 A1
20040186814 Chalermkraivuth et al. Sep 2004 A1
20040199448 Chalermkraivuth et al. Oct 2004 A1
20040199452 Johnston et al. Oct 2004 A1
20040205149 Dillon et al. Oct 2004 A1
20050005145 Teixeira Jan 2005 A1
20050027634 Gershon Feb 2005 A1
20050033672 Lasry et al. Feb 2005 A1
20050044344 Stevens Feb 2005 A1
20050080649 Alvarez et al. Apr 2005 A1
20050086520 Dharmapurikar et al. Apr 2005 A1
20050091142 Renton et al. Apr 2005 A1
20050097027 Kavanaugh May 2005 A1
20050108518 Pandya May 2005 A1
20050131790 Benzschawel et al. Jun 2005 A1
20050135608 Zheng Jun 2005 A1
20050187844 Chalermkraivuth et al. Aug 2005 A1
20050187845 Eklund et al. Aug 2005 A1
20050187846 Subbu et al. Aug 2005 A1
20050187847 Bonissone et al. Aug 2005 A1
20050187848 Bonissone et al. Aug 2005 A1
20050187849 Bollapragada et al. Aug 2005 A1
20050190787 Kuik et al. Sep 2005 A1
20050195832 Dharmapurikar et al. Sep 2005 A1
20050197938 Davie et al. Sep 2005 A1
20050197939 Davie et al. Sep 2005 A1
20050197948 Davie et al. Sep 2005 A1
20050203938 Hoeft et al. Sep 2005 A1
20050216384 Partlow et al. Sep 2005 A1
20050226165 Pope et al. Oct 2005 A1
20050228735 Duquette Oct 2005 A1
20050229254 Singh et al. Oct 2005 A1
20050240510 Schweickert et al. Oct 2005 A1
20050243824 Abbazia et al. Nov 2005 A1
20050267836 Crosthwaite et al. Dec 2005 A1
20050283423 Moser et al. Dec 2005 A1
20050283743 Mulholland et al. Dec 2005 A1
20060020536 Renton et al. Jan 2006 A1
20060020715 Jungck Jan 2006 A1
20060026090 Balabon Feb 2006 A1
20060031154 Noviello et al. Feb 2006 A1
20060031156 Noviello et al. Feb 2006 A1
20060047636 Mohania et al. Mar 2006 A1
20060053295 Madhusudan et al. Mar 2006 A1
20060059064 Glinberg et al. Mar 2006 A1
20060059065 Glinberg et al. Mar 2006 A1
20060059066 Glinberg et al. Mar 2006 A1
20060059067 Glinberg et al. Mar 2006 A1
20060059068 Glinberg et al. Mar 2006 A1
20060059069 Glinberg et al. Mar 2006 A1
20060059083 Friesen et al. Mar 2006 A1
20060075404 Rosu et al. Apr 2006 A1
20060129745 Thiel et al. Jun 2006 A1
20060143099 Partlow et al. Jun 2006 A1
20060215695 Olderdissen Sep 2006 A1
20060242123 Williams, Jr. Oct 2006 A1
20060259407 Rosenthal et al. Nov 2006 A1
20060259417 Marynowski et al. Nov 2006 A1
20060269148 Farber et al. Nov 2006 A1
20060282281 Egetoft Dec 2006 A1
20060282369 White Dec 2006 A1
20060292292 Brightman et al. Dec 2006 A1
20060294059 Chamberlain et al. Dec 2006 A1
20070011183 Langseth et al. Jan 2007 A1
20070011687 Ilik et al. Jan 2007 A1
20070025351 Cohen Feb 2007 A1
20070027788 Bandman Feb 2007 A1
20070061231 Kim-E Mar 2007 A1
20070061241 Jovanovic et al. Mar 2007 A1
20070067108 Buhler et al. Mar 2007 A1
20070067481 Sharma et al. Mar 2007 A1
20070078837 Indeck et al. Apr 2007 A1
20070094199 Deshpande et al. Apr 2007 A1
20070112837 Houh et al. May 2007 A1
20070118457 Peterffy et al. May 2007 A1
20070118500 Indeck et al. May 2007 A1
20070130140 Cytron et al. Jun 2007 A1
20070130188 Moon et al. Jun 2007 A1
20070156574 Marynowski et al. Jul 2007 A1
20070174841 Chamberlain et al. Jul 2007 A1
20070179935 Lee et al. Aug 2007 A1
20070209068 Ansari et al. Sep 2007 A1
20070237327 Taylor et al. Oct 2007 A1
20070244859 Trippe et al. Oct 2007 A1
20070260602 Taylor Nov 2007 A1
20070277036 Chamberlain et al. Nov 2007 A1
20070294157 Singla Dec 2007 A1
20070294162 Borkovec Dec 2007 A1
20080077793 Tan et al. Mar 2008 A1
20080082502 Gupta Apr 2008 A1
20080084573 Horowitz et al. Apr 2008 A1
20080086274 Chamberlain et al. Apr 2008 A1
20080097893 Walsky et al. Apr 2008 A1
20080104542 Cohen et al. May 2008 A1
20080109413 Indeck et al. May 2008 A1
20080114724 Indeck et al. May 2008 A1
20080114725 Indeck et al. May 2008 A1
20080114760 Indeck et al. May 2008 A1
20080126320 Indeck et al. May 2008 A1
20080133453 Indeck et al. Jun 2008 A1
20080133519 Indeck et al. Jun 2008 A1
20080162378 Levine et al. Jul 2008 A1
20080189251 Branscome et al. Aug 2008 A1
20080243675 Parsons Oct 2008 A1
20080253395 Pandya Oct 2008 A1
20080266302 Andre et al. Oct 2008 A1
20080275805 Hecht Nov 2008 A1
20090019219 Magklis et al. Jan 2009 A1
20090037514 Lankford et al. Feb 2009 A1
20090182683 Taylor et al. Jul 2009 A1
20090183174 Subhraveti Jul 2009 A1
20090262741 Jungck et al. Oct 2009 A1
20090287628 Indeck et al. Nov 2009 A1
20100005036 Kraus et al. Jan 2010 A1
20100106976 Aciicmez et al. Apr 2010 A1
20100174770 Pandya Jul 2010 A1
20100306479 Ezzat Dec 2010 A1
20110066832 Casselman et al. Mar 2011 A1
20110125960 Casselman May 2011 A1
20110178911 Parsons et al. Jul 2011 A1
20110178912 Parsons et al. Jul 2011 A1
20110178917 Parsons et al. Jul 2011 A1
20110178918 Parsons et al. Jul 2011 A1
20110178919 Parsons et al. Jul 2011 A1
20110178957 Parsons et al. Jul 2011 A1
20110179050 Parsons et al. Jul 2011 A1
20110184844 Parsons et al. Jul 2011 A1
20110199243 Fallon et al. Aug 2011 A1
20110231446 Buhler et al. Sep 2011 A1
20110246353 Kraus et al. Oct 2011 A1
20110252008 Chamberlain et al. Oct 2011 A1
20110291615 Pandya et al. Dec 2011 A1
20110295967 Wang et al. Dec 2011 A1
20120089496 Taylor et al. Apr 2012 A1
20120089497 Taylor et al. Apr 2012 A1
20120095893 Taylor et al. Apr 2012 A1
20120109849 Chamberlain et al. May 2012 A1
20120110316 Chamberlain et al. May 2012 A1
20120116998 Indeck et al. May 2012 A1
20120117610 Pandya May 2012 A1
20120130922 Indeck et al. May 2012 A1
20120179590 Borkovec et al. Jul 2012 A1
20120215801 Indeck et al. Aug 2012 A1
20120216019 Bower et al. Aug 2012 A1
20120246052 Taylor et al. Sep 2012 A1
20130018835 Pandya Jan 2013 A1
20130151458 Indeck et al. Jun 2013 A1
20130290163 Parsons et al. Oct 2013 A1
20140180903 Parsons et al. Jun 2014 A1
20140180904 Parsons et al. Jun 2014 A1
20140180905 Parsons et al. Jun 2014 A1
20140181133 Parsons et al. Jun 2014 A1
20140310717 Chamberlain et al. Oct 2014 A1
20150023501 Taylor et al. Jan 2015 A1
20150052148 Indeck et al. Feb 2015 A1
20150055776 Taylor et al. Feb 2015 A1
20150317277 Wang et al. Nov 2015 A1
20160070583 Chamberlain et al. Mar 2016 A1
20160154867 Chen et al. Jun 2016 A1
20160328470 Indeck et al. Nov 2016 A1
20170102950 Chamberlain et al. Apr 2017 A1
20190139138 Parsons et al. May 2019 A1
20190304016 Parsons et al. Oct 2019 A1
20200111163 Parsons et al. Apr 2020 A1
Foreign Referenced Citations (62)
Number Date Country
0573991 Dec 1993 EP
0880088 Nov 1996 EP
0851358 Jul 1998 EP
0887723 Dec 1998 EP
0911738 Apr 1999 EP
09145544 Jun 1997 JP
09-269901 Oct 1997 JP
11-259559 Sep 1999 JP
11316765 Nov 1999 JP
2000286715 Oct 2000 JP
2001268071 Sep 2001 JP
2001283000 Oct 2001 JP
2002101089 Apr 2002 JP
2002117232 Apr 2002 JP
2002269343 Sep 2002 JP
2002352070 Dec 2002 JP
2003-036360 Feb 2003 JP
2003256660 Sep 2003 JP
2006293852 Oct 2006 JP
4180644 Nov 2008 JP
199010910 Sep 1990 WO
199409443 Apr 1994 WO
199737735 Oct 1997 WO
2000041136 Jul 2000 WO
2001022425 Mar 2001 WO
0135216 May 2001 WO
200172106 Oct 2001 WO
2001080082 Oct 2001 WO
2001080558 Oct 2001 WO
0190890 Nov 2001 WO
2002061525 Aug 2002 WO
2003100650 Apr 2003 WO
2003036845 May 2003 WO
2003100662 Dec 2003 WO
2003104943 Dec 2003 WO
2004017604 Feb 2004 WO
2004042560 May 2004 WO
2004042561 May 2004 WO
2004042562 May 2004 WO
2004042574 May 2004 WO
2005017708 Feb 2005 WO
2005026925 Mar 2005 WO
2005048134 May 2005 WO
2005081855 Sep 2005 WO
2005114339 Dec 2005 WO
2006023948 Mar 2006 WO
2006059312 Jun 2006 WO
2006060571 Jun 2006 WO
2006096324 Sep 2006 WO
2007016078 Feb 2007 WO
2007064685 Jun 2007 WO
2007074903 Jul 2007 WO
2007079095 Jul 2007 WO
2007087507 Aug 2007 WO
2007127336 Nov 2007 WO
2007149378 Dec 2007 WO
2008022036 Feb 2008 WO
2008036197 Mar 2008 WO
2008073824 Jun 2008 WO
2009089467 Jul 2009 WO
2009140363 Nov 2009 WO
2010077829 Jul 2010 WO
Non-Patent Literature Citations (198)
Entry
Schmerken, I. (2005). Direct from the source: Brokerage firm eSpeed equities taps direct ECN order book data feeds through a hosted solution. Wall Street & Technology, 23(12), 17(3). Retrieved from https://dialog.proquest.com/professional/docview/675556677?accountid=131444 on Oct. 22, 2020 (Year: 2005).
Cuppu and Jacob, “Organizational Design Trade-Offs at the DRAM, Memory Bus and Memory Controller Level: Initial Results,” Technical Report UMB-SCA-1999-2, Univ. of Maryland Systems & Computer Architecture Group, Nov. 1999, pp. 1-10.
Currid, “TCP Offload to the Rescue”, Networks, Jun. 14, 2004, 16 pages, vol. 2, No. 3.
Denoyer et al., “HMM-based Passage Models for Document Classification and Ranking”, Proceedings of ECIR-01, 23rd European Colloquim Information Retrieval Research, Darmstatd, DE, pp. 126-135, 2001.
Dharmapurikar et al., “Deep Packet Inspection Using Parallel Bloom Filters,” IEEE Micro, Jan.-Feb., 2004, vol. 24, Issue: 1, pp. 52-61.
Dharmapurikar et al., “Deep Packet Inspection Using Parallel Bloom Filters,” Symposium on High Performance Interconnects (Hotl), Stanford, California, 2003, pp. 44-51.
Dharmapurikar et al., “Design and Implementation of a String Matching System for Network Intrusion Detection using FPGA-based Bloom Filters”, Proc. of 12th Annual IEEE Symposium on Field Programmable Custom Computing Machines, 2004, pp. 1-10.
Dharmapurikar et al., “Longest Prefix Matching Using Bloom Filters,” SIGCOMM, 2003, pp. 201-212.
Dharmapurikar et al., “Robust TCP Stream Reassembly in the Presence of Adversaries”, Proc. of the 14th Conference on USENIX Security Symposium—vol. 14, 16 pages, Baltimore, MD, 2005; http://www.icir.org/vern/papers/TcpReassembly/TCPReassembly.pdf.
Dharmapurikar, “Fast and Scalable Pattern Matching for Content Filtering”, ACM, ANCS 05, 2005, pp. 183-192.
Ebeling et al., “RaPiD—Reconfigurable Pipelined Datapath”, University of Washington, Dept. of Computer Science and Engineering, Sep. 23, 1996, Seattle, WA.
Exegy Inc., “Exegy and HyperFeed to Unveil Exelerate TP at SIA Conference”, Release Date: Jun. 20, 2006, downloaded from http://news.thomasnet.com/companystory/488004 on Jun. 19, 2007, 4 pages.
Exegy Inc., “First Exegy Ticker Plant Deployed”, Release Date: Oct. 17, 2006, downloaded from http://news.thomasnet.com/companystory/496530 on Jun. 19, 2007, 5 pages.
Feldman, “High Frequency Traders Get Boost From FPGA Acceleration”, Jun. 8, 2007, downloaded from http://www.hpcwire.com/hpc.1600113.html on Jun. 19, 2007, 4 pages.
Franklin et al., “An Architecture for Fast Processing of Large Unstructured Data Sets.” Proc. of 22nd Int'l Conf. on Computer Design, Oct. 2004, pp. 280-287.
Franklin et al., “Assisting Network Intrusion Detection with Reconfigurable Hardware”, Symposium on Field-Programmable Custom Computing Machines (FCCM 2002), Apr. 2002, Napa, California.
Fu et al., “The FPX KCPSM Module: An Embedded, Reconfigurable Active Processing Module for the Field Programmable Port Extender (FPX)”, Washington University, Department of Computer Science, Technical Report WUCS-01-14, Jul. 2001.
Gaughan, “Data Streaming: Very Low Overhead Communication for Fine-Grained Multicomputing”, 1995, <http://dx.doi.org/ 10.1109/SPDP.1995.530727>, 3 pages.
Gavrila et al., “Multi-feature Hierarchical Template Matching Using Distance Transforms”, IEEE, Aug. 16-20, 1998, vol. 1, pp. 439-444.
Gokhale et al., “Reconfigurable Computing: Accelerating Computation With Field-Programmable Gate Arrays”, 2005, pp. 1-3, 7, 11-15, 39, 92-93, Springer.
Gokhale et al., “Reconfigurable Computing: Accelerating Computation with Field-Programmable Gate Arrays”, Springer, 2005, pp. 1-36.
Gokhale et al., “Reconfigurable Computing: Accelerating Computation with Field-Programmable Gate Arrays”, Springer, 2005, pp. 1-54, 92-96.
Google Search Results Page for “field programmable gate array financial calculation stock market” over dates of Jan. 1, 1990-May 21, 2002, 1 page.
Gunther et al., “Assessing Document Relevance with Run-Time Reconfigurable Machines”, IEEE Symposium on FPGAs for Custom Computing Machines, 1996, pp. 10-17, Proceedings, Napa Valley, CA.
Gupta et al., “High-Speed Implementations of Rule-Based Systems,” ACM Transactions on Computer Systems, May 1989, pp. 119-146, vol. 7, Issue 2.
Gupta et al., “PMM: A Parallel Architecture for Production Systems,” Proceedings of the IEEE, Apr. 1992, pp. 693-696, vol. 2.
Gyang, “NCBI BLASTN Stage 1 in Reconfigurable Hardware,” Technical Report WUCSE-2005-30, Aug. 2004, Department of Computer Science and Engineering, Washington University, St. Louis, MO.
Halaas et al., “A Recursive MISD Architecture for Pattern Matching”, IEEE Transactions on Very Large Scale Integration, vol. 12, No. 7, pp. 727-734, Jul. 2004.
Harris, “Pete's Blog: Can FPGAs Overcome the FUD?”, Low-Latency.com, May 14, 2007, URL: http://www.a-teamgroup.com/article/pete-blog-can-fpgas-overcome-the-fud/.
Hauck et al., “Software Technologies for Reconfigurable Systems”, Northwestern University, Dept. of ECE, Technical Report, 1996.
Hayes, “Computer Architecture and Organization”, Second Edition, 1988, pp. 448-459, McGraw-Hill, Inc.
Hezel et al., “FPGA-Based Template Matching Using Distance Transforms”, Proceedings of the 10th Annual IEEE Symposium on Field-Programmable Custom Computing Machines, Apr. 22, 2002, pp. 89-97, IEEE Computer Society, USA.
Hirsch, “Tech Predictions for 2008”, Reconfigurable Computing, Jan. 16, 2008, URL: http://fpgacomputing.blogspot.com/2008_01_01_archive_html.
Hoinville, et al. “Spatial Noise Phenomena of Longitudinal Magnetic Recording Media”, IEEE Transactions on Magnetics, vol. 28, No. 6, Nov. 1992.
Hollaar, “Hardware Systems for Text Information Retrieval”, Proceedings of the Sixth Annual International ACM Sigir Conference on Research and Development in Information Retrieval, Jun. 6-8, 1983, pp. 3-9, Baltimore, Maryland, USA.
Howe, Data Analysis for Database Design Third Edition, 2001, 335 pages, Butterworth-Heinemann.
Hutchings et al., “Assisting Network Intrusion Detection with Reconfigurable Hardware”, FCCM 2002: 10th Annual IEEE Symposium on Field-Programmable Custom Computing Machines, 2002.
Ibrahim et al., “Lecture Notes in Computer Science: Database and Expert Systems Applications”, 2000, p. 769, vol. 1873, Springer.
International Preliminary Report on Patentability (Chapter I) for PCT/US2009/067935 dated Jun. 30, 2011.
International Search Report and Written Opinion for PCT/US2009/067935 dated Apr. 30, 2010.
Jacobson et al., “RFC 1072: TCP Extensions for Long-Delay Paths”, Oct. 1988.
Jacobson et al., “tcpdump—dump traffic on a network”, Jun. 30, 1997, online at www.cse.cuhk.edu.hk/˜cslui/CEG4430/tcpdump.ps.gz.
Jones et al., “A Probabilistic Model of Information Retrieval: Development and Status”, Information Processing and Management, Aug. 1998, 76 pages.
Koloniari et al., “Content-Based Routing of Path Queries in Peer-to-Peer Systems”, pp. 1-19, E. Bertino et al. (Eds.): EDBT 2004, LNCS 2992, pp. 29-47, 2004, copyright by Springer-Verlag, Germany.
Krishnamurthy et al., “Biosequence Similarity Search on the Mercury System”, Proceedings of the 15th IEEE International Conference on Application-Specific Systems, Architectures, and Processors (ASAP04), Sep. 2004, pp. 365-375.
Office Action for CA Application 2820898 dated Aug. 20, 2018.
Office Action for EP Application 11847815.5 dated Dec. 21, 2018.
Prosecution History for U.S. Appl. No. 14/315,571, now U.S. Pat. No. 10,062,115, filed Jun. 26, 2014.
Roesch, “Snort—Lightweight Intrusion Detection for Networks”, Proceedings of LISA '99: 13th Systems Administration Conference; Nov. 7-12, 1999; pp. 229-238; USENIX Association, Seattle, WA USA.
Russ et al., Non-Intrusive Built-In Self-Test for FPGA and MCM Applications, Aug. 8-10, 1995, IEEE, 480-485.
Sachin Tandon, “A Programmable Architecture for Real-Time Derivative Trading”, Master's Thesis, University of Edinburgh, 2003.
Schmerken, “With Hyperfeed Litigation Pending, Exegy Launches Low-Latency Ticker Plant”, in Wall Street & Technology Blog, Mar. 20, 2007, pp. 1-2.
Schmit, “Incremental Reconfiguration for Pipelined Applications”, FPGAs for Custom Computing Machines, Proceedings, The 5th Annual IEEE Symposium, Dept. of ECE, Carnegie Mellon University, Apr. 16-18, 1997, pp. 47-55, Pittsburgh, PA.
Schuehler et al., “Architecture for a Hardware Based, TCP/IP Content Scanning System”, IEEE Micro, 24(1):62-69, Jan.-Feb. 2004, USA.
Schuehler et al., “TCP-Splitter: A TCP/IP Flow Monitor in Reconfigurable Hardware”, Hot Interconnects 10 (Hotl-10), Stanford, CA, Aug. 21-23, 2002, pp. 127-131.
Seki et al., “High Speed Computation of Shogi With FPGA”, Proceedings of 58th Convention Architecture, Software Science, Engineering, Mar. 9, 1999, pp. 1-133-1-134.
Shah, “Understanding Network Processors”, Version 1.0, University of California-Berkeley, Sep. 4, 2001.
Shalunov et al., “Bulk TCP Use and Performance on Internet 2”, ACM SIGCOMM Internet Measurement Workshop, 2001.
Shasha et al., “Database Tuning”, 2003, pp. 280-284, Morgan Kaufmann Publishers.
Shirazi et al., “Quantitative Analysis of FPGA-based Database Searching”, Journal of VLSI Signal Processing Systems for Signal, Image, and Video Technology, May 2001, pp. 85-96, vol. 28, No. 1/2, Kluwer Academic Publishers, Dordrecht, NL.
Sidhu et al., “Fast Regular Expression Matching Using FPGAs”, IEEE Symposium on Field Programmable Custom Computing Machines (FCCM 2001), Apr. 2001.
Sidhu et al., “String Matching on Multicontext FPGAs Using Self-Reconfiguration”, FPGA '99: Proceedings of the 1999 ACM/SIGDA 7th International Symposium on Field Programmable Gate Arrays, Feb. 1999, pp. 217-226.
Singh et al., “The EarlyBird System for Real-Time Detection on Unknown Worms”, Technical report CS2003-0761, Aug. 2003.
Skiena et al., “Programming Challenges: The Programming Contest Training Manual”, 2003, pp. 30-31, Springer.
Sourdis and Pnevmatikatos, “Fast, Large-Scale String Match for a 10Gbps FPGA-based Network Intrusion Detection System”, 13th International Conference on Field Programmable Logic and Applications, 2003.
Srini, “DFS-SuperMPx: Low-Cost Parallel Processing System for Machine Vision and Image Processing”, 1995, Retrieved from https://dialog.proquest.com/professional/docview/828015663?accountid=142257, 3 pages.
Steinbach et al., “A Comparison of Document Clustering Techniques”, KDD Workshop on Text Mining, 2000.
Sugawara et al., “Over 10Gbps String Matching Mechanism for Multi-Stream Packet Scanning Systems”, Field Programmable Logic and Application Lecture Notes in Computer Science, 2004, pp. 484-493, vol. 3203.
Tan et al., “A High Throughput String Matching Architecture for Intrusion Detection and Prevention”, ISCA 2005: 32nd Annual International Symposium on Computer Architecture, pp. 112-122, 2005.
Tau et al., “A First Generation DPGA Implementation”, Third Canadian Workshop of Field-Programmable Devices, May 1995, pp. 138-143.
Taylor et al., “Dynamic Hardware Plugins (DHP): Exploiting Reconfigurable Hardware for High-Performance Programmable Routers”, Computer Networks, 38(3): 295-310 (16), Feb. 21, 2002, and online at http://www.cc.gatech.edu/classes/AY2007/cs8803hpc_fall/papers/phplugins.pdf.
Taylor et al., “Generalized RAD Module Interface Specification of the Field Programmable Port Extender (FPX) Version 2”, Washington University, Department of Computer Science, Technical Report, Jul. 5, 2001, pp. 1-10.
Taylor et al., “Modular Design Techniques for the FPX”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Taylor et al., “Scalable Packet Classification using Distributed Crossproducting of Field Labels”, Proceedings of IEEE Infocom, Mar. 2005, pp. 1-12, vol. 20, No. 1.
Taylor, “Models, Algorithms, and Architectures for Scalable Packet Classification”, doctoral thesis, Department of Computer Science and Engineering, Washington University, St. Louis, MO, Aug. 2004, pp. 1-201.
Thomson Reuters, “Mellanox InfiniBand Accelerates the Exegy Ticker Plant at Major Exchanges”, Jul. 22, 2008, URL: http://www.reuters.com/article/pressRelease/idUS125385+22-Jul-2008+BW20080722.
Uluski et al., “Characterizing Antivirus Workload Execution”, SIGARCH Comput. Archit. News, vol. 33, No. 1, pp. 90-98, Mar. 2005.
U.S. Appl. No. 61/570,670, filed Dec. 14, 2011 (Taylor et al.).
Villasenor et al., “Configurable Computing Solutions for Automatic Target Recognition”, FPGAS for Custom Computing Machines, 1996, Proceedings, IEEE Symposium on Napa Valley, CA, Apr. 17-19, 1996, pp. 70-79, 1996 IEEE, Napa Valley, CA, Los Alamitos, CA, USA.
Waldvogel et al., “Scalable High-Speed Prefix Matching”, ACM Transactions on Computer Systems, Nov. 2001, pp. 440-482, vol. 19, No. 4.
Ward et al., “Dynamically Reconfigurable Computing: A Novel Computation Technology with Potential to Improve National Security Capabilities”, May 15, 2003, A White Paper Prepared by Star Bridge Systems, Inc. [retrieved Dec. 12, 2006]. Retrieved from the Internet: <URL: http://www.starbridgesystems.com/resources/whitepapers/Dynamically%20Reconfigurable%20Computing.pdf.
Weaver et al., “Very Fast Containment of Scanning Worms”, Proc. USENIX Security Symposium 2004, San Diego, CA, Aug. 2004, located at http://www.icsi.berkely.edu/˜nweaver/containment/containment.pdf.
Web-Pop (Professional Options Package) (www.pmpublishing.com).
West et al., “An FPGA-Based Search Engine for Unstructured Database”, Proc. of 2nd Workshop on Application Specific Processors, Dec. 2003, San Diego, CA.
Wooster et al., “HTTPDUMP Network HTTP Packet Snooper”, Apr. 25, 1996.
Worboys, “GIS: A Computing Perspective”, 1995, pp. 245-247, 287, Taylor & Francis Ltd.
Yamaguchi et al., “High Speed Homology Search with FPGAs”, Proceedings Pacific Symposium on Biocomputing, Jan. 3-7, 2002, pp. 271-282, vol. 7, Online, Lihue, Hawaii, USA.
Yan et al., “Enhancing Collaborative Spam Detection with Bloom Filters”, 2006, IEEE, pp. 414-425.
Yoshitani et al., “Performance Evaluation of Parallel Volume Rendering Machine Re Volver/C40”, Study Report of Information Processing Society, Mar. 5, 1999, pp. 79-84, vol. 99, No. 21.
Ziv et al., “A Universal Algorithm for Sequential Data Compression”, IEEE Trans. Inform. Theory, IT-23(3): 337-343 (1997).
“A Reconfigurable Computing Model for Biological Research Application of Smith-Waterman Analysis to Bacterial Genomes” A White Paper Prepared by Star Bridge Systems, Inc. [retrieved Dec. 12, 2006]. Retrieved from the Internet: <URL: http://www.starbridgesystems.com/resources/whitepapers/Smith%20 Waterman%20Whitepaper.pdf.
“ACTIV Financial Announces Hardware Based Market Data Feed Processing Strategy”, for Release on Apr. 2, 2007, 2 pages.
“ACTIV Financial Delivers Accelerated Market Data Feed”, Apr. 6, 2007, byline of Apr. 2, 2007, downloaded from http://hpcwire.com/hpc.1346816.html on Jun. 19, 2007, 3 pages.
“DRC, Exegy Announce Joint Development Agreement”, Jun. 8, 2007, byline of Jun. 4, 2007; downloaded from http://www.hpcwire.com/hpc/1595644.html on Jun. 19, 2007, 3 pages.
“Lucent Technologies Delivers “PayloadPlus” Network Processors for Programmable, MultiProtocol, OC-48c Processing”, Lucent Technologies Press Release, downloaded from http://www.lucent.com/press/1000/0010320.meb.html on Mar. 21, 2002.
“Overview, Field Programmable Port Extender”, Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002, pp. 1-4.
“Payload Plus™ Agere System Interface”, Agere Systems Product Brief, Jun. 2001, downloaded from Internet, Jan. 2002, pp. 1-6.
“RFC793: Transmission Control Protocol, Darpa Internet Program, Protocol Specification”, Sep. 1981.
“Technology Overview”, Data Search Systems Incorporated, downloaded from the http://www.datasearchsystems.com/tech.htm on Apr. 19, 2004.
“The Field-Programmable Port Extender (FPX)”, downloaded from http://www.arl.wustl.edu/arl/ in Mar. 2002.
Adachi, “Yoku Wakaru Kinyu Sakimono Torihiki” (guidance for financial futures trading), Nippon Jitsugyo Publishing, Aug. 30, 1997, pp. 113-115 and pp. 221-227.
Aldwairi et al., “Configurable String Matching Hardware for Speeding up Intrusion Detection”, SIRARCH Comput. Archit. News, vol. 33, No. 1, pp. 99-107, Mar. 2005.
Amanuma et al., “A FPGA Architecture for High Speed Computation”, Proceedings of 60th Convention Architecture, Software Science, Engineering, Mar. 14, 2000, pp. 1-163-1-164, Information Processing Society, Japan.
Anerousis et al., “Using the AT&T Labs PacketScope for Internet Measurement, Design, and Performance Analysis”, Network and Distributed Systems Research Laboratory, AT&T Labs-Research, Florham, Park, NJ, Oct. 1997.
Anonymous, “Method for Allocating Computer Disk Space to a File of Known Size”, IBM Technical Disclosure Bulletin, vol. 27, No. 10B, Mar. 1, 1985, New York.
Arnold et al., “The Splash 2 Processor and Applications”, Proceedings 1993 IEEE International Conference on Computer Design: VLSI in Computers and Processors (ICCD '93), Oct. 3, 1993, pp. 482-485, IEEE Computer Society, Cambridge, MA USA.
Artan et al., “Multi-packet Signature Detection using Prefix Bloom Filters”, 2005, IEEE, pp. 1811-1816.
Asami et al., “Improvement of DES Key Search on FPGA-Based Parallel Machine “Rash””, Proceedings of Information Processing Society, Aug. 15, 2000, pp. 50-57, vol. 41, No. SIG5 (HPS1), Japan.
Baboescu et al., “Scalable Packet Classification,” SIGCOMM'01, Aug. 27-31, 2001, pp. 199-210, San Diego, California, USA; http://www.ecse.rpi.edu/homepages/shivkuma/teaching/sp2001/readings/baboescu-pkt-classification.pdf.
Babz, “Empower Your Virtual Instruments”, Electronic Musician, 2005, 10(21), Retrieved from https://dialog.proquest.com/professional/docview/668423692?accountid=142257, 1 page.
Baer, “Computer Systems Architecture”, 1980, pp. 262-265; Computer Science Press, Potomac, Maryland.
Baeza-Yates et al., “New and Faster Filters for Multiple Approximate String Matching”, Random Structures and Algorithms (RSA), Jan. 2002, pp. 23-49, vol. 20, No. 1.
Baker et al., “High-throughput Linked-Pattern Matching for Intrusion Detection Systems”, ANCS 2005: Proceedings of the 2005 Symposium on Architecture for Networking and Communications Systems, pp. 193-202, ACM Press, 2005.
Baker et al., “Time and Area Efficient Pattern Matching on FPGAs”, ACM, Feb. 22-24, 2004, pp. 223-232.
Barone-Adesi et al., “Efficient Analytic Approximation of American Option Values”, Journal of Finance, vol. 42, No. 2 (Jun. 1987), pp. 301-320.
Batory, “Modeling the Storage Architectures of Commercial Database Systems”, ACM Transactions on Database Systems, Dec. 1985, pp. 463-528, vol. 10, issue 4.
Behrens et al., “BLASTN Redundancy Filter in Reprogrammable Hardware,” Final Project Submission, Fall 2003, Department of Computer Science and Engineering, Washington University.
Berk, “JLex: A lexical analyzer generator for Java™”, downloaded from http://www.cs.princeton.edu/˜appel/modern/java/Jlex/ in Jan. 2002, pp. 1-18.
Bianchi et al., “Improved Queueing Analysis of Shared Buffer Switching Networks”, ACM, Aug. 1993, pp. 482-490.
Bloom, “Space/Time Trade-offs in Flash Coding With Allowable Errors”, Communications of the ACM, Jul. 1970, pp. 122-426, vol. 13, No. 7, Computer Usage Company, Newton Upper Falls, Massachusetts, USA.
Braun et al., “Layered Protocol Wrappers for Internet Packet Processing in Reconfigurable Hardware”, Proceedings of Hot Interconnects 9 (Hotl-9) Stanford, CA, Aug. 22-24, 2001, pp. 93-98.
Braun et al., “Protocol Wrappers for Layered Network Packet Processing in Reconfigurable Hardware”, IEEE Micro, Jan.-Feb. 2002, pp. 66-74.
Brodie et al., “Dynamic Reconfigurable Computing”, in Proc. of 9th Military and Aerospace Programmable Logic Devices International Conference, Sep. 2006.
Cavnar et al., “N-Gram-Based Text Categorization”, Proceedings of SDAIR-94, 3rd Annual Symposium on Document Analysis and Information Retrieval, Las Vegas, pp. 161-175, 1994.
Celko, “Joe Celko's Data & Databases: Concepts in Practice”, 1999, pp. 72-74, Morgan Kaufmann Publishers.
Chamberlain et al., “Achieving Real Data Throughput for an FPGA Co-Processor on Commodity Server Platforms”, Proc. of 1st Workshop on Building Block Engine Architectures for Computers and Networks, Oct. 2004, Boston, MA.
Chamberlain et al., “The Mercury System: Embedding Computation Into Disk Drives”, 7th High Performance Embedded Computing Workshop, Sep. 2003, Boston, MA.
Chamberlain et al., “The Mercury System: Exploiting Truly Fast Hardware for Data Search”, Proc. of Workshop on Storage Network Architecture and Parallel I/Os, Sep. 2003, New Orleans, LA.
Cho et al., “Deep Packet Filter with Dedicated Logic and Read Only Memories”, 12th Annual IEEE Symposium on Field-Programmable Custom Computing Machines, Apr. 2004.
Choi et al., “Design of a Flexible Open Platform for High Performance Active Networks”, Allerton Conference, 1999, Champaign, IL.
Cholleti, “Storage Allocation in Bounded Time”, MS Thesis, Dept. of Computer Science and Engineering, Washington Univeristy, St. Louis, MO (Dec. 2002). Available as Washington University Technical Report WUCSE-2003-2.
Clark et al., “Scalable Pattern Matching for High Speed Networks”, Proceedings of the 12th Annual IEEE Symposium on Field-Programmable Custom Computing Machines, 2004; FCCM 2004, Apr. 20-23, 2004; pp. 249-257; IEEE Computer Society; Cambridge, MA USA.
Cloutier et al., “VIP: An FPGA-Based Processor for Image Processing and Neural Networks”, Proceedings of Fifth International Conference on Microelectronics for Neural Networks, Feb. 12, 1996, pp. 330-336, Los Alamitos, California.
Compton et al., “Configurable Computing: A Survey of Systems and Software”, Technical Report, Northwestern University, Dept. of ECE, 1999.
Compton et al., “Reconfigurable Computing: A Survey of Systems and Software”, Technical Report, Northwestern University, Dept. of ECE, 1999, presented by Yi-Gang Tai.
Compton et al., “Reconfigurable Computing: A Survey of Systems and Software”, University of Washington, ACM Computing Surveys, Jun. 2, 2002, pp. 171-210, vol. 34 No. 2, <http://www.idi.ntnu.no/emner/tdt22/2011/reconfig.pdf>.
Cong et al., “An Optional Technology Mapping Algorithm for Delay Optimization in Lookup-Table Based FPGA Designs”, IEEE, 1992, pp. 48-53.
Corbet et al., Linux Device Drivers: Where the Kernel Meets the Hardware, O'Reilly, Feb. 2005, pp. 19-20, 412-414, and 441, 3rd Edition.
Crosman, “Who Will Cure Your Data Latency?”, Storage & Servers, Jun. 20, 2007, URL: http://www.networkcomputing.com/article/printFullArticleSrc.jhtml?article ID=199905630.
Kunitomo, Database First Edition, Jun. 28, 2008, pp. 174-175, Juse Press, Ltd., Japan.
Lancaster et al., “Acceleration of Ungapped Extension in Mercury BLAST”, Seventh (7th) Workshop on Media and Streaming Processors, Nov. 12, 2005, Thirty-Eighth (38th) International Symposium on Microarchitecture (MICRO-38), Barcelona, Spain.
Li et al., “Large-Scale IP Traceback in High-Speed Internet: Practical Techniques and Theoretical Foundation”, Proceedings of the 2004 IEEE Symposium on Security and Privacy, 2004, pp. 1-15.
Lin et al., “Real-Time Image Template Matching Based on Systolic Array Processor”, International Journal of Electronics; Dec. 1, 1992; pp. 1165-1176; vol. 73, No. 6; London, Great Britain.
Lockwood et al., “Field Programmable Port Extender (FPX) for Distributed Routing and Queuing”, ACM International Symposium on Field Programmable Gate Arrays (FPGA 2000), Monterey, CA, Feb. 2000, pp. 137-144.
Lockwood et al., “Hello, World: A Simple Application for the Field Programmable Port Extender (FPX)”, Washington University, Department of Computer Science, Technical Report WUCS-00-12, Jul. 11, 2000.
Lockwood et al., “Parallel FPGA Programming over Backplane Chassis”, Washington University, Department of Computer Science, Technical Report WUCS-00-11, Jun. 12, 2000.
Lockwood et al., “Reprogrammable Network Packet Processing on the Field Programmable Port Extender (FPX)”, ACM International Symposium on Field Programmable Gate Arrays (FPGA 2001), Monterey, CA, Feb. 2001, pp. 87-93.
Lockwood, “An Open Platform for Development of Network Processing Modules in Reprogrammable Hardware”, IEC DesignCon 2001, Santa Clara, CA, Jan. 2001, Paper WB-19.
Lockwood, “Building Networks with Reprogrammable Hardware”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Evolvable Internet Hardware Platforms”, NASA/DoD Workshop on Evolvable Hardware (EHW'01), Long Beach, CA, Jul. 12-14, 2001, pp. 271-279.
Lockwood, “Hardware Laboratory Configuration”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Introduction”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Platform and Methodology for Teaching Design of Hardware Modules in Internet Routers and Firewalls”, IEEE Computer Society International Conference on Microelectronic Systems Education (MSE'2001), Las Vegas, NV, Jun. 17-18, 2001, pp. 56-57.
Lockwood, “Protocol Processing on the FPX”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Simulation and Synthesis”, Field Programmable Port Extender: Jan. 2002 Gigabit Workshop Tutorial, Washington University, St. Louis, MO, Jan. 3-4, 2002.
Lockwood, “Simulation of the Hello World Application for the Field-Programmable Port Extender (FPX)”, Washington University, Applied Research Lab, Spring 2001 Gigabits Kits Workshop.
Madhusudan, “Design of a System for Real-Time Worm Detection”, Hot Interconnects, pp. 77-83, Stanford, CA, Aug. 2004, found at http://www.hoti.org/hoti12/program/papers/2004/paper4.2.pdf.
Madhusudan, “Design of a System for Real-Time Worm Detection”, Power Point Presentation in Support of Master's Thesis, Washington Univ., Dept. of Computer Science and Engineering, St. Louis, MO, Aug. 2004.
Minutes of the Oral Proceedings for EP Patent Application No. 03729000.4 dated Jul. 12, 2010.
Mosanya et al., “A FPGA-Based Hardware Implementation of Generalized Profile Search Using Online Arithmetic”, ACM/Sigda International Symposium on Field Programmable Gate Arrays (FPGA '99), Feb. 21-23, 1999, pp. 101-111, Monterey, CA, USA.
Moscola et al., “FPGrep and FPSed: Regular Expression Search and Substitution for Packet Streaming in Field Programmable Hardware”, Dept. of Computer Science, Applied Research Lab, Washington University, Jan. 8, 2002, unpublished, pp. 1-19, St. Louis, MO.
Moscola et al., “FPSed: A Streaming Content Search-and-Replace Module for an Internet Firewall”, Proc. of Hot Interconnects, 11th Symposium on High Performance Interconnects, pp. 122-129, Aug. 20, 2003.
Moscola, “FPGrep and FPSed: Packet Payload Processors for Managing the Flow of Digital Content on Local Area Networks and the Internet”, Master's Thesis, Sever Institute of Technology, Washington University, St. Louis, MO, Aug. 2003.
Motwani et al., “Randomized Algorithms”, 1995, pp. 215-216, Cambridge University Press.
Mueller, “Upgrading and Repairing PCs, 15th Anniversary Edition”, 2004, pp. 63-66, 188, Que.
Navarro, “A Guided Tour to Approximate String Matching”, ACM Computing Surveys, vol. 33, No. 1, Mar. 2001, pp. 31-88.
Notice of Allowance for U.S. Appl. No. 11/932,652 dated Jan. 19, 2011.
Nunez et al., “The X-MatchLITE FPGA-Based Data Compressor”, Euromicro Conference 1999, Proceedings, Italy, Sep. 8-10, 1999, pp. 126-132, Los Alamitos, CA.
NVidia Developer Zone, “GPU Gems 2—Chapter 45, Options Pricing on the GPU”; Mar. 7-11, 2005, 20 pages.
Nwodoh et al., “A Processing System for Real-Time Holographic Video Computation”, Reconfigurable Technology: FPGAs for Computing and Application, Proceedings for the SPIE, Sep. 1999, Boston, pp. 129-140, vol. 3844.
Office Action for CA Application 2744746 dated Feb. 5, 2016.
Office Action for CA Application 2744746 dated Mar. 10, 2017.
Office Action for EP Application 9836816.0 dated Jan. 26, 2016.
Office Action for JP Application 2011-540957 dated Feb. 18, 2014.
Office Action for U.S. Appl. No. 13/132,408 dated Jun. 26, 2017.
Pramanik et al., “A Hardware Pattern Matching Algorithm on a Dataflow”; Computer Journal; Jul. 1, 1985; pp. 264-269; vol. 28, No. 3; Oxford University Press, Surrey, Great Britain.
Prosecution History for U.S. Appl. No. 13/076,968, filed Mar. 31, 2011 (Parsons et al.).
Prosecution History for U.S. Appl. No. 13/077,036, filed Mar. 31, 2011 (Parsons et al.).
Prosecution History for U.S. Appl. No. 13/132,408, filed Dec. 21, 2011 (Taylor et al.).
Prosecution History for U.S. Appl. No. 13/154,804, filed Jun. 7, 2011 (Taylor et al.).
Prosecution History for U.S. Appl. No. 13/154,824, filed Jun. 7, 2011, now U.S. Pat. No. 8,768,805 filed Jun. 7, 2015 (Taylor et al.).
Ramakrishna et al., “A Performance Study of Hashing Functions for Hardware Applications”, Int. Conf. on Computing and Information, May 1994, pp. 1621-1636, vol. 1, No. 1.
Ramakrishna et al., “Efficient Hardware Flashing Functions for High Performance Computers”, IEEE Transactions on Computers, Dec. 1997, vol. 46, No. 12.
Ratha et al., “Convolution on Splash 2”, Proceedings of IEEE Symposium on FPGAS for Custom Computing Machines, Apr. 19, 1995, pp. 204-213, Los Alamitos, California.
Response to Office Action for U.S. Appl. No. 13/132,408 dated Jun. 26, 2017.
“Accelerating High-Performance Computing with FPGAs”, Altera White Paper, pp. 1-8, Oct. 2007.
“Curriculum Vitae of Bernard S. Donefer” for Inter Partes Review Petition re U.S. Pat. No. 10,062,115; pp. 1-3; dated May 29, 2020.
“Data Records”, Newton's Telecom Dictionary, 21st Ed., 2009.
“Declaration of Bernard S. Donefer” for Inter Partes Review Petition re U.S. Pat. No. 10,062,115; pp. 1-145; dated May 29, 2020.
“Lava Trading Deploys Foundry”, Foundry Networks Press Release, pp. 1-3, Apr. 7, 2004.
“Petition for Inter Partes Review” re U.S. Pat. No. 10,062,115, pp. 1-92; dated May 29, 2020.
“Record”, Newton's Telecom Dictionary, 21st Ed., 2005.
“XDP Integrated Feed Client Specification”, Version 2.3a, pp. 1-31, Oct. 25, 2019.
Actel White Paper, “Flash FPGAs in the Value-Based Market”, 2005, 9 pages.
Gokhale et al., “Reconfigurable Computing: Accelerating Computation with Field-Programmable Gate Arrays”, Springer, 2005, 247 pages.
Summons to Attend Oral Proceedings for EP Application 09836816.0 dated May 14, 2019.
U.S. Appl. No. 60/814,796, filed Jun. 19, 2006.
“Decision Granting Institution of Inter Partes Review” for U.S. Pat. No. 10,062,115; IPR2020-01010, dated Dec. 7, 2020, pp. 1-30.
Related Publications (1)
Number Date Country
20180365766 A1 Dec 2018 US
Provisional Applications (1)
Number Date Country
61122673 Dec 2008 US
Divisions (2)
Number Date Country
Parent 13154804 Jun 2011 US
Child 14315571 US
Parent PCT/US2009/067935 Dec 2009 US
Child 13154804 US
Continuations (1)
Number Date Country
Parent 14315571 Jun 2014 US
Child 16111530 US