This application is the national phase under 35 U.S.C. § 371 of PCT International Application No. PCT/NO00/00155 which has an International filing date of May 10, 2000, which designated the United States of America.
The present invention concerns a search engine with two-dimensional linearly scalable parallel architecture for searching a collection of text documents D, wherein the documents can be divided into a number of partitions d1, d2, . . . dn, wherein the collection of documents D is preprocessed in a text filtration system such that a preprocessed document collection Dp is obtained and corresponding preprocessed partitions dp1, dp2, . . . dpn, wherein an index I can be generated from the document collection D such that for each previous preprocessed partition dp1, dp2, . . . dpn a corresponding index i1, i2, . . . in is obtained, wherein searching a partition d of the document collection D takes place with a partition-dependent data set dp,k comprising both the preprocessed partition dpk and the corresponding index ik, with 1≦k≦n, and wherein the search engine comprises data processing units which form sets of nodes connected in a network.
Most prior art search engines work with large data set and employ powerful computers to perform the search. However, searching is a partitionable data processing problem, and this fact can be used to partition a search problem into a large number of specific queries and let each query be processed simultaneously on a commensurate number of processors connected in parallel in a network. Particularly searching can be regarded as a binary partitionable data processing problem, and hence a binary tree network is used for establishing a multiprocessor architecture such as disclosed for instance in U.S. Pat. No. 4,860,201 (Stolfo & al.) and international patent application PCT/NO99/00308 which belongs to the applicant and hereby is incorporated by reference. The present applicant has developed proprietary technologies for searching within regular text documents. These technologies are i.a. based on a search system and a method for searching as described in international patent application PCT/NO99/00233 which belongs to the applicant and hereby is incorporated by reference. The search system is based on efficient core search algorithms which may be used in the search engine according to the invention.
However, it has become increasingly important to cater for a growing number of documents to be searched and also to be able to handle an increased traffic load, i.e. the number of queries per second which shall be processed by the search system. This, apart from the ability to handle a large number of queries simultaneously on processor level, implies that a search engine should be implemented with an architecture that allows for preferably linear scalability in two dimensions, viz. both with regard to the data volume and the performance, i.e. the ability to handle a very large number of queries per second. Considering the development of the World Wide Web, a scalability problem in the search engine architecture will be extremely important as there presently is an enormous growth rate in both the number of documents and the number of users on the Internet.
Prior art search engine solutions for Internet are able to scale to a certain level, but almost always this is achieved in a manner that requires a high cost increase of the search engine system relative to the growth in data volume or data traffic. Very often the system costs scale as the square of the data volume or the traffic, a doubling of the data volume thus leading to quadrupled system costs. Furthermore all the major Internet search engines presently are based on very expensive server technology, often coupled with brute computing force-approaches and accompanied with disadvantages such as slow server turnaround, requirements for special hardware to provide fault tolerance etc. The system costs can e.g. be measured as the amount of hardware required to implement a search engine solution or the actual aggregated price of the system.
It is thus a main object of the present invention to provide a search engine with a multilevel data and functional parallelism, such that large volumes of data can be searched efficiently and very fast by a large number of users simultaneously.
Particularly it is a further object of the invention to provide a parallel architecture for implementing a search engine with a multilevel data and functional parallelism.
Yet a further object of the present invention is to provide a parallel architecture which is linearly scalable in two dimensions, i.e. with regard to both data volume and performance, that is the query rate.
The above-mentioned objects and further features and advantages are provided with a search engine according to the invention which is characterized in that the first set of nodes comprises a dispatch nodes, a second set of nodes comprises b search nodes, a third set of nodes comprises g indexing nodes, and an optional fourth set of nodes comprises e acquisition nodes, that the dispatch nodes are connected in a multilevel configuration in the network, that the search nodes are grouped in columns which are connected in parallel in the network between the dispatch nodes and an indexing node, that the dispatch nodes are adapted to process search queries and search answers, the search queries being dispatched further to all search nodes and in case the acquisition nodes are not present, the search answers being returned to the dispatch nodes and therein being combined to a final search result, that the search nodes each are adapted to contain search software, that at least some of the search nodes additionally comprise at least one search processor module, that the indexing nodes are adapted for generally generating indexes i for the search software and optionally for generating partition-dependent data sets dp,k to search nodes which comprise a search processor module, that in case acquisition nodes are present, these are connected in a multilevel configuration in the network similar to that of the dispatch node, and adapted for gathering answers to search queries and outputting a final result thereof, thus relieving the dispatch nodes of this task, and that the two-dimensional linear scaling respectively takes place by scaling of the data volume through an increase in the number of partitions d and scaling of performance through replication of one or more partitions d.
According to the invention are advantageously the multilevel configuration of the dispatch nodes and the optional acquisition nodes network provided by hierarchical tree structures, and the multilevel configuration of the optional acquisition nodes is then preferably a mirror image of the multilevel configuration of the dispatch nodes, the hierarchical tree structures preferably being binary tree structures.
According to the invention each of the search nodes advantageously comprises a search software module.
Further, according to the invention at least some of the search nodes comprises at least one dedicated search processor module, each dedicated search processor module being realized with one or more dedicated processor chips, which is adapted for parallel handling of a number of search queries. In this connection it is preferred that the dedicated search processor chips are provided in the search processor modules in y processor groups, each with z search processor chips and being connected with and adapted to receive data from a memory assigned to the processor group.
According to the invention the increase in the number of partitions in the scaling of the data volume is advantageously implemented by an increase in the number of search node groups or columns. In this connection the increase in the number of partitions can preferably be accompanied by a corresponding increase in the number of dispatch nodes and, in case, also in the number of acquisition nodes, and optionally also by an increase in the number of index nodes.
According to the invention the replication of one or more partition in the scaling of performance is advantageously implemented by an increase in the number of search nodes in each group or column.
Finally it is according to the invention advantageous that the separate node sets each is implemented over one or more workstations connected in a data communications network.
The search engine according to the invention shall now be described in terms of non-limiting exemplary embodiments and with reference to the accompanying drawings, in which
Searching a large collection of independent documents is a highly parallel task. The search engine according to the invention employs parallelism on different levels as shall be discussed in the following.
The search engine according to the present invention searches a document collection of documents D. The documents can be divided into n partitions d1, d2, . . . , dn. Each document collection D, or partition d of a document collection can be preprocessed for use in a hardware text filtering system, for instance implemented by dedicated hardware like the applicant's so-called Pattern Matching Chip (PMC) which is disclosed in the applicant's international patent application No. PCT/NO99/00344 which hereby is incorporated by reference. The preprocessed document collection is denoted Dp and the corresponding preprocessed document collection partitions dp are denoted dp1, dp2, . . . , dpn.
Software-based search systems require an index generated from the document collection. The index is denoted I and the indexes corresponding to the document collection partitions dp1, dp2, . . . dpn, are denoted i1, i2, . . . in.
The data set needed for searching a partition d of the document collection D, is called the partition-required or partition-dependent data set. In a software only system (SW system) the data set is the index ik, while in the systems with hardware (SW/HW systems), the data set also includes the preprocessed document collection partition dp,k with the corresponding index ik, where 1≦k≦n.
The essentially software-implemented partitioning and preprocessing operations can be rendered schematically as *(D)→*(d1, . . . , dn)→*(dp1, . . . , dpn)→dpk, where *(D) denotes a partitioning operation on the input D, *(d1, . . . , dn) a filtering operation, e.g. indexing, on d1, . . . , dn and dp,k of course is the partition-dependent data set, which in an SW system only shall be the index ik, and with 1≦k≦n.
A search engine is implemented on a cluster of workstations that are connected using a high performance interconnect bus. The not shown workstations then constitute the server of the search system (search server). The workstations implement nodes of the search server. The nodes perform different tasks and are according to the invention implemented as set out below.
The nodes can be regarded as virtual nodes distributed among the workstations, but in a SW/HW search engine the dedicated search processing hardware must be physically present in some workstations in order to support the hardware-based search nodes. The search node software may then still be distributed. Also, some search nodes in SW/HW search engine may comprise software only, and optionally be distributed over more than one workstation.
A first embodiment of the search engine according to the invention is shown in
A second embodiment of the search engine according to the invention is shown in
It is to be understood that singular workstations may implement a specific type of nodes only, on alternatively more than one type of nodes. In other words, the different types of nodes may be distributed over the cluster of workstations. Hence the architecture shown in
The nodes shall now be discussed in more detail, starting with the search nodes which are central to the search engine according to the invention.
A search node Nβ holds as mentioned a portion of the entire data set dp. The search node has both a software search engine SW, and optionally a number of PMC modules M, as shown in
A search node may be equipped with a number x of PMC modules M for very fast searching, as shown in
A pattern matching chip PMC can process a data volume of tc bytes per second. Assuming that the memory modules are capable of delivering ty bytes per second to the pattern matching chips PMCs, a PMC can search through the data volume of Tc bytes, Tc=min{tc, ty}t, in the given time t.
As shown in
When x modules M are provided in a search node Nβ, these PMC modules M can search through an amount of data equal to Tr=Ty·x=min{tc, ty}txy—since no PMC modules search through the same data, the number of concurrent queries is still zq.
Thus the total query rate of the PMC modules in a search node can be expressed as
where Tr denotes the total data volume on a node. The search node performance can now be calculated.
Given that the PMC modules N (or any hardware equivalent) has a query rate of rHW and that the search software on a search node Nβ has a query rate of rSW, the total query rate rΣ of a search node Ns can be expressed as
rΣ=rt,HW(1−φSW)+rt,SWφSW (2)
where φSW denotes the percentage of queries q that will be executed in software. The actual value of φSW is dynamically updated at runtime from a statistical model.
The dispatch nodes Nα receive all the queries, and resend them to all the search nodes Nβ. The answers from the different search nodes Nβ are merged and in case the dispatch nodes Nα functions as acquisition nodes, a complete answer is returned.
The indexing nodes Nγ collect documents and create prebuilt indexes for the search software on the different search nodes Nβ. Hence the indexing nodes Nγ can be incorporated in the search nodes Nβ with appropriate indexing software in the latter. The hardware is based on scanning through the entire collection of raw data, but some preprocessing and filtering of the raw data can be done in the indexing nodes Nγ.
Concerning the interconnect and data traffic, some general observations can be made based on the following considerations.
Different types of interconnect can be used for connecting the nodes. For a lower end system, a regular 100 Mbit Fast Ethernet will for instance handle the traffic.
Traffic on the interconnect between the nodes can be divided into two categories:
A typical query will transfer a query string from the dispatch node Nα to the search nodes Nβ. Then the search nodes Nβ will reply with a sequence of documents matching the query. Optionally Nα shall also be able to query the search node for the URL strings for the document, but this is considered immaterial in the present context.
The architecture of the search engine according to the invention can, based on the above considerations, now easily be scaled in two dimensions, viz. the data volume and the performance dimensions respectively,
Data volume scaling is achieved by adding more data set partitions d, in other words more groups or columns S of search nodes Nβ are added. Also the number of indexing nodes Nγ and dispatch nodes Nα can be increased as necessary in order to handle more data set partitions d.
Performance scaling can be achieved in the search engine architecture by replicating data set partitions d with a corresponding increase in the number of search nodes Nβ, by increasing the number as illustrated in
hs denoting the scaling factor. The group Sβj contains the search nodes Nβj, Nβj+1, Nβj+2 and Nβj+3 as rendered in
Scaling the data volume may cause the number of search nodes Nβ receiving queries broadcast from a dispatch node Nα growing quite large. The architecture solves this problem by using several levels λ of dispatch nodes Nα—this is illustrated in
The schematic layout of a scalable search engine architecture according to the invention is shown in
An additional important benefit of a search engine according to the invention with the architecture scalable as herein disclosed, is that the query response time is essentially independent of the catalogue size, as each query is executed in parallel on all search nodes Nβ and that the architecture is inherently fault tolerant, such that faults in the individual nodes will not result in a system breakdown, only temporary reduce the performance until the fault is corrected.
Moreover, the in principle unlimited linear scalability of the data volume and the traffic volume which can be provided in a search engine according to the invention, contrasts sharply with prior art search engines, wherein the search cost typically increases exponentially with the data or traffic volume increase, and wherein the maximum capacity of the prior art search engines typically will be reached at low to moderate volumes. With the search engine according to the invention the cost will scale linearly with the increase in capacity at most, depending actually on whether the capacity increase is provided by adding SW search nodes only or also SW/HW search nodes. Finally the search engine according to the invention offers the advantage that each node in practice can be implemented with standard low cost commercially available PCs, but alternatively also with more expensive UNIX-based servers such as for instance the Sun or Alpha computers as currently available.
Number | Date | Country | Kind |
---|---|---|---|
19992269 | May 1999 | NO | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/NO00/00155 | 5/10/2000 | WO | 00 | 2/16/2001 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO00/68834 | 11/16/2000 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
4860201 | Stolfo et al. | Aug 1989 | A |
5388214 | Leiserson et al. | Feb 1995 | A |
5694593 | Baclawski | Dec 1997 | A |
5717912 | Millett et al. | Feb 1998 | A |
5742806 | Reiner et al. | Apr 1998 | A |
5878409 | Baru et al. | Mar 1999 | A |
5960194 | Choy et al. | Sep 1999 | A |
5978791 | Farber et al. | Nov 1999 | A |
6026394 | Tsuchida et al. | Feb 2000 | A |
6112198 | Lohman et al. | Aug 2000 | A |
6370527 | Singhal | Apr 2002 | B1 |
6424970 | Arakawa et al. | Jul 2002 | B1 |
6549519 | Michels et al. | Apr 2003 | B1 |
6567806 | Tsuchida et al. | May 2003 | B1 |
20010011268 | Tsuchida et al. | Aug 2001 | A1 |