The present invention relates in general to concept and term scoring and clustering and, in particular, to a system and method for scoring concepts in a document set.
Large collections of documents have become increasingly available in electronically stored form due, in part, to the widespread adoption of computer-automated information and decision support systems. At the same time, electronically stored document collections have increasingly complemented and often supplanted traditional forms of printed communications. Electronically stored documents present several significant advantages over traditional printed formats, including efficient storage, rapid searchability, and facilitating immediate communication and publication over networking means, including the Internet.
From a pragmatic standpoint, the availability of electronically stored document collections has presented both a treasure and a curse to those seeking information discovery and retrieval. These types of document collections have expanded to include various forms of information classes, such as word processing documents, electronic mail, Worldwide Web (or simply “Web”) pages, spreadsheets, databases, and the like. And although now available in a highly searchable format, information embedded in documents stored in an electronic format must generally still be “mined” at a semantic level to discover and retrieve the data contained within. Mining out the semantic content of a document collection is essential to certain fields of endeavor, such as during the discovery phase of litigation. However, efficiently discovering and extracting such embedded semantic information can be an intractable problem, particularly when the size of the collection of documents is large.
Text mining is at the core of the information discovery process, and is described in D. Sullivan, “Document Warehousing and Text Mining, Techniques for Improving Business Operations, Marketing, and Sales,” Chs. 1-3, Wiley Computer Publishing (2001), the disclosure of which is incorporated by reference. Text mining involves the compiling, organizing and analyzing of document collections to support identification of types of information contained in the documents and to discover relationships between relevant facts. However, identifying relevant information can be difficult. First, extracting relevant content requires a high degree of precision and recall. Precision is the measure of how well the documents returned in response to a query actually address the query criteria. Recall is the measure of what should have been returned by the query. Typically, the broader and less structured the documents, the lower the degree of precision and recall. Second, analyzing an unstructured document collection without the benefit of a priori knowledge in the form of keywords and indices can present a potentially intractable problem space. Finally, synonymy and polysemy can cloud and confuse extracted content. Synonymy refers to multiple words having the same meaning and polysemy refers to a single word with multiple meanings. Fine-grained text mining must reconcile synonymy and polysemy to yield meaningful results.
Text mining is a significant first step in the overall process of discovering semantic meanings within a document collection. A further problem involves classifying the documents within a collection with respect to ad hoc categories of interest. For instance, during the discovery phase of litigation, documents must often be categorized into distinct groups, such as “relevant,” “non-relevant,” and “privileged.” Generally, the various documents falling into each group share certain characteristics, which can often be expressed as concepts and terms.
Similarly, categorizing the documents themselves into groups of related documents may be necessary as an aid to post-text mining document analysis. Text mining creates a multi-dimensional problem space that can be difficult to intuitively comprehend based on the presence of concepts and terms within the document collection overlapping by various degrees. Data visualization tools are available to display groups or “clusters” of documents, such as described in commonly-assigned U.S. Pat. No. 6,888,548, issued May 3, 2005, and U.S. Pat. No. 6,778,995, issued Aug. 17, 2004, and U.S. Pat. No. 7,271,804, issued Sep. 18, 2007, pending, the disclosures of which are incorporated by reference. Data visualization tools enable a user to rapidly comprehend and pare down the potential search field within a document collection, based on extracted concepts and terms.
In the prior art, text mining is performed in two ways. First, syntactic searching provides a brute force approach to analyzing and extracting content based on literal textual attributes found in each document. Syntactic searching includes keyword and proximate keyword searching as well as rule-based searching through Boolean relationships. Syntactic searching relies on predefined indices of keywords and stop words to locate relevant information. However, there are several ways to express any given concept. Accordingly, syntactic searching can fail to yield satisfactory results due to incomplete indices and poorly structured search criteria.
A more advanced prior art approach uses a vector space model to search for underlying meanings in a document collection. The vector space model employs a geometric representation of documents using word vectors. Individual keywords are mapped into vectors in multi-dimensional space along axes representative of query search terms. Significant terms are assigned a relative weight and semantic content is extracted based on threshold filters. Although substantially overcoming the shortcomings of syntactic searching, the multivariant and multidimensional nature of the vector space model can lead to a computationally intractable problem space. As well, the vector space model fails to resolve the problems of synonymy and polysemy.
Therefore, there is a need for an approach to identifying semantic information within a document collection based on extracted concepts and terms. Preferably, such an approach would assign a score to each concept and term based on the inherent characteristics of each document and the overall document set.
There is a further need for an approach to clustering documents within a document collection with respect to similarities reflected by the scores assigned to the concepts and terms. Preferably, such an approach would accept a set of candidate seed documents for evaluation and initial clustering.
The present invention provides a system and method for scoring and clustering documents based on extracted concepts and terms. Canonical concepts are formed from concepts and terms extracted from a set of documents and the frequencies of occurrences and reference counts of the concepts and terms are determined. Each concept and term is then scored based on frequency, concept weight, structural weight, and corpus weight. The scores are compressed and assigned to normalized score vectors for each of the documents. A similarity between each normalized score vector is determined, preferably as a cosine value. A set of candidate seed documents is evaluated to select a set of seed documents as initial cluster centers based on relative similarity between the assigned normalized score vectors for each of the candidate seed documents. The remaining non-seed documents are evaluated against the cluster centers also based on relative similarity and are grouped into clusters based oil a best fit, subject to a minimum fit criterion.
An embodiment provides a system and method for scoring concepts in a document set. A set of documents is maintained. Concepts including two or more terms extracted from the document set are identified. Each document having one or more of the concepts is designated as a candidate seed document. A score is calculated for each of the concepts identified within each candidate seed document based on a frequency of occurrence, concept weight, structural weight, and corpus weight. A vector is formed for each candidate seed document including the concepts located in that candidate seed document and the associated concept scores. The vector for each candidate seed document is compared with a center of one or more clusters each having thematically-related documents. At least one of the candidate seed documents that is sufficiently distinct from the other candidate seed documents is selected as a seed document for a new cluster. Each of the unselected candidate seed documents is placed into one of the clusters having a most similar cluster center.
Still other embodiments of the present invention will become readily apparent to those skilled in the art from the following detailed description, wherein are described embodiments of the invention by way of illustrating the best mode contemplated for carrying out the invention. As will be realized, the invention is capable of other and different embodiments and its several details are capable of modifications in various obvious respects, all without departing from the spirit and the scope of the present invention. Accordingly, the drawings and detailed description are to be regarded as illustrative in nature and not as restrictive.
The document analyzer 31 analyzes documents retrieved from a plurality of local sources. The local sources include documents 17 maintained in a storage device 16 coupled to a local server 15 and documents 20 maintained in a storage device 19 coupled to a local client 18. The local server 15 and local client 18 are interconnected to the production system 11 over an intranetwork 21. In addition, the document analyzer 31 can identify and retrieve documents from remote sources over an internetwork 22, including the Internet, through a gateway 23 interfaced to the intranetwork 21. The remote sources include documents 26 maintained in a storage device 25 coupled to a remote server 24 and documents 29 maintained in a storage device 28 coupled to a remote client 27.
The individual documents 17, 20, 26, 29 include all forms and types of structured and unstructured data, including electronic message stores, such as word processing documents, electronic mail (email) folders, Web pages, and graphical or multimedia data. Notwithstanding, the documents could be in the form of organized data, such as stored in a spreadsheet or database.
In the described embodiment, the individual documents 17, 20, 26, 29 include electronic message folders, such as maintained by the Outlook and Outlook Express products, licensed by Microsoft Corporation, Redmond, Wash. The database is an SQL-based relational database, such as the Oracle database management system, release 8, licensed by Oracle Corporation, Redwood Shores, Calif.
The individual computer systems, including production system 11, server 15, client 18, remote server 24 and remote client 27, are general purpose, programmed digital computing devices consisting of a central processing unit (CPU), random access memory (RAM), non-volatile secondary storage, such as a hard drive or CD ROM drive, network interfaces, and peripheral devices, including user interfacing means, such as a keyboard and display. Program code, including software programs, and data are loaded into the RAM for execution and processing by the CPU and results are generated for display, output, transmittal, or storage.
Document Analyzer
The scoring module 42 generates scores 52 for each of the concepts and terms, based on frequencies 53, concept weights 54, structural weights 55, and corpus weights 56, as further described below with reference to
The clustering module 43 forms clusters 58 of the documents 14 using the similarities of concepts and terms between the normalized score vectors 57, as further described below with reference to
The display and visualization module 44 complements the operations performed by the document analyzer 31 by presenting visual representations of the information extracted from the documents 14. The display and visualization module 44 generates a concept graph 61 of concept references determined over all documents 14, as further described below with reference to
Each module is a computer program, procedure or module written as source code in a conventional programming language, such as the C++ programming language, and is presented for execution by the CPU as object or byte code, as is known in the art. The various implementations of the source code and object and byte codes can be held on a computer-readable storage medium or embodied on a transmission medium in a carrier wave. The document analyzer 31 operates in accordance with a sequence of process steps, as further described below with reference to
Document Scoring
Document Clustering
Method Overview
As a preliminary step, the set of documents 14 to be analyzed is preprocessed (block 81) to identify terms and to extract concepts 65 and terms 66, as further described below with reference to
Document Parsing
The set of documents 14 maintained in the storage device 13 is processed in an iterative processing loop (blocks 91-99). During each iteration (block 91), each document 14 is retrieved from the storage device 13 and converted into a document record 48 (block 92) maintained in the database 30, as further described below with reference to
Preliminarily, each document 14 may be preprocessed (block 93) to remove extraneous formatting characters, such as hard returns or angle brackets, often used to embed previous email messages. Preprocessing maximizes syntactical extraction of desired terms and phrases without altering any semantic contents.
The global stop concept cache 45 contains a set of globally-applicable stop concepts used to suppress generic terms, such as “late,” “more,” “good,” or any user-defined stop concepts, which are suppressed to emphasize other important concepts in specific review contexts. In the described embodiment, the global stop concept cache 45 is generated dynamically after document analysis as document review progresses. Other forms of term and concept exclusion could be provided, as would be recognized by one skilled in the art.
Next, terms within the documents 14 are identified (block 94). Terms are defined on the basis of extracted noun phrases, although individual nouns or tri-grams (word triples) could be used in lieu of noun phrases. In the described embodiment, the noun phrases are extracted using the LinguistX product licensed by Inxight Software, Inc., Santa Clara, Calif. The identified phrases consist of regular nouns, as well as proper nouns or adjectives.
Next, the phrases are normalized (block 95) and used to identify canonical concepts (block 96). Unless indicated otherwise, the term “concepts” refers to canonical concepts as stored in a concept record 49 and applies equally to both concepts 65 and terms 66. Canonical concepts include the concepts 65 and terms 66 preferably processed into word stem form. In addition, the individual terms 66 comprising each concept 65 are converted to uniform lower case type and are alphabetized. By way of example, the sentence, “I went to the Schools of Business,” would yield the canonical concept “business, school.” Similarly, the sentence, “He went to Business School,” would yield the same canonical concept “business, school.” Other forms of canonical concepts could be used, including alternate word forms and arrangements, as would be recognized by one skilled in the art.
The canonical concepts are then used to build concept records 49 and term records 50 (block 97), as further described below with reference to
Document Record Schema
Concept Record Schema
Associated Concept Record Schema
Content Record Schema
Document Scoring Routine
As an initial step, each concept 56 and term 66 is individually scored (block 131), as further described below with reference to
For example, assume a normalized score vector 57 for a first document A is {right arrow over (S)}A={(5, 0.5), (120, 0.75)} and a normalized score vector 57 for another document B is {right arrow over (S)}B={(3, 0.4), (5, 0.75), (47, 0.15)}. Document A has scores corresponding to concepts ‘5’ and ‘120’ and Document B has scores corresponding to concepts ‘3,’ ‘5’ and ‘47.’ Thus, these documents only have concept ‘5’ in common.
An inner product of the normalized score vector 57 for the current document 14 is calculated against the normalized score vectors 57 of each other document 14 among corresponding dimensions (block 134) by iterating through the paired values in the normalized score vector 57 to identify commonly occurring concepts 65 and terms 66. Cosine cos σ is equivalent to the inner products between two normalized vectors. The cosine cos σ provides a measure of relative similarity or dissimilarity between the concepts 65 and terms 66 occurring in each document 14 and can therefore serve as a form of similarity metric, as would be recognized by one skilled in the art. In the described embodiment, the cosine cos σ is calculated in accordance with the equation:
where cos σAB comprises the similarity for between the document A and the document B, {right arrow over (S)}A comprises a score vector 57 for document A, and {right arrow over (S)}B comprises a score vector 57 for document B. Other forms of determining a relative similarity metric are feasible, as would be recognized by one skilled in the art. Processing continues with the next document 14 (block 135), after which the routine returns.
Concept and Term Scoring Routine
A score 52 is calculated for each concept 65 and term 66 in an iterative processing loop (block 141-147). During each iteration (block 141), a score 52 is calculated as follows. First, a concept weight 54 is determined for the concept 65 or term 66 (block 142). The concept weight 54 reflects the specificity of the meaning of a single concept 65 or term 66.
In the described embodiment, each concept weight 54 is based on the number of individual terms 66 that make up the concept 65 or term 66. Each concept weight 54 is calculated in accordance with the following equation:
where cwij comprises the concept weight and tij comprises a number of terms for occurrence j of each such concept i. The specificity of the meaning of a single concept 65 increases as the number of terms 66 occurring in the concept 65 increases. Intuitively, three to four terms 66 per concept 65 have proven more useful than other numbers of terms for differentiating between documents 14. Conversely, long concepts having in excess of five or more terms 66 tend to reflect parsing errors or are too specific for effective clustering.
Next, a structural weight 55 is determined for the concept 65 or term 66 (block 143). Each structural weight 55 reflects a varying degree of significance assigned to the concept 65 or term 66 based on structural location within the document 14. For example, subject lines in electronic mail (email) messages are assigned more importance than signature blocks.
In the described embodiment, each structural weight 55 is determined in accordance with the equation:
where swij comprises the structural weight for occurrence j of each such concept i. Other assignments of structural weight based on the location or arrangement of a concept 65 or term 66 occurrence within a document 14 are feasible, as would be recognized by one skilled in the art.
Next, a corpus weight is determined for the concept 65 or term 66 (block 144). The corpus weight 56 inversely weighs the reference count of the occurrences of each concept 65 or term 66 within a given document 14. The overall goal of forming clusters 58 is to group those documents 14 having similar content. Accordingly, the reference count of each concept 65 and term 66 can be used to differentiate document similarities. However, frequently referenced concepts 65 and terms 66 can dilute the differentiating measure of the reference counts and are ineffective in grouping similar documents. The reference counts for infrequently referenced concepts 65 and terms 66 also lack appreciable meaning as a differentiating measure except when evaluating clusters 58 for a small document set.
In the described embodiment, each corpus weight 56 is determined in accordance with the equation:
where rwij comprises the corpus weight, rij comprises a reference count for occurrence j of each such concept i, T comprises a total number of reference counts of documents in the document set, and M comprises a maximum reference count of documents in the document set. A value of 10% is used to indicate the maximum reference count at which a score contribution is discounted, although other limits could be used, as would be recognized by one skilled in the art.
Next, the actual score 52 for each concept 65 and term 66 is determined (block 145). Note each concept 65 and term 66 could occur one or more times within the same document 14 and could be assigned different structural weights 55 based on structural locations within the document 14. Each score 52 represents the relative weight afforded to each concept 65 and term 66 with respect to a particular document 14.
In the described embodiment, each score 52 is calculated in accordance with the equation:
where Si comprises the score 52, fij comprises the frequency 53, 0≦cwij≦1 comprises the concept weight 54, 0<swij≦1 comprises the structural weight 55, and 0<rwij≦1 comprises the corpus weight 56 for occurrence j of concept i within a given document 14. Finally, the score 52 is compressed (block 146) to minimize the skewing caused by concepts 65 and terms 66 occurring too frequently.
In the described embodiment, each compressed score is determined in accordance with the equation:
Si′=log(Si+1)
where Si′ comprises the compressed score 52 for each such concept i. Logarithmical compression provides effective linear vector representation of those documents 14 having a large body of content. Other forms of score compression could be used, as would be recognized by one skilled in the art.
Processing continues with the next concept 65 or term 66 (block 147), after which the routine returns.
Concept Reference Frequencies Graph
Cluster Forming Routine
The routine proceeds in two phases. During the first phase (blocks 161-169), seed candidate documents 60 are evaluated to identify a set of seed documents 59. During the second phase (blocks 170-176), non-seed documents 78 are evaluated and grouped into clusters 58 based on a best-fit criterion.
First, candidate seed documents 60 are identified (block 161) and ordered by category (block 162). In the described embodiment, the candidate seed documents 60 are selected based on a subjective evaluation of the documents 14 and are assigned into generalized categories, such as “responsive,” “non-responsive,” or “privileged.” Other forms of classification and categorization are feasible, as would be recognized by one skilled in the art.
Next, the candidate seed documents 60 are ordered within each category based on length (block 163). Each candidate seed document 60 is then processed in an iterative processing loop (blocks 164-169) as follows. The similarity between each current candidate seed document 60 and the cluster centers 58, based on seed documents already selected 59, is determined (block 165) as the cosine cos σ of the normalized score vectors 57 for the candidate seed documents 60 being compared. Only those candidate seed documents 60 that are sufficiently distinct from all cluster centers 58 (block 166) are selected as seed documents 59 (block 167). In the described embodiment, a range of 0.10 to 0.25 is used, although other ranges and spatial values could be used, as would be recognized by one skilled in the art.
If the candidate seed documents 60 being compared are not sufficiently distinct (block 166), the candidate seed document 60 is grouped into a cluster 58 with the most similar cluster center 58 to which the candidate seed document 60 was compared (block 168). Processing continues with the next candidate seed document 60 (block 169).
In the second phase, each non-seed document 78 is iteratively processed in an iterative processing loop (blocks 170-176) as follows. The non-seed documents 78 are simply those documents 14 other than the seed documents 60. Again, the similarity between each current non-seed document 78 and each of the cluster centers based on the seed documents 59 is determined (block 171) as the cosine cos σ of the normalized score vectors 57 for each of the non-seed documents 78. A best fit between the current non-seed document 78 and the cluster centers 58 is found subject to a minimum fit criterion (block 172). In the described embodiment, a minimum fit criterion of 0.25 is used, although other minimum fit criteria could be used, as would be recognized by one skilled in the art. If a best fit is found (block 173), the current non-seed document 78 is grouped into the cluster 58 having the best fit (block 175). Otherwise, the current non-seed document 78 is grouped into a miscellaneous cluster (block 174). Processing continues with the next non-seed document 78 (block 176). Finally, a dynamic threshold is applied to each cluster 58 (block 177), as further described below with reference to
Referring back to
Upon completion of the computation of similarity calculations for each document 202, the standard deviation of all documents 202 from the center 203 of the current cluster 201 is determined and a dynamic threshold 204 is set (block 185). In the described embodiment, a dynamic threshold 204 of ±1.2 standard deviations is used, although other dynamic thresholds 204 could also be used, as would be recognized by one skilled in the art. Next, those documents 202 in the current cluster 201, which are outside of the dynamic threshold 204, that is, outlier documents 205, are identified (block 186) and are processed in an iterative processing loop (blocks 187-193) as follows. The similarity between each outlier document 205 and each of the cluster centers is determined (block 188) based on the cosine cos σ of the normalized score vectors 57 for each of the outlier documents 205. A best fit between the outlier document 205 and the cluster centers is found subject to a minimum fit criterion and the dynamic threshold 204 (block 189). In the described embodiment, a minimum fit criterion of 0.25 is used, although other minimum fit criteria could be used, as would be recognized by one skilled in the art. The dynamic threshold 204 used to rescale each cluster-to-document similarity, which enables comparisons of similarities across all available clusters, is calculated in accordance with the equation:
where similaritynew comprises a new similarity, similarityold comprises the old similarity and threshold comprises the dynamic threshold 204.
If a best fit is found (block 190), the outlier document 205 is grouped into the cluster 58. Otherwise, the outlier document 205 is grouped into a miscellaneous cluster (block 191). Processing continues with the next outlier document 205 (block 192) and the next cluster 201 (block 193), after which the routine returns.
While the invention has been particularly shown and described as referenced to the embodiments thereof, those skilled in the art will understand that the foregoing and other changes in form and detail may be made therein without departing from the spirit and scope of the invention.
This patent application is a continuation of U.S. patent application Ser. No. 10/626,984, filed Jul. 25, 2003 now U.S. Pat. No. 7,610,313, issued on Oct. 27, 2009, the priority date of which is claimed and the disclosure of which is incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
3416150 | Lindberg | Dec 1968 | A |
3426210 | Agin | Feb 1969 | A |
3668658 | Flores et al. | Jun 1972 | A |
4893253 | Lodder | Jan 1990 | A |
5056021 | Ausborn | Oct 1991 | A |
5121338 | Lodder | Jun 1992 | A |
5133067 | Hara et al. | Jul 1992 | A |
5278980 | Pedersen et al. | Jan 1994 | A |
5371673 | Fan | Dec 1994 | A |
5442778 | Pedersen et al. | Aug 1995 | A |
5477451 | Brown et al. | Dec 1995 | A |
5488725 | Turtle et al. | Jan 1996 | A |
5524177 | Suzuoka | Jun 1996 | A |
5528735 | Strasnick et al. | Jun 1996 | A |
5619632 | Lamping et al. | Apr 1997 | A |
5619709 | Caid et al. | Apr 1997 | A |
5635929 | Rabowsky et al. | Jun 1997 | A |
5649193 | Sumita et al. | Jul 1997 | A |
5675819 | Schuetze | Oct 1997 | A |
5696962 | Kupiec | Dec 1997 | A |
5737734 | Schultz | Apr 1998 | A |
5754938 | Herz et al. | May 1998 | A |
5794178 | Caid et al. | Aug 1998 | A |
5794236 | Mehrle | Aug 1998 | A |
5799276 | Komissarchik et al. | Aug 1998 | A |
5819258 | Vaithyanathan et al. | Oct 1998 | A |
5835905 | Pirolli et al. | Nov 1998 | A |
5842203 | D'Elena et al. | Nov 1998 | A |
5844991 | Hochberg et al. | Dec 1998 | A |
5857179 | Vaithyanathan et al. | Jan 1999 | A |
5860136 | Fenner | Jan 1999 | A |
5862325 | Reed et al. | Jan 1999 | A |
5864846 | Voorhees et al. | Jan 1999 | A |
5864871 | Kitain et al. | Jan 1999 | A |
5867799 | Lang et al. | Feb 1999 | A |
5870740 | Rose et al. | Feb 1999 | A |
5895470 | Pirolli et al. | Apr 1999 | A |
5909677 | Broder et al. | Jun 1999 | A |
5915024 | Kitaori et al. | Jun 1999 | A |
5920854 | Kirsch et al. | Jul 1999 | A |
5940821 | Wical | Aug 1999 | A |
5943669 | Numata | Aug 1999 | A |
5950146 | Vapnik | Sep 1999 | A |
5950189 | Cohen et al. | Sep 1999 | A |
5966126 | Szabo | Oct 1999 | A |
5987446 | Corey et al. | Nov 1999 | A |
6006221 | Liddy et al. | Dec 1999 | A |
6012053 | Pant et al. | Jan 2000 | A |
6026397 | Sheppard | Feb 2000 | A |
6038574 | Pitkow et al. | Mar 2000 | A |
6070133 | Brewster et al. | May 2000 | A |
6089742 | Warmerdam et al. | Jul 2000 | A |
6094649 | Bowen et al. | Jul 2000 | A |
6100901 | Mohda et al. | Aug 2000 | A |
6119124 | Broder et al. | Sep 2000 | A |
6122628 | Castelli et al. | Sep 2000 | A |
6137499 | Tesler | Oct 2000 | A |
6137545 | Patel et al. | Oct 2000 | A |
6137911 | Zhilyaev | Oct 2000 | A |
6148102 | Stolin | Nov 2000 | A |
6154219 | Wiley et al. | Nov 2000 | A |
6167368 | Wacholder | Dec 2000 | A |
6173275 | Caid et al. | Jan 2001 | B1 |
6202064 | Julliard | Mar 2001 | B1 |
6216123 | Robertson et al. | Apr 2001 | B1 |
6243713 | Nelson et al. | Jun 2001 | B1 |
6243724 | Mander et al. | Jun 2001 | B1 |
6260038 | Martin et al. | Jul 2001 | B1 |
6326962 | Szabo | Dec 2001 | B1 |
6338062 | Liu | Jan 2002 | B1 |
6345243 | Clark | Feb 2002 | B1 |
6349296 | Broder et al. | Feb 2002 | B1 |
6349307 | Chen | Feb 2002 | B1 |
6360227 | Aggarwal et al. | Mar 2002 | B1 |
6363374 | Corston-Oliver et al. | Mar 2002 | B1 |
6377287 | Hao et al. | Apr 2002 | B1 |
6381601 | Fujiwara et al. | Apr 2002 | B1 |
6389433 | Bolonsky et al. | May 2002 | B1 |
6389436 | Chakrabarti et al. | May 2002 | B1 |
6408294 | Getchius et al. | Jun 2002 | B1 |
6414677 | Robertson et al. | Jul 2002 | B1 |
6415283 | Conklin | Jul 2002 | B1 |
6418431 | Mahajan et al. | Jul 2002 | B1 |
6421709 | McCormick et al. | Jul 2002 | B1 |
6438537 | Netz et al. | Aug 2002 | B1 |
6438564 | Morton et al. | Aug 2002 | B1 |
6442592 | Alumbaugh et al. | Aug 2002 | B1 |
6446061 | Doerre et al. | Sep 2002 | B1 |
6449612 | Bradley et al. | Sep 2002 | B1 |
6453327 | Nielsen | Sep 2002 | B1 |
6460034 | Wical | Oct 2002 | B1 |
6470307 | Turney | Oct 2002 | B1 |
6480843 | Li | Nov 2002 | B2 |
6480885 | Olivier | Nov 2002 | B1 |
6484168 | Pennock et al. | Nov 2002 | B1 |
6484196 | Maurille | Nov 2002 | B1 |
6493703 | Knight et al. | Dec 2002 | B1 |
6496822 | Rosenfelt et al. | Dec 2002 | B2 |
6502081 | Wiltshire, Jr. et al. | Dec 2002 | B1 |
6507847 | Fleischman | Jan 2003 | B1 |
6510406 | Marchisio | Jan 2003 | B1 |
6519580 | Johnson et al. | Feb 2003 | B1 |
6523026 | Gillis | Feb 2003 | B1 |
6523063 | Miller et al. | Feb 2003 | B1 |
6542889 | Aggarwal et al. | Apr 2003 | B1 |
6544123 | Tanaka et al. | Apr 2003 | B1 |
6549957 | Hanson et al. | Apr 2003 | B1 |
6560597 | Dhillon et al. | May 2003 | B1 |
6571225 | Oles et al. | May 2003 | B1 |
6584564 | Olkin et al. | Jun 2003 | B2 |
6594658 | Woods | Jul 2003 | B2 |
6598054 | Schuetze et al. | Jul 2003 | B2 |
6606625 | Muslea et al. | Aug 2003 | B1 |
6611825 | Billheimer et al. | Aug 2003 | B1 |
6628304 | Mitchell et al. | Sep 2003 | B2 |
6629097 | Keith | Sep 2003 | B1 |
6651057 | Jin et al. | Nov 2003 | B1 |
6654739 | Apte et al. | Nov 2003 | B1 |
6658423 | Pugh et al. | Dec 2003 | B1 |
6675159 | Lin et al. | Jan 2004 | B1 |
6675164 | Kamath et al. | Jan 2004 | B2 |
6678705 | Berchtold et al. | Jan 2004 | B1 |
6684205 | Modha et al. | Jan 2004 | B1 |
6697998 | Damerau et al. | Feb 2004 | B1 |
6701305 | Holt et al. | Mar 2004 | B1 |
6711585 | Copperman et al. | Mar 2004 | B1 |
6714929 | Micaelian et al. | Mar 2004 | B1 |
6735578 | Shetty et al. | May 2004 | B2 |
6738759 | Wheeler et al. | May 2004 | B1 |
6747646 | Gueziec et al. | Jun 2004 | B2 |
6751628 | Coady | Jun 2004 | B2 |
6757646 | Marchisio | Jun 2004 | B2 |
6785679 | Dane et al. | Aug 2004 | B1 |
6804665 | Kreulen et al. | Oct 2004 | B2 |
6816175 | Hamp et al. | Nov 2004 | B1 |
6819344 | Robbins | Nov 2004 | B2 |
6823333 | McGreevy | Nov 2004 | B2 |
6841321 | Matsumoto et al. | Jan 2005 | B2 |
6847966 | Sommer et al. | Jan 2005 | B1 |
6862710 | Marchisio | Mar 2005 | B1 |
6879332 | Decombe | Apr 2005 | B2 |
6883001 | Abe | Apr 2005 | B2 |
6886010 | Kostoff | Apr 2005 | B2 |
6888584 | Suzuki et al. | May 2005 | B2 |
6915308 | Evans et al. | Jul 2005 | B1 |
6922699 | Schuetze et al. | Jul 2005 | B2 |
6941325 | Benitez et al. | Sep 2005 | B1 |
6970881 | Mohan et al. | Nov 2005 | B1 |
6978274 | Gallivan et al. | Dec 2005 | B1 |
6978419 | Kantrowitz | Dec 2005 | B1 |
6990238 | Saffer et al. | Jan 2006 | B1 |
6996575 | Cox et al. | Feb 2006 | B2 |
7003551 | Malik | Feb 2006 | B2 |
7020645 | Bisbee et al. | Mar 2006 | B2 |
7051017 | Marchisio | May 2006 | B2 |
7054870 | Holbrook | May 2006 | B2 |
7080320 | Ono | Jul 2006 | B2 |
7096431 | Tambata et al. | Aug 2006 | B2 |
7099819 | Sakai et al. | Aug 2006 | B2 |
7117246 | Christenson et al. | Oct 2006 | B2 |
7130807 | Mikurak | Oct 2006 | B1 |
7137075 | Hoshito et al. | Nov 2006 | B2 |
7146361 | Broder et al. | Dec 2006 | B2 |
7155668 | Holland et al. | Dec 2006 | B2 |
7158957 | Joseph et al. | Jan 2007 | B2 |
7188107 | Moon et al. | Mar 2007 | B2 |
7188117 | Farahat et al. | Mar 2007 | B2 |
7194458 | Micaelian et al. | Mar 2007 | B1 |
7194483 | Mohan et al. | Mar 2007 | B1 |
7197497 | Cossock | Mar 2007 | B2 |
7209949 | Mousseau et al. | Apr 2007 | B2 |
7233940 | Bamberger et al. | Jun 2007 | B2 |
7240199 | Tomkow | Jul 2007 | B2 |
7246113 | Cheetham et al. | Jul 2007 | B2 |
7251637 | Caid et al. | Jul 2007 | B1 |
7266365 | Ferguson et al. | Sep 2007 | B2 |
7266545 | Bergman et al. | Sep 2007 | B2 |
7269598 | Marchisio | Sep 2007 | B2 |
7271801 | Toyozawa et al. | Sep 2007 | B2 |
7277919 | Donoho et al. | Oct 2007 | B1 |
7308451 | Lamping et al. | Dec 2007 | B1 |
7325127 | Olkin et al. | Jan 2008 | B2 |
7353204 | Liu | Apr 2008 | B2 |
7363243 | Arnett et al. | Apr 2008 | B2 |
7366759 | Trevithick et al. | Apr 2008 | B2 |
7379913 | Steele et al. | May 2008 | B2 |
7383282 | Whitehead et al. | Jun 2008 | B2 |
7401087 | Copperman et al. | Jul 2008 | B2 |
7412462 | Margolus et al. | Aug 2008 | B2 |
7418397 | Kojima et al. | Aug 2008 | B2 |
7444356 | Calistri-Yeh et al. | Oct 2008 | B2 |
7457948 | Bilicksa et al. | Nov 2008 | B1 |
7472110 | Achlioptas | Dec 2008 | B2 |
7490092 | Sibley et al. | Feb 2009 | B2 |
7523349 | Barras | Apr 2009 | B2 |
7571177 | Damle | Aug 2009 | B2 |
7607083 | Gong et al. | Oct 2009 | B2 |
7647345 | Trepess et al. | Jan 2010 | B2 |
7698167 | Batham et al. | Apr 2010 | B2 |
7971150 | Raskutti et al. | Jun 2011 | B2 |
20020002556 | Yoshida et al. | Jan 2002 | A1 |
20020032735 | Burnstein et al. | Mar 2002 | A1 |
20020065912 | Catchpole et al. | May 2002 | A1 |
20020078090 | Hwang et al. | Jun 2002 | A1 |
20020122543 | Rowen | Sep 2002 | A1 |
20020156763 | Marchisio | Oct 2002 | A1 |
20020184193 | Cohen | Dec 2002 | A1 |
20030130991 | Reijerse et al. | Jul 2003 | A1 |
20030172048 | Kauffman | Sep 2003 | A1 |
20040024755 | Rickard | Feb 2004 | A1 |
20040034633 | Rickard | Feb 2004 | A1 |
20040205578 | Wolff et al. | Oct 2004 | A1 |
20040215608 | Gourlay | Oct 2004 | A1 |
20040243556 | Ferrucci et al. | Dec 2004 | A1 |
20050004949 | Trepess et al. | Jan 2005 | A1 |
20070020642 | Deng et al. | Jan 2007 | A1 |
Number | Date | Country |
---|---|---|
1 024 437 | Aug 2000 | EP |
1024437 | Aug 2000 | EP |
1049030 | Nov 2000 | EP |
0886227 | Oct 2003 | EP |
0067162 | Nov 2000 | WO |
WO 03052627 | Jun 2003 | WO |
WO 03060766 | Jul 2003 | WO |
Entry |
---|
Artero et al., “Viz3D: Effective Exploratory Visualization of Large Multidimensional Data Sets,” IEEE Computer Graphics and Image Processing 2004, pp. 340-347 (Oct. 20, 2004). |
Wang et al., “Learning text classifier using the domain concept hierarchy,” Communications, Circuits and Systems and West Sino Expositions, IEEE 2002 International Conference on Jun. 29-Jul. 1, 2002, Piscataway, NJ, USA, IEEE, vol. 2, pp. 1230-1234. |
Vance Faber, “Clustering and the Continuous K-Means Algorithm,” Los Alamos Science, The Laboratory, Los Alamos, NM, US, No. 22, Jan. 1, 1994, pp. 138-144. |
D. Sullivan, “Document Warehousing And Text Mining, Techniques For Improving Business Operations, Marketing, And Sales,” Chs. 1-3, Wiley Computer Publishing (2001). |
Linhui, Jiang, “K-Mean Algorithm: Iterative Partitioning Clustering Algorithm,” http://www.cs.regina.ca/˜linhui/K—mean—algorithm.html, (2001) Computer Science Department, University of Regina, Saskatchewan, Canada. |
Kanugo et al., “The Analysis Of A Simple K-Means Clustering Algorithm,” pp. 100-109, PROC 16th annual symposium of computational geometry (May 2000). |
Pelleg et al., “Accelerating Exact K-Means Algorithms With Geometric Reasoning,” pp. 277-281, CONF on Knowledge Discovery in Data, PROC fifth ACM SIGKDD (1999). |
Jain et al., “Data Clustering: A Review,” vol. 31, No. 3, ACM Computing surveys, (Sep. 1999). |
F. Can., “Incremental Clustering For Dynamic Information Processing,” ACM Transactions On Information Systems, vol. 11, No. 1, pp. 143-164 (Apr. 1993). |
Christina Yip Chung et al., “Thematic Mapping—From Unstructured Documents To Taxonomies,” CIKM'02, Nov. 4-9, 2002, pp. 608-610, ACM, McLean, Virginia, USA. |
Hiroyuki Kawano, “Overview of Mondou Web Search Engine Using Text Mining And Information Visualizing Technologies,” IEEE, 2001, pp. 234-241. |
James Osborn et al., “JUSTICE: A Judicial Search Tool Using Intelligent Concept Extraction,” ICAIL-99, 1999, pp. 173-181, ACM. |
Chen An et al., “Fuzzy Concept Graph And Application In Web Document Clustering,” 2001, pp. 101-106, IEEE. |
Baeza-Yates et al., “Modern Information Retrieval,” Ch. 2 “Modeling,” Modern Information Retrieval, Harlow: Addison-Wesley, Great Britain 1999, pp. 18-71 (1999). |
Slaney et al., “Multimedia Edges: Finding Hierarchy in all Dimensions” PROC. 9-th ACM Intl. Conf. on Multimedia, pp. 29-40, ISBN.1-58113-394-4, Sep. 30, 2001, XP002295016 OTTAWA (Sep. 3, 2001). |
T. Kohonen, “Self-Organizing Maps,” Ch. 1-2, Springer-Verlag (3rd ed.) (2001). |
Anna Sachinopoulou, “Multidimensional Visualization,” Technical Research Centre of Finland, ESPOO 2001, VTT Research Notes 2114, pp. 1-37 (2001). |
Salton et al., “Estended Boolean Information Retrieval,” Communications of the Association for Computing Machinery, ACM, New York, NY, USA, vol. 26, No. 12, pp. 1022-1036, XP000670417 (Nov. 1, 1983). |
Jiang Linhui, “K-Mean Algorithm: Iterative Partitioning Clustering Algorithm,” http://www.cs.regina.ca/-linhui/K.sub.--mean.sub.--algorithm.html, (2001) Computer Science Department, University of Regina, Saskatchewan, Canada (2001). |
Kazumasa Ozawa, “A Stratificational Overlapping Cluster Scheme,” Information Science Center, Osaka Electro-Communication University, Neyagawa-shi, Osaka 572, Japan, Pattern Recognition, vol. 18, pp. 279-286 (1985). |
Bier et al. “Toolglass and Magic Lenses: The See-Through Interface”, Computer Graphics Proceedings, Proceedings of Siggraph Annual International Conference on Computer Graphics and Interactive Techniques, pp. 73-80, XP000879378 (Aug. 1993). |
Davison et al., “Brute Force Estimation of the Number of Human Genes Using EST Clustering as a Measure,” IBM Journal of Research & Development, vol. 45, pp. 439-447 (May 2001). |
Lam et al., “A Sliding Window Technique for Word Recognition,” SPIE, vol. 2422, pp. 38-46, Center of Excellence for Document Analysis and Recognition, State University of New Yrok at Baffalo, NY, USA (1995). |
Eades et al. “Multilevel Visualization of Clustered Graphs,” Department of Computer Science and Software Engineering, University if Newcastle, Australia, Proceedings of Graph Drawing '96, Lecture Notes in Computer Science, NR. 1190, Sep. 18, 1996-Sep. 20, 1996, pp. 101-112, Berkeley, CA, USA, ISBN: 3-540-62495-3 (Sep. 18, 1996). |
http://em-ntserver.unl.edu/Math/mathweb/vecors/vectors.html © 1997. |
North et al. “A Taxonomy of Multiple Window Coordinations,” Institute for Systems Research & Department of Computer Science, University of Maryland, Maryland, USA, http://www.cs.umd.edu/localphp/hcil/tech-reports-search.php?number=97-18 (1997). |
R.E. Horn, “Communication Units, Morphology, and Syntax,” Visual Language: Global Communication for the 21st Century, 1998, Ch. 3, pp. 51-92, MacroVU Press, Bainbridge Island, Washington, USA. |
B.B. Hubbard, “The World According the Wavelet: The Story of a Mathematical Technique in the Making,” AK Peters (2nd ed.), pp. 227-229, Massachusetts, USA (1998). |
Whiting et al., “Image Quantization: Statistics and Modeling,” SPIE Conference of Physics of Medical Imaging, San Diego, CA, USA , vol. 3336, pp. 260-271 (Feb. 1998). |
Miller et al., “Topic Islands: A Wavelet Based Text Visualization System,” Proceedings of the IEEE Visualization Conference. 1998, pp. 189-196. |
Fekete et al., “Excentric Labeling: Dynamic Neighborhood Labeling For Data Visualization,” CHI 1999 Conference Proceedings Human Factors In Computing Systems, Pittsburgh, PA, pp. 512-519 (May 15-20, 1999). |
Estivill-Castro et al. “Amoeba: Hierarchical Clustering Based On Spatial Proximity Using Delaunaty Diagram”, Department of Computer Science, The University of Newcastle, Australia, 1999 ACM Sigmod International Conference on Management of Data, vol. 28, No. 2, Jun. 1-Jun. 3, 1999 pp. 49-60, Philadelphia, PA, USA (Jun. 1, 1999). |
Lio et al., “Funding Pathogenicity Islands And Gene Transfer Events in Genome Data,” Bioinformatics, vol. 16, pp. 932-940, Department of Zoology, University of Cambridge, UK (Jan. 25, 2000). |
Kurimo M., “Fast Latent Semantic Indexing of Spoken Documents by Using Self-Organizing Maps” IEEE International Conference on Accoustics, Speech, And Signal Processing, vol. 6, pp. 2425-2428 (Jun. 2000). |
Magarshak, Greg., Theory & Practice. Issue 01. May 17, 2000. http://www.flipcode.com/articles/tp.sub.--issue01-pf.shtml (May 17, 2000). |
Strehl et al., “Cluster Ensembles—A Knowledge Reuse Framework for Combining Partitioning,” Journal of Machine Learning Research, MIT Press, Cambridge, MA, US, ISSN: 1533-7928, vol. 3, No. 12, pp. 583-617, XP002390603 (Dec. 2002). |
Bernard et al.: “Labeled Radial Drawing of Data Structures” Proceedings of the Seventh International Conference on Information Visualization, Infovis. IEEE Symposium, Jul. 16-18, 2003, Piscataway, NJ, USA, IEEE, Jul. 16, 2003, pp. 479-484, XP010648809, ISBN: 0-7695-1988-1 (Jul. 16, 2003). |
Rauber et al., “Text Mining in the SOMLib Digital Library System: The Representation of Topics and Genres,” Applied Intelligence 18, pp. 271-293, 2003 Kluwer Academic Publishers (2003). |
Boukhelifa et al., “A Model and Software System for Coordinated and Multiple Views in Exploratory Visualization,” Information Visualization, No. 2, pp. 258-269, GB (2003). |
Number | Date | Country | |
---|---|---|---|
20100049708 A1 | Feb 2010 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10626984 | Jul 2003 | US |
Child | 12606171 | US |