This application relates in general to using electronically stored information as a reference point and, in particular, to a system and method for displaying relationships between electronically stored information to provide classification suggestions via injection.
Historically, document review during the discovery phase of litigation and for other types of legal matters, such as due diligence and regulatory compliance, have been conducted manually. During document review, individual reviewers, generally licensed attorneys, are assigned sets of documents for coding. A reviewer must carefully study each document and categorize the document by assigning a code or other marker from a set of descriptive classifications, such as “privileged,” “responsive,” and “non-responsive.” The classifications affect the disposition of each document, including admissibility into evidence. During discovery, document review can potentially affect the outcome of the underlying legal matter, so consistent and accurate results are crucial.
Manual document review is tedious and time-consuming. Marking documents is solely at the discretion of each reviewer and inconsistent results may occur due to misunderstanding, time pressures, fatigue, or other factors. A large volume of documents reviewed, often with only limited time, can create a loss of mental focus and a loss of purpose for the resultant classification. Each new reviewer also faces a steep learning curve to become familiar with the legal matter, coding categories, and review techniques.
Currently, with the increasingly widespread movement to electronically stored information (ESI), manual document review is no longer practicable. The often exponential growth of ESI exceeds the bounds reasonable for conventional manual human review and underscores the need for computer-assisted ESI review tools.
Conventional ESI review tools have proven inadequate to providing efficient, accurate, and consistent results. For example, DiscoverReady LLC, a Delaware limited liability company, conducts semi-automated document review through multiple passes over a document set in ESI form. During the first pass, documents are grouped by category and basic codes are assigned. Subsequent passes refine and further assign codings. Multiple pass review also requires a priori project-specific knowledge engineering, which is useful for only the single project, thereby losing the benefit of any inferred knowledge or know-how for use in other review projects.
Thus, there remains a need for a system and method for increasing the efficiency of document review that bootstraps knowledge gained from other reviews while ultimately ensuring independent reviewer discretion.
Document review efficiency can be increased by identifying relationships between reference ESI and uncoded ESI and providing a suggestion for classification based on the relationships. A set of clusters including uncoded ESI is obtained. The uncoded ESI for a cluster are compared to a set of reference ESI. Those reference ESI most similar to the uncoded ESI are identified and inserted into the cluster. The relationship between the inserted reference ESI and uncoded ESI for the cluster are visually depicted and provide a suggestion regarding classification of the uncoded ESI.
An embodiment provides a system and method for identifying relationships between electronically stored information to provide a classification suggestion via injection. A reference set of electronically stored information items, each associated with a classification code, is designated. Clusters of uncoded electronically stored information items are designated. One or more of the uncoded electronically stored information items from at least one cluster is compared to the reference set. At least one of the electronically stored information items in the reference set is identified as similar to the one or more uncoded electronically stored information items. The similar electronically stored information items are injected into the at least one cluster. Relationships are visually depicted between the uncoded electronically stored information items and the similar electronically stored information items in the at least one cluster as suggestions for classifying the uncoded electronically stored information items.
A further embodiment provides a system and method for injecting reference documents into a cluster set as suggestions for classifying uncoded documents. A set of reference documents, each associated with a classification code, is designated. Clusters of uncoded reference documents are also designated. One or more of the uncoded documents in at least one of the clusters are compared with the reference document set. The reference documents that satisfy a similarity threshold with the one or more uncoded documents are selected. The selected reference documents are injected into the at least one cluster. Relationships are displayed between the uncoded documents and the selected reference documents in the at least one cluster as suggestions for classifying the uncoded documents.
Still other embodiments of the present invention will become readily apparent to those skilled in the art from the following detailed description, wherein are described embodiments by way of illustrating the best mode contemplated for carrying out the invention. As will be realized, the invention is capable of other and different embodiments and its several details are capable of modifications in various obvious respects, all without departing from the spirit and the scope of the present invention. Accordingly, the drawings and detailed description are to be regarded as illustrative in nature and not as restrictive.
The ever-increasing volume of ESI underlies the need for automating document review for improved consistency and throughput. Previously classified ESI offer knowledge gleaned from earlier work in similar legal projects, as well as a reference point for classifying uncoded ESI.
Providing Suggestions Using Reference Documents
Reference ESI is previously classified by content and can be injected into clusters of uncoded, that is unclassified. ESI to influence classification of the uncoded ESI. Specifically, relationships between an uncoded ESI and the reference ESI in terms of semantic similarity or distinction can be used as an aid in providing suggestions for classifying uncoded ESI.
Complete ESI review requires a support environment within which classification can be performed.
The backend server 11 is coupled to an intranetwork 21 and executes a workbench software suite 31 for providing a user interface framework for automated document management, processing, analysis, and classification. In a further embodiment, the backend server 11 can be accessed via an internetwork 22. The workbench software suite 31 includes a document mapper 32 that includes a clustering engine 33, similarity searcher 34, classifier 35, and display generator 36. Other workbench suite modules are possible.
The clustering engine 33 performs efficient document scoring and clustering of uncoded documents, such as described in commonly-assigned U.S. Pat. No. 7,610,313, the disclosure of which is incorporated by reference. Clusters of uncoded documents 14 can be organized along vectors, known as spines, based on a similarity of the clusters. The similarity can be expressed in terms of distance. Document clustering is further discussed below with reference to
The document mapper 32 operates on uncoded documents 14a, which can be retrieved from the storage 13, as well as a plurality of local and remote sources. The local and remote sources can also store the reference documents 14b. The local sources include documents 17 maintained in a storage device 16 coupled to a local server 15 and documents 20 maintained in a storage device 19 coupled to a local client 18. The local server 15 and local client 18 are interconnected to the backend server 11 and the work client 12 over the intranetwork 21. In addition, the document mapper 32 can identify and retrieve documents from remote sources over the internetwork 22, including the Internet, through a gateway 23 interfaced to the intranetwork 21. The remote sources include documents 26 maintained in a storage device 25 coupled to a remote server 24 and documents 29 maintained in a storage device 28 coupled to a remote client 27. Other document sources, either local or remote, are possible.
The individual documents 14a, 14b, 17, 20, 26, 29 include all forms and types of structured and unstructured ESI, including electronic message stores, word processing documents, electronic mail (email) folders, Web pages, and graphical or multimedia data. Notwithstanding, the documents could be in the form of structurally organized data, such as stored in a spreadsheet or database.
In one embodiment, the individual documents 14a, 14b, 17, 20, 26, 29 can include electronic message folders storing email and attachments, such as maintained by the Outlook and Outlook Express products, licensed by Microsoft Corporation, Redmond, Wash. The database can be an SQL-based relational database, such as the Oracle database management system, release 8, licensed by Oracle Corporation, Redwood Shores, Calif.
The individual documents can be designated and stored as uncoded documents or reference documents. The reference documents are initially uncoded documents that can be selected from the corpus or other source of uncoded documents and subsequently classified. The reference documents assist in providing suggestions for classification of the remaining uncoded documents in the corpus based on visual relationships between the uncoded documents and reference documents. The reviewer can classify one or more of the remaining uncoded documents by assigning a classification code based on the relationships. In a further embodiment, the reference documents can be used as a training set to form machine-generated suggestions for classifying the remaining uncoded documents, as further described below with reference to
The reference documents are representative of the document corpus for a review project in which data organization or classification is desired or a subset of the document corpus. A set of reference documents can be generated for each document review project or alternatively, the reference documents can be selected from a previously conducted document review project that is related to the current document review project. Guided review assists a reviewer in building a reference document set representative of the corpus for use in classifying uncoded documents. During guided review, uncoded documents that are dissimilar to all other uncoded documents in the corpus are identified based on a similarity threshold. Other methods for determining dissimilarity are possible. Identifying the dissimilar documents provides a group of uncoded documents that is representative of the corpus for a document review project. Each identified dissimilar document is then classified by assigning a particular classification code based on the content of the document to generate a set of reference documents for the document review project. Guided review can be performed by a reviewer, a machine, or a combination of the reviewer and machine.
Other methods for generating a reference document set for a document review project using guided review are possible, including clustering. For example, a set of uncoded document to be classified is clustered, as described in commonly-assigned U.S. Pat. No. 7,610,313, the disclosure of which is incorporated by reference. A plurality of the clustered uncoded documents are selected based on selection criteria, such as cluster centers or sample clusters. The cluster centers can be used to identify uncoded documents in a cluster that are most similar or dissimilar to the cluster center. The identified uncoded documents are then selected for classification by assigning codes. After classification, the previously uncoded documents represent a reference set. In a further example, sample clusters can be used to generate a reference set by selecting one or more sample clusters based on cluster relation criteria, such as size, content, similarity, or dissimilarity. The uncoded documents in the selected sample clusters are then assigned classification codes. The classified documents represent a reference document set for the document review project. Other methods for selecting uncoded documents for use as a reference set are possible.
The document corpus for a document review project can be divided into subsets of uncoded documents, which are each provided as an assignment to a particular reviewer. To maintain consistency, the same classification codes can be used across all assignments in the document review project. The classification codes can be determined using taxonomy generation, during which a list of classification codes can be provided by a reviewer or determined automatically. For purposes of legal discovery, the classification codes used to classify uncoded documents can include “privileged,” “responsive,” or “non-responsive.” Other codes are possible. A “privileged” document contains information that is protected by a privilege, meaning that the document should not be disclosed to an opposing party. Disclosing a “privileged” document can result in an unintentional waiver of the subject matter. A “responsive” document contains information that is related to a legal matter on which the document review project is based and a “non-responsive” document includes information that is not related to the legal matter.
Utilizing reference documents to assist in classifying uncoded documents, clusters, or spines can be performed by the system 10, which includes individual computer systems, such as the backend server 11, work server 12, server 15, client 18, remote server 24 and remote client 27. The individual computer systems are general purpose, programmed digital computing devices consisting of a central processing unit (CPU), random access memory (RAM), non-volatile secondary storage, such as a hard drive or CD ROM drive, network interfaces, and peripheral devices, including user interfacing means, such as a keyboard and display. The various implementations of the source code and object and byte codes can be held on a computer-readable storage medium, such as a floppy disk, hard drive, digital video disk (DVD), random access memory (RAM), read-only memory (ROM) and similar storage mediums. For example, program code, including software programs, and data are loaded into the RAM for execution and processing by the CPU and results are generated for display, output, transmittal, or storage.
Identifying the reference documents for use as classification suggestions includes a comparison of the uncoded documents and the reference documents.
Together, reference documents injected into the clusters represent a subset of reference documents specific to that cluster set. The clusters of uncoded documents and inserted reference documents can be displayed to visually depict relationships (block 45) between the uncoded documents in the cluster and the inserted reference documents. The relationships can provide a suggestion for use by an individual reviewer, for classifying that cluster. Determining relationships between the reference documents and uncoded documents to identify classification suggestions is further discussed below with reference to
Obtaining Clusters
The corpus of uncoded documents for a review project can be divided into assignments using assignment criteria, such as custodian or source of the uncoded documents, content, document type, and date. Other criteria are possible. Each assignment is assigned to an individual reviewer for analysis. The assignments can be separately clustered or alternatively, all of the uncoded documents in the document corpus can be clustered together. The content of each uncoded document within the corpus can be converted into a set of tokens, which are word-level or character-level n-grams, raw terms, concepts, or entities. Other tokens are possible.
An n-gram is a predetermined number of items selected from a source. The items can include syllables, letters, or words, as well as other items. A raw term is a term that has not been processed or manipulated. Concepts typically include nouns and noun phrases obtained through part-of-speech tagging that have a common semantic meaning. Entities further refine nouns and noun phrases into people, places, and things, such as meetings, animals, relationships, and various other objects. Entities can be extracted using entity extraction techniques known in the field. Clustering of the uncoded documents can be based on cluster criteria, such as the similarity of tokens, including n-grams, raw terms, concepts, entities, email addresses, or other metadata.
Clustering provides groupings of related uncoded documents.
As an initial step for generating score vectors, each token for an uncoded document is individually scored. Next, a normalized score vector is created for the uncoded document by identifying paired values, consisting of a token occurring in that document and the scores for that token. The paired values are ordered along a vector to generate the score vector. The paired values can be ordered based on tokens, including concepts or frequency, as well as other factors. For example, assume a normalized score vector for a first uncoded document A is {right arrow over (S)}A={(5, 0.5), (120, 0.75)} and a normalized score vector for another uncoded document B is {right arrow over (S)}B={(3, 0.4), (5, 0.75), (47, 0.15)}. Document A has scores corresponding to tokens ‘5’ and ‘120’ and Document B has scores corresponding to tokens ‘3’, ‘5’ and ‘47.’ Thus, these uncoded documents only have token ‘5’ in common. Once generated, the score vectors can be compared to determine similarity or dissimilarity between the corresponding uncoded documents during clustering.
The routine for forming clusters proceeds in two phases. During the first phase (blocks 53-58), uncoded documents are evaluated to identify a set of seed documents, which can be used to form new clusters. During the second phase (blocks 60-66), the uncoded documents not previously placed are evaluated and grouped into existing clusters based on a best-fit criterion.
Initially, a single cluster is generated with one or more uncoded documents as seed documents and additional clusters of uncoded documents are added. Each cluster is represented by a cluster center that is associated with a score vector, which is representative of the tokens in all the documents for that cluster. In the following discussion relating to
During clustering, the uncoded documents are identified (block 51) and ordered by length (block 52). The uncoded documents can include all uncoded documents in a corpus or can include only those uncoded documents for a single assignment. Each uncoded document is then processed in an iterative processing loop (blocks 53-58) as follows. The similarity between each uncoded document and the cluster centers, based on uncoded documents already clustered, is determined (block 54) as the cosine (cos) σ of the score vectors for the uncoded documents and cluster being compared. The cos σ provides a measure of relative similarity or dissimilarity between tokens, including the concepts, in the uncoded documents and is equivalent to the inner products between the score vectors for the uncoded document and cluster center.
In the described embodiment, the cos σ is calculated in accordance with the equation:
where cos σAB comprises the similarity metric between uncoded document A and cluster center B, {right arrow over (S)}A comprises a score vector for the uncoded document A, and {right arrow over (S)}B comprises a score vector for the cluster center B. Other forms of determining similarity using a distance metric are feasible, as would be recognized by one skilled in the art. An example includes using Euclidean distance.
Only those uncoded documents that are sufficiently distinct from all cluster centers (block 55) are selected as seed documents for forming new clusters (block 56). If the uncoded documents being compared are not sufficiently distinct (block 55), each uncoded document is then grouped into a cluster with the most similar cluster center (block 57). Processing continues with the next uncoded document (block 58).
In the second phase, each uncoded document not previously placed is iteratively processed in an iterative processing loop (blocks 60-66) as follows. Again, the similarity between each remaining uncoded document and each cluster center is determined based on a distance (block 61) as the cos σ of the normalized score vectors for the remaining uncoded document and the cluster center. A best fit between the remaining uncoded document and one of the cluster centers can be found subject to a minimum fit criterion (block 62). In the described embodiment, a minimum fit criterion of 0.25 is used, although other minimum fit criteria could be used. If a best fit is found (block 63), the remaining uncoded document is grouped into the cluster having the best fit (block 65). Otherwise, the remaining uncoded document is grouped into a miscellaneous cluster (block 64). Processing continues with the next remaining uncoded document (block 66). Finally, a dynamic threshold can be applied to each cluster (block 67) to evaluate and strengthen document membership in a particular cluster. The dynamic threshold is applied based on a cluster-by-cluster basis, as described in commonly-assigned U.S. Pat. No. 7,610,313, the disclosure of which is incorporated by reference. The routine then returns. Other methods and processes for forming clusters are possible.
Identifying Similar Reference Documents
Once a cluster set is obtained, one or more uncoded documents associated with a cluster are compared to a set of reference documents to identify a subset of the reference documents that are similar. The similarity is determined based on a similarity metric, which can include a distance metric. The similarity metric can be determined as the cos σ of the score vectors for the reference documents and clusters associated with the one or more uncoded documents. The one or more uncoded documents can be selected based on a cluster measure.
Identifying similar reference documents using the cluster center measure 71 includes determining a cluster center for each cluster, comparing one or more of the cluster centers to a set of reference documents, and identifying the reference documents that satisfy a threshold similarity with the particular cluster center. More specifically, the score vector for the cluster center is compared to score vectors associated with each reference document as cos σ of the score vectors for the reference document and the cluster center. The score vector for the cluster is based on the cluster center, which considers the score vectors for all the uncoded documents in that cluster. The sample cluster measure 72 includes generating a sample of one or more uncoded documents in a single cluster that is representative of that cluster. The number of uncoded documents in the sample can be defined by the reviewer, set as a default, or determined automatically. Once generated, a score vector is calculated for the sample by comparing the score vectors for the individual uncoded documents selected for inclusion in the sample and identifying the most common concepts shared by the selected documents. The most common concepts and associated weights for the samples are positioned along a score vector, which is representative of the sample of uncoded documents for the cluster. The cluster center and sample cluster measure 73 includes comparing both the cluster center score vector and the sample score vector for a cluster to identify reference documents that are similar to the uncoded documents in that cluster.
Further, similar reference documents can be identified based on a spine, which includes those clusters that share one or more tokens, such as concepts, and are arranged linearly along a vector. The cluster spines are generated as described in commonly-assigned U.S. Pat. No. 7,271,804, the disclosure of which is incorporated by reference. Also, the cluster spines can be positioned in relation to other cluster spines, as described in commonly-assigned U.S. Pat. No. 7,610,313, issued Oct. 27, 2009, the disclosure of which is incorporated by reference. Organizing the clusters into spines and groups of cluster spines provides an individual reviewer with a display that presents the uncoded documents and reference documents according to theme while maximizing the number of relationships depicted between the documents. Each theme can include one or more concepts defining a semantic meaning.
The spine cluster measure 74 involves generating a score vector for a spine by comparing the score vectors for the clusters positioned along that spine and identifying the most common concepts shared by the clusters. The most common concepts and associated scores are positioned along a vector to form a spine score vector. The spine score vector is compared with the score vectors of the reference documents in the set to identify similar reference documents.
The measure of similarity determined between the reference documents and selected uncoded documents can be calculated as cos σ of the corresponding score vectors. However, other similarity calculations are possible. The similarity calculations can be applied to a threshold and those references documents that satisfy the threshold can be selected as the most similar. The most similar reference documents selected for a cluster can be the same or different from the most similar reference documents for the other clusters. Although four types of similarity metrics are described above, other similarity metrics are possible.
Upon identification, the similar reference documents for a cluster are injected into that cluster to provide relationships between the similar reference documents and uncoded documents. Identifying the most similar reference documents and injecting those documents can occur cluster-by-cluster or for all the clusters simultaneously. The number of similar reference documents selected for injection can be defined by the reviewer, set as a default, or determined automatically. Other determinations for the number of similar reference documents are possible. The similar reference documents can provide hints or suggestions to a reviewer regarding how to classify the uncoded documents based on the relationships.
Displaying the Reference Documents
The clusters of uncoded documents and inserted reference documents can be provided as a display to the reviewer.
The display 81 can be manipulated by a individual reviewer via a compass 82, which enables the reviewer to navigate, explore, and search the clusters 83 and spines 86 appearing within the compass 82, as further described in commonly-assigned U.S. Pat. No. 7,356,777, the disclosure of which is incorporated by reference. Visually, the compass 82 emphasizes clusters 83 located within the compass 82, while deemphasizing clusters 83 appearing outside of the compass 82.
Spine labels 89 appear outside of the compass 82 at an end of each cluster spine 86 to connect the outermost cluster of the cluster spine 86 to the closest point along the periphery of the compass 82. In one embodiment, the spine labels 89 are placed without overlap and circumferentially around the compass 82. Each spine label 89 corresponds to one or more concepts that most closely describe the cluster spines 86 appearing within the compass 82. Additionally, the cluster concepts for each of the spine labels 89 can appear in a concepts list (not shown) also provided in the display. Toolbar buttons 87 located at the top of the display 81 enable a user to execute specific commands for the composition of the spine groups displayed. A set of pull down menus 88 provides further control over the placement and manipulation of clusters 83 and cluster spines 86 within the display 81. Other types of controls and functions are possible.
A document guide 90 can be placed in the display 81. The document guide 90 can include a “Selected” field, a “Search Results” field, and details regarding the numbers of uncoded documents and reference documents provided in the display. The number of uncoded documents includes all uncoded documents within a corpus of documents for a review project or within an assignment for the project. The number of reference documents includes the total number of reference documents selected for injection into the cluster set. The “Selected” field in the document guide 90 provides a number of documents within one or more clusters selected by the reviewer. The reviewer can select a cluster by “double clicking” the visual representation of that cluster using a mouse. The “Search Results” field provides a number of uncoded documents and reference documents that include a particular search term identified by the reviewer in a search query box 92.
In one embodiment, a garbage can 91 is provided to remove tokens, such as cluster concepts from consideration in the current set of clusters 83. Removed cluster concepts prevent those concepts from affecting future clustering, as may occur when a reviewer considers a concept irrelevant to the clusters 83.
The display 81 provides a visual representation of the relationships between thematically related documents, including uncoded documents and injected reference documents. The uncoded documents and injected reference documents located within a cluster or spine can be compared based on characteristics, such as the assigned classification codes of the reference documents, a number of reference documents associated with each classification code, and a number of different classification codes, to identify relationships between the uncoded documents and injected reference documents. The reviewer can use the displayed relationships as suggestions for classifying the uncoded documents. For example,
Alternatively, the three reference documents can be classified as “non-responsive,” instead of “privileged” as in the previous example.
A further example can include a combination of “privileged” and “non-responsive” reference documents. For example,
Additionally, the reference documents can also provide suggestions for classifying clusters and spines. The suggestions provided for classifying a cluster can include factors, such as a presence or absence of classified documents with different classification codes within the cluster and a quantity of the classified documents associated with each classification code in the cluster. The classified documents can include reference documents and newly classified uncoded documents. The classification code assigned to the cluster is representative of the documents in that cluster and can be the same as or different from one or more classified documents within the cluster. Further, the suggestions provided for classifying a spine include factors, such as a presence or absence of classified documents with different classification codes within the clusters located along the spine and a quantity of the classified documents for each classification code. Other suggestions for classifying documents, clusters, and spines are possible.
Classifying Uncoded Documents
The display of relationships between the uncoded documents and reference documents provides suggestions to an individual reviewer. The suggestions can indicate a need for manual review of the uncoded documents, when review may be unnecessary, and hints for classifying the uncoded documents. Additional information can be provided to assist the reviewer in making classification decisions for the uncoded documents, such as a machine-generated confidence level associated with a suggested classification code, as described in commonly-assigned U.S. Patent Application Publication No. 011/0029525, published Feb. 3, 2011, pending, the disclosure of which is incorporated by reference.
The machine-generated suggestion for classification and associated confidence level can be determined by a classifier.
The neighborhood of x-reference documents is determined separately for each selected uncoded document and can include one or more injected reference documents within that cluster. During neighborhood generation, the x-number of reference documents in a neighborhood can first be determined automatically or by an individual reviewer. Next, the x-number of reference documents nearest in distance to the selected uncoded document are identified. Finally, the identified x-number of reference documents are provided as the neighborhood for the selected uncoded document. In a further embodiment, the x-number of reference documents are defined for each classification code, rather than across all classification codes. Once generated, the x-number of reference documents in the neighborhood and the selected uncoded document are analyzed by the classifier to provide a machine-generated classification suggestion (block 103). A confidence level for the suggested classification is also provided (block 104).
The analysis of the selected uncoded document and x-number of reference documents can be based on one or more routines performed by the classifier, such as a nearest neighbor (NN) classifier. The routines for determining a suggested classification code for an uncoded document include a minimum distance classification measure, also known as closest neighbor, minimum average distance classification measure, maximum count classification measure, and distance weighted maximum count classification measure. The minimum distance classification measure includes identifying a neighbor that is the closest distance to the selected uncoded document and assigning the classification code of the closest neighbor as the suggested classification code for the selected uncoded document. The closest neighbor is determined by comparing score vectors for the selected uncoded document with each of the x-number reference documents in the neighborhood as the cos σ to determine a distance metric. The distance metrics for the x-number of reference documents are compared to identify the reference document closest to the selected uncoded document as the closest neighbor.
The minimum average distance classification measure includes calculating an average distance of the reference documents in a cluster for each classification code. The classification code of the reference documents having the closest average distance to the selected uncoded document is assigned as the suggested classification code. The maximum count classification measure, also known as the voting classification measure, includes counting a number of reference documents within the cluster for each classification code and assigning a count or “vote” to the reference documents based on the assigned classification code. The classification code with the highest number of reference documents or “votes” is assigned to the selected uncoded document as the suggested classification. The distance weighted maximum count classification measure includes identifying a count of all reference documents within the cluster for each classification code and determining a distance between the selected uncoded document and each of the reference documents. Each count assigned to the reference documents is weighted based on the distance of the reference document from the selected uncoded document. The classification code with the highest count, after consideration of the weight, is assigned to the selected uncoded document as the suggested classification.
The x-NN classifier provides the machine-generate classification code with a confidence level that can be presented as an absolute value or percentage. Other confidence level measures are possible. The reviewer can use the suggested classification code and confidence level to assign a classification to the selected uncoded document. Alternatively, the x-NN classifier can automatically assign the suggested classification. In one embodiment, the x-NN classifier only assigns an uncoded document with the suggested classification code if the confidence level is above a threshold value, which can be set by the reviewer or the x-NN classifier.
As briefly described above, classification can also occur on a cluster or spine level. For instance, for cluster classification, a cluster is selected and a score vector for the center of the cluster is determined as described above with reference to
Throughout the process of identifying similar reference documents and injecting the reference documents into a cluster to provide a classification suggestion, the reviewer can retain control over many aspects, such as a source of the reference documents and a number of similar reference documents to be selected.
The reference source parameter 112 allows the reviewer to identify one or more sources of the reference documents. The sources can include all previously classified reference documents in a document review project, all reference documents for which the associated classification has been verified, all reference documents that have been analyzed or all reference documents in a particular binder. The binder can include categories of reference documents, such as reference documents that are particular to the document review project or that are related to a prior document review project. The category filter parameter 113 allows the reviewer to generate and display the set of reference documents using only those reference documents associated with a particular classification code. The target parameter 114 allows the reviewer to select a target for injection of the similar reference documents. Options available for the target parameter 114 can include an assignment, all clusters, select clusters, all spines, select spines, all documents, and select documents. The assignment can be represented as a cluster set; however, other representations are possible, including a file hierarchy and a list of documents, such as an email folder, as described in commonly-assigned U.S. Pat. No. 7,404,151, the disclosure of which is incorporated by reference
The action parameter 115 allows the reviewer to define display options for the injected reference documents. The display options can include injecting the similar reference documents into a map display of the clusters, displaying the similar reference documents in the map until reclustering occurs, displaying the injected reference documents in the map, and not displaying the injected reference documents in the map. Using the automatic parameter 116, the reviewer can define a time for injection of the similar reference documents. The timing options can include injecting the similar reference documents upon opening of an assignment, upon reclustering, or upon changing the selection of the target. The reviewer can specify a threshold number of similar reference documents to be injected in each cluster or spine via the similarity option 117. The number selected by a reviewer is an upper threshold since a lesser number of similar reference documents may be identified for injecting into a cluster or spine. Additionally, the reviewer can use the similarity option 117 to set a value for determining whether a reference document is sufficiently similar to the uncoded documents.
Further, the reviewer can select a location within the cluster for injection of the similar reference documents via the cluster site parameter 118. Options for cluster site injection can include the cluster centroid. Other cluster sites are possible. The user-selectable options for each preference can be compiled as a list of injection commands 119 for use in the injection process. Other user selectable parameters, options, and actions are possible.
The clustering of uncoded documents and injection of similar reference documents in the clusters has been described in relation to documents; however, in a further embodiment, the cluster and injection process can be applied to tokens. For example, uncoded tokens are clustered and similar reference tokens are injected into the clusters and displayed to provide classification suggestions based on relationships between the uncoded tokens and similar reference tokens. The uncoded documents can then be classified based on the classified tokens. In one embodiment, the tokens include concepts, n-grams, raw terms, and entities.
While the invention has been particularly shown and described as referenced to the embodiments thereof, those skilled in the art will understand that the foregoing and other changes in form and detail may be made therein without departing from the spirit and scope.
This non-provisional patent application claims priority under 35 U.S.C. §119(e) to U.S. Provisional Patent Application Ser. No. 61/229,216, filed Jul. 28, 2009, and U.S. Provisional Patent Application Ser. No. 61/236,490, filed Aug. 24, 2009, the disclosures of which are incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
3416150 | Lindberg | Dec 1968 | A |
3426210 | Agin | Feb 1969 | A |
3668658 | Flores et al. | Jun 1972 | A |
4893253 | Lodder | Jan 1990 | A |
5056021 | Ausborn | Oct 1991 | A |
5121338 | Lodder | Jun 1992 | A |
5133067 | Hara et al. | Jul 1992 | A |
5278980 | Pedersen et al. | Jan 1994 | A |
5371673 | Fan | Dec 1994 | A |
5442778 | Pedersen et al. | Aug 1995 | A |
5477451 | Brown et al. | Dec 1995 | A |
5488725 | Turtle et al. | Jan 1996 | A |
5524177 | Suzuoka | Jun 1996 | A |
5528735 | Strasnick et al. | Jun 1996 | A |
5619632 | Lamping et al. | Apr 1997 | A |
5619709 | Caid et al. | Apr 1997 | A |
5635929 | Rabowsky et al. | Jun 1997 | A |
5649193 | Sumita et al. | Jul 1997 | A |
5675819 | Schuetze | Oct 1997 | A |
5696962 | Kupiec | Dec 1997 | A |
5737734 | Schultz | Apr 1998 | A |
5754938 | Herz et al. | May 1998 | A |
5794236 | Mehrle | Aug 1998 | A |
5799276 | Komissarchik et al. | Aug 1998 | A |
5819258 | Vaithyanathan et al. | Oct 1998 | A |
5842203 | D'Elena et al. | Nov 1998 | A |
5844991 | Hochberg et al. | Dec 1998 | A |
5857179 | Vaithyanathan et al. | Jan 1999 | A |
5860136 | Fenner | Jan 1999 | A |
5862325 | Reed et al. | Jan 1999 | A |
5864846 | Voorhees et al. | Jan 1999 | A |
5864871 | Kitain et al. | Jan 1999 | A |
5867799 | Lang et al. | Feb 1999 | A |
5870740 | Rose et al. | Feb 1999 | A |
5909677 | Broder et al. | Jun 1999 | A |
5915024 | Kitaori et al. | Jun 1999 | A |
5920854 | Kirsch et al. | Jul 1999 | A |
5924105 | Punch et al. | Jul 1999 | A |
5940821 | Wical | Aug 1999 | A |
5950146 | Vapnik | Sep 1999 | A |
5950189 | Cohen et al. | Sep 1999 | A |
5966126 | Szabo | Oct 1999 | A |
5987446 | Corey et al. | Nov 1999 | A |
6006221 | Liddy et al. | Dec 1999 | A |
6012053 | Pant et al. | Jan 2000 | A |
6026397 | Sheppard | Feb 2000 | A |
6038574 | Pitkow et al. | Mar 2000 | A |
6070133 | Brewster et al. | May 2000 | A |
6089742 | Warmerdam et al. | Jul 2000 | A |
6092059 | Straforini et al. | Jul 2000 | A |
6094649 | Bowen et al. | Jul 2000 | A |
6100901 | Mohda et al. | Aug 2000 | A |
6119124 | Broder et al. | Sep 2000 | A |
6122628 | Castelli et al. | Sep 2000 | A |
6137499 | Tesler | Oct 2000 | A |
6137545 | Patel et al. | Oct 2000 | A |
6137911 | Zhilyaev | Oct 2000 | A |
6148102 | Stolin | Nov 2000 | A |
6154219 | Wiley et al. | Nov 2000 | A |
6167368 | Wacholder | Dec 2000 | A |
6173275 | Caid et al. | Jan 2001 | B1 |
6202064 | Julliard | Mar 2001 | B1 |
6216123 | Robertson et al. | Apr 2001 | B1 |
6243713 | Nelson et al. | Jun 2001 | B1 |
6243724 | Mander et al. | Jun 2001 | B1 |
6260038 | Martin et al. | Jul 2001 | B1 |
6326962 | Szabo | Dec 2001 | B1 |
6338062 | Liu | Jan 2002 | B1 |
6345243 | Clark | Feb 2002 | B1 |
6349296 | Broder et al. | Feb 2002 | B1 |
6349307 | Chen | Feb 2002 | B1 |
6360227 | Aggarwal et al. | Mar 2002 | B1 |
6363374 | Corston-Oliver et al. | Mar 2002 | B1 |
6377287 | Hao et al. | Apr 2002 | B1 |
6381601 | Fujiwara et al. | Apr 2002 | B1 |
6389433 | Bolonsky et al. | May 2002 | B1 |
6389436 | Chakrabarti et al. | May 2002 | B1 |
6408294 | Getchius et al. | Jun 2002 | B1 |
6414677 | Robertson et al. | Jul 2002 | B1 |
6415283 | Conklin | Jul 2002 | B1 |
6418431 | Mahajan et al. | Jul 2002 | B1 |
6421709 | McCormick et al. | Jul 2002 | B1 |
6438537 | Netz et al. | Aug 2002 | B1 |
6438564 | Morton et al. | Aug 2002 | B1 |
6442592 | Alumbaugh et al. | Aug 2002 | B1 |
6446061 | Doerre et al. | Sep 2002 | B1 |
6449612 | Bradley et al. | Sep 2002 | B1 |
6453327 | Nielsen | Sep 2002 | B1 |
6460034 | Wical | Oct 2002 | B1 |
6470307 | Turney | Oct 2002 | B1 |
6480843 | Li | Nov 2002 | B2 |
6480885 | Olivier | Nov 2002 | B1 |
6484168 | Pennock et al. | Nov 2002 | B1 |
6484196 | Maurille | Nov 2002 | B1 |
6493703 | Knight et al. | Dec 2002 | B1 |
6496822 | Rosenfelt et al. | Dec 2002 | B2 |
6502081 | Wiltshire, Jr. et al. | Dec 2002 | B1 |
6507847 | Fleischman | Jan 2003 | B1 |
6510406 | Marchisio | Jan 2003 | B1 |
6519580 | Johnson et al. | Feb 2003 | B1 |
6523026 | Gillis | Feb 2003 | B1 |
6523063 | Miller et al. | Feb 2003 | B1 |
6542889 | Aggarwal et al. | Apr 2003 | B1 |
6544123 | Tanaka et al. | Apr 2003 | B1 |
6549957 | Hanson et al. | Apr 2003 | B1 |
6560597 | Dhillon et al. | May 2003 | B1 |
6571225 | Oles et al. | May 2003 | B1 |
6584564 | Olkin et al. | Jun 2003 | B2 |
6594658 | Woods | Jul 2003 | B2 |
6598054 | Schuetze et al. | Jul 2003 | B2 |
6606625 | Muslea et al. | Aug 2003 | B1 |
6611825 | Billheimer et al. | Aug 2003 | B1 |
6628304 | Mitchell et al. | Sep 2003 | B2 |
6629097 | Keith | Sep 2003 | B1 |
6640009 | Zlotnick | Oct 2003 | B2 |
6651057 | Jin et al. | Nov 2003 | B1 |
6654739 | Apte et al. | Nov 2003 | B1 |
6658423 | Pugh et al. | Dec 2003 | B1 |
6675159 | Lin et al. | Jan 2004 | B1 |
6675164 | Kamath et al. | Jan 2004 | B2 |
6678705 | Berchtold et al. | Jan 2004 | B1 |
6684205 | Modha et al. | Jan 2004 | B1 |
6697998 | Damerau et al. | Feb 2004 | B1 |
6701305 | Holt et al. | Mar 2004 | B1 |
6711585 | Copperman et al. | Mar 2004 | B1 |
6714929 | Micaelian et al. | Mar 2004 | B1 |
6735578 | Shetty et al. | May 2004 | B2 |
6738759 | Wheeler et al. | May 2004 | B1 |
6747646 | Gueziec et al. | Jun 2004 | B2 |
6751628 | Coady | Jun 2004 | B2 |
6757646 | Marchisio | Jun 2004 | B2 |
6785679 | Dane et al. | Aug 2004 | B1 |
6804665 | Kreulen et al. | Oct 2004 | B2 |
6816175 | Hamp et al. | Nov 2004 | B1 |
6819344 | Robbins | Nov 2004 | B2 |
6823333 | McGreevy | Nov 2004 | B2 |
6841321 | Matsumoto et al. | Jan 2005 | B2 |
6847966 | Sommer et al. | Jan 2005 | B1 |
6862710 | Marchisio | Mar 2005 | B1 |
6879332 | Decombe | Apr 2005 | B2 |
6883001 | Abe | Apr 2005 | B2 |
6886010 | Kostoff | Apr 2005 | B2 |
6888584 | Suzuki et al. | May 2005 | B2 |
6915308 | Evans et al. | Jul 2005 | B1 |
6922699 | Schuetze et al. | Jul 2005 | B2 |
6941325 | Benitez et al. | Sep 2005 | B1 |
6970881 | Mohan et al. | Nov 2005 | B1 |
6978419 | Kantrowitz | Dec 2005 | B1 |
6990238 | Saffer et al. | Jan 2006 | B1 |
6993535 | Bolle et al. | Jan 2006 | B2 |
6996575 | Cox et al. | Feb 2006 | B2 |
7003551 | Malik | Feb 2006 | B2 |
7013435 | Gallo et al. | Mar 2006 | B2 |
7020645 | Bisbee et al. | Mar 2006 | B2 |
7039856 | Peairs et al. | May 2006 | B2 |
7051017 | Marchisio | May 2006 | B2 |
7054870 | Holbrook | May 2006 | B2 |
7080320 | Ono | Jul 2006 | B2 |
7096431 | Tambata et al. | Aug 2006 | B2 |
7099819 | Sakai et al. | Aug 2006 | B2 |
7107266 | Breyman et al. | Sep 2006 | B1 |
7117151 | Iwahashi et al. | Oct 2006 | B2 |
7117246 | Christenson et al. | Oct 2006 | B2 |
7130807 | Mikurak | Oct 2006 | B1 |
7137075 | Hoshito et al. | Nov 2006 | B2 |
7139739 | Agrafiotis et al. | Nov 2006 | B2 |
7146361 | Broder et al. | Dec 2006 | B2 |
7155668 | Holland et al. | Dec 2006 | B2 |
7188107 | Moon et al. | Mar 2007 | B2 |
7188117 | Farahat et al. | Mar 2007 | B2 |
7194458 | Micaelian et al. | Mar 2007 | B1 |
7194483 | Mohan et al. | Mar 2007 | B1 |
7197497 | Cossock | Mar 2007 | B2 |
7209949 | Mousseau et al. | Apr 2007 | B2 |
7233886 | Wegerich et al. | Jun 2007 | B2 |
7233940 | Bamberger et al. | Jun 2007 | B2 |
7239986 | Golub et al. | Jul 2007 | B2 |
7240199 | Tomkow | Jul 2007 | B2 |
7246113 | Cheetham et al. | Jul 2007 | B2 |
7251637 | Caid et al. | Jul 2007 | B1 |
7266365 | Ferguson et al. | Sep 2007 | B2 |
7266545 | Bergman et al. | Sep 2007 | B2 |
7269598 | Marchisio | Sep 2007 | B2 |
7271801 | Toyozawa et al. | Sep 2007 | B2 |
7277919 | Donoho et al. | Oct 2007 | B1 |
7325127 | Olkin et al. | Jan 2008 | B2 |
7353204 | Liu | Apr 2008 | B2 |
7359894 | Liebman et al. | Apr 2008 | B1 |
7363243 | Arnett et al. | Apr 2008 | B2 |
7366759 | Trevithick et al. | Apr 2008 | B2 |
7373612 | Risch et al. | May 2008 | B2 |
7376635 | Porcari et al. | May 2008 | B1 |
7379913 | Steele et al. | May 2008 | B2 |
7383282 | Whitehead et al. | Jun 2008 | B2 |
7401087 | Copperman et al. | Jul 2008 | B2 |
7412462 | Margolus et al. | Aug 2008 | B2 |
7418397 | Kojima et al. | Aug 2008 | B2 |
7430688 | Matsuno et al. | Sep 2008 | B2 |
7430717 | Spangler | Sep 2008 | B1 |
7433893 | Lowry | Oct 2008 | B2 |
7440662 | Antona et al. | Oct 2008 | B2 |
7444356 | Calistri-Yeh et al. | Oct 2008 | B2 |
7457948 | Bilicksa et al. | Nov 2008 | B1 |
7472110 | Achlioptas | Dec 2008 | B2 |
7490092 | Morton et al. | Feb 2009 | B2 |
7509256 | Iwahashi et al. | Mar 2009 | B2 |
7516419 | Petro et al. | Apr 2009 | B2 |
7523349 | Barras | Apr 2009 | B2 |
7558769 | Scott et al. | Jul 2009 | B2 |
7571177 | Damle | Aug 2009 | B2 |
7574409 | Patinkin | Aug 2009 | B2 |
7584221 | Robertson et al. | Sep 2009 | B2 |
7639868 | Regli et al. | Dec 2009 | B1 |
7640219 | Perrizo | Dec 2009 | B2 |
7647345 | Trepess et al. | Jan 2010 | B2 |
7668376 | Lin et al. | Feb 2010 | B2 |
7698167 | Batham et al. | Apr 2010 | B2 |
7716223 | Haveliwala et al. | May 2010 | B2 |
7761447 | Brill et al. | Jul 2010 | B2 |
7801841 | Mishra et al. | Sep 2010 | B2 |
7885901 | Hull et al. | Feb 2011 | B2 |
7971150 | Raskutti et al. | Jun 2011 | B2 |
8010466 | Patinkin | Aug 2011 | B2 |
8010534 | Roitblat et al. | Aug 2011 | B2 |
8165974 | Privault et al. | Apr 2012 | B2 |
20020032735 | Burnstein et al. | Mar 2002 | A1 |
20020065912 | Catchpole et al. | May 2002 | A1 |
20020078044 | Song et al. | Jun 2002 | A1 |
20020078090 | Hwang et al. | Jun 2002 | A1 |
20020122543 | Rowen | Sep 2002 | A1 |
20020184193 | Cohen | Dec 2002 | A1 |
20030046311 | Baidya et al. | Mar 2003 | A1 |
20030130991 | Reijerse et al. | Jul 2003 | A1 |
20030172048 | Kauffman | Sep 2003 | A1 |
20030174179 | Suermondt et al. | Sep 2003 | A1 |
20040024739 | Copperman et al. | Feb 2004 | A1 |
20040024755 | Rickard | Feb 2004 | A1 |
20040034633 | Rickard | Feb 2004 | A1 |
20040205482 | Basu | Oct 2004 | A1 |
20040205578 | Wolf et al. | Oct 2004 | A1 |
20040215608 | Gourlay | Oct 2004 | A1 |
20040243556 | Ferrucci et al. | Dec 2004 | A1 |
20050022106 | Kawai et al. | Jan 2005 | A1 |
20050025357 | Landwehr et al. | Feb 2005 | A1 |
20050097435 | Prakash et al. | May 2005 | A1 |
20050171772 | Iwahashi et al. | Aug 2005 | A1 |
20050203924 | Rosenberg | Sep 2005 | A1 |
20050283473 | Rousso et al. | Dec 2005 | A1 |
20060008151 | Lin et al. | Jan 2006 | A1 |
20060021009 | Lunt | Jan 2006 | A1 |
20060053382 | Gardner et al. | Mar 2006 | A1 |
20060122974 | Perisic | Jun 2006 | A1 |
20060122997 | Lin | Jun 2006 | A1 |
20070020642 | Deng et al. | Jan 2007 | A1 |
20070043774 | Davis et al. | Feb 2007 | A1 |
20070044032 | Mollitor et al. | Feb 2007 | A1 |
20070109297 | Borchardt et al. | May 2007 | A1 |
20070112758 | Livaditis | May 2007 | A1 |
20070150801 | Chidlovskii et al. | Jun 2007 | A1 |
20070214133 | Liberty et al. | Sep 2007 | A1 |
20070288445 | Kraftsow | Dec 2007 | A1 |
20080005081 | Green et al. | Jan 2008 | A1 |
20080140643 | Ismalon | Jun 2008 | A1 |
20080183855 | Agarwal et al. | Jul 2008 | A1 |
20080189273 | Kraftsow | Aug 2008 | A1 |
20080228675 | Daffy et al. | Sep 2008 | A1 |
20090041329 | Nordell et al. | Feb 2009 | A1 |
20090043797 | Dorie | Feb 2009 | A1 |
20090049017 | Gross | Feb 2009 | A1 |
20090097733 | Hero et al. | Apr 2009 | A1 |
20090106239 | Getner et al. | Apr 2009 | A1 |
20090222444 | Chowdhury et al. | Sep 2009 | A1 |
20090228499 | Schmidtler et al. | Sep 2009 | A1 |
20090228811 | Adams et al. | Sep 2009 | A1 |
20100100539 | Davis et al. | Apr 2010 | A1 |
20100198802 | Kraftsow | Aug 2010 | A1 |
20100250477 | Yadav | Sep 2010 | A1 |
20100262571 | Schmidtler et al. | Oct 2010 | A1 |
20100268661 | Levy et al. | Oct 2010 | A1 |
20120124034 | Jing et al. | May 2012 | A1 |
Number | Date | Country |
---|---|---|
0886227 | Dec 1998 | EP |
1024437 | Aug 2000 | EP |
1049030 | Nov 2000 | EP |
WO 0067162 | Nov 2000 | WO |
03052627 | Jun 2003 | WO |
03060766 | Jul 2003 | WO |
WO 2005073881 | Aug 2005 | WO |
2006008733 | Jan 2006 | WO |
Entry |
---|
Eades et al. “Multilevel Visualization of Clustered Graphs,” Department of Computer Science and Software Engineering, University if Newcastle, Australia, Proceedings of Graph Drawing '96, Lecture Notes in Computer Science, NR. 1190, (Sep. 1996). |
Eades et al., “Orthogonal Grid Drawing of Clustered Graphs,” Department of Computer Science, the University of Newcastle, Australia, Technical Report 96-04, [Online] 1996, Retrieved from the internet: URL: http://citeseer.ist.psu.edu/eades96ort hogonal.html (1996). |
Estivill-Castro et al. “Amoeba: Hierarchical Clustering Based on Spatial Proximity Using Delaunaty Diagram”, Department of Computer Science, The University of Newcastle, Australia, 1999 ACM Sigmod International Conference on Management of Data, vol. 28, No. 2, Jun. 1999, pp. 49-60, Philadelphia, PA, USA (Jun. 1999). |
Ryall et al., “An Interactive Constraint-Based System for Drawing Graphs,” UIST '97 Proceedings of the 10th Annual ACM Symposium on User Interface Software and Technology, pp. 97-104 (1997). |
Anna Sachinopoulou, “Multidimensional Visualization,” Technical Research Centre of Finland, ESPOO 2001, VTT Research Notes 2114, pp. 1-37 (2001). |
B.B. Hubbard, “The World According the Wavelet: The Story of a Mathematical Technique in the Making,” AK Peters (2nd ed.), pp. 227-229, Massachusetts, USA (1998). |
Baeza-Yates et al., “Modern Information Retrieval,” Ch. 2 “Modeling,” Modern Information Retrieval, Harlow: Addison-Wesley, Great Britain 1999, pp. 18-71 (1999). |
Bernard et al.: “Labeled Radial Drawing of Data Structures” Proceedings of the Seventh International Conference on Information Visualization, Infovis. IEEE Symposium, Jul. 16-18, 2003, Piscataway, NJ, USA, IEEE, Jul. 16, 2003, pp. 479-484, XP010648809, IS. |
Bier et al. “Toolglass and Magic Lenses: The See-Through Interface”, Computer Graphics Proceedings, Proceedings of Siggraph Annual International Conference on Computer Graphics and Interactive Techniques, pp. 73-80, XP000879378 (Aug. 1993). |
Boukhelifa et al., “A Model and Software System for Coordinated and Multiple Views in Exploratory Visualization,” Information Visualization, No. 2, pp. 258-269, GB (2003). |
C. Yip Chung et al., “Thematic Mapping—From Unstructured Documents to Taxonomies,” CIKM'02, Nov. 4-9, 2002, pp. 608-610, ACM, McLean, Virginia, USA (Nov. 4, 2002). |
Chen An et al., “Fuzzy Concept Graph and Application in Web Document Clustering,” IEEE, pp. 101-106 (2001). |
Davison et al., “Brute Force Estimation of the Number of Human Genes Using EST Clustering as a Measure,” IBM Journal of Research & Development, vol. 45, pp. 439-447 (May 2001). |
Eades et al. “Multilevel Visualization of Clustered Graphs,” Department of Computer Science and Software Engineering, University of Newcastle, Australia, Proceedings of Graph Drawing '96, Lecture Notes in Computer Science, NR. 1190, Sep. 18, 1996—Se. |
Eades et al., “Orthogonal Grid Drawing of Clustered Graphs,” Department of Computer Science, the University of Newcastle, Australia, Technical Report 96-04, [Online] 1996, Retrieved from the internet: URL:http://citeseer.ist.psu.edu/eades96ort hogonal.ht. |
Estivill-Castro et al. “Amoeba: Hierarchical Clustering Based on Spatial Proximity Using Delaunaty Diagram”, Department of Computer Science, The University of Newcastle, Australia, 1999 ACM Sigmod International Conference on Management of Data, vol. 28, N. |
F. Can, Incremental Clustering for Dynamic Information Processing: ACM Transactions on Information Systems, ACM, New York, NY, US, vol. 11, No. 2, pp. 143-164, XP-002308022 (Apr. 1993). |
Fekete et al., “Excentric Labeling: Dynamic Neighborhood Labeling for Data Visualization,” CHI 1999 Conference Proceedings Human Factors in Computing Systems, Pittsburgh, PA, pp. 512-519 (May 15-20, 1999). |
http://em-ntserver.unl.edu/Math/mathweb/vecors/vectors.html © 1997. |
Inxight VizServer, “Speeds and Simplifies the Exploration and Sharing of Information”, www.inxight.com/products/vizserver, copyright 2005. |
Jain et al., “Data Clustering: A Review,” ACM Computing Surveys, vol. 31, No. 3, Sep. 1999, pp. 264-323, New York, NY, USA (Sep. 1999). |
James Osborn et al., “Justice: A Jidicial Search Tool Using Intelligent Cencept Extraction,” Department of Computer Science and Software Engineering, University of Melbourne, Australia, ICAIL-99, 1999, pp. 173-181, ACM (1999). |
Jiang Linhui, “K-Mean Algorithm: Iterative Partitioning Clustering Algorithm,” http://www.cs.regina.ca/-linhui/K.sub.--mean.sub.--algorithm.html, (2001) Computer Science Department, University of Regina, Saskatchewan, Canada (2001). |
Kanungo et al., “The Analysis of a Simple K-Means Clustering Algorithm,” pp. 100-109, PROC 16th annual symposium of computational geometry (May 2000). |
Kawano, Hiroyuki., “Overview of Mondou Web Search Engine Using Text Mining and Information Visualizing Technologies,” IEEE, 2001, pp. 234-241. |
Kazumasa Ozawa, “A Stratificational Overlapping Cluster Scheme,” Information Science Center, Osaka Electro-Communication University, Neyagawa-shi, Osaka 572, Japan, Pattern Recognition, vol. 18, pp. 279-286 (1985). |
Kohonen, T., “Self-Organizing Maps,” Ch. 1-2, Springer-Verlag (3rd ed.) (2001). |
Kurimo M., “Fast Latent Semantic Indexing of Spoken Documents by Using Self-Organizing Maps” IEEE International Conference on Accoustics, Speech, and Signal Processing, vol. 6, pp. 2425-2428 (Jun. 2000). |
Lam et al., “A Sliding Window Technique for Word Recognition,” SPIE, vol. 2422, pp. 38-46, Center of Excellence for Document Analysis and Recognition, State University of New Yrok at Baffalo, NY, USA (1995). |
Lio et al., “Funding Pathogenicity Islands and Gene Transfer Events in Genome Data,” Bioinformatics, vol. 16, pp. 932-940, Department of Zoology, University of Cambridge, UK (Jan. 25, 2000). |
Artero et al., “Viz3D: Effective Exploratory Visualization of Large Multidimensional Data Sets,” IEEE Computer Graphics and Image Processing, pp. 340-347 (Oct. 20, 2004). |
Magarshak, Greg., Theory & Practice. Issue 01. May 17, 2000. http://www.flipcode.com/articles/tp.sub.--issue01-pf.shtml (May 17, 2000). |
Maria Cristin Ferreira de Oliveira et al., “From Visual Data Exploration to Visual Data Mining: A Survey,” Jul.-Sep. 2003, IEEE Transactions on Visualization and Computer Graphics, vol. 9, No. 3, pp. 378-394 (Jul. 2003). |
Rauber et al., “Text Mining in the SOMLib Digital Library System: The Representation of Topics and Genres,” Applied Intelligence 18, pp. 271-293, 2003 Kluwer Academic Publishers (2003). |
Miller et al., “Topic Islands: A Wavelet Based Text Visualization System,” Proceedings of the IEEE Visualization Conference. 1998, pp. 189-196. |
North et al. “A Taxonomy of Multiple Window Coordinations,” Institute for Systems Research & Department of Computer Science, University of Maryland, Maryland, USA, http://www.cs.umd.edu/localphp/hcil/tech-reports-search.php?number=97-18 (1997). |
Shuldberg et al., “Distilling Information from Text: The EDS TemplateFiller System,” Journal of the American Society for Information Science, vol. 44, pp. 493-507 (1993). |
Pelleg et al., “Accelerating Exact K-Means Algorithms With Geometric Reasoning,” pp. 277-281, CONF on Knowledge Discovery in Data, PROC fifth ACM SIGKDD (1999). |
R.E. Horn, “Communication Units, Morphology, and Syntax,” Visual Language: Global Communication for the 21st Century, 1998, Ch. 3, pp. 51-92, MacroVU Press, Bainbridge Island, Washington, USA. |
Slaney, M., et al., “Multimedia Edges: Finding Hierarchy in all Dimensions” PROC. 9-th ACM Intl. Conf. on Multimedia, pp. 29-40, ISBN. 1-58113-394-4, Sep. 30, 2001, XP002295016 Ottawa (Sep. 3, 2001). |
Strehl et al., “Cluster Ensembles—A Knowledge Reuse Framework for Combining Partitioning,” Journal of Machine Learning Research, MIT Press, Cambridge, MA, US, ISSN: 1533-7928, vol. 3, No. 12, pp. 583-617, XP002390603 (Dec. 2002). |
Sullivan, Dan., “Document Warehousing and Text Mining: Techniques for Improving Business Operations, Marketing and Sales,” Ch. 1-3, John Wiley & Sons, New York, NY (2001). |
V. Faber, “Clustering and the Continuous K-Means Algorithm,” Los Alamos Science, The Laboratory, Los Alamos, NM, US, No. 22, Jan. 1, 1994, pp. 138-144 (Jan. 1, 1994). |
Wang et al., “Learning text classifier using the domain concept hierarchy,” Communications, Circuits and Systems and West Sino Expositions, IEEE 2002 International Conference on Jun. 29-Jul. 1, 2002, Piscataway, NJ, USA, IEEE, vol. 2, pp. 1230-1234 (2002). |
Whiting et al., “Image Quantization: Statistics and Modeling,” SPIE Conference of Physics of Medical Imaging, San Diego, CA, USA , vol. 3336, pp. 260-271 (Feb. 1998). |
O'Neill et al., “DISCO: Intelligent Help for Document Review,” 12th International Conference on Artificial Intelligence and Law, Barcelona, Spain, Jun. 8, 2009, pp. 1-10, ICAIL 2009, Association for Computing Machinery, Red Hook, New York (Online); XP 002607216. |
McNee, “Meeting User Information Needs in Recommender Systems,” Ph.D. Dissertation, University of Minnesota-Twin Cities, Jun. 2006. |
S.S. Weng, C.K. Liu, “Using text classification and multiple concepts to answer e-mails.” Expert Systems with Applications, 26 (2004), pp. 529-543. |
Number | Date | Country | |
---|---|---|---|
20110029536 A1 | Feb 2011 | US |
Number | Date | Country | |
---|---|---|---|
61229216 | Jul 2009 | US | |
61236490 | Aug 2009 | US |