The present disclosure relates in general to data storage and more specifically to a method for performing automated discovery of new topics in a corpus.
As storage and availability of data grows, a large amount of time is spent identifying data relationships for discovery of new topics. Conventionally, the discovery of new topics is oftentimes performed manually by repetitive work leading to wasting valuable time of users.
Information can have great value. Assembling and maintaining a database to store information involves real costs, such as the costs to acquire information, the costs associated with physical assets used to house, secure, and make the information available, and labor costs to manage the information.
As computer processors are becoming more powerful, it would be particularly useful to save the time that an individual conventionally spends discovering new topics and identifying relationship criteria with existing models, or between the source and the target.
Oftentimes there are simple transformations, or complex topic identification across a large corpus of documents from any subject domain, requiring a lot of user's time for discovery of relationships associated with existing data.
Thus, there is a need for a simple and flexible method which assists users in connection with performing automated discovery of new topics, employing a new topic database for comparison with the existing topics for new application environments.
Embodiments of the present disclosure provide a method for performing automated discovery of new topics from unlimited documents related to any subject domain, employing a multi-component extension of Latent Dirichlet Allocation (MC-LDA) topic models, to discover related topics in a corpus. The resulting data may contain millions of term vectors from any subject domain identifying the most distinguished co-occurring topics that users may be interested in, which may be employed to create a Master Topic Model.
In accordance with one aspect of the present disclosure, the method for automated discovery of new topics may include multiple topic identification models with different number of term vectors and other parameters. For example a topic identification model with 64 term vectors may provide a broader topic scope, while models with 256, 1024, or 16K term vectors may provide more specific fine-grained topics.
According to another embodiment, a new data may contain a large number of entities/topics in a database, which may be used periodically to parse and extract data from topics that users may be interested in. This method may identify term vectors to change detection using term vector differences with no correlation in the Master Topic Model to compare and measure the significance of these changes, based on established thresholds to identify the similarity of the topics found by comparing one by one with topics from Periodic New model.
The present disclosure may provide a method for automated discovery of new topics in a corpus, using new content and comparing it to the existing model for periodically building new topic ID model database compressed into the smallest memory footprint possible, for providing fuzzy indexing to allow query-time linking on massive data sets, providing reliability and fault-tolerance through data, which may prevent software and hardware redundancy.
In one embodiment, a method comprises automatically extracting, by a database source computer, from a document corpus, data associated with a plurality of co-occurring topics; in response to automatically extracting the plurality of co-occurring topics, extracting, by a synchronizing framework computer, a plurality of topic identifies from the plurality of co-occurring topics; creating, by the synchronizing framework computer, a master topic computer model for the document corpus from a first plurality of term vectors; creating, by the synchronizing framework computer, a periodic new topic computer model by comparing topic significance among the plurality of topic identifiers, the periodic new topic computer model including a second plurality of term vectors; and selecting, by the synchronizing framework computer, one or more new topics by identifying one or more term vectors from the second plurality of term vectors in the periodic new topic computer model that have no correlation with the first plurality of term vectors in the master topic computer model.
In another embodiment, a system comprises a database source computer module configured to extract data associated with a plurality of co-occurring topics in a document corpus; and a synchronizing framework computer module configured to: (a) extract a plurality of topic identifies from the plurality of co-occurring topics; (b) create a master topic computer model for the document corpus from a first plurality of term vectors; (c) create a periodic new topic computer model by comparing topic significance among the plurality of topic identifiers, the periodic new topic computer model including a second plurality of term vectors; and (d) select one or more new topics by identifying one or more term vectors from the second plurality of term vectors in the periodic new topic computer model that have no correlation with the first plurality of term vectors in the master topic computer model.
In another embodiment, a non-transitory computer readable medium having stored thereon computer executable instructions executed by a processor comprises automatically extracting, by a processor executing a database source computer module, from a document corpus data associated with a plurality of co-occurring topics; in response to automatically extracting the plurality of co-occurring topics, extracting, by the processor executing a synchronizing framework computer module, a plurality of topic identifies from the plurality of co-occurring topics; creating, by the processor executing the synchronizing framework computer, a master topic computer model for the document corpus from a first plurality of term vectors; creating, by the processor executing the synchronizing framework computer, a periodic new topic computer model by comparing topic significance among the plurality of topic identifiers, the periodic new topic computer model including a second plurality of term vectors; and selecting, by the processor executing the synchronizing framework computer, one or more new topics by identifying one or more term vectors from the second plurality of term vectors in the periodic new topic computer model that have no correlation with the first plurality of term vectors in the master topic computer model.
Numerous other aspects, features, and benefits of the present disclosure may be made apparent from the following detailed description taken together with the drawing features.
The present disclosure can be better understood by referring to the following figures. The components in the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the disclosure. In the figures, reference numerals designate corresponding parts throughout the different views.
The present disclosure is here described in detail with reference to embodiments illustrated in the drawings, which form a part hereof. Other embodiments may be used and/or other changes may be made without departing from the spirit or scope of the present disclosure. The illustrative embodiments described in the detailed description are not meant to be limiting of the subject matter presented herein.
As used herein, the following terms have the following definitions:
“Parse” refers to analyzing the source code of a computer program to make sure that it is structurally correct before it is compiled and turned into machine code.
“Term vector” refers to an algebraic model for representing text documents (and any objects, in general) as vectors of identifiers, such as, for example, index terms. It is used in information filtering, information retrieval, indexing, and relevancy rankings.
“Database” refers to any system including any combination of clusters and modules suitable for storing one or more collections and suitable to process one or more queries.
“Document” refers to a discrete electronic representation of information having a start and end.
“Multi-Document” refers to a document with its tokens, different types of named entities, and key phrases organized into separate “bag-of-surface-forms” components.
“Corpus” refers to a collection of one or more documents.
“Feature” refers to any information which is at least partially derived from a document.
“Cluster” refers to a collection of features.
“Memory” refers to any hardware component suitable for storing information and retrieving said information at a sufficiently high speed.
“Module” refers to a computer software and/or hardware component suitable for carrying out one or more defined tasks.
“Topic” refers to a set of thematic information which is at least partially derived from a corpus.
“Query” refers to a request to retrieve information from one or more suitable databases.
Various aspects of the present disclosure describe a system and method for automated discovery of new topics in a corpus based on a concept of co-occurring topics from different pre-built topic models. These different topic models are built with different levels of granularity of topics, vocabulary and converging parameters, thus providing a vertical hierarchy/scalability over a specific domain of interest. Embodiments of the present disclosure extend the conventional LDA topic modeling to support multi component LDA, where each component is treated as conditionally-independent, given document topic proportions. These components can include features like terms, key phrases, entities, facts, among others. Thus, this approach provides a concept of horizontal scalability of the topic models over a specific domain. The combination of the vertical vocabulary and horizontal feature selection in the pre-built topic models, provides varied dimensions of co-occurring topics, which on appropriate clustering and differential training via MEMDB can produce new topics. These new topics would not exists in the pre-built topic models to begin with, but could be discovered by running the documents in parallel across all the pre-built topic models.
Embodiments of the present disclosure describe a computer executed method for automated discovery of new topics that may facilitate the automated determination of relationships of corresponding term vectors from any subject domain identifying the most distinguished co-occurring topics that users may be interested in, which may be employed to create a Master Topic Model.
According to an embodiment, a term vector component may be a search component configured to return information about documents. In the term vector space model of information retrieval, the documents are modeled as vectors in a high-dimensional space of millions of terms. The terms are derived from words and phrases in the document, which are weighted by their importance within the document and within the corpus of documents. Each document's vector seeks to represent the document in a “vector space,” allowing comparison with vectors derived from other sources, for example, queries, or other documents. Term vectors may be used as the basis of successful algorithms for document ranking, document filtering, document clustering, and relevance feedback.
The embodiments recite a procedure for automated discovery of new topics in a corpus based on a concept of co-occurring topics from different pre-built topic models. These different topic models are built with different levels of granularity of topics, vocabulary, and converging parameters, thereby providing a vertical hierarchy/scalability over a specific domain of interest. The embodiments can extend LDA topic modeling to support multi-component LDA, where each component is treated as conditionally-independent, given document topic proportions. These components can include features, such as terms, key phrases, entities, facts, etc. Thus, this approach can provide a concept of horizontal scalability of the topic models over a specific domain. The combination of the vertical vocabulary and horizontal feature selection in the pre-built topic models provides varied dimensions of co-occurring topics, which on appropriate clustering and differential training via an in-memory database can produce new topics. These new topics would not exist in the pre-built topic models, to begin with, but could be discovered by running the documents in parallel across all the pre-built topic models.
A System for Automated Discovery of New Topics
In accordance with one aspect of the present disclosure, the system for automated discovery of new topics may include one or more central servers having a plurality of special purpose software and hardware computer modules, including a database source module 102 which may contain a large number of entities/topics that users may be interested in. The resulting data may contain a large number of term vectors from any subject domain identifying the most distinguished co-occurring topics that users may be interested in, which may be employed to implement a Master Topic Model computer module 104.
Although the system architecture 100 includes a single database source module 102 and a single destination in-memory database module 112, it is to be understood and appreciated that the novel functionality of a system and method for automatic discovery of new topics may be employed with any number of sources and/or destination components, which may be remotely located and accessed.
Embodiments of the present disclosure may be directed to a system and method for automated discovery of new topics, which may include multiple topic identification models with different numbers of term vectors and other parameters. For example, a topic identification model with 64 term vectors may provide a broader topic scope, while models with 256, 1024, or 16K term vectors may provide more specific fine-grained topics. Each topic or document may be analyzed on co-occurring topics across models to discover related topics characterized by a particular set of term vectors, making each individual word exchangeable, having good probabilities of generating new term vectors facilitating the automated discovery of new topics.
According to principles of the present disclosure, the system and method for automated discovery of new topics periodically may use new data in the database source module 102 to select data of interest or item feature.
This information may be used periodically to parse and extract data from topics that users may be interested in, to compare all term vectors from Master Topic Model module 104 with no correlation with term vectors of Periodic New Model module 106 employing a Detector of Term Vector Differences module 108. The system measures the significances of the changes by comparing each term vector one by one, selecting the more specific term vectors that do not correlate or have similarities with Master Topic Model 104, employing different methods or any suitable method existing for this type of comparison.
An embodiment of the present disclosure may include a synchronization frame work computer module 110 which may be a framework of data collection interfaces that may communicate with database source computer module 102 and pull data items that may contain relevant information to a project. Employing this process may generate a new set of topics to produce from zero to unlimited number of topics, which may be added to Master Topic Model 104 for periodical updates of automated discovery of new topics in a corpus, using the new content and comparing it to the existing model for periodical building of new topic ID model in-memory database 112. The in-memory database 112 may be compressed into the smallest memory footprint possible for providing fuzzy indexing to allow query-time linking on massive data-sets, providing reliability and fault tolerance through data for automated discovery of new topics in a corpus.
The actual software code or specialized control hardware used to implement these systems, modules and methods are not limiting the invention. Thus, the operation and behavior of the systems, modules and methods were described without reference to the specific software code, being understood that software and control hardware may be designed to implement the systems, modules and methods based on the description herein.
A Method for Automated Discovery of New Topics
As may be seen in
In step 208, the method may periodically run a new set of data to select topics of interest from a very large collection of co-occurring entities extracted from a document corpus of the targeted domain. This new data may be analyzed to discover a relationship between data elements. In addition, topic identifiers may be extracted to improve precision for creation of a Periodic New Model, step 210, using a Detector of Term Vectors Differences module 108 of the synchronizing framework module 110 to compare and measure the significance of topics based on established thresholds, for periodically building new topic ID models using new content to identify the similarity of topics found. In step 212, term vectors from Periodic New Model having no correlation with term vectors of Master Topic Model are identified, where all term vectors are compared one by one with topics from Master Topic Model. In step 214, all differences may change detection of term vector differences.
The next step 216 involves the addition of selected topics to Master Topic Model, which, in step 218, may be used to periodically build a new topics ID model to compress data into smallest memory possible configured to fit into in-memory database 112. In embodiments, the in-memory database 112 may have an advanced searching and imbedded record linking capabilities to provide fuzzy indexing, matching and match scores and non-exclusionary searching to provide in-database analytics and to allow query-time linking on massive data-set for automated discovery of new topics.
The foregoing method descriptions and the process flow diagrams are provided merely as illustrative examples and are not intended to require or imply that the steps of the various embodiments must be performed in the order presented. As will be appreciated by one of skill in the art the steps in the foregoing embodiments may be performed in any order. Words such as “then,” “next,” etc. may not be intended to limit the order of the steps; these words are simply used to guide the reader through the description of the methods. Although process flow diagrams may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be re-arranged. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination may correspond to a return of the function to the calling function or the main function.
The various illustrative logical blocks, modules, circuits, and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
Embodiments implemented in computer software may be implemented in software, firmware, middleware, microcode, hardware description languages, or any combination thereof. A code segment or machine-executable instructions may represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a class, or any combination of instructions, data structures, or program statements. A code segment may be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, or memory contents. Information, arguments, parameters, data, etc. may be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, token passing, network transmission, etc.
The actual software code or specialized control hardware used to implement these systems and methods is not limiting of the invention. Thus, the operation and behavior of the systems and methods were described without reference to the specific software code being understood that software and control hardware can be designed to implement the systems and methods based on the description herein.
When implemented in software, the functions may be stored as one or more instructions or code on a non-transitory computer-readable or processor-readable storage medium. The steps of a method or algorithm disclosed herein may be embodied in a processor-executable software module which may reside on a computer-readable or processor-readable storage medium. A non-transitory computer-readable or processor-readable media includes both computer storage media and tangible storage media that facilitate transfer of a computer program from one place to another. A non-transitory processor-readable storage media may be any available media that may be accessed by a computer. By way of example, and not limitation, such non-transitory processor-readable media may comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other tangible storage medium that may be used to store desired program code in the form of instructions or data structures and that may be accessed by a computer or processor. Disk and disc, as used herein, include compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk, and blu-ray disc where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. Combinations of the above should also be included within the scope of computer-readable media. Additionally, the operations of a method or algorithm may reside as one or any combination or set of codes and/or instructions on a non-transitory processor-readable medium and/or computer-readable medium, which may be incorporated into a computer program product.
The preceding description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the following claims and the principles and novel features disclosed.
This application is a continuation of U.S. patent application Ser. No. 14/873,635, entitled “Method of Automated Discovery of New Topics,” filed Oct. 2, 2015, which is a continuation of U.S. patent application Ser. No. 14/558,076, entitled “Method for Automated Discovery of New Topics,” filed on Dec. 2, 2014, which is a non-provisional patent application that claims the benefit of U.S. Provisional Application No. 61/910,763, entitled “Method for Automated Discovery of New Topics,” filed Dec. 2, 2013, each of which are hereby incorporated by reference herein in their entirety. This application is related to U.S. application Ser. No. 14/557,794, entitled “Method for Disambiguating Features in Unstructured Text,” filed Dec. 2, 2014; U.S. application Ser. No. 14/558,300, entitled “Event Detection Through Text Analysis Using Trained Event Template Models,” filed Dec. 2, 2014; and U.S. application Ser. No. 14/557,906, entitled “Method of Automated Discovery of Topic Relatedness,” filed Dec. 2, 2014; each of which are hereby incorporated by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
6128660 | Grimm et al. | Oct 2000 | A |
6178529 | Short et al. | Jan 2001 | B1 |
6266781 | Chung et al. | Jul 2001 | B1 |
6353926 | Parthesarathy et al. | Mar 2002 | B1 |
6738759 | Wheeler et al. | May 2004 | B1 |
6832373 | O'Neill | Dec 2004 | B2 |
6832737 | Karlsson et al. | Dec 2004 | B2 |
7058846 | Kelkar et al. | Jun 2006 | B1 |
7370323 | Marinelli et al. | May 2008 | B2 |
7421478 | Muchow | Sep 2008 | B1 |
7447940 | Peddada | Nov 2008 | B2 |
7543174 | van Rietschote et al. | Jun 2009 | B1 |
7681075 | Havemose et al. | Mar 2010 | B2 |
7818615 | Krajewski et al. | Oct 2010 | B2 |
7899871 | Kumar et al. | Mar 2011 | B1 |
8055933 | Jaehde et al. | Nov 2011 | B2 |
8090717 | Bharat et al. | Jan 2012 | B1 |
8122026 | Laroco et al. | Feb 2012 | B1 |
8122047 | Kanigsberg et al. | Feb 2012 | B2 |
8341622 | Eatough | Dec 2012 | B1 |
8345998 | Malik et al. | Jan 2013 | B2 |
8356036 | Bechtel et al. | Jan 2013 | B2 |
8375073 | Jain | Feb 2013 | B1 |
8423522 | Lang et al. | Apr 2013 | B2 |
8429256 | Vidal et al. | Apr 2013 | B2 |
8645298 | Hennig et al. | Feb 2014 | B2 |
8726267 | Li et al. | May 2014 | B2 |
8782018 | Shim et al. | Jul 2014 | B2 |
8972396 | Zhang et al. | Mar 2015 | B1 |
8995717 | Cheng et al. | Mar 2015 | B2 |
9009153 | Khan et al. | Apr 2015 | B2 |
9025892 | Lightner et al. | May 2015 | B1 |
9032387 | Hill et al. | May 2015 | B1 |
20010037398 | Chao et al. | Nov 2001 | A1 |
20020031260 | Thawonmas et al. | Mar 2002 | A1 |
20020165847 | McCartney et al. | Nov 2002 | A1 |
20020174138 | Nakamura | Nov 2002 | A1 |
20030028869 | Drake et al. | Feb 2003 | A1 |
20030112792 | Cranor et al. | Jun 2003 | A1 |
20030158839 | Faybishenko et al. | Aug 2003 | A1 |
20030182282 | Ripley | Sep 2003 | A1 |
20040010502 | Bomfim et al. | Jan 2004 | A1 |
20040027349 | Landau et al. | Feb 2004 | A1 |
20040049478 | Jasper et al. | Mar 2004 | A1 |
20040143571 | Bjornson et al. | Jul 2004 | A1 |
20040153869 | Marinelli et al. | Aug 2004 | A1 |
20040205064 | Zhou et al. | Oct 2004 | A1 |
20040215755 | O'Neill | Oct 2004 | A1 |
20040243645 | Broder et al. | Dec 2004 | A1 |
20050192994 | Caldwell et al. | Sep 2005 | A1 |
20060101081 | Lin et al. | May 2006 | A1 |
20060122978 | Brill et al. | Jun 2006 | A1 |
20060294071 | Weare et al. | Dec 2006 | A1 |
20070005639 | Gaussier et al. | Jan 2007 | A1 |
20070005654 | Schachar et al. | Jan 2007 | A1 |
20070073708 | Smith et al. | Mar 2007 | A1 |
20070156748 | Emam et al. | Jul 2007 | A1 |
20070174167 | Natella et al. | Jul 2007 | A1 |
20070203693 | Estes | Aug 2007 | A1 |
20070203924 | Guha et al. | Aug 2007 | A1 |
20070240152 | Li et al. | Oct 2007 | A1 |
20070250501 | Grubb et al. | Oct 2007 | A1 |
20070250519 | Fineberg et al. | Oct 2007 | A1 |
20070282959 | Stern | Dec 2007 | A1 |
20080010683 | Baddour et al. | Jan 2008 | A1 |
20080027920 | Schipunov et al. | Jan 2008 | A1 |
20080077570 | Tang et al. | Mar 2008 | A1 |
20090019013 | Tareen et al. | Jan 2009 | A1 |
20090043792 | Barsness et al. | Feb 2009 | A1 |
20090049038 | Gross | Feb 2009 | A1 |
20090089626 | Gotch et al. | Apr 2009 | A1 |
20090094484 | Son et al. | Apr 2009 | A1 |
20090144609 | Liang et al. | Jun 2009 | A1 |
20090216734 | Aghajanyan et al. | Aug 2009 | A1 |
20090222395 | Light et al. | Sep 2009 | A1 |
20090240682 | Balmin et al. | Sep 2009 | A1 |
20090292660 | Behal et al. | Nov 2009 | A1 |
20090299999 | Loui et al. | Dec 2009 | A1 |
20090322756 | Robertson et al. | Dec 2009 | A1 |
20100077001 | Vogel et al. | Mar 2010 | A1 |
20100138931 | Thorley et al. | Jun 2010 | A1 |
20100161566 | Adair et al. | Jun 2010 | A1 |
20100223264 | Brucker et al. | Sep 2010 | A1 |
20100235311 | Cao et al. | Sep 2010 | A1 |
20100274785 | Procopiuc et al. | Oct 2010 | A1 |
20110047167 | Caceres | Feb 2011 | A1 |
20110071975 | Friedlander et al. | Mar 2011 | A1 |
20110093471 | Brockway et al. | Apr 2011 | A1 |
20110119243 | Diamond et al. | May 2011 | A1 |
20110125764 | Carmel et al. | May 2011 | A1 |
20110161333 | Langseth et al. | Jun 2011 | A1 |
20110282888 | Koperski et al. | Nov 2011 | A1 |
20110296390 | Vidal et al. | Dec 2011 | A1 |
20110296397 | Vidal et al. | Dec 2011 | A1 |
20110302163 | Rhinelander | Dec 2011 | A1 |
20120016875 | Jin et al. | Jan 2012 | A1 |
20120030220 | Edwards et al. | Feb 2012 | A1 |
20120059839 | Andrade et al. | Mar 2012 | A1 |
20120102121 | Wu et al. | Apr 2012 | A1 |
20120117069 | Kawanishi et al. | May 2012 | A1 |
20120131139 | Siripurapu et al. | May 2012 | A1 |
20120143911 | Liebald et al. | Jun 2012 | A1 |
20120246154 | Duan et al. | Sep 2012 | A1 |
20120310934 | Peh et al. | Dec 2012 | A1 |
20120323839 | Kiciman et al. | Dec 2012 | A1 |
20130036076 | Yang et al. | Feb 2013 | A1 |
20130132405 | Bestgen et al. | May 2013 | A1 |
20130166480 | Popescu et al. | Jun 2013 | A1 |
20130166547 | Pasumarthi et al. | Jun 2013 | A1 |
20130290232 | Tsytsarau et al. | Oct 2013 | A1 |
20130303198 | Sadasivam et al. | Nov 2013 | A1 |
20130325660 | Callaway | Dec 2013 | A1 |
20130326325 | De et al. | Dec 2013 | A1 |
20140013233 | Ahlberg et al. | Jan 2014 | A1 |
20140022100 | Fallon et al. | Jan 2014 | A1 |
20140046921 | Bau | Feb 2014 | A1 |
20140089237 | Adibi | Mar 2014 | A1 |
20140156634 | Buchmann et al. | Jun 2014 | A1 |
20140244550 | Jin et al. | Aug 2014 | A1 |
20140255003 | Abramson et al. | Sep 2014 | A1 |
20140280183 | Brown et al. | Sep 2014 | A1 |
20140351233 | Crupi et al. | Nov 2014 | A1 |
20150074037 | Sarferaz | Mar 2015 | A1 |
20150154079 | Lightner et al. | Jun 2015 | A1 |
20150154264 | Lightner et al. | Jun 2015 | A1 |
20150154297 | Lightner et al. | Jun 2015 | A1 |
Number | Date | Country |
---|---|---|
2013003770 | Jan 2013 | WO |
Entry |
---|
Jorg Becker, Dominik Kuropka, Witlod Ambramowicz, Gary Klein (eds.), “Topic-based Vector Space Model,” Business Information Systems, Proceedings of BIS 2003, Colorado Springs, USA. |
International Search Report and Written Opinion dated Mar. 6, 2015 corresponding to International Patent Application No. PCT/US2014/067993, 9 pages. |
International Search Report and Written Opinion dated Mar. 10, 2015 corresponding to International Patent Application No. PCT/US2014/067999, 10 pages. |
International Search Report and Written Opinion of the International Searching Authority dated Apr. 15, 2015 corresponding to International Patent Application No. PCT/US2014/068002, 10 pages. |
International Search Report and Written Opinion dated Feb. 24, 2015 corresponding to International Patent Application No. PCT/US2014/067918, 10 pages. |
International Search Report and Written Opinion dated Mar. 3, 2015 corresponding to International Patent Application No. PCT/US2014/067921, 10 pages. |
International Search Report and Written Opinion of the International Searching Authority dated Apr. 15, 2015, corresponding to International Patent Application No. PCT/2014/067994, 9 pages. |
Blei, D.M., et al., “Latent Dririchlet Allocation,” Journal of Machine Learning Research, vol. 3, 2003, pp. 993-1022. |
Chuang, S.L., et al., A Practical Web-based Approach to Generating Topic Hierarchy for Text Segments,' Institute of Information Science, Academia Sinica, Taiwan, R.O.C., 2004, pp. 127-136. |
Wang, C., et al., “Automatic Online News Issue Construction in Web Environment,” Tsinghuya Univ., Beijing, China, 2008, pp. 457-466. |
Tunkelang, D., “Faceted Search,” Morgan & Claypool Publ., 2009, pp. i-79. |
Schuth, A., et al., “University of Amsterdam Data Centric Ad Hoc and Faceted Search Runs,” ISLA, 2012, pp. 155-160. |
Tools, Search Query Suggestions using ElasticSearch via Shingle Filter and Facets, Nov. 2012, pp. 1-12. |
Vizard, The Rise of In-Memory Databases, Jul. 13, 2012, 3 pages. |
Number | Date | Country | |
---|---|---|---|
20160042276 A1 | Feb 2016 | US |
Number | Date | Country | |
---|---|---|---|
61910763 | Dec 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14873635 | Oct 2015 | US |
Child | 14919631 | US | |
Parent | 14558076 | Dec 2014 | US |
Child | 14873635 | US |