Systems and methods for key phrase characterization of documents

Information

  • Patent Grant
  • 10162887
  • Patent Number
    10,162,887
  • Date Filed
    Monday, April 10, 2017
    8 years ago
  • Date Issued
    Tuesday, December 25, 2018
    6 years ago
Abstract
Systems and methods are disclosed for key phrase characterization of documents. In accordance with one implementation, a method is provided for key phrase characterization of documents. The method includes obtaining a first plurality of documents based at least on a user input, obtaining a statistical model based at least on the user input, and obtaining, from content of the first plurality of documents, a plurality of segments. The method also includes determining statistical significance of the plurality of segments based at least on the statistical model and the content, and providing for display a representative segment from the plurality of segments, the representative segment being determined based at least on the statistical significance.
Description
BACKGROUND

Vast amounts of data are readily available to readers, analysts, and researchers today, on the one hand allowing them to perform more complicated and detailed data analyses than ever, but on the other hand making it more difficult to quickly sort through the data. Automatically characterizing the data in a concise and informative way can help users to identify data that is most relevant for their particular needs.





BRIEF DESCRIPTION OF THE DRAWINGS

Reference will now be made to the accompanying drawings, which illustrate exemplary embodiments of the present disclosure. In the drawings:



FIG. 1 is a block diagram of an exemplary electronic device, consistent with embodiments of the present disclosure;



FIG. 2 is a flowchart of an exemplary method for generating a statistical model, consistent with embodiments of the present disclosure;



FIG. 3 is a flowchart of an exemplary method for key phrase characterization of documents, consistent with embodiments of the present disclosure; and



FIGS. 4A-4C illustrate an exemplary display showing exemplary user interactions, consistent with embodiments of the present disclosure.





DESCRIPTION OF EXEMPLARY EMBODIMENTS

Reference will now be made in detail to several exemplary embodiments of the present disclosure, including those illustrated in the accompanying drawings. Whenever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts.


Embodiments disclosed herein are directed to, among other things, systems and methods that automatically characterize documents by identifying key phrases describing the contents of those documents. More specifically, the systems and methods can, among other things, obtain a first plurality of documents based at least on a user input; obtain a statistical model based at least on the user input; obtain, from content of the first plurality of documents, a plurality of segments; determine statistical significance of the plurality of segments based at least on the statistical model and the content; and provide for display a representative segment from the plurality of segments, the representative segment being determined based at least on the statistical significance.


According to some embodiments, the operations, techniques, and/or components described herein can be implemented by an electronic device, which can include one or more special-purpose computing devices. The special-purpose computing devices can be hard-wired to perform the operations, techniques, and/or components described herein, or can include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the operations, techniques and/or components described herein, or can include one or more hardware processors programmed to perform such features of the present disclosure pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices can also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the technique and other features of the present disclosure. The special-purpose computing devices can be desktop computer systems, portable computer systems, handheld devices, networking devices, or any other device that incorporates hard-wired and/or program logic to implement the techniques and other features of the present disclosure.


The one or more special-purpose computing devices can be generally controlled and coordinated by operating system software, such as iOS, Android, Blackberry, Chrome OS, Windows XP, Windows Vista, Windows 7, Windows 8, Windows Server, Windows CE, Unix, Linux, SunOS, Solaris, VxWorks, or other compatible operating systems. In other embodiments, the computing device can be controlled by a proprietary operating system. Operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface functionality, such as a graphical user interface (“GUI”), among other things.


By way of example, FIG. 1 is a block diagram that illustrates an implementation of an electronic device 110, which, as described above, can comprise one or more electronic devices. Electronic device 110 includes a bus 102 or other communication mechanism for communicating information, and one or more hardware processors 104, coupled with bus 102 for processing information. One or more hardware processors 104 can be, for example, one or more microprocessors.


Electronic device 110 also includes a main memory 106, such as a random access memory (RAM) or other dynamic storage device, coupled to bus 102 for storing information and instructions to be executed by processor 104. Main memory 106 also can be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 104. Such instructions, when stored in non-transitory storage media accessible to one or more processors 104, render electronic device 110 into a special-purpose machine that is customized to perform the operations specified in the instructions.


Electronic device 110 further includes a read only memory (ROM) 108 or other static storage device coupled to bus 102 for storing static information and instructions for processor 104. A storage device 150, such as a magnetic disk, optical disk, or USB thumb drive (Flash drive), etc., is provided and coupled to bus 102 for storing information and instructions.


Electronic device 110 can be coupled via bus 102 to a display 112, such as a cathode ray tube (CRT), an LCD display, or a touchscreen, for displaying information to a computer user. An input device 114, including alphanumeric and other keys, is coupled to bus 102 for communicating information and command selections to processor 104. Another type of user input device is cursor control 116, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 104 and for controlling cursor movement on display 112. The input device typically has two degrees of freedom in two axes, a first axis (for example, x) and a second axis (for example, y), that allows the device to specify positions in a plane. In some embodiments, the same direction information and command selections as cursor control may be implemented via receiving touches on a touch screen without a cursor.


Electronic device 110 can include a user interface module to implement a GUI that may be stored in a mass storage device as executable software codes that are executed by the one or more computing devices. This and other modules may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.


In general, the word “module,” as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, Lua, C, and C++. A software module can be compiled and linked into an executable program, installed in a dynamic link library, or written in an interpreted programming language such as, for example, BASIC, Perl, Python, or Pig. It will be appreciated that software modules can be callable from other modules or from themselves, and/or can be invoked in response to detected events or interrupts. Software modules configured for execution on computing devices can be provided on a computer readable medium, such as a compact disc, digital video disc, flash drive, magnetic disc, or any other tangible medium, or as a digital download (and can be originally stored in a compressed or installable format that requires installation, decompression, or decryption prior to execution). Such software code can be stored, partially or fully, on a memory device of the executing computing device, for execution by the computing device. Software instructions can be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules can be comprised of connected logic units, such as gates and flip-flops, and/or can be comprised of programmable units, such as programmable gate arrays or processors. The modules or computing device functionality described herein are preferably implemented as software modules, but can be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage.


Electronic device 110 can implement the techniques and other features described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the electronic device causes or programs electronic device 110 to be a special-purpose machine. According to some embodiments, the techniques and other features described herein are performed by electronic device 110 in response to one or more processors 104 executing one or more sequences of one or more instructions contained in main memory 106. Such instructions can be read into main memory 106 from another storage medium, such as storage device 150. Execution of the sequences of instructions contained in main memory 106 causes processor 104 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry can be used in place of or in combination with software instructions.


The term “non-transitory media” as used herein refers to any media storing data and/or instructions that cause a machine to operate in a specific fashion. Such non-transitory media can comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 150. Volatile media includes dynamic memory, such as main memory 106. Common forms of non-transitory media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge, and networked versions of the same.


Non-transitory media is distinct from, but can be used in conjunction with, transmission media. Transmission media participates in transferring information between storage media. For example, transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 102. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.


Various forms of media can be involved in carrying one or more sequences of one or more instructions to processor 104 for execution. For example, the instructions can initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to electronic device 110 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 102. Bus 102 carries the data to main memory 106, from which processor 104 retrieves and executes the instructions. The instructions received by main memory 106 can optionally be stored on storage device 150 either before or after execution by processor 104.


Electronic device 110 also includes a communication interface 118 coupled to bus 102. Communication interface 118 provides a two-way data communication coupling to a network link 120 that is connected to a local network 122. For example, communication interface 118 can be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 118 can be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links can also be implemented. In any such implementation, communication interface 118 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.


Network link 120 typically provides data communication through one or more networks to other data devices. For example, network link 120 can provide a connection through local network 122 to a host computer 124 or to data equipment operated by an Internet Service Provider (ISP) 126. ISP 126 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 128. Local network 122 and Internet 128 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 120 and through communication interface 118, which carry the digital data to and from electronic device 110, are example forms of transmission media.


Electronic device 110 can send messages and receive data, including program code, through the network(s), network link 120 and communication interface 118. In the Internet example, a server 130 might transmit a requested code for an application program through Internet 128, ISP 126, local network 122 and communication interface 118.


The received code can be executed by processor 104 as it is received, and/or stored in storage device 150, or other non-volatile storage for later execution.



FIG. 2 shows a flowchart representing an exemplary method 200 for generating a statistical model representing a plurality of documents. In some embodiments, method 200 can be performed by a client application (e.g., a web browser, a plug-in to a web browser, a standalone executable application, etc.) running on a client device, by a server (e.g., a web server), or it can have some steps or parts thereof executed on the client device, and some steps or parts thereof executed on the server. Thus, method 200 can be performed by one or more electronic devices, such as electronic device 110. And while method 200 and the other following embodiments described herein can be performed by multiple electronic devices each having one or more processors, for purposes of simplicity and without limitation, these embodiments will be explained with respect to a single electronic device (e.g., electronic device 110). While the flowchart discloses the following steps in a particular order, it is appreciated that at least some of the steps can be moved, modified, or deleted where appropriate, consistent with the teachings of the present disclosure.


Referring to FIG. 2, at step 210, the electronic device obtains a plurality (a corpus) of documents. The electronic device can obtain the documents from one or more databases that can be stored locally at the electronic device and/or on one or more remote devices. The documents can be of any type of digital format (e.g., HTML, PDF, MS Word, or any other format containing digital text) and can originate from different public or private sources, for example, from Internet webpages, library archives, proprietary subscription-based archives such as IEEE technical articles or Dow Jones's collection of news articles, or any other source of information. The documents can also be obtained by searching the Internet, and include, for example, some or all search results returned by a search engine in response to a user search query. In some embodiments, the documents can be associated with a date, such as the date of publication, the date of the most recent update, and so forth.


In some embodiments, the documents can also be associated with one or more entities, such as one or more companies or industries. For example, a document can be associated with tags, codes, or other types of metadata that describe one or more product, companies or industries, such as companies or industries discussed in the documents. As an example, a news article discussing a new iPhone® device being released by Apple Inc., can be pre-tagged (e.g., automatically or manually by a person) with one or more tags such as “iPhone,” “Apple,” “AAPL,” “consumer electronics,” and the like.


In some embodiments, the electronic device obtains at step 210 all documents that are associated with at least one common entity. For example, the electronic device can obtain all documents that are associated with the company Apple Inc., for example, by obtaining all documents tagged with one of the tags associated with Apple Inc., such as “AAPL,” “Apple” “Apple, Inc.,” etc. The particular common entity can be arbitrarily selected by the user, or it can be automatically selected by the electronic device. For example, the electronic device can automatically run method 200 several times, each time obtaining, at step 210, all documents associated with the next entity from a list of entities.


At step 220, the electronic device can obtain the content of the documents obtained at step 210. In some embodiments, for reasons of performance, the obtained content can include the headlines or titles of the documents, but not contain the entire text (the body) of the document. In other embodiments, in addition or instead of the headlines or titles of the documents, the obtained content can include the entire text (the body) of the documents. In yet other embodiments, the obtained content can also contain additional text associated with the documents, such as meta data, user comments, or any other additional text associated with the documents.


At step 230, the electronic device can segment the obtained contents of each document, e.g., break down the contents of each document into segments. Each segment can include, for example, one or more characters, which can include parts of words, single words (e.g., tokens), multiple words (e.g., n-grams), or entire phrases or clauses, such as noun phrases or clauses. For example, if document's content includes a headline “Apple releases a new iPhone device in October,” the electronic device can break this content into single-word segments (or tokens) “Apple,” “releases,” “a,” “new,” “iPhone,” “device,” “in,” “October.” As another example, the electronic device can break this content into bigram segments, e.g., “Apple releases,” “new iPhone,” “iPhone device,” “in October.” As seen from the last example, the electronic device can, in some embodiments, omit some content (e.g., articles, prepositions, etc.), and in some embodiments the same content can be included in more than one segment (e.g., the word “iPhone” in the above example appears in segments “new iPhone” and “iPhone device”).


In some embodiments, before or after breaking the content into segments, the electronic device can also perform various processing of the contents. For example, the electronic device can remove some words from the contents (e.g., articles and prepositions), normalize the contents by reducing each word to its morphological root or lemma, and so forth.


At step 240, the electronic device can generate a statistical model based on the frequencies at which the different segments obtained at step 230 occur within the contents of the documents. For example, the statistical model can contain a number associated with each segment, the number reflecting the number of times the segment was found across all contents of all the documents obtained at 210. In other words, the statistical model can include information reflecting, for various segments, how frequent or rare that segment is found within the plurality of documents. For example, the model is likely to indicate English words like “a,” “the,” “has,” “is,” etc., as frequently occurring words. Moreover, if the plurality of documents obtained at step 210 were all associated with a common entity (e.g., “Apple Inc.”), the model is also likely to indicate the segments related to that entity (e.g., “Apple,” “Steve Jobs,” etc.) as relatively frequent, because these segments are more likely to occur in this plurality of documents than in pluralities of documents not necessarily associated with the same entity.


In some embodiments, the statistical model can store, in addition to the segments and their frequencies, statistical information for sequences of subsequent segments. For example, the statistical model can store frequency information for the segment “new,” for the segment “iPhone,” and also frequency information of the segment “new” being followed by the segment “iPhone.” As discussed above, segments can include more than one word (e.g., bigrams), and therefore in some embodiments a bigram “new iPhone” can be a single segment for which the statistical model calculates an independent frequency.


At step 250, the electronic device can store the statistical model in a data storage, for example, in a volatile memory (e.g., in main memory 106), in a non-volatile memory (e.g., in storage device 150), or in both. In some embodiments, the statistical model can be stored in cache memory to allow quick access and update in the future.


As discussed above, the electronic device can perform method many times, each time for a different plurality of documents. Accordingly, the electronic device can generate and store more than one statistical model. In some embodiments, each statistical model can be associated with a different plurality of documents. For example, if a particular plurality of documents is associated with one or more entities (e.g., with Apple Inc., and/or with the consumer electronics industry) the statistical model generated based on that plurality of documents can be stored in association with those entities. Thus, the electronic device can maintain a set of statistical models, and maintain, for example, a look-up table associating one or ore entities with one or ore statistical models. Each statistical model can be associated with one or more entities, and each entity can be associated with either exactly one statistical model, or, in some embodiments, more than one statistical model. In some embodiments, the electronic device can store, in addition to or instead of statistical models associated with particular entities, a default statistical model that is not necessarily associated with any particular entity and that was generated, for example, based on all documents available to the electronic device.



FIG. 3 shows a flowchart representing an exemplary method 300 for key phrase characterization of documents. In some embodiments, method 300 can be performed by a client application (e.g., a web browser, a plug-in to a web browser, a standalone executable application, etc.) running on a client device, by a server (e.g., a web server), or it can have some steps or parts thereof executed on the client device, and some steps or parts thereof executed on the server. Thus, method 300 can be performed by one or more electronic devices, such as electronic device 110. And while method 300 and the other following embodiments described herein can be performed by multiple electronic devices each having one or more processors, for purposes of simplicity and without limitation, these embodiments will be explained with respect to a single electronic device (e.g., electronic device 110). While the flowchart discloses the following steps in a particular order, it is appreciated that at least some of the steps can be moved, modified, or deleted where appropriate, consistent with the teachings of the present disclosure.


At step 310, the electronic device can obtain a user input. The user input can include any input that identifies or defines, either directly or indirectly, a set of documents that are of interest to the user. User input can also include selection or specification, by the user, of an entity (e.g., a product, a company, an industry, etc.) for example, using that entity's name, symbol, stock ticker, etc. In some embodiments, user input can include selection or specification, by the user, of multiple entities. In some embodiments, user input can also specify desired (or undesired) date ranges, geographical regions, document sizes, document sources, or other types of filters indicating the scope of desired documents. In some embodiments, the user input can include a search query from the user, such as “Apple news for April 2014.”


In some embodiments, as illustrated in FIG. 4A, user input can include user's selection of a stock ticker 420 and a desired date range 440. In some embodiments, selecting a stock ticker can cause the electronic device to present a stock chart 410 corresponding to the selected stock ticker and showing, for example, changes in stock prices and trading volume. The user can then select the date range directly from the stock chart. For example, the user can identify an unusual spike in stock price or in trading volume and select a date range 430 around that spike.


In some embodiments, the date range can be initially set by the electronic device to a predetermined date range, allowing the user the modify it if necessary. For example, the electronic device can preset the date range to include a predetermined number of days, weeks, months, etc., ending with the current date. In some embodiments, the electronic device can preset the date range in accordance with prior date ranges used by the user. In some embodiments, the user can select, instead of whole date ranges, time intervals in resolution of hours, minutes, seconds, etc.


At step 320, the electronic device can obtain a set of one or more documents based on the user input received at step 310. For example, if user input specifies a particular entity (e.g., the company Apple Inc.), the electronic device can obtain, from all the documents available to the electronic device, a set of documents associated with the particular entity, such as a set of documents tagged with a tag “Apple Inc.,” “Apple,” or “AAPL.” If the user input also specifies a date range (e.g., Jun. 15, 2007-Jul. 16, 2007, as in the example of FIG. 4A), the electronic device can limit the obtained documents only to documents corresponding to (e.g., published on) dates that fall within the specified date range. Similarly, the electronic device can limit the scope of the obtained documents based on any other types of filters specified by the user.


In some embodiments, step 310 can be omitted, and the electronic device can obtain the set of documents without relying on user input. In some embodiments, when step 310 is omitted, or when it is performed but the user input fails to clearly define the desired scope of documents, the electronic device can obtain a predetermined (e.g., default) set of documents, for example, all documents available to the electronic device.


At step 330, the electronic device can dynamically select a statistical model based on the user input. In some embodiments, the electronic device can select, from a plurality of statistical models, a statistical model that best corresponds to the user input. The plurality of statistical models can be pre-generated, for example, by method 300, and stored, for example, in a data storage (e.g., cache, volatile, and/or non-volatile memory) accessible by the electronic device.


As discussed above, in some embodiments, each statistical model can correspond to and be stored in association with one or more entities. For example, the electronic device can maintain a look-up table associating one or more entities with one or more statistical models. Accordingly, in some embodiments, if the user input indicates a particular entity, the electronic device can obtain a statistical model that best corresponds to the particular entity, for example, a statistical model that was generated based on a plurality of documents associated with the particular entity. To illustrate, if the user input indicates that the user is interested in documents related to Apple Inc., (e.g., the user inputs a search query “Apple news for April 2014” or selects a stock-price graph corresponding to stock ticker AAPL) the electronic device can automatically obtain a statistical model that was generated based on a plurality of documents tagged with “Apple Inc.,” “Apple,” or “AAPL” Thus, in some embodiments, the set of documents obtained at step 320 can be a subset of the plurality of documents obtained at step 210 of method 200, based on which the statistical model was generated.


In some embodiments, if the electronic device cannot obtain a statistical model that was generated based on a plurality of documents associated with a particular entity specified by the user, the electronic device can obtain, instead, a statistical model that was generated based on a plurality of documents associated with another entity that is related to the particular entity. For example, if the user-specified entity is Apple Inc., and the electronic device cannot obtain a statistical model associated with Apple Inc. (e.g., if it has not been generated or is inaccessible) the electronic device can obtain, instead, a statistical model associated with the consumer-electronics industry, because that is an industry associated with Apple Inc. In some embodiments, the electronic device can access a designated database to determine which entities are associated with each other.


In some embodiments, if the electronic device cannot obtain a statistical model associated with the user-specified entity, the electronic device can generate such a statistical model (e.g., using method 200) in real time, after receiving user input at step 310. In some embodiments, the electronic device can generate a new statistical model in real time based on the user input, by using one or more previously generated statistical models. For example, if the documents obtained at step 210 were search results obtained based on user search query related to more than one distinct concept (e.g., “Apple versus Samsung”) the electronic device can generate a new statistical model by merging two previously generated models (e.g., a model associated with Apple, Inc. and a model associated with Samsung Group).


In some embodiments, if the electronic device cannot obtain a statistical model associated with the user-specified entity, the electronic device can obtain a predetermined (default) statistical model that is not necessarily associated with any particular entity and that was generated, for example, based on all documents available to the electronic device.


While step 330 is listed in the exemplary flowchart of FIG. 3 as being performed before step 340, it is appreciated that step 330 can be performed after step 340, or, generally, at any time before step 360.


At step 340, the electronic device can obtain the contents of the set of documents obtained at step 320. In some embodiments, for reasons of performance, the obtained content can include the headlines or titles of the documents, but not contain the entire text (the body) of the document. In other embodiments, in addition or instead of the headlines or titles of the documents, the obtained content can include the entire text (the body) of the documents. In yet other embodiments, the obtained content can also contain additional text associated with the documents, such as meta data, user comments, or any other additional text associated with the documents.


At step 350, the electronic device can segment the contents of the set of documents, e.g., break down the contents of each document into segments. Each segment can include one or more characters, which can include parts of words, single words (e.g., tokens), multiple words (e.g., n-grams), or entire phrases or clauses, such as noun phrases or clauses. For example, if document's content includes a headline “Apple releases a new iPhone device in October,” the electronic device can break this content into single-word segments “Apple,” “releases,” “a,” “new,” “iPhone,” “device,” “in,” “October.” As another example, the electronic device can break this content into bigram segments, e.g., “Apple releases,” “new iPhone,” “iPhone device,” “in October.” As seen from the last example, the electronic device can, in some embodiments, omit some content (e.g., articles, prepositions, etc.), and in some embodiments the same content can be included in more than one segment (e.g., the word “iPhone” in the above example appears in segments “new iPhone” and “iPhone device”).


In some embodiments, before or after breaking the content into segments, the electronic device can also perform various processing of the contents. For example, the electronic device can remove some words from the contents (e.g., articles and prepositions), normalize the contents by reducing each word in the to its morphological root or lemma, and so forth.


While steps 340 and 350 of method 300 for key phrase characterization of documents can be same or similar to steps 220 and 230 of method 200 for generating a statistical model, respectively, it is appreciated that steps 340 and 350 do not have to be same or similar to steps 220 and 230, respectively. In some embodiments, however, to achieve better results in terms of quality and/or performance, the electronic device can chose to use the same segmentation and processing methods at step 350 as the segmentation and processing methods performed at step 220.


At step 360, the electronic device can calculate the statistical significance of the segments obtained at step 350 based on the statistical model obtained at step 330. In some embodiments, the electronic device calculates the statistical significance of a segment by comparing the frequency with which the segment occurred in a particular document (or a set of documents) with the frequency of the segment as indicated in the statistical model. Because the statistical model can be generated based on a large corpus of documents, it can serve as an estimation of a typical or average frequency of a given segment within that corpus. Accordingly, by comparing the frequency of a given segment within a particular document to its frequency in the statistical model, the electronic device can determine whether the segment's frequency in the particular document is higher than average, which may be a good indication that the particular document is specifically directed to the concept or the subject matter described by the segment. For example, a segment “new iPhone” can have a frequency A in the particular document and a frequency B in the statistical model. If frequency A is higher than frequency B (e.g., higher by a predetermined threshold), i.e., the particular document mentions the term “new iPhone” more than usual, the electronic device can determine that the segment is statistically significant meaning that there is high likelihood that the particular document discusses a new iPhone® or some other closely related subject matter.


In some embodiments, the electronic device can calculate a statistical significance value for each segment obtained from the contents of a given document, where the calculation is based at least on the segment's frequency within the contents of the given document and the segment's frequency in the statistical model obtained based on user input in step 330. For example, the statistical significance value can be directly proportional to the segment's frequency within the contents of the given document and inversely proportional to the segment's frequency in the statistical model.


At step 370, the electronic device can determine one or more representative segments for each of the set of documents obtained at step 320. For example, the electronic device can determine, for each document, N segments having highest statistical significant values among segments of that document's contents, where N is an integer greater or equal to one. As discussed above, segments having higher statistical significance values are more likely to represent or correspond to the subject matter discussed in the document.


At step 380, the electronic device can determine, among all the representative segments obtained for the set of documents, one or more repetitive representative segments. For example, the electronic device can identify one or more groups of M or more identical or substantially identical representative segments, where M is a predetermined threshold that can be an integer greater or equal to two. For example, if the set of documents obtained at step 320 includes a thousand documents, and M=20, the electronic device can identify any groups of twenty or more documents having identical or substantially identical representative segments. In some embodiments, two segments can be considered substantially identical if the edit distance (e.g., Levenshtein distance) between them is small, e.g., lower than a predetermined threshold, and/or when the segments are synonyms of each other. Repetitive representative segments can indicate an important subject or event, such as a news event, that is described in multiple documents, such as multiple news sources.


In some embodiments, step 380 can be omitted, and each representative segment can be considered as a repetitive representative segment for the purposes of the following discussion.


At step 390, the electronic device can display (e.g., on display 112) or provide for display the one or more repetitive representative segments. For instance, in the example illustrated in FIG. 4A, the electronic device determines and displays in window 450 two repetitive representative segments: “Friday” and “Apple iPhone,” In some embodiments, as illustrated in FIGS. 4A and 4B, the electronic device can also allow the user to select one of the repetitive representative segments (e.g., “Apple iPhone” and “Friday,” respectively) and responsive to the user's selection, display (e.g., in an adjacent window 460) the contents of and/or links to the documents that contained the selected repetitive representative segment.


In some embodiments, to provide additional context to the segments and thereby improve readability, the electronic device can display or provide for display at step 390 a key phrase that includes the repetitive segment. The key phrase can include, for example, a complete sentence, a noun phrase, a verb phrase, or any other meaningful combination of words. The key phrase can be obtained by the electronic device, for example, from the contents of one of the documents having the repetitive representative segment. For example, if the repetitive representative segment “Apple iPhone” appeared in one of the documents (represented by the segment) as part of the phrase “AT&T stores getting ready for release of Apple iPhone,” the electronic device can display or provide for display (e.g., in key phrase window 470) the key phrase “release of Apple iPhone,” as illustrated in FIG. 4C. As another example, if the repetitive representative segment “Friday” appeared in one of the documents (represented by the segment) as part of the phrase “Consumers, industry eagerly await iPhone as Friday launch approaches,” the electronic device can display or provide for display the key phrase “Friday launch approaches,” as illustrated in FIG. 4C. In some embodiments, if the key phrase containing the repetitive representative segment is different in different documents represented by the segment, the electronic device can select, among the different key phrases, an “average” key phrase, i.e., the key phrase that is most similar to the other key phrases, as measured based on edit distance, based on the number of common words contained in the phrase, or any other suitable algorithm.


Allowing the user to select a particular date range, for example, around unusual changes in stock price or trading volume, and displaying repetitive segments or key phrases representing the documents (e.g., news articles) within that date range, allows the user to quickly determine which events may have caused the unusual changes.


Embodiments of the present disclosure have been described herein with reference to numerous specific details that can vary from implementation to implementation. Certain adaptations and modifications of the described embodiments can be made. Other embodiments can be apparent to those skilled in the art from consideration of the specification and practice of the embodiments disclosed herein. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the present disclosure being indicated by the following claims. It is also intended that the sequence of steps shown in figures are only for illustrative purposes and are not intended to be limited to any particular sequence of steps. As such, it is appreciated that these steps can be performed in a different order while implementing the exemplary methods or processes disclosed herein.

Claims
  • 1. An electronic device comprising: one or more computer-readable storage media configured to store instructions; andone or more processors configured to execute the instructions to cause the electronic device to: receive a user input indicative of an entity and a search query;identify a statistical model associated with the entity, wherein the statistical model is determined based on a first plurality of documents associated with the entity, the statistical model indicative at least of frequencies of one or more words within the first plurality of documents;identify, responsive to the user input, a second plurality of documents, at least partially different than the first plurality of documents, corresponding to the search query and the indicated entity;identify, for each of the second plurality of documents, one or more segments;apply the identified statistical model to each of the identified segments to determine, for each of the second plurality of documents, a statistical significance of segments identified in the documents, the statistical significance indicative of frequencies of the one or more words in the segment compared to the frequencies of the one or more words indicated in the statistical model;and provide for display at least a representative segment having a highest statistical significance and a link to the document containing the representative segment.
  • 2. The electronic device of claim 1, wherein the second plurality of documents is a subset of the first plurality of documents.
  • 3. The electronic device of claim 1, wherein the entity identifies a company, and the first plurality of documents are associated with an industry associated with the company.
  • 4. The electronic device of claim 1, wherein the user input comprises user's selection of an area on a stock chart, and wherein the first plurality of documents have publication dates corresponding to a date range that corresponds to the area on the stock chart.
  • 5. The electronic device of claim 1, wherein providing the representative segment for display comprises providing for display a phrase from the representative segment.
  • 6. The electronic device of claim 1, wherein the one or more processors are further configured to execute the instructions to cause the electronic device to: after providing the representative segment for display, receive a selection input associated with the representative segment; andresponsive to the selection input, provide for display the contents of one or more of the second plurality of documents.
  • 7. The electronic device of claim 1, wherein statistical significance of segments is further based on a comparison of the statistical model and content of the second plurality of documents.
  • 8. A method comprising: by a computing system comprising a hardware computer processor and non-transitory storage medium storing software instructions,receive a user input indicative of an entity and a search query;identify a statistical model associated with the entity, wherein the statistical model is determined based on a first plurality of documents associated with the entity, the statistical model indicative at least of frequencies of one or more words within the first plurality of documents;identify a second plurality of documents, at least partially different than the first plurality of documents, corresponding to the search query and the indicated entity;identify, for each of the second plurality of documents, one or more segments;apply the identified statistical model to each of the identified segments to determine, for each of the second plurality of documents, a statistical significance of segments identified in the documents, the statistical significance indicative of frequencies of the one or more words in the segment compared to the frequencies of the one or more words indicated in the statistical model; andprovide for display at least a representative segment having a highest statistical significance and a link to the document containing the representative segment.
  • 9. The method of claim 8, wherein the second plurality of documents is a subset of the first plurality of documents.
  • 10. The method of claim 8, wherein the entity identifies a company, and the first plurality of documents are associated with an industry associated with the company.
  • 11. The method of claim 8, further comprising: after providing the representative segment for display, receiving a selection input associated with the representative segment; andresponsive to the selection input, providing for display the contents of one or more of the second plurality of documents.
  • 12. The method of claim 8, wherein statistical significance of segments is further based on comparing the statistical model and content of the second plurality of documents.
  • 13. A non-transitory computer-readable medium storing a set of instructions that are executable by one or more electronic devices, each having one or more processors, to cause the one or more electronic devices to perform a method, the method comprising: receiving a user input indicative of an entity and a search query;identifying a statistical model associated with the entity, wherein the statistical model is determined based on a first plurality of documents associated with the entity, the statistical model indicative at least of frequencies of one or more words within the first plurality of documents;identifying a second plurality of documents, at least partially different than the first plurality of documents, corresponding to the search query and the indicated entity;identifying, for each of the second plurality of documents, one or more segments;applying the identified statistical model to each of the identified segments to determine, for each of the second plurality of documents, a statistical significance of segments identified in the documents, the statistical significance indicative of frequencies of the one or more words in the segment compared to the frequencies of the one or more words indicated in the statistical model; andproviding for display at least a representative segment having a highest statistical significance and a link to the document containing the representative segment.
  • 14. The non-transitory computer-readable medium of claim 13, wherein the second plurality of documents is a subset of the first plurality of documents.
  • 15. The non-transitory computer-readable medium of claim 13, wherein the entity identifies a company, and the first plurality of documents are associated with an industry associated with the company.
  • 16. The non-transitory computer-readable medium of claim 13, wherein the user input comprises user's selection of an area on a stock chart, and wherein the first plurality of documents have publication dates corresponding to a date range that corresponds to the area on the stock chart.
  • 17. The non-transitory computer-readable medium of claim 13, wherein providing the representative segment for display comprises providing for display a phrase from the representative segment.
REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. Non-Provisional patent application Ser. No. 14/319,765, filed on Jun. 30, 2014, the disclosure of which is expressly incorporated herein by reference in its entirety.

US Referenced Citations (812)
Number Name Date Kind
5109399 Thompson Apr 1992 A
5329108 Lamoure Jul 1994 A
5632009 Rao et al. May 1997 A
5632987 Rao et al. May 1997 A
5670987 Doi et al. Sep 1997 A
5781704 Rossmo Jul 1998 A
5798769 Chiu et al. Aug 1998 A
5845300 Corner et al. Dec 1998 A
5978475 Schneier et al. Nov 1999 A
6057757 Arrowsmith et al. May 2000 A
6091956 Hollenberg Jul 2000 A
6161098 Wallman Dec 2000 A
6219053 Tachibana et al. Apr 2001 B1
6232971 Haynes May 2001 B1
6247019 Davies Jun 2001 B1
6253203 O'Flaherty et al. Jun 2001 B1
6279018 Kudrolli et al. Aug 2001 B1
6341310 Leshem et al. Jan 2002 B1
6366933 Ball et al. Apr 2002 B1
6369835 Lin Apr 2002 B1
6374251 Fayyad et al. Apr 2002 B1
6385619 Eichstaedt et al. May 2002 B1
6430305 Decker Aug 2002 B1
6456997 Shukla Sep 2002 B1
6549944 Weinberg et al. Apr 2003 B1
6560620 Ching May 2003 B1
6567936 Yang et al. May 2003 B1
6581068 Bensoussan et al. Jun 2003 B1
6594672 Lampson et al. Jul 2003 B1
6631496 Li et al. Oct 2003 B1
6642945 Sharpe Nov 2003 B1
6674434 Chojnacki et al. Jan 2004 B1
6714936 Nevin, III Mar 2004 B1
6725240 Asad et al. Apr 2004 B1
6775675 Nwabueze et al. Aug 2004 B1
6807569 Bhimani et al. Oct 2004 B1
6820135 Dingman Nov 2004 B1
6828920 Owen et al. Dec 2004 B2
6839745 Dingari et al. Jan 2005 B1
6877137 Rivette et al. Apr 2005 B1
6976210 Silva et al. Dec 2005 B1
6978419 Kantrowitz Dec 2005 B1
6980984 Huffman Dec 2005 B1
6985950 Hanson et al. Jan 2006 B1
7017046 Doyle et al. Mar 2006 B2
7036085 Barros Apr 2006 B2
7043702 Chi et al. May 2006 B2
7055110 Kupka et al. May 2006 B2
7069586 Winneg et al. Jun 2006 B1
7139800 Bellotti et al. Nov 2006 B2
7158878 Rasmussen et al. Jan 2007 B2
7162475 Ackerman Jan 2007 B2
7168039 Bertram Jan 2007 B2
7171427 Witkowski et al. Jan 2007 B2
7225468 Waisman et al. May 2007 B2
7269786 Malloy et al. Sep 2007 B1
7278105 Kitts Oct 2007 B1
7290698 Poslinski et al. Nov 2007 B2
7333998 Heckerman et al. Feb 2008 B2
7370047 Gorman May 2008 B2
7373669 Eisen May 2008 B2
7379811 Rasmussen et al. May 2008 B2
7379903 Caballero et al. May 2008 B2
7426654 Adams, Jr. et al. Sep 2008 B2
7451397 Weber et al. Nov 2008 B2
7454466 Bellotti et al. Nov 2008 B2
7461077 Greenwood Dec 2008 B1
7467375 Tondreau et al. Dec 2008 B2
7487139 Fraleigh et al. Feb 2009 B2
7496567 Steichen Feb 2009 B1
7502786 Liu et al. Mar 2009 B2
7525422 Bishop et al. Apr 2009 B2
7529727 Arning et al. May 2009 B2
7529734 Dirisala May 2009 B2
7558677 Jones Jul 2009 B2
7574409 Patinkin Aug 2009 B2
7574428 Leiserowitz et al. Aug 2009 B2
7579965 Bucholz Aug 2009 B2
7596285 Brown et al. Sep 2009 B2
7614006 Molander Nov 2009 B2
7617232 Gabbert et al. Nov 2009 B2
7620628 Kapur et al. Nov 2009 B2
7627812 Chamberlain et al. Dec 2009 B2
7634717 Chamberlain et al. Dec 2009 B2
7703021 Flam Apr 2010 B1
7706817 Bamrah et al. Apr 2010 B2
7712049 Williams et al. May 2010 B2
7716077 Mikurak May 2010 B1
7725530 Sah et al. May 2010 B2
7725547 Albertson et al. May 2010 B2
7730082 Sah et al. Jun 2010 B2
7730109 Rohrs et al. Jun 2010 B2
7756843 Palmer Jul 2010 B1
7770032 Nesta et al. Aug 2010 B2
7770100 Chamberlain et al. Aug 2010 B2
7783658 Bayliss Aug 2010 B1
7801871 Gosnell Sep 2010 B2
7805457 Viola et al. Sep 2010 B1
7809703 Balabhadrapatruni et al. Oct 2010 B2
7814102 Miller et al. Oct 2010 B2
7818658 Chen Oct 2010 B2
7870493 Pall et al. Jan 2011 B2
7894984 Rasmussen et al. Feb 2011 B2
7899611 Downs et al. Mar 2011 B2
7899796 Borthwick et al. Mar 2011 B1
7917376 Bellin et al. Mar 2011 B2
7920963 Jouline et al. Apr 2011 B2
7933862 Chamberlain et al. Apr 2011 B2
7941321 Greenstein et al. May 2011 B2
7962281 Rasmussen et al. Jun 2011 B2
7962495 Jain et al. Jun 2011 B2
7962848 Bertram Jun 2011 B2
7970240 Chao et al. Jun 2011 B1
7971150 Raskutti et al. Jun 2011 B2
7984374 Caro et al. Jul 2011 B2
8001465 Kudrolli et al. Aug 2011 B2
8001482 Bhattiprolu et al. Aug 2011 B2
8010545 Stefik et al. Aug 2011 B2
8015487 Roy et al. Sep 2011 B2
8024778 Cash et al. Sep 2011 B2
8036632 Cona et al. Oct 2011 B1
8036971 Aymeloglu et al. Oct 2011 B2
8037046 Udezue et al. Oct 2011 B2
8046283 Burns Oct 2011 B2
8046362 Bayliss Oct 2011 B2
8054756 Chand et al. Nov 2011 B2
8103543 Zwicky Jan 2012 B1
8134457 Velipasalar et al. Mar 2012 B2
8135679 Bayliss Mar 2012 B2
8135719 Bayliss Mar 2012 B2
8145703 Frishert et al. Mar 2012 B2
8181253 Zaitsev et al. May 2012 B1
8185819 Sah et al. May 2012 B2
8190893 Benson et al. May 2012 B2
8196184 Amirov et al. Jun 2012 B2
8214361 Sandler et al. Jul 2012 B1
8214490 Vos et al. Jul 2012 B1
8214764 Gemmell et al. Jul 2012 B2
8225201 Michael Jul 2012 B2
8229902 Vishniac et al. Jul 2012 B2
8229947 Fujinaga Jul 2012 B2
8230333 Decherd et al. Jul 2012 B2
8239668 Chen et al. Aug 2012 B1
8266168 Bayliss Sep 2012 B2
8271461 Pike et al. Sep 2012 B2
8280880 Aymeloglu et al. Oct 2012 B1
8290838 Thakur et al. Oct 2012 B1
8290926 Ozzie et al. Oct 2012 B2
8290942 Jones et al. Oct 2012 B2
8301464 Cave et al. Oct 2012 B1
8301904 Gryaznov Oct 2012 B1
8302855 Ma et al. Nov 2012 B2
8312367 Foster Nov 2012 B2
8312546 Alme Nov 2012 B2
8321943 Walters et al. Nov 2012 B1
8347398 Weber Jan 2013 B1
8352881 Champion et al. Jan 2013 B2
8368695 Howell et al. Feb 2013 B2
8386377 Xiong et al. Feb 2013 B1
8397171 Klassen et al. Mar 2013 B2
8412707 Mianji Apr 2013 B1
8447674 Choudhuri et al. May 2013 B2
8447722 Abuja et al. May 2013 B1
8452790 Mianji May 2013 B1
8463036 Ramesh et al. Jun 2013 B1
8473454 Evanitsky et al. Jun 2013 B2
8484115 Aymeloglu et al. Jul 2013 B2
8484168 Bayliss Jul 2013 B2
8489331 Kopf et al. Jul 2013 B2
8489641 Seefeld et al. Jul 2013 B1
8495077 Bayliss Jul 2013 B2
8498969 Bayliss Jul 2013 B2
8498984 Hwang et al. Jul 2013 B1
8510743 Hackborn et al. Aug 2013 B2
8514082 Cova et al. Aug 2013 B2
8515207 Chau Aug 2013 B2
8554579 Tribble et al. Oct 2013 B2
8554653 Falkenborg et al. Oct 2013 B2
8554709 Goodson et al. Oct 2013 B2
8560413 Quarterman Oct 2013 B1
8577911 Stepinski et al. Nov 2013 B1
8589273 Creeden et al. Nov 2013 B2
8595234 Siripurapu et al. Nov 2013 B2
8600872 Yan Dec 2013 B1
8620641 Farnsworth et al. Dec 2013 B2
8639757 Zang et al. Jan 2014 B1
8646030 Williamson et al. Feb 2014 B2
8676857 Adams et al. Mar 2014 B1
8683322 Cooper Mar 2014 B1
8688573 Rukonic et al. Apr 2014 B1
8689108 Duffield et al. Apr 2014 B1
8713467 Goldenberg et al. Apr 2014 B1
8726379 Stiansen et al. May 2014 B1
8739278 Varghese May 2014 B2
8742934 Sarpy et al. Jun 2014 B1
8744890 Bernier et al. Jun 2014 B1
8745516 Mason et al. Jun 2014 B2
8769412 Gill et al. Jul 2014 B2
8781169 Jackson et al. Jul 2014 B2
8782794 Ramcharran Jul 2014 B2
8787939 Papakipos et al. Jul 2014 B2
8788405 Sprague et al. Jul 2014 B1
8788407 Singh et al. Jul 2014 B1
8799799 Cervelli et al. Aug 2014 B1
8806355 Twiss et al. Aug 2014 B2
8812960 Sun et al. Aug 2014 B1
8818892 Sprague et al. Aug 2014 B1
8830322 Nerayoff et al. Sep 2014 B2
8832594 Thompson et al. Sep 2014 B1
8868537 Colgrove et al. Oct 2014 B1
8917274 Ma et al. Dec 2014 B2
8924388 Elliot et al. Dec 2014 B2
8924389 Elliot et al. Dec 2014 B2
8924872 Bogomolov et al. Dec 2014 B1
8931043 Cooper et al. Jan 2015 B2
8937619 Sharma et al. Jan 2015 B2
8938686 Erenrich et al. Jan 2015 B1
8949164 Mohler Feb 2015 B1
9009171 Grossman et al. Apr 2015 B1
9009827 Albertson et al. Apr 2015 B1
9021260 Falk et al. Apr 2015 B1
9021384 Beard et al. Apr 2015 B1
9043696 Meiklejohn et al. May 2015 B1
9043894 Dennison et al. May 2015 B1
9049117 Nucci et al. Jun 2015 B1
9069842 Melby Jun 2015 B2
9100428 Visbal Aug 2015 B1
9111281 Stibel et al. Aug 2015 B2
9116975 Shankar et al. Aug 2015 B2
9129219 Robertson et al. Sep 2015 B1
9135658 Sprague et al. Sep 2015 B2
9146954 Boe et al. Sep 2015 B1
9165299 Stowe et al. Oct 2015 B1
9171334 Visbal et al. Oct 2015 B1
9177344 Singh et al. Nov 2015 B1
9202249 Cohen et al. Dec 2015 B1
9230280 Maag et al. Jan 2016 B1
9256664 Chakerian et al. Feb 2016 B2
9335897 Goldenberg May 2016 B2
9338013 Castellucci et al. May 2016 B2
9619557 Kesin et al. Apr 2017 B2
20020033848 Sciammarella et al. Mar 2002 A1
20020065708 Senay et al. May 2002 A1
20020091707 Keller Jul 2002 A1
20020095360 Joao Jul 2002 A1
20020095655 Shulman Jul 2002 A1
20020103705 Brady Aug 2002 A1
20020112157 Doyle et al. Aug 2002 A1
20020116120 Ruiz et al. Aug 2002 A1
20020147805 Leshem et al. Oct 2002 A1
20020174201 Ramer et al. Nov 2002 A1
20020194119 Wright et al. Dec 2002 A1
20030028560 Kudrolli et al. Feb 2003 A1
20030033228 Bosworth-Davies et al. Feb 2003 A1
20030036848 Sheha et al. Feb 2003 A1
20030039948 Donahue Feb 2003 A1
20030074368 Schuetze et al. Apr 2003 A1
20030097330 Hillmer et al. May 2003 A1
20030126102 Borthwick Jul 2003 A1
20030140106 Raguseo Jul 2003 A1
20030144868 MacIntyre et al. Jul 2003 A1
20030163352 Surpin et al. Aug 2003 A1
20030225755 Iwayama et al. Dec 2003 A1
20030229848 Arend et al. Dec 2003 A1
20040032432 Baynger Feb 2004 A1
20040034570 Davis Feb 2004 A1
20040064256 Barinek et al. Apr 2004 A1
20040085318 Hassler et al. May 2004 A1
20040095349 Bito et al. May 2004 A1
20040111410 Burgoon et al. Jun 2004 A1
20040111480 Yue Jun 2004 A1
20040123139 Aiello et al. Jun 2004 A1
20040126840 Cheng et al. Jul 2004 A1
20040143602 Ruiz et al. Jul 2004 A1
20040143796 Lerner et al. Jul 2004 A1
20040153418 Hanweck Aug 2004 A1
20040163039 McPherson et al. Aug 2004 A1
20040193600 Kaasten et al. Sep 2004 A1
20040205524 Richter et al. Oct 2004 A1
20040221223 Yu et al. Nov 2004 A1
20040230577 Kawatani Nov 2004 A1
20040236688 Bozeman Nov 2004 A1
20040250124 Chesla et al. Dec 2004 A1
20040260702 Cragun et al. Dec 2004 A1
20040267746 Marcjan et al. Dec 2004 A1
20050010472 Quatse et al. Jan 2005 A1
20050027705 Sadri et al. Feb 2005 A1
20050028094 Allyn Feb 2005 A1
20050039119 Parks et al. Feb 2005 A1
20050065811 Chu et al. Mar 2005 A1
20050078858 Yao et al. Apr 2005 A1
20050080769 Gemmell Apr 2005 A1
20050086207 Heuer et al. Apr 2005 A1
20050108063 Madill et al. May 2005 A1
20050125715 Franco et al. Jun 2005 A1
20050154628 Eckart et al. Jul 2005 A1
20050154769 Eckart et al. Jul 2005 A1
20050157662 Bingham et al. Jul 2005 A1
20050162523 Darrell et al. Jul 2005 A1
20050166144 Gross Jul 2005 A1
20050180330 Shapiro Aug 2005 A1
20050182764 Evans Aug 2005 A1
20050182793 Keenan et al. Aug 2005 A1
20050183005 Denoue et al. Aug 2005 A1
20050204009 Hazarika et al. Sep 2005 A1
20050210409 Jou Sep 2005 A1
20050222928 Steier et al. Oct 2005 A1
20050229256 Banzhof Oct 2005 A2
20050246327 Yeung et al. Nov 2005 A1
20050251786 Citron et al. Nov 2005 A1
20050262556 Waisman et al. Nov 2005 A1
20050275638 Kolmykov-Zotov et al. Dec 2005 A1
20060026120 Carolan et al. Feb 2006 A1
20060026170 Kreitler et al. Feb 2006 A1
20060031928 Conley et al. Feb 2006 A1
20060059139 Robinson Mar 2006 A1
20060069912 Zheng et al. Mar 2006 A1
20060074881 Vembu et al. Apr 2006 A1
20060080283 Shipman Apr 2006 A1
20060080619 Carlson et al. Apr 2006 A1
20060093222 Saffer et al. May 2006 A1
20060095521 Patinkin May 2006 A1
20060129746 Porter Jun 2006 A1
20060139375 Rasmussen et al. Jun 2006 A1
20060142949 Helt Jun 2006 A1
20060143034 Rothermel Jun 2006 A1
20060143075 Carr et al. Jun 2006 A1
20060143079 Basak et al. Jun 2006 A1
20060149596 Surpin et al. Jul 2006 A1
20060179003 Steele et al. Aug 2006 A1
20060184889 Molander Aug 2006 A1
20060203337 White Sep 2006 A1
20060209085 Wong et al. Sep 2006 A1
20060212931 Shull et al. Sep 2006 A1
20060218637 Thomas et al. Sep 2006 A1
20060241974 Chao et al. Oct 2006 A1
20060242040 Rader Oct 2006 A1
20060242630 Koike et al. Oct 2006 A1
20060265747 Judge Nov 2006 A1
20060271277 Hu et al. Nov 2006 A1
20060279630 Aggaraal et al. Dec 2006 A1
20070000999 Kubo et al. Jan 2007 A1
20070011150 Frank Jan 2007 A1
20070011304 Error Jan 2007 A1
20070016363 Huang et al. Jan 2007 A1
20070038646 Thota Feb 2007 A1
20070038962 Fuchs et al. Feb 2007 A1
20070057966 Ohno et al. Mar 2007 A1
20070078832 Ott et al. Apr 2007 A1
20070083541 Fraleigh et al. Apr 2007 A1
20070094389 Nussey et al. Apr 2007 A1
20070106582 Baker et al. May 2007 A1
20070112867 Evans et al. May 2007 A1
20070143851 Nicodemus Jun 2007 A1
20070150369 Zivin Jun 2007 A1
20070150801 Chidlovskii et al. Jun 2007 A1
20070156673 Maga Jul 2007 A1
20070162454 D'Albora et al. Jul 2007 A1
20070174760 Chamberlain et al. Jul 2007 A1
20070185867 Maga Aug 2007 A1
20070192122 Routson et al. Aug 2007 A1
20070192265 Chopin et al. Aug 2007 A1
20070198571 Ferguson et al. Aug 2007 A1
20070208497 Downs et al. Sep 2007 A1
20070208498 Barker et al. Sep 2007 A1
20070208736 Tanigawa et al. Sep 2007 A1
20070233709 Abnous Oct 2007 A1
20070240062 Christena et al. Oct 2007 A1
20070266336 Nojima et al. Nov 2007 A1
20070284433 Domenica et al. Dec 2007 A1
20070294200 Au Dec 2007 A1
20070294643 Kyle Dec 2007 A1
20070294766 Mir et al. Dec 2007 A1
20080040275 Paulsen et al. Feb 2008 A1
20080040684 Crump Feb 2008 A1
20080051989 Welsh Feb 2008 A1
20080052142 Bailey et al. Feb 2008 A1
20080065655 Chakravarthy et al. Mar 2008 A1
20080077597 Butler Mar 2008 A1
20080077642 Carbone et al. Mar 2008 A1
20080082486 Lermant et al. Apr 2008 A1
20080104019 Nath May 2008 A1
20080104063 Gallivan et al. May 2008 A1
20080104407 Horne et al. May 2008 A1
20080126951 Sood et al. May 2008 A1
20080133567 Ames et al. Jun 2008 A1
20080148398 Mezack et al. Jun 2008 A1
20080155440 Trevor et al. Jun 2008 A1
20080162616 Gross et al. Jul 2008 A1
20080195417 Surpin et al. Aug 2008 A1
20080195608 Clover Aug 2008 A1
20080201580 Savitzky et al. Aug 2008 A1
20080208735 Balet et al. Aug 2008 A1
20080222295 Robinson et al. Sep 2008 A1
20080222706 Renaud et al. Sep 2008 A1
20080229422 Hudis et al. Sep 2008 A1
20080249983 Meisels et al. Oct 2008 A1
20080255973 El Wade et al. Oct 2008 A1
20080263468 Cappione et al. Oct 2008 A1
20080267107 Rosenberg Oct 2008 A1
20080276167 Michael Nov 2008 A1
20080278311 Grange et al. Nov 2008 A1
20080288306 MacIntyre et al. Nov 2008 A1
20080288425 Posse et al. Nov 2008 A1
20080294663 Heinley et al. Nov 2008 A1
20080301643 Appleton et al. Dec 2008 A1
20080313132 Hao et al. Dec 2008 A1
20090002492 Velipasalar et al. Jan 2009 A1
20090018940 Wang et al. Jan 2009 A1
20090024505 Patel et al. Jan 2009 A1
20090027418 Maru et al. Jan 2009 A1
20090030915 Winter et al. Jan 2009 A1
20090044279 Crawford et al. Feb 2009 A1
20090055251 Shah et al. Feb 2009 A1
20090082997 Tokman et al. Mar 2009 A1
20090083184 Eisen Mar 2009 A1
20090088964 Schaaf et al. Apr 2009 A1
20090094270 Alirez et al. Apr 2009 A1
20090103442 Douville Apr 2009 A1
20090106178 Chu Apr 2009 A1
20090112745 Stefanescu Apr 2009 A1
20090119309 Gibson et al. May 2009 A1
20090125359 Knapic May 2009 A1
20090125369 Kloostra et al. May 2009 A1
20090125459 Norton et al. May 2009 A1
20090132921 Hwangbo et al. May 2009 A1
20090132953 Reed et al. May 2009 A1
20090143052 Bates et al. Jun 2009 A1
20090144262 White et al. Jun 2009 A1
20090144274 Fraleigh et al. Jun 2009 A1
20090157732 Hao et al. Jun 2009 A1
20090164934 Bhattiprolu et al. Jun 2009 A1
20090171939 Athsani et al. Jul 2009 A1
20090172511 Decherd et al. Jul 2009 A1
20090172821 Daira et al. Jul 2009 A1
20090177962 Gusmorino et al. Jul 2009 A1
20090179892 Tsuda et al. Jul 2009 A1
20090187464 Bai et al. Jul 2009 A1
20090187546 Whyte et al. Jul 2009 A1
20090187548 Ji et al. Jul 2009 A1
20090192957 Subramanian et al. Jul 2009 A1
20090222400 Kupershmidt et al. Sep 2009 A1
20090222759 Drieschner Sep 2009 A1
20090222760 Halverson et al. Sep 2009 A1
20090228701 Lin Sep 2009 A1
20090234720 George et al. Sep 2009 A1
20090249244 Robinson et al. Oct 2009 A1
20090254842 Leacock et al. Oct 2009 A1
20090254970 Agarwal et al. Oct 2009 A1
20090254971 Herz Oct 2009 A1
20090271343 Vaiciulis et al. Oct 2009 A1
20090281839 Lynn et al. Nov 2009 A1
20090287470 Farnsworth et al. Nov 2009 A1
20090292626 Oxford Nov 2009 A1
20090307049 Elliott et al. Dec 2009 A1
20090313463 Pang et al. Dec 2009 A1
20090318775 Michelson et al. Dec 2009 A1
20090319418 Herz Dec 2009 A1
20090319891 MacKinlay Dec 2009 A1
20090328222 Heiman et al. Dec 2009 A1
20100011282 Dollard et al. Jan 2010 A1
20100030722 Goodson et al. Feb 2010 A1
20100031141 Summers et al. Feb 2010 A1
20100042922 Bradateanu et al. Feb 2010 A1
20100057622 Faith et al. Mar 2010 A1
20100057716 Stefik et al. Mar 2010 A1
20100070523 Delgo et al. Mar 2010 A1
20100070842 Aymeloglu et al. Mar 2010 A1
20100070845 Facemire et al. Mar 2010 A1
20100070897 Aymeloglu et al. Mar 2010 A1
20100077481 Polyakov et al. Mar 2010 A1
20100077483 Stolfo et al. Mar 2010 A1
20100098318 Anderson Apr 2010 A1
20100100963 Mahaffey Apr 2010 A1
20100103124 Kruzeniski et al. Apr 2010 A1
20100106611 Paulsen et al. Apr 2010 A1
20100106752 Eckardt et al. Apr 2010 A1
20100114887 Conway et al. May 2010 A1
20100122152 Chamberlain et al. May 2010 A1
20100125546 Barrett et al. May 2010 A1
20100131457 Heimendinger May 2010 A1
20100131502 Fordham May 2010 A1
20100161735 Sharma Jun 2010 A1
20100162176 Dunton Jun 2010 A1
20100169237 Howard et al. Jul 2010 A1
20100179831 Brown et al. Jul 2010 A1
20100185691 Irmak et al. Jul 2010 A1
20100191563 Schlaifer et al. Jul 2010 A1
20100198684 Eraker et al. Aug 2010 A1
20100199225 Coleman et al. Aug 2010 A1
20100211535 Rosenberger Aug 2010 A1
20100228812 Uomini Sep 2010 A1
20100235915 Memon et al. Sep 2010 A1
20100250412 Wagner Sep 2010 A1
20100262688 Hussain et al. Oct 2010 A1
20100280857 Liu et al. Nov 2010 A1
20100293174 Bennett et al. Nov 2010 A1
20100306029 Jolley Dec 2010 A1
20100306713 Geisner et al. Dec 2010 A1
20100312837 Bodapati et al. Dec 2010 A1
20100313119 Baldwin et al. Dec 2010 A1
20100318924 Frankel et al. Dec 2010 A1
20100321399 Ellren et al. Dec 2010 A1
20100325526 Ellis et al. Dec 2010 A1
20100325581 Finkelstein et al. Dec 2010 A1
20100330801 Rouh Dec 2010 A1
20110004498 Readshaw Jan 2011 A1
20110029526 Knight et al. Feb 2011 A1
20110047159 Baid et al. Feb 2011 A1
20110060753 Shaked et al. Mar 2011 A1
20110060910 Gormish et al. Mar 2011 A1
20110061013 Bilicki et al. Mar 2011 A1
20110066933 Ludwig Mar 2011 A1
20110074811 Hanson et al. Mar 2011 A1
20110078055 Faribault et al. Mar 2011 A1
20110078173 Seligmann et al. Mar 2011 A1
20110087519 Fordyce, III et al. Apr 2011 A1
20110093327 Fordyce, III et al. Apr 2011 A1
20110099133 Chang et al. Apr 2011 A1
20110117878 Barash et al. May 2011 A1
20110119100 Ruhl et al. May 2011 A1
20110131122 Griffin et al. Jun 2011 A1
20110131547 Elaasar Jun 2011 A1
20110137766 Rasmussen et al. Jun 2011 A1
20110153334 Horne et al. Jun 2011 A1
20110161096 Buehler et al. Jun 2011 A1
20110135316 Reid et al. Jul 2011 A1
20110167105 Rarnakrishnan et al. Jul 2011 A1
20110167493 Song et al. Jul 2011 A1
20110170799 Carrino et al. Jul 2011 A1
20110173032 Payne et al. Jul 2011 A1
20110173093 Psota et al. Jul 2011 A1
20110178842 Rane et al. Jul 2011 A1
20110181598 O'Neall et al. Jul 2011 A1
20110202555 Cordover et al. Aug 2011 A1
20110208565 Ross et al. Aug 2011 A1
20110208724 Jones et al. Aug 2011 A1
20110213655 Henkin Sep 2011 A1
20110218934 Elser Sep 2011 A1
20110218955 Tang Sep 2011 A1
20110219321 Gonzalez et al. Sep 2011 A1
20110219450 McDougal et al. Sep 2011 A1
20110225198 Edwards et al. Sep 2011 A1
20110225650 Margolies et al. Sep 2011 A1
20110231223 Winters Sep 2011 A1
20110235413 Wu Sep 2011 A1
20110238495 Kang Sep 2011 A1
20110238510 Rowen et al. Sep 2011 A1
20110238553 Raj et al. Sep 2011 A1
20110238570 Li et al. Sep 2011 A1
20110246229 Pacha Oct 2011 A1
20110251951 Kolkowitz Oct 2011 A1
20110258158 Resende et al. Oct 2011 A1
20110270604 Qi et al. Nov 2011 A1
20110270705 Parker Nov 2011 A1
20110270834 Sokolan et al. Nov 2011 A1
20110289397 Eastmond et al. Nov 2011 A1
20110289407 Naik et al. Nov 2011 A1
20110289420 Morioka et al. Nov 2011 A1
20110291851 Whisenant Dec 2011 A1
20110295649 Fine Dec 2011 A1
20110307382 Siegel et al. Dec 2011 A1
20110310005 Chen et al. Dec 2011 A1
20110314007 Dassa et al. Dec 2011 A1
20110314024 Chang et al. Dec 2011 A1
20120004904 Shin Jan 2012 A1
20120011238 Rathod Jan 2012 A1
20120011245 Gillette et al. Jan 2012 A1
20120019559 Siler et al. Jan 2012 A1
20120022945 Falkenborg et al. Jan 2012 A1
20120036013 Neuhaus et al. Feb 2012 A1
20120036434 Oberstein Feb 2012 A1
20120050293 Carlhian et al. Mar 2012 A1
20120054284 Rakshit Mar 2012 A1
20120059853 Jagota Mar 2012 A1
20120066166 Curbera et al. Mar 2012 A1
20120066296 Appleton et al. Mar 2012 A1
20120072825 Sherkin et al. Mar 2012 A1
20120075324 Cardno et al. Mar 2012 A1
20120079363 Folting et al. Mar 2012 A1
20120079592 Pandrangi Mar 2012 A1
20120084117 Tavares et al. Apr 2012 A1
20120084118 Bai et al. Apr 2012 A1
20120084135 Nissan et al. Apr 2012 A1
20120084287 Lakshminarayan et al. Apr 2012 A1
20120084866 Stolfo Apr 2012 A1
20120106801 Jackson May 2012 A1
20120110674 Belani et al. May 2012 A1
20120117082 Koperda et al. May 2012 A1
20120131512 Takeuchi et al. May 2012 A1
20120137235 TS et al. May 2012 A1
20120144335 Abeln et al. Jun 2012 A1
20120158527 Cannelongo et al. Jun 2012 A1
20120159307 Chung et al. Jun 2012 A1
20120159362 Brown et al. Jun 2012 A1
20120159399 Bastide et al. Jun 2012 A1
20120169593 Mak et al. Jul 2012 A1
20120170847 Tsukidate Jul 2012 A1
20120173381 Smith Jul 2012 A1
20120173985 Peppel Jul 2012 A1
20120180002 Campbell et al. Jul 2012 A1
20120191502 Gross Jul 2012 A1
20120196557 Reich et al. Aug 2012 A1
20120196558 Reich et al. Aug 2012 A1
20120197651 Robinson et al. Aug 2012 A1
20120203708 Psota et al. Aug 2012 A1
20120208636 Feige Aug 2012 A1
20120215784 King et al. Aug 2012 A1
20120215898 Shah et al. Aug 2012 A1
20120218305 Patterson et al. Aug 2012 A1
20120221511 Gibson et al. Aug 2012 A1
20120221553 Wittmer et al. Aug 2012 A1
20120221580 Barney Aug 2012 A1
20120226523 Weiss et al. Sep 2012 A1
20120245976 Kumar et al. Sep 2012 A1
20120246148 Dror Sep 2012 A1
20120254129 Wheeler et al. Oct 2012 A1
20120284345 Costenaro et al. Nov 2012 A1
20120284791 Miller et al. Nov 2012 A1
20120290879 Shibuya et al. Nov 2012 A1
20120296907 Long et al. Nov 2012 A1
20120310831 Harris et al. Dec 2012 A1
20120310838 Harris et al. Dec 2012 A1
20120311684 Paulsen et al. Dec 2012 A1
20120323888 Osann, Jr. Dec 2012 A1
20120330801 McDougal et al. Dec 2012 A1
20120330973 Ghuneim et al. Dec 2012 A1
20130006426 Healey et al. Jan 2013 A1
20130006655 Van Arkel et al. Jan 2013 A1
20130006668 Van Arkel et al. Jan 2013 A1
20130006725 Simanek et al. Jan 2013 A1
20130006916 McBride et al. Jan 2013 A1
20130016106 Yip et al. Jan 2013 A1
20130018796 Kolhatkar et al. Jan 2013 A1
20130019306 Lagar-Cavilla et al. Jan 2013 A1
20130024268 Manickavelu Jan 2013 A1
20130024307 Fuerstenberg et al. Jan 2013 A1
20130024339 Choudhuri et al. Jan 2013 A1
20130046635 Grigg et al. Feb 2013 A1
20130046842 Muntz et al. Feb 2013 A1
20130050217 Armitage Feb 2013 A1
20130054306 Bhalla Feb 2013 A1
20130057551 Ebert et al. Mar 2013 A1
20130060786 Serrano et al. Mar 2013 A1
20130061169 Pearcy et al. Mar 2013 A1
20130073377 Heath Mar 2013 A1
20130073454 Busch Mar 2013 A1
20130078943 Biage et al. Mar 2013 A1
20130086482 Parsons Apr 2013 A1
20130096988 Grossman et al. Apr 2013 A1
20130097482 Marantz et al. Apr 2013 A1
20130097709 Basavapatna et al. Apr 2013 A1
20130101159 Rosen Apr 2013 A1
20130110746 Ahn May 2013 A1
20130110822 Ikeda et al. May 2013 A1
20130110876 Meijer et al. May 2013 A1
20130110877 Bonham et al. May 2013 A1
20130111320 Campbell et al. May 2013 A1
20130117011 Ahmed et al. May 2013 A1
20130117651 Waldman et al. May 2013 A1
20130132381 Chakrabarti May 2013 A1
20130139268 An et al. May 2013 A1
20130150004 Rosen Jun 2013 A1
20130151148 Parundekar et al. Jun 2013 A1
20130151388 Falkenborg et al. Jun 2013 A1
20130151453 Bhanot et al. Jun 2013 A1
20130157234 Gulli et al. Jun 2013 A1
20130160120 Malaviya et al. Jun 2013 A1
20130166348 Scotto Jun 2013 A1
20130166480 Popescu et al. Jun 2013 A1
20130166550 Buchmann et al. Jun 2013 A1
20130176321 Mitchell et al. Jul 2013 A1
20130179420 Park et al. Jul 2013 A1
20130185245 Anderson et al. Jul 2013 A1
20130185307 El-Yaniv et al. Jul 2013 A1
20130197925 Blue Aug 2013 A1
20130208565 Castellanos et al. Aug 2013 A1
20130211985 Clark et al. Aug 2013 A1
20130218879 Park et al. Aug 2013 A1
20130224696 Wolfe et al. Aug 2013 A1
20130225212 Khan Aug 2013 A1
20130226318 Procyk Aug 2013 A1
20130226953 Markovich et al. Aug 2013 A1
20130232045 Tai Sep 2013 A1
20130238616 Rose et al. Sep 2013 A1
20130239217 Kindler et al. Sep 2013 A1
20130246170 Gross et al. Sep 2013 A1
20130246537 Gaddala Sep 2013 A1
20130246597 Iizawa et al. Sep 2013 A1
20130251233 Yang et al. Sep 2013 A1
20130262527 Hunter et al. Oct 2013 A1
20130263019 Castellanos et al. Oct 2013 A1
20130267207 Hao et al. Oct 2013 A1
20130268520 Fisher et al. Oct 2013 A1
20130276799 Davidson Oct 2013 A1
20130279757 Kephart Oct 2013 A1
20130282696 John et al. Oct 2013 A1
20130290011 Lynn et al. Oct 2013 A1
20130290825 Arndt et al. Oct 2013 A1
20130297619 Chandrasekaran et al. Nov 2013 A1
20130304770 Boero et al. Nov 2013 A1
20130311375 Priebatsch Nov 2013 A1
20130318594 Hoy et al. Nov 2013 A1
20130325859 Porter Dec 2013 A1
20130339218 Subramanian et al. Dec 2013 A1
20140006109 Callioni et al. Jan 2014 A1
20140012796 Petersen et al. Jan 2014 A1
20140013451 Kulka et al. Jan 2014 A1
20140019936 Cohanoff Jan 2014 A1
20140032506 Hoey et al. Jan 2014 A1
20140033010 Richardt et al. Jan 2014 A1
20140040371 Gurevich et al. Feb 2014 A1
20140047319 Eberlein Feb 2014 A1
20140047357 Alfaro et al. Feb 2014 A1
20140058763 Zizzamia et al. Feb 2014 A1
20140058914 Song et al. Feb 2014 A1
20140059038 McPherson et al. Feb 2014 A1
20140059683 Ashley Feb 2014 A1
20140067611 Adachi et al. Mar 2014 A1
20140068487 Steiger et al. Mar 2014 A1
20140074855 Zhao et al. Mar 2014 A1
20140081652 Klindworth Mar 2014 A1
20140095273 Tang et al. Apr 2014 A1
20140095509 Patton Apr 2014 A1
20140108068 Williams Apr 2014 A1
20140108380 Gotz et al. Apr 2014 A1
20140108985 Scott et al. Apr 2014 A1
20140122456 Dies May 2014 A1
20140122501 Shen et al. May 2014 A1
20140123279 Bishop et al. May 2014 A1
20140129261 Bothwell May 2014 A1
20140136285 Carvalho May 2014 A1
20140143009 Brice et al. May 2014 A1
20140149130 Getchius May 2014 A1
20140149272 Hirani et al. May 2014 A1
20140149417 Kim May 2014 A1
20140149436 Bahrami et al. May 2014 A1
20140156527 Grigg et al. Jun 2014 A1
20140157172 Peery et al. Jun 2014 A1
20140164502 Khodorenko et al. Jun 2014 A1
20140173712 Ferdinand Jun 2014 A1
20140173738 Condry et al. Jun 2014 A1
20140188895 Wang et al. Jul 2014 A1
20140189536 Lange et al. Jul 2014 A1
20140195515 Baker et al. Jul 2014 A1
20140195887 Ellis et al. Jul 2014 A1
20140214579 Shen et al. Jul 2014 A1
20140222521 Chait Aug 2014 A1
20140222793 Sadkin et al. Aug 2014 A1
20140229422 Jain et al. Aug 2014 A1
20140229554 Grunin et al. Aug 2014 A1
20140244388 Manouchehri et al. Aug 2014 A1
20140258246 Lo Faro et al. Sep 2014 A1
20140267294 Ma Sep 2014 A1
20140267295 Sharma Sep 2014 A1
20140279824 Tamayo Sep 2014 A1
20140283107 Walton et al. Sep 2014 A1
20140310266 Greenfield Oct 2014 A1
20140310282 Sprague et al. Oct 2014 A1
20140316911 Gross Oct 2014 A1
20140333651 Cervelli et al. Nov 2014 A1
20140337772 Cervelii et al. Nov 2014 A1
20140344230 Krause et al. Nov 2014 A1
20140351070 Christner et al. Nov 2014 A1
20140358829 Hurwitz Dec 2014 A1
20140366132 Stiansen et al. Dec 2014 A1
20150019394 Unser et al. Jan 2015 A1
20150039565 Lucas Feb 2015 A1
20150046870 Goldenberg et al. Feb 2015 A1
20150073929 Psota et al. Mar 2015 A1
20150073954 Braff Mar 2015 A1
20150081803 Dick et al. Mar 2015 A1
20150089424 Duffield et al. Mar 2015 A1
20150095773 Gonsalves et al. Apr 2015 A1
20150100897 Sun et al. Apr 2015 A1
20150100907 Erenrich et al. Apr 2015 A1
20150106170 Bonica Apr 2015 A1
20150106379 Elliot et al. Apr 2015 A1
20150128274 Giokas May 2015 A1
20150134599 Banerjee et al. May 2015 A1
20150134666 Gattiker et al. May 2015 A1
20150135256 Hoy et al. May 2015 A1
20150169709 Kara et al. Jun 2015 A1
20150169726 Kara et al. Jun 2015 A1
20150170077 Kara et al. Jun 2015 A1
20150178825 Huerta Jun 2015 A1
20150178877 Bogomolov et al. Jun 2015 A1
20150186532 Agarwal et al. Jul 2015 A1
20150186821 Wang et al. Jul 2015 A1
20150187036 Wang et al. Jul 2015 A1
20150188715 Castelluci et al. Jul 2015 A1
20150188872 White Jul 2015 A1
20150227295 Meiklejohn et al. Aug 2015 A1
20150227518 Kallan Aug 2015 A1
20150229664 Hawthorn et al. Aug 2015 A1
20150235334 Wang et al. Aug 2015 A1
20150242401 Liu Aug 2015 A1
20150248563 Alfarano et al. Sep 2015 A1
20150261847 Ducott et al. Sep 2015 A1
20150324868 Kaftan et al. Nov 2015 A1
20150326601 Grondin et al. Nov 2015 A1
20150338233 Cervelli et al. Nov 2015 A1
20150347903 Saxena et al. Dec 2015 A1
20150370888 Fonseca et al. Dec 2015 A1
20150378996 Kesin et al. Dec 2015 A1
20150379413 Robertson et al. Dec 2015 A1
20160004764 Chakerian et al. Jan 2016 A1
20160004864 Falk et al. Jan 2016 A1
20160028759 Visbal Jan 2016 A1
20160034470 Sprague et al. Feb 2016 A1
20160034545 Shankar et al. Feb 2016 A1
20160048937 Mathura et al. Feb 2016 A1
Foreign Referenced Citations (53)
Number Date Country
2014250678 May 2015 AU
101729531 Jun 2010 CN
102546446 Jul 2012 CN
103167093 Jun 2013 CN
103281301 Sep 2013 CN
102054015 May 2014 CN
102014204827 Sep 2014 DE
102014204830 Sep 2014 DE
102014204834 Sep 2014 DE
102014215621 Feb 2015 DE
1672527 Jun 2006 EP
1962222 Aug 2008 EP
2487610 Aug 2012 EP
2551799 Jan 2013 EP
2835745 Feb 2015 EP
2835770 Feb 2015 EP
2838039 Feb 2015 EP
2846241 Mar 2015 EP
2851852 Mar 2015 EP
2858014 Apr 2015 EP
2858018 Apr 2015 EP
2863326 Apr 2015 EP
2863346 Apr 2015 EP
2869211 May 2015 EP
2881868 Jun 2015 EP
2884439 Jun 2015 EP
2884440 Jun 2015 EP
2889814 Jul 2015 EP
2891992 Jul 2015 EP
2892197 Jul 2015 EP
2897051 Jul 2015 EP
2911078 Aug 2015 EP
2911100 Aug 2015 EP
2963578 Jan 2016 EP
2963595 Jan 2016 EP
2985974 Feb 2016 EP
2516155 Jan 2015 GB
2518745 Apr 2015 GB
2012778 Nov 2014 NL
2013306 Feb 2015 NL
2011642 Aug 2015 NL
624557 Dec 2014 NZ
WO 0009529 Feb 2000 WO
WO 02065353 Aug 2002 WO
WO 2005010685 Feb 2005 WO
WO 2005104736 Nov 2005 WO
WO 2005116851 Dec 2005 WO
WO 2008064207 May 2008 WO
WO 2009061501 May 2009 WO
WO 2010000014 Jan 2010 WO
WO 2010030913 Mar 2010 WO
WO 2013010157 Jan 2013 WO
WO 2013102892 Jul 2013 WO
Non-Patent Literature Citations (331)
Entry
US 8,712,906, 04/2014, Sprague et al. (withdrawn)
“A First Look: Predicting Market Demand for Food Retail using a Huff Analysis,” TRF Policy Solutions, Jul. 2012, pp. 30.
“A Quick Guide to UniProtKB Swiss-Prot & TrEMBL,” Sep. 2011, pp. 2.
Acklen, Laura, “Absolute Beginner's Guide to Microsoft Word 2003,” Dec. 24, 2003, pp. 15-18, 34-41, 308-316.
Ananiev et al., “The New Modality API,” http://web.archive.org/web/20061211011958/http://java.sun.com/developer/technicalArticles/J2SE/Desktop/javase6/modality/ Jan. 21, 2006, pp. 8.
Bluttman et al., “Excel Formulas and Functions for Dummies,” 2005, Wiley Publishing, Inc., pp. 230, 284-286.
Bugzilla@Mozilla, “Bug 18726—[Feature] Long-click means of invoking contextual menus not supported,” http://bugzilla.mozilla.org/show.sub.--bug.cgi?id=18726 printed Jun. 13, 2013 in 11 pages.
Canese et al., “Chapter 2: PubMed: The Bibliographic Database,” The NCBI Handbook, Oct. 2002, pp. 1-10.
Chen et al., “Bringing Order to the Web: Automatically Categorizing Search Results,” CHI 2000, Proceedings of the SIGCHI conference on Human Factors in Computing Systems, Apr. 1-6, 2000, The Hague, The Netherlands, pp. 145-152.
Conner, Nancy, “Google Apps: The Missing Manual,” May 1 , 2008, pp. 15.
Delcher et al., “Identifying Bacterial Genes and Endosymbiont DNA with Glimmer,” Bioinformatics, vol. 23, No. 6, 2007, pp. 673-679.
Dramowicz, Ela, “Retail Trade Area Analysis Using the Huff Model,” Directions Magazine, Jul. 2, 2005 in 10 pages, http://www.directionsmag.com/articles/retail-trade-area-analysis-using-th- e-huff-model/123411.
GIS-NET 3 Public—Department of Regional Planning. Planning & Zoning Information for Unincorporated LA County. Retrieved Oct. 2, 2013 from http://gis.planning.lacounty.gov;/GIS-NET3.sub.--Public/Viewer.html.
Goswami, Gautam, “Quite Writly Said!,” One Brick at a Time. Aug. 21 2005, pp. 7.
Griffith, Daniel A., “A Generalized Huff Model,” Geographical Analysis, Apr. 1982, vol. 14, No. 2, pp. 135-144.
Hansen et al. “Analyzing Social Media Networks with NodeXL: Insights from a Connected World”, Chapter 4, pp. 53-67 and Chapter 10, pp. 143-164, published Sep. 2010.
Hibbert et al., “Prediction of Shopping Behavior Using a Huff Model Within a GIS Framework,” Healthy Eating in Context, Mar. 18, 2011, pp. 16.
Kahan et al., “Annotea: an open RDF infrastructure for shared WEB annotation”, Computer Networks 39, pp. 589-608, 2002.
Keylines.com, “An Introduction to KeyLines and Network Visualization,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-White-Paper.p- df> downloaded May 12, 2014 in 8 pages.
Keylines.com, “KeyLines Datasheet,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-datasheet.pdf- > dowloaded May 12, 2014 in 2 pages.
Keylines.com, “Visualizing Threats: Improved Cyber Security Through Network Visualization,” Apr. 2014, <http://keylines.com/wp-content/uploads/2014/04/Visualizing-Threats1.p- df> downloaded May 12, 2014 in 10 pages.
Kitts, Paul, “Chapter 14: Genome Assembly and Annotation Process,” The NCBI Handbook, Oct. 2002, pp. 1-21.
Liu, Tianshun, “Combining GIS and the Huff Model to Analyze Suitable Locations for a New Asian Supermarket in the Minneapolis and St. Paul, Minnesota USA,” Papers in Resource Analysis, 2012, vol. 14, pp. 8.
Madden, Tom, “Chapter 16: The BLAST Sequence Analysis Tool,” The NCBI Handbook, Oct. 2002, pp. 1-15.
Manno et al., “Introducing Collaboration in Single-user Applications through the Centralized Control Architecture,” 2010, pp. 10.
Manske, “File Saving Dialogs,” <http://www.mozilla.org/editor/ui.sub.--specs/FileSaveDialogs.html>- , Jan. 20, 1999, pp. 7.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.bing.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.google.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.yahoo.com.
Microsoft—Developer Network, “Getting Started with VBA in Word 2010,” Apr. 2010, <http://msdn.microsoft./com/en-us/library/ff604039%28v=office.14%29.asp- x> as printed Apr. 4, 2014 in 17 pages.
Microsoft Office—Visio, “About connecting shapes,” <http://office.microsoft.com/en-us/visio-help/about-connecting shapes-- HP085050369.aspx> printed Aug. 4, 2011 in 6 pages.
Microsoft Office—Visio, “Add and glue connectors with the Connector tool,” <http://office.microsoft.com/en-us/visio-help/add-and-glue-conn- ectors-with-the-connector-tool-HA010048532.aspx?CTT=1> printed Aug. 4, 2011 in 1 page.
Mizrachi, Ilene, “Chapter 1: GenBank: The Nukeotide Sequence Database,” The NCBI Handbook, Oct. 2002, pp. 1-14.
Palmas et al., “An Edge-Bunding Layout for Interactive Parallel Coordinates” 2014 IEEE Pacific Visualization Symposium, 2014, pp. 57-64.
Rouse, Margaret, “OLAP Cube,”0 <http://searchdatamanagement.techtarget.com/definition/OLAP-cube>, Apr. 28, 2012, pp. 16
Sigrist, et al., “PROSITE, a Protein Domain Database for Functional Characterization and Annotation,” Nucleic Acid Research, 2010, vol. 38, pp. D161-D166.
Sirotkin et al.., “Chapter 13: The Processing of Biological Sequence Data at NCBI,” The NCBI Handbook, Oct. 2002, pp. 1-11.
“The FASTA Progam Package,” fasta-35.3.4, Mar. 25, 2011, pp. 29.
Vose et al., “Help File for ModelRisk Version 5,” 2007, Vose Software, pp. 349-353. [Uploaded in 2 Parts].
Issue Notification for U.S. Appl. No. 13/917,571 dated Aug. 5, 2014.
Notice of Allowance for U.S. Appl. No. 14/102,394 dated Aug. 25, 2014.
Notice of Allowance for U.S. Appl. No. 14/108,187 dated Aug. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/135,289 dated Oct. 14, 2014.
Notice of Allowance for U.S. Appl. No. 14/268,964 dated Dec. 3, 2014.
Official Communication for New Zealand Patent Application No. 624557 dated May 14, 2014.
Official Communication for New Zealand Patent Application No. 628585 dated Aug. 26, 2014.
Official Communication for European Patent Application No. 14158861.6 dated Jun. 16, 2014.
Official Communication for New Zealand Patent Application No. 622517 dated Apr. 3, 2014.
Official Communication for New Zealand Patent Application No. 628263 dated Aug. 12, 2014
Official Communication for Great Britain Patent Application No. 1404457.2 dated Aug. 14, 2014.
Official Communication for New Zealand Patent Application No. 627962 dated Aug. 5, 2014.
Official Communication for European Patent Application No. 14159464.8 dated Jul. 31, 2014.
Official Communication for European Patent Application No. 14159464.8 dated Oct. 8, 2014.
Official Communication for New Zealand Patent Application No. 628840 dated Aug. 28, 2014.
Official Communication in New Zealand Application No. 623495 dated Aug. 19, 2014.
Official Communication for United Kingdom Patent Application No. 1408025.3 dated Nov. 6, 2014.
Official Communication for New Zealand Patent Application No. 622513 dated Apr. 3, 2014.
Official Communication for New Zealand Patent Application No. 628161 dated Aug. 25, 2014.
Official Communication for U.S. Appl. No. 14/289,596 dated Jul. 18, 2014.
Official Communication for U.S. Appl. No. 14/289,599 dated Jul. 22, 2014.
Official Communication for U.S. Appl. No. 14/225,160 dated Jul. 29, 2014.
Official Communication for U.S. Appl. No. 14/268,964 dated Sep. 3, 2014.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 2, 2014.
Official Communication for U.S. Appl. No. 14/225,160 dated Oct. 22, 2014.
Boyce, Jim, “Microsoft Outlook 2010 Inside Out,” Aug. 1, 2010, retrieved from the internet https://capdtron.files.wordpress.com/2013/01/outlook-2010-inside.sub.--ou- t.pdf.
Chung, Chin-Wan, “Dataplex: An Access to Heterogeneous Distributed Databases,” Communications of the ACM, Association for Computing Machinery, Inc., vol. 33, No. 1, Jan. 1, 1990, pp. 70-80.
Definition “Identify” downloaded Jan. 22, 2015, 1 page.
Definition “Overlay” downloaded Jan. 22, 2015, 1 page.
Hardesty, “Privacy Challenges: Analysis: It's Surprisingly Easy to Identify Individuals from Credit-Card Metadata,” MIT News on Campus and Around the World, MIT News Office, Jan. 29, 2015, 3 pages.
Hogue et al., “Thresher: Automating the Unwrapping of Semantic Content from the World Wide Web,” 14.sup.th International Conference on World Wide Web, WWW 2005: Chiba, Japan, May 10-14, 2005, pp. 86-95.
Nierman, “Evaluating Structural Similarity in XML Documents,” 2002, 6 pages.
Olanoff, Drew, “Deep Dive with the New Google Maps for Desktop with Google Earth Integration, It's More than Just a Utility,” May 15, 2013, pp. 1-6, retrieved from the internet: http://web.archive.org/web/20130515230641/http://techcrunch.com/2013/05/1- 5/deep-dive-with-the-new-google-maps-for-desktop-with-google-earth-integra- tion-its-more-than-just-a-utility/.
Wikipedia, “Federated Database System,” Sep. 7, 2013, retrieved from the Internet on Jan. 27, 2015 http://en.wikipedia.org/w/index.php?title=Federated.sub.--database.sub.--- system&oldid=571954221.
Yang et al., “HTML Page Analysis Based on Visual Cues,” 2001, pp. 859-864.
Official Communication for Australian Patent Application No. 2014201511 dated Feb. 27, 2015.
Official Communication for Australian Patent Application No. 2014202442 dated Mar. 19, 2015.
Official Communication for European Patent Application No. 14180281.9 dated Jan. 26, 2015.
Official Communication for European Patent Application No. 14180142.3 dated Feb. 6, 2015.
Official Communication for European Patent Application No. 14187996.5 dated Feb. 12, 2015.
Official Communication for European Patent Application No, 14186225.0 dated Feb. 13, 2015.
Official Communication for European Patent Application No. 14189344.6 dated Feb. 20, 2015.
Official Communication for European Patent Application No. 14189347.9 dated Mar. 4, 2015.
Official Communication for European Patent Application No. 14199182.8 dated Mar. 13, 2015.
Official Communication for European Patent Application No. 14180321.3 dated Apr. 17, 2015.
Official Communication for Great Britain Patent Application No. 1404574.4 dated Dec. 18, 2014.
Official Communication for Great Britain Patent Appiication No. 1411984.6 dated Dec. 22, 2014.
Official Communication for Great Britain Patent Application No. 1413935.6 dated Jan. 27, 2015.
Notice of Allowance for U.S. Appl. No. 14/192,767 dated Dec. 16, 2014.
Notice of Allowance for U.S. Appl. No. 14/294,098 dated Dec. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/473,860 dated Jan. 5, 2015.
Notice of Allowance for U.S. Appl. No. 14/616,080 dated Apr. 2, 2015.
Notice of Allowance for U.S. Appl. No. 14/486,991 dated May 1, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Jan. 26, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Feb. 5, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Feb. 18, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 20, 2015.
Official Communication for U.S. Appl. No. 14/473,552 dated Feb. 24, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Feb. 27, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Mar. 4, 2015.
Official Communication for U.S. Appl. No. 14/486,991 dated Mar. 10, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 11, 2015.
Official Communication for U.S. Appl. No. 14/148,568 dated Mar. 26, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 13/247,987 dated Apr. 2, 2015.
Restriction Requirement for U.S. Appl. No. 13/839,026 dated Apr. 2, 2015.
“A Word About Banks and the Laundering of Drug Money,” Aug. 18, 2012, http://www.golemxiv.co.uk/2012/08/a-word-about-banks-and-the-laundering-o- f-drug-money/.
AMNET, “5 Great Tools for Visualizing Your Twitter Followers,” posted Aug. 4, 2010, http://www.amnetblog.com/component/content/article/115-5-grate-t- ools-for-visualizing-your-twitter-followers.html.
Celik, Tantek, “CSS Basic User Interface Module Level 3 (CSS3 UI),” Section 8 Resizing and Overflow, Jan. 17, 2012, retrieved from internet http://www.w3.org/TR/2012/WD-css3-ui-20120117/#resizing-amp-overflow retrieved on May 18, 2015.
Huang et al., “Systematic and Integrative Analysis of Large Gene Lists Using DAVID Bioinformatics Resources,” Nature Protocols, 4.1, 2008, 44-57.
Li et al., “Interactive Multimodal Search on Mobile Device,” IEEE Transactions on Multimedia, vol. 15, No. 3, Apr. 1, 2013, pp. 594-607.
“Potential Money Laundering Warning Signs,” snapshot taken 2003, https://web.archive.org/web/20030816090055/http:/finsolinc.com/ANTI-MONEY-%20LAUNDERING%20TRAINING%20GUIDES.pdf.
“Refresh CSS Ellipsis When Resizing Container—Stack Overflow,” Jul. 31, 2013, retrieved from internet http://stackoverflow.com/questions/17964681/refresh-css-ellipsis-when-res- izing-container, retrieved on May 18, 2015.
Thompson, Mick, “Getting Started with GEO,” Getting Started with GEO, Jul. 26, 2011.
Umagandhi et al., “Search Query Recommendations Using Hybrid User Profile with Query Logs,” International Journal of Computer Applications, vol. 80, No. 10, Oct. 1, 2013, pp. 7-18.
Official Communication for Australian Patent Application No. 2014213553 dated May 7, 2015.
Official Communication for Australian Patent Application No. 2014210604 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014210614 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014250678 dated Jun. 17, 2015.
Notice of Acceptance for Australian Patent Application No. 2014250678 dated Oct. 7, 2015.
Official Communication for European Patent Application No. 14159464.8 dated Sep. 22, 2014.
Official Communication for European Patent Application No. 14197879.1 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14197895.7 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14189802.3 dated May 11, 2015.
Official Communication for European Patent Application No. 14191540.5 dated May 27, 2015.
Official Communication for European Patent Application No. 14180432.8 dated Jun. 23, 2015.
Official Communication for European Patent Application No, 14187739.9 dated Jul 6, 2015.
Official Communication for European Patent Application No. 15155846.7 dated Jul. 8, 2015.
Official Communication for Netherlands Patent Application No. 2013306 dated Apr. 24, 2015.
Notice of Allowance for U.S. Appl. No. 14/225,084 dated May 4. 2015.
Notice of Allowance for U.S. Appl. No. 14/504,103 dated May 18, 2015.
Notice of Allowance for U.S. Appl. No. 14/473,552 dated Jul. 24, 2015.
Notice of Allowance for U.S. Appl. No. 14/148,568 dated Aug. 26, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Apr. 30, 2015.
Official Communication for U.S. Appl. No. 14/196,814 dated May 5, 2015.
Official Communication for U.S. Appl. No. 14/305,154 dated May 15, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated May 18, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated May 20, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated May 29, 2015.
Official Communication for U.S. Appl. No. 13/835,688 dated Jun. 17, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 12/556,318 dated Jul. 2, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Jul. 6, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated Jul. 24, 2015.
Official Communication for U.S. Appl. No. 13/839,026 dated Aug. 4, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Aug. 6, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Aug. 7, 2015.
Official Communication for U.S. Appl. No. 14/490,612 dated Aug. 18, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated Aug. 19, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 2, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated Sep. 4, 2015.
Official Communication for U.S. Appl. No. 14/631,633 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/726,353 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 11, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 14, 2015.
Official Communication for U.S. Appl. No. 13/247,987 dated Sep. 22, 2015.
Official Communication for U.S. Appl. No. 14/813,749 dated Sep. 28, 2015.
Official Communication for U.S. Appl. No. 14/134,558 dated Oct. 7, 2015.
About 80 Minutes, “Palantir in a Number of Parts—Part 6—Graph,” Mar. 21, 2013, pp. 1-6, retrieved from the internet http://about80minutes.blogspot.nl/2013/03/palantir-in-number-of-parts-par- t-6-graph.html retrieved on Aug. 18, 2015.
Appacts, “Smart Thinking for Super Apps,” <http://www.appacts.com> Printed Jul. 18, 2013 in 4 pages.
Apsalar, “Data Powered Mobile Advertising,” “Free Mobile App Analytics” and various analytics related screen shots <http://apsalar.com> Printed Jul. 18, 2013 in 8 pages.
Capptain—Pilot Your Apps, <http://www.capptain.com> Printed Jul. 18, 2013 in 6 pages.
Chaudhuri et al., “An Overview of Business Intelligence Technology,” Communications of the ACM, Aug. 2011, vol. 54, No. 8.
Cohn et al., “Semi-supervised Clustering with User Feedback,” Constrained Clustering: Advances in Algorithms, Theory, and Applications 4.1, 2003, pp. 17-32.
Countly Mobile Analytics, <http://count.ly/> Printed Jul. 18, 2013 in 9 pages.
Distimo—App Analytics, <http://www.distimo.com/app-analytics> Printed Jul. 18, 2013 in 5 pages
Flurry Analytics, <http://www.flurry.com/> Printed Jul. 18, 2013 in 14 pages.
Gesher, Ari, “Palantir Screenshots in the Wild: Swing Sightings,” The Palantir Blog, Sep. 11, 2007, pp. 1-12, retrieved from the internet https://www.palantir.com/2007/09/palantir-screenshots/ retrieved on Aug. 18, 2015.
Google Analytics Official Website—Web Analytics & Reporting, <http://www.google.com/analytics.index.html> Printed Jul. 18, 2013 in 22 pages.
Gorr et al., “Crime Hot Spot Forecasting: Modeling and Comparative Evaluation,” Grant 98-IJ-CX-K005, May 6, 2002, 37 pages.
Gu et al., “Record Linkage: Current Practice and Future Directions,” Jan. 15, 2004, pp. 32.
Hua et al., “A Multi-attribute Data Structure with Parallel Bloom Filters for Network Services” HiPC 2006, LNCS 4297, pp. 277-288, 2006.
“HunchLab: Heat Map and Kernel Density Calculation for Crime Analysis,” Azavea Journal, printed from www.azavea.com/blogs/newsletter/v4i4/kernel-density-capabilities-added-to- -hunchlab/ on Sep. 9, 2014, 2 pages.
Kontagent Mobile Analytics, <http://www.kontagent.com/> Printed Jul. 18, 2013 in 9 pages.
Localytics—Mobile App Marketing & Analytics, <http://www.localytics.com/> Printed Jul. 18, 2013 in 12 pages.
Mixpanel—Mobile Analytics, <https://mixpanel.com/> Printed Jul. 18, 2013 in 13 pages.
Open Web Analytics (OWA), <http://www.openwebanalytics.com/> Printed Jul. 19, 2013 in 5 pages.
Palantir Technolgies, “Palantir Labs—Timeline,” Oct. 1, 2010, retrieved from the internet https://www.youtube.com/watch?v=JCgDW5bru9M retrieved on Aug. 19, 2015.
Piwik—Free Web Analytics Software. <http://piwik.org/> Printed Jul. 19, 2013 in18 pages.
StatCounter—Free Invisible Web Tracker, Hit Counter and Web Stats, <http://statcounter.com/> Printed Jul. 19, 2013 in 17 pages.
TestFlight—Beta Testing on the Fly, <http://testflightapp.com/> Printed Jul. 18, 2013 in 3 pages.
trak.io, <http://trak.io/> printed Jul. 18, 2013 in 3 pages.
UserMetrix, <http://usermetrix.com/android-analytics> printed Jul. 18, 2013 in 3 pages.
Valentini et al., “Ensembles of Learning Machines,” M. Marinaro and R. Tagliaferri (Eds.): WIRN VIETRI 2002, LNCS 2486, pp. 3-20.
Wang et al., “Research on a Clustering Data De-Duplication Mechanism Based on Bloom Filter,” IEEE 2010, 5 pages.
Wikipedia, “Multimap,” Jan. 1, 2013, https://en.wikipedia.org/w/index.php?title=Multimap&oldid=530800748.
Wright et al., “Palantir Technologies VAST 2010 Challenge Text Records—Investigations into Arms Dealing,” Oct. 29, 2010, pp. 1-10, retrieved from the internet http://hcil2.cs.umd.edu/newvarepository/VAST%20Challenge%202010/challenge- s/MC1%20-%20investigations%20into%20Arms%20Dealing/entries/Palantir%20Tech- nologies/ retrieved on Aug. 20, 2015.
Notice of Allowance for U.S. Appl. No. 14/479,863 dated Mar. 31, 2015.
Notice of Allowance for U.S. Appl. No. 14/319,161 dated May 4, 2015.
Notice of Allowance for U.S. Appl. No. 14/323,935 dated Oct. 1 , 2015.
Notice of Allowance for U.S. Appl. No. 14/552,336 dated Nov. 3, 2015.
Notice of Allowance for U.S. Appl. No. 14/326,738 dated Nov. 18, 2015.
Notice of Allowance for U.S. Appl. No. 14/746,671 dated Jan. 21, 2016.
Notice of Allowance for U.S. Appl. No. 14/858,647 dated Mar. 4, 2016.
Notice of Allowance for U.S. Appl. No. 13/247,987 dated Mar. 17, 2016.
Official Communication for U.S. Appl. No. 14/451,221 dated Oct. 21, 2014.
Official Communication for U.S. Appl. No. 14/463,615 dated Nov. 13, 2014.
Official Communication for U.S. Appl. No. 13/827,491 dated Dec. 1 , 2014.
Official Communication for U.S. Appl. No. 14/479,863 dated Dec. 26, 2014.
Official Communication for U.S. Appl. No. 14/319,161 dated Jan. 23, 2015.
Official Communication for U.S. Appl. No. 14/483,527 dated Jan. 28, 2015.
Official Communication for U.S. Appl. No. 14/463,615 dated Jan. 28, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 4, 2015.
Official Communication for U.S. Appl. No. 14/571,098 dated Mar. 11, 2015.
Official Communication for U.S. Appl. No. 14/463,615 dated May 21, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Jun. 16, 2015.
Official Communication for U.S. Appi. No. 14/571,098 dated Aug. 5, 2015.
Official Communication for U.S. Appl. No. 14/571,098 dated Aug. 24, 2015.
Official Communication for U.S. Appl. No. 14/463,615 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/562,524 dated Sep. 14, 2015.
Official Communication for U.S. Appl. No. 14/746,671 dated Sep. 28, 2015.
Official Communication for U.S. Appl. No. 14/141,252 dated Oct. 8, 2015.
Official Communication for U.S. Appl. No. 13/827,491 dated Oct. 9, 2015.
Official Communication for U.S. Appl. No. 14/483,527 dated Oct. 28, 2015.
Official Communication for U.S. Appl. No. 14/676,621 dated Oct. 29, 2015.
Official Communication for U.S. Appl. No. 14/571,098 dated Nov. 10, 2015.
Official Communication for U.S. Appl. No. 14/562,524 dated Nov. 10, 2015.
Official Communication for U.S. Appl. No. 14/746,671 dated Nov. 12, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Nov. 16, 2015.
Official Communication for U.S. Appl. No. 14/842,734 dated Nov. 19, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Dec. 3, 2015.
Official Communication for U.S. Appl. No. 14/463,615 dated Dec. 9, 2015.
Official Communication for U.S. Appl. No. 14/300,447 dated Dec. 10, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Dec. 21, 2015.
Official Communication for U.S. Appl. No. 14/874,690 dated Dec. 21, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Dec. 24, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Dec. 24, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Jan. 4, 2016.
Official Communication for U.S. Appl. No. 14/225,160 dated Jan. 25, 2016.
Official Communication for U.S. Appl. No. 14/645,304 dated Jan. 25, 2016.
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 1, 2016.
Official Communication for U.S. Appl. No. 14/306,154 dated Feb. 1, 2016.
Official Communication for U.S. Appl. No. 14/929,584 dated Feb. 4, 2016.
Official Communication for U.S. Appl. No. 14/571,098 dated Feb. 23, 2016.
Official Communication for U.S. Appl. No. 14/948,009 dated Feb. 25, 2016.
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 26, 2016.
Official Communication for U.S. Appl. No. 14/306,138 dated Mar. 17, 2016.
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 17, 2016.
Official Communication for New Zealand Patent Application No. 622473 dated Mar. 27, 2014.
Official Communication for New Zealand Patent Application No. 622473 dated Jun. 19, 2014.
Official Communication for Great Britain Patent Application No. 1404499.4 dated Aug. 20, 2014.
Official Communication for Great Britain Patent Application No. 1404486.1 dated Aug. 27, 2010.
Official Communication for Great Britain Patent Application No. 1404489.5 dated Aug. 27, 2014.
Official Communicationor Great Britain Patent Application No. 1404499.4 dated Sep. 29, 2014.
Official Communication for Great Britain Patent Application No. 1404489.5 dated Oct. 6, 2014.
Official Communication for European Patent Application No. 14197938.5 dated Apr. 23, 2015.
Official Communication for European Patent Application No. 14200298.9 dated May 13, 2015.
Official Communication for Great Britain Patent Application No. 1404486.1 dated May 21, 2015.
Official Communication for Great Britain Patent Application No. 1404489.5 dated May 21, 2015.
Official Communication for European Patent Application No. 14200246.8 dated May 29, 2015.
Official Communication for Great Britain Patent Application No. 1404499.4 dated Jun. 11, 2015.
Official Communication for European Patent Application No. 15165244.3 dated Aug. 27, 2015.
Official Communication for Netherlands Patents Application No. 2012421 dated Sep. 18, 2015.
Official Communication for Netherlands Patents Application No. 2012417 dated Sep. 18, 2015.
Official Communication for Netherlands Patent Application 2012438 dated Sep. 21, 2015.
Official Communication for European Patent Application No. 15181419.1 dated Sep. 29, 2015.
Official Communication for European Patent Application No. 15184764.7 dated Dec. 14, 2015.
Official Communication for European Patent Application No. 14187996.5 dated Feb. 19, 2016.
“Using Whois Based Geolocation and Google Maps API for Support Cybercrime Investigations,” http://wseas.us/e-library/conferences/2013/Dubrovnik/TELECIRC/TELECIRC-32.pdf.
Notice of Allowance for U.S. Appl. No. 14/139,628 dated Jun. 24, 2015.
Notice of Allowance for U.S. Appl. No. 14/139,640 dated Jun. 17, 2015.
Notice of Allowance for U.S. Appl. No. 14/139,713 dated Jun. 12, 2015.
Notice of Allowance for U.S. Appl. No. 14/264,445 dated May 14, 2015.
Notice of Allowance for U.S. Appl. No. 14/278,963 dated Sep. 2, 2015.
Official Communication for U.S. Appl. No. 13/827,491 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 14/141,252 dated Apr. 14, 2016.
Official Communication for U.S. Appl. No. 14/148,568 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 10, 2014.
Official Communication for U.S. Appl. No. 14/225,160 dated Feb. 11, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Aug. 12, 2015.
Official Communication for U.S. Appl. No. 14/251,485 dated Oct. 1, 2015.
Official Communication for U.S. Appl. No. 14/264,445 dated Apr. 17, 2015.
Official Communication for U.S. Appl. No. 14/278,963 dated Jan. 30, 2015.
Official Communication for U.S. Appl. No. 14/294,098 dated Aug. 15, 2014.
Official Communication for U.S. Appl. No. 14/294,098 dated Nov. 6, 2014.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 23, 2014.
Official Communication for U.S. Appl. No. 14/306,147 dated Feb. 19, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/306,154 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Nov. 28, 2014.
Official Communication for U.S. Appl. No. 14/326,738 dated Dec. 2, 2014.
Official Communication for U.S. Appl. No. 14/326,738 dated Jul. 31, 2015.
Official Communication for U.S. Appl. No. 14/483,527 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 14/518,757 dated Dec. 1, 2015.
Official Communication for U.S. Appl. No. 14/518,757 dated Apr. 2, 2015.
Official Communication for U.S. Appl. No. 14/518,757 dated Jul. 20, 2015.
Official Communication for U.S. Appl. No. 14/552,336 dated Jul. 20, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated Dec. 9, 2015.
Official Communication for U.S. Appl. No. 14/581,920 dated Mar. 1, 2016.
Official Communication for U.S. Appl. No. 14/639,606 dated Oct. 16, 2015.
Official Communication for U.S. Appl. No. 14/676,621 dated Jul. 30, 2015.
Official Communication for U.S. Appl. No. 14/726,353 dated Mar. 1, 2016.
Official Communication for U.S. Appl. No. 14/857,071 dated Mar. 2, 2016.
Restriction Requirement for U.S. Appl. No. 14/857,071 dated Dec. 11, 2015.
Bhuyan et al., “Network Anomaly Detection: Methods, Systems and Tools,” First Quarter 2014, IEEE.
FireEye, <http://www.fireeye.com/> Printed Jun. 30, 2014 in 2 pages.
Lee et al., “A Data Mining and CIDF Based Approach for Detecting Novel and Distributed Intrusions,” Lecture Notes in Computer Science, vol. 1907 Nov. 11, 2000, pp. 49-65.
VirusTotal—About, <http://www.virustotal.com/en/about/> Printed Jun. 30, 2014 in 8 pages.
Zheng et al., “GOEAST: a web-based software toolkit for Gene Ontology enrichment analysis,” Nucleic acids research 36.suppl 2 (2008): pp. W385-W363.
Notice of Allowance for U.S. Appl. No. 14/223,918 dated Jan. 6, 2016.
Notice of Allowance for U.S. Appl. No. 14/319,765 dated Nov. 25, 2016.
Notice of Allowance for U.S. Appl. No. 14/823,935 dated Apr. 25, 2016.
Notice of Allowance for U.S. Appl. No. 14/970,317 dated May 26, 2016.
Notice of Allowance for U.S. Appl. No. 15/053,155 dated Sep. 26, 2017.
Notice of Allowance for U.S. Appl. No. 15/053,177 dated Sep. 15, 2017.
Official Communication for European Patent Application No. 14199180.2 dated Jun. 22, 2015.
Official Communication for European Patent Application No. 14199180.2 dated Aug. 31, 2015.
Official Communication for European Patent Application No. 14200298.9 dated Mar. 15, 2016.
Official Communication for European Patent Application No. 15175106.2 dated Nov. 5, 2015.
Official Communication for European Patent Application No. 15175151.8 dated Nov. 25, 2015.
Official Communication for European Patent Application No. 15180985.2 dated Jan. 15, 2016.
Official Communication for U.S. Appl. No. 14/141,252 dated Nov. 28, 2017.
Official Communication for U.S. Appl. No. 14/223,918 dated Jun. 8, 2015.
Official Communication for U.S. Appl. No. 14/280,490 dated Jul. 24, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Aug. 1, 2016.
Official Communication for U.S. Appl. No. 14/319,765 dated Nov. 25, 2014.
Official Communication for U.S. Appl. No. 14/490,612 dated Jan. 27, 2015.
Official Communication for U.S. Appl. No. 14/490,612 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/731,312 dated Apr. 14, 2016.
Official Communication for U.S. Appl. No. 14/823,935 dated Dec. 4, 2015.
Official Communication for U.S. Appl. No. 14/923,712 dated Feb. 12, 2016.
Official Communication for U.S. Appl. No. 14/970,317 dated Mar. 21, 2016.
Official Communication for U.S. Appl. No. 14/982,699 dated Mar. 25, 2016.
Official Communication for U.S. Appl. No. 15/053,155 dated Sep. 13, 2017.
Official Communication for U.S. Appl. No. 15/053,155 dated Mar. 7, 2017.
Official Communication for U.S. Appl. No. 15/053,177 dated Mar. 7, 2017.
Official Communication for U.S. Appl. No. 15/071,064 dated Jun. 16, 2016.
Related Publications (1)
Number Date Country
20170277780 A1 Sep 2017 US
Continuations (1)
Number Date Country
Parent 14319765 Jun 2014 US
Child 15483731 US