Focused probabilistic entity resolution from multiple data sources

Information

  • Patent Grant
  • 10437840
  • Patent Number
    10,437,840
  • Date Filed
    Friday, August 19, 2016
    8 years ago
  • Date Issued
    Tuesday, October 8, 2019
    4 years ago
  • CPC
  • Field of Search
    • CPC
    • G06F17/30011
    • G06F17/30699
    • G06F17/30604
    • G06F11/1435
    • G06F17/30
    • G06F16/93
    • G06F16/335
    • G06F16/288
    • G06F16/00
    • G06F16/90335
    • G06F16/907
  • International Classifications
    • G06F7/02
    • G06F16/00
    • G06F16/2457
    • G06F3/0484
    • G06F16/2455
    • G06F16/93
    • G06F16/335
    • Term Extension
      320
Abstract
Various systems and methods are provided for performing soft entity resolution. A plurality of data objects are retrieved from a plurality of data stores to create aggregated data objects for one or more entities. One or more retrieved data objects may be associated with the same entity, based at least in part upon one or more attribute types and attribute values of the data objects. In response to a determination that the one or more of the retrieved data objects should be associated with the same entity, metadata is generated that associates the data objects with the entity, the metadata being stored separately from the data objects, such that the underlying data objects remain unchanged. In addition, one or more additional attributes may be determined for the entity, based upon the data objects associated with the entity.
Description
TECHNICAL FIELD

The present disclosure relates to systems, methods, and techniques for entity resolution, and more specifically to performing entity resolution on data objects retrieved from a plurality of different data sources.


BACKGROUND

In many applications that involve the analysis or handling of a large number of entities, it is often beneficial to be able to focus attention and resources on the most relevant entities instead of taking a more diffuse, unfocused approach. For example, in a humanitarian aid or fraud detection context, it may be beneficial to be able to focus resources based on desired factors.


SUMMARY

The systems, methods, and devices described herein each have several aspects, no single one of which is solely responsible for its desirable attributes. Without limiting the scope of this disclosure, several non-limiting features will now be discussed briefly.


The systems, methods, and devices of the present disclosure may provide, among other features, soft entity resolution schemes in which large amounts of entity and event data, objects, features, and/or metadata are efficiently presented to a user on an interactive user interface. In various embodiments, an interactive soft entity resolution system may enable rapid and deep analysis of various entities, events, and their associated attributes by the user. In some embodiments, data objects from different data sources may be retrieved and associated with entities and/or events, allowing the user to be able to access more complete data for particular entities or events. Metadata may be generated associating different data objects from different data stores to entities and events, while preserving the content of the underlying data objects. One or more scores may be determined for each entity or event, based upon attributes values associated with the entity or event. A user interface may be generated that displays entities or events to the user, based upon the generated scores and one or more user attributes, such that the user is presented with the entities or events that are most relevant to the user.


It has been noted that design of computer user interfaces “that are useable and easily learned by humans is a non-trivial problem for software developers.” (Dillon, A. (2003) User Interface Design. MacMillan Encyclopedia of Cognitive Science, Vol. 4, London: MacMillan, 453-458.) The present disclosure describes various embodiments of interactive and dynamic user interfaces that are the result of significant development. This non-trivial development has resulted in the user interfaces described herein which may provide significant cognitive and ergonomic efficiencies and advantages over previous systems. The interactive and dynamic user interfaces include improved human-computer interactions that may provide reduced mental workloads, improved decision-making, reduced work stress, and/or the like, for a user. For example, user interaction with the interactive user interfaces described herein may provide a display of entities or events of greatest relevant to the user, each entity or event comprising data from a plurality of data objects from different data sources, and may enable a user to more quickly and accurately access, navigate, assess, and digest the data associated with particular entities or events than previous systems, without having to wade through the different data from various data sources.


Further, the interactive and dynamic user interfaces described herein are enabled by innovations in efficient interactions between the user interfaces and underlying systems and components. For example, disclosed herein are improved methods of performing entity resolution on data objects from different data sources without changing the underlying data objects, and displaying to the user data on entities or events that are of most relevance to the user. The interactions and presentation of data via the interactive user interfaces described herein may accordingly provide cognitive and ergonomic efficiencies and advantages over previous systems.


Various embodiments of the present disclosure provide improvements to various technologies and technological fields. For example, existing data resolution and user interface systems may be limited in various ways, and various embodiments of the disclosure provide significant improvements over such technology. Additionally, various embodiments of the present disclosure are inextricably tied to computer technology. In particular, various embodiments rely on detection of user inputs via graphical user interfaces, retrieval of data from different data sources, creation and storage of metadata associated with different data objects, and/or the like. Such features and others are intimately tied to, and enabled by, computer technology, and would not exist except for computer technology. For example, the interactions with displayed data described below in reference to various embodiments cannot reasonably be performed by humans alone, without the computer technology upon which they are implemented. Further, the implementation of the various embodiments of the present disclosure via computer technology enables many of the advantages described herein, including more efficient interaction with, and presentation of, data relating to various entities and events.


In an embodiment, a computer system is disclosed comprising one or more computer readable storage mediums configured to store one or more software modules including computer executable instructions and one or more computer processors in communication with the one or more computer readable storage mediums. The one or more computer processors may be configured to execute the one or more software modules in order to cause the computer system to retrieve a plurality of data objects from a plurality of data stores, the data stores comprising at least a first data store containing data objects corresponding to entities, and a second data store containing data objects corresponding to events. The one or more processors may be further configured to cause the computer system to create aggregated data objects for one or more entities. Creating an aggregated data object for an entity comprises determining whether one or more of the retrieved data objects should be associated with the same entity, based at least in part upon one or more attribute types and attribute values associated with the one or more data objects, in response to a determination that the one or more of the retrieved data objects should be associated with the same entity, generating metadata associating the one or more data objects with the entity, wherein the metadata is stored separately from the one or more data objects, such that the one or more data objects are unchanged, and determining one or more additional attributes for the entity, based upon the data objects associated with the entity.


In some embodiments, the one or more computer processors are further configured to identify at least a portion of the attributes associated with the one or more entities as scoring factors; generate one or more scores for the one or more entities, based at least in part upon one or more values for the scoring factors for the entities determined using the data objects and additional attributes associated with the entities; and generate user interface information for displaying at least a portion of the entities, based at least in part upon the generated scores for the entities.


In some embodiments, generating scores for the one or more entities comprises, for an entity of the one or more entities, processing the one or more scoring factors associated with the entity using a scoring model; and generating a score for the entity using the scoring model.


In some embodiments, the scoring model comprises one or more weights associated with the one or more scoring factors.


In some embodiments, the one or more additional attributes comprises a history attribute generated based at least in part upon one or more event data objects associated with the entity.


In some embodiments, determining whether one or more of the retrieved data objects should be associated with the same entity comprises determining a confidence value based at least in part upon the one or more attribute types and attribute values associated with the one or more data objects, wherein the data objects are associated with the same entity if the confidence value exceeds a threshold confidence.


In some embodiments, the user interface information for displaying at least a portion of the entities is further based at least in part upon one or more user attributes. The one or more user attributes may comprise a user location.


In some embodiments, the one or more computer processors are further configured to: receive an indication of an event; associate the event with one or more scoring factors; in substantially real-time, update the one or more scoring factors based at least in part upon the event; re-generate scores for at least some of the one or more entities, based upon the updated scoring factors; and generate user interface information for displaying at least a portion of the entities, based at least in part upon the re-generated scores.


In some embodiments, the one or more scoring factors are updated based upon a relationship between an entity attribute and an attribute value associated with the event.


In some embodiments, the user interface information further comprises information for displaying a basis for a change in a score of a displayed entity, based upon the updated scoring factors.


Additional embodiments of the disclosure are described below in reference to the appended claims, which may serve as an additional summary of the disclosure.


In various embodiments, computer-implemented methods are disclosed in which, under control of one or more hardware computing devices configured with specific computer executable instructions, one or more aspects of the above-described embodiments (including one or more aspects of the appended claims) are implemented and/or performed.


In various embodiments, non-transitory computer-readable storage mediums storing software instructions are disclosed, wherein, in response to execution by a computing system having one or more hardware processors, the software instructions configure the computing system to perform operations comprising one or more aspects of the above-described embodiments (including one or more aspects of the appended claims).


Further, as described herein, various embodiments of the system may be configured and/or designed to generate user interface data useable for rendering the various interactive user interfaces described. The user interface data may be used by the system, and/or another computer system, device, and/or software program (for example, a browser program), to render the interactive user interfaces. The interactive user interfaces may be displayed on, for example, electronic displays (including, for example, touch-enabled displays).





BRIEF DESCRIPTION OF THE DRAWINGS

The following drawings and the associated descriptions are provided to illustrate embodiments of the present disclosure and do not limit the scope of the claims. Aspects and many of the attendant advantages of this disclosure will become more readily appreciated as the same become better understood by reference to the following detailed description, when taken in conjunction with the accompanying drawings, wherein:



FIG. 1 is a block diagram illustrating an example data analysis system, according to an embodiment.



FIG. 2 illustrates a block diagram of various types of data sources that may be used to perform probabilistic entity resolution, in accordance with some embodiments.



FIG. 3 illustrates a flowchart of a process for associating a plurality of data objects with a single entity, in accordance with some embodiments.



FIG. 4 illustrates a block diagram of a system for performing probabilistic entity resolution on data objects from a plurality of data sources, and assessing scores or other relevant factors of the entities, in accordance with some embodiments



FIG. 5 illustrates a flowchart of a process for performing entity resolution and determining scores for entities, in accordance with some embodiments.



FIG. 6A illustrates an example user interfaces that may be generated to display entities and their associated scores.



FIG. 6B illustrates a user interface that may be used to add additional entities to the displayed entity list.



FIG. 7 illustrates a flowchart of a process for displaying a listing of entities to a user at a user interface, in accordance with some embodiments.



FIG. 8 illustrates a flowchart of a process for updating scores in response to inputted events, in accordance with some embodiments.



FIG. 9A illustrates an embodiment of a database system using an ontology.



FIG. 9B illustrates an embodiment of a system for creating data in a data store using a dynamic ontology.



FIG. 9C illustrates a sample user interface using relationships described in a data store using a dynamic ontology.



FIG. 10 illustrates a computer system with which certain methods discussed herein may be implemented.





Throughout the drawings, reference numbers may be reused to indicate correspondence between referenced elements. Nevertheless, use of different numbers does not necessarily indicate a lack of correspondence between elements. And, conversely, reuse of a number does not necessarily indicate that the elements are the same.


DETAILED DESCRIPTION

Definitions


In order to facilitate an understanding of the systems and methods discussed herein, a number of terms are defined below. The terms defined below, as well as other terms used herein, should be construed to include the provided definitions, the ordinary and customary meaning of the terms, and/or any other implied meaning for the respective terms. Thus, the definitions below do not limit the meaning of these terms, but only provide exemplary definitions.


Ontology: Stored information that provides a data model for storage of data in one or more databases. For example, the stored data may comprise definitions for object types and property types for data in a database, and how objects and properties may be related.


Data Store: Any computer readable storage medium and/or device (or collection of data storage mediums and/or devices). Examples of data stores include, but are not limited to, optical disks (e.g., CD-ROM, DVD-ROM, etc.), magnetic disks (e.g., hard disks, floppy disks, etc.), memory circuits (e.g., solid state drives, random-access memory (RAM), etc.), and/or the like. Another example of a data store is a hosted storage environment that includes a collection of physical data storage devices that may be remotely accessible and may be rapidly provisioned as needed (commonly referred to as “cloud” storage).


Database: Any data structure (and/or combinations of multiple data structures) for storing and/or organizing data, including, but not limited to, relational databases (e.g., Oracle databases, mySQL databases, etc.), non-relational databases (e.g., NoSQL databases, etc.), in-memory databases, spreadsheets, as comma separated values (CSV) files, eXtendible markup language (XML) files, TeXT (TXT) files, flat files, spreadsheet files, and/or any other widely used or proprietary format for data storage. Databases are typically stored in one or more data stores. Accordingly, each database referred to herein (e.g., in the description herein and/or the figures of the present application) is to be understood as being stored in one or more data stores.


Entity: A person, a place, an organization, a market instrument, or other noun.


Data Object or Object: A data container for information representing specific things in the world that have a number of definable properties. For example, a data object can represent an entity such as a person, a place, an organization, a market instrument, or other noun. A data object can represent an event that happens at a point in time or for a duration. A data object can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object may be associated with a unique identifier that uniquely identifies the data object. The object's attributes (e.g. metadata about the object) may be represented in one or more properties.


Object Type: Type of a data object (e.g., Person, Event, or Document). Object types may be defined by an ontology and may be modified or updated to include additional object types. An object definition (e.g., in an ontology) may include how the object is related to other objects, such as being a sub-object type of another object type (e.g. an agent may be a sub-object type of a person object type), and the properties the object type may have.


Properties: Attributes of a data object that represent individual data items. At a minimum, each property of a data object has a property type and a value or values.


Property Type: The type of data a property is, such as a string, an integer, or a double. Property types may include complex property types, such as a series data values associated with timed ticks (e.g. a time series), etc.


Property Value: The value associated with a property, which is of the type indicated in the property type associated with the property. A property may have multiple values.


Link: A connection between two data objects, based on, for example, a relationship, an event, and/or matching properties. Links may be directional, such as one representing a payment from person A to B, or bidirectional.


Link Set: Set of multiple links that are shared between two or more data objects.


Overview


Various embodiments of the present disclosure provide improvements to various technologies and technological fields. For example, various aspects of the embodiments can be used to generate holistic entity-centric profiles for individual entities based upon data gathered from multiple different types of data sources, generate user interfaces allowing users to receive information on the most relevant entities, and update entity data in response to current events.


By aggregating different data objects associated with an entity or event into a single aggregated data object corresponding to the entity or event, entity or event data may be better organized for assessing relevant attributes associated with the entity (e.g., in a fraud detection or humanitarian aid context, a risk or need associated with the entity). In some embodiments, the data objects may be aggregated even if there is not complete certainty that the data objects are associated with the same entity or event. On the other hand, in some embodiments, the original underlying data objects may be preserved, such that they can continue to be used for other applications where it may not be appropriate to aggregate and associate the data objects into a single aggregated data object. As such, the usefulness and functionality of the stored data objects may be improved.


In addition, in some embodiments, customized listing of aggregated data objects may be provided to different users, based upon one or more entity or event metrics (e.g., scores) and one or more user attributes. This allows the system to provide to different users listings that will be most relevant to each user. Furthermore, updating entity metrics in response to inputted events may improve the ability of the system to provide users with the most up-to-date and relevant information.


Not necessarily all such advantages can be achieved in accordance with any particular embodiment of the invention. Thus, the invention can be embodied or carried out in a manner that achieves or optimizes one advantage or group of advantages as taught herein without necessarily achieving other advantages as can be taught or suggested herein.



FIG. 1 is a block diagram illustrating an example data analysis system 100, according to an embodiment. As shown in the embodiment of FIG. 1, the data analysis system 100 includes an application server 115 running on a server computing system 110, a client 135 running on a client computer system 130 (also referred to as a user device), and at least one data store 140. Further, the client 135, application server 115, and data store 140 may communicate over a network 150, for example, to access data sources 160.


The application server 115 may include a resolution engine 120, a metrics engine 122, and a workflow engine 124. The resolution engine 120, metrics engine 122, and workflow engine 124 may be software modules as described below in reference to FIGS. 2-8. According to an embodiment, the resolution engine 120 is configured to read data from a variety of data sources 160 to obtain a plurality of different data objects. The accessed data objects from data sources 160 may be of different types. For example, the data objects may include entity data objects, event data objects, and/or the like. The resolution engine 120 may use the retrieved data objects to generate one or more aggregated data objects for one or more entities and/or events. For example, the resolution engine 120 may analyze the retrieved data objects and generate additional metadata associating the data objects with particular entities and/or events. In some embodiments, data generated by the resolution engine may be stored on the server computing system 110 and/or on the data store 140. The operations of the resolution engine 120 are discussed in greater detail below in reference to FIGS. 2-4.


In some embodiments, the metrics engine 122 may be configured to score entities and/or events according to one or more criteria (e.g., risk criteria, need criteria, and/or the like). In some embodiments, the metrics engine 122 may analyze and weigh one or more attributes values or combinations of attributes values associated with an entity (e.g., provided by a holistic entity object generated by the resolution engine 120) or an event.


The application server may further include a workflow engine 124. The workflow engine may generate and/or provide the various user interfaces of the data analysis system. For example, the workflow engine 124 may present various data generated by the resolution engine 120 and/or the metrics engine 122. For example, the workflow engine 124 may present a user with a list of one or more entities, based upon the determined scores associated with the entities. In some embodiments, the workflow engine 124 may be used to generate one or more user interfaces that may be displayed to a user at a client computer system 130 (e.g., as a web service).


The client 135 may represent one or more software applications or modules configured to present data and translate input, from a user, into requests for data analyses by the application server 115. In one embodiment, the client 135 and the application server 115 may be embodied in the same software module and/or may be included in the same computing system. However, several clients 135 may execute on the client computer 130, and/or several clients 135 on several client computers 130 may interact with the application server 115. In some embodiments, client computer 130 may correspond to a personal computer, a laptop, a mobile device (e.g., a mobile phone, tablet, and/or the like), etc.


In one embodiment, the client 135 may be a browser accessing a web service. In various embodiments, a component of the system, for example the workflow engine 124, may generate user interfaces (for example, that may be transmitted to a display or browser and displayed to a user) and/or may generate instructions or code useable to generate a display and/or user interface (for example, that may be transmitted to a display or browser where a user interface may be generated and displayed to a user).


While the client 135 and application server 115 are shown running on distinct computing systems, the client 135 and application server 115 may run on the same computing system. Further, the entity resolution engine 120, the risk engine 122, and/or the workflow engine 124 may run on separate applications servers 115, on separate server computing systems, or some combination thereof.


In an embodiment, the data sources 160 provide data available to the entity resolution engine 120, risk engine 122, and/or workflow engine 124, for example data objects associated with one or more entities or events. Such data sources may include relational data sources, web services data, XML data, and the like. Further, such data sources may include a variety of information and data, for example, entity information, event information, financial information, computer network-related data, and/or computer-related activity data, among others. Data sources 160 is included to be representative of a variety of data available to the server computer system 110 over network 150, as well as locally available data sources.


The data store 140 may be a Relational Database Management System (RDBMS) that stores the data as rows in relational tables. The term “database,” as used herein, may refer to a database (e.g., RDBMS or SQL database), or may refer to any other data structure, such as, for example a comma separated values (CSV), extensible markup language (XML), text (TXT) file, flat file, spreadsheet file, and/or any other widely used or proprietary format. While the data store 140 is shown as a distinct computing system, the data store 140 may operate on the same server computing system 110 as the application server 115.


Data Object Resolution


In many applications (e.g., humanitarian aid or fraud detection), it is often beneficial to be able to focus resources on particular entities (e.g., refugees, accounts, etc.) or events (e.g., money laundering) based on a scoring metric, e.g., greatest need, highest risk. For example, in a humanitarian aid context, certain entities (e.g., individuals, families, geographic areas) may be determined to have greater need for aid based upon one or more identified factors. These factors may be based upon attributes associated with the entity, such as previous aid history, demographic information, geographic region, and/or the like. In some embodiments, a score or metric indicating a level of need may be calculated for an entity based upon the factors associated with the entity, allowing a user (e.g., a humanitarian aid worker) to quickly identify which entities are most in need, and be able to allocate resources and attention accordingly.


In order to assess each entity or event based on a scoring metric (e.g., have the highest risk scores, need scores), an aggregated data object aggregating available data and attributes associated with each entity or event may be constructed. In addition, in some embodiments, patterns of events may be identified and used to generate one or more additional attributes. For example, the presence of multiple events within a certain period of time may be used to determine a scoring or ranking for the multiple events or one or more entities associated with the events. However, in many cases, data relating to an entity or event oftentimes may be fragmented and scattered between many different data sources and organized in different ways. In some embodiments, different databases may be maintained by different organizations or departments of an organization. As such, it may often be difficult to determine which data should be associated with which entities or events.



FIG. 2 illustrates a block diagram of various types of data sources that may be used to aggregated entity or event data objects, in accordance with some embodiments. In some embodiments, the data sources may include one or more data sources. For example, an entity information data base 202 may contain data objects relating to a plurality of entities. For example, in the context of humanitarian aid, each entity may correspond to an individual, family, group, and/or other entity that may potentially be in need of aid. In a fraud detection context, each entity may correspond to an account, an individual or organization associated with one or more accounts, and/or the like. Each entity in the entity information database may be associated with one or more attributes corresponding to different types of identification information. These may include names, account numbers, known addresses, physical descriptions, associated geographical areas, and/or the like. In some embodiments, an entity of the entity information DB 202 may be associated with values for some but not all types of identification information that may be associated with the entity. For example, a particular entity corresponding to a family in potential need of aid may be associated with one or more physical description attributes, but whose names or address are unknown. In some embodiments, different types of entities may be stored in different data stores.


In some embodiments, the data sources may also include event data sources. In some embodiments, an entity may be associated with one or more event data objects. For example, in a humanitarian aid context, an aid database 204 may contain entries relating to aid events, each entry specifying one or more attributes related to the aid event, such as time of aid, type of aid provided, recipient of aid, cost of aid, and/or the like. As such, an aid event may be associated with one or more entities (e.g., as providers or recipients of aid).


In some embodiments, different types of events may be organized and stored in different data sources. For example, in addition to an aid database 204, a humanitarian aid organization may have access to other databases, such as a natural disaster database 206 comprising information on natural disaster events (e.g., containing attributes such as type of disaster, geographic region affected by disaster, date of disaster, and/or the like), a field report database 208 comprising information on field report event, and/or the like.


In some embodiments, the data sources may also include other types of data sources 210 (e.g., data sources containing data objects relating to other types of entities).


In some embodiments, in order to form an aggregated data object 212 for an entity or event, data from these different sources may be retrieved and aggregated in accordance with a resolution scheme (e.g., at the resolution engine 120). The resolution engine 120 may be configured to identify which data objects from the accessed data sources 202-210 are associated with each entity, as well as determine one or more additional attributes for an entity based upon the data objects associated with the entity. By having the available data for each entity aggregated into a single data object, a score can be calculated for the entities or events based upon their related factors, properties, etc.


In some embodiments, a financial institution (e.g., a bank, a lender, and/or the like) may monitor one or more entities (e.g., individuals or organizations) for fraud, e.g., money laundering. Each entity may be associated with one or more risk factors indicating a likelihood that the entity may have committed fraud (e.g., transfers for unusual amounts of money, transactions with known fraudsters, fraud history, credit history, number of financial accounts owned, and/or the like). Accessed data sources may comprise one or more data sources containing individual or organization data, one or more data sources containing data on a plurality of accounts, one or more data sources containing transaction data, and/or the like.


In another embodiment, one or more of the techniques disclosed herein may be utilized by an insurance agency or underwriter to identify at risk insurance policies. Different entities corresponding to companies or other organizations may be analyzed to determine a risk score, based upon one or more risk factors (e.g., claim history, geographic location, industry, and/or the like). Aggregated data objects may be formed based upon data objects retrieved from a client data store, a claims data store, and/or the like. While the examples disclosed below will refer primarily to certain applications for ease of explanation (e.g., fraud detection context, humanitarian aid context, etc.), it is understood that the techniques disclosed herein may be applied to many other applications.


Probabilistic Object Resolution


As discussed above, a particular entity or an event may be associated with a plurality of different data objects distributed across a plurality of different data stores. For example, an entity object may be associated with a first data object from an entity data store, a second data object from an event data store, and a third data object from another type of data store (e.g., another entity data store).


In addition, in some embodiments, an entity may be associated with multiple data objects from the same data source. For example, a particular entity may be associated with a plurality of different entity objects in a first database. This may occur due to the entity being observed on multiple occasions, without the knowledge that the entities involved with each occasion are actually the same entity. For example, in the context of humanitarian aid and natural disaster relief, there may be multiple reports for entities (e.g., families, neighborhoods, etc.) in need of aid due to a natural disaster that may actually be directed to the same entity. Similarly, in some embodiments, multiple aid agencies or other sources may record events (e.g., natural disaster events) that may actually refer to the same event. In some embodiments, some data objects from a particular data source may not have complete data that may be associated with an entity. For example, a particular entity data object may be associated with a given neighborhood, but not have the street name or address number.


In addition, in some embodiments, a particular entity may be associated with multiple events (e.g., the money launderer/fraudster was described in multiple suspicious activity reports, and/or the like). For example, a fraudster may be associated with multiple suspicious activity report event data objects from a financial fraud database. Similarly, a family or other entity residing in a particular area may be associated with multiple events associated with deployment of humanitarian aid.


In some embodiments, attributes associated with a plurality of different data objects (from the same data source or from different data sources) may be compared, in order to determine whether different data objects can be associated with the same entity or event. In some embodiments, a confidence level is determined, wherein the confidence level indicates whether the data objects should be associated with the same entity. As discussed above, data objects may be associated with one or more attributes that may be used to associate the data object with a particular entity or event. For example, an entity data object may be associated with one or more attributes that may be used to identify the entity (e.g., name, description, identification numbers, and/or the like). An event data object may be associated with attributes that may be used to identify the event (e.g., time of event, type of event, and/or the like) and/or an entity associated with the event (e.g., an entity that caused the event, participated in the event, was affected by the event, and/or like). In some embodiments, data objects may not always have a complete set of attributes for which to identify an event or entity (e.g., a reporter/personnel recording a field report of an event may have observed a physical description an entity associated with the event, but was unable to record information on other attributes of the entity, such as name or identification numbers).


Attributes associated with the data objects may be compared to determine if different data objects (e.g., corresponding to entities, events, and/or the like) should be associated with the same aggregated data object corresponding to a particular entity or event. In some embodiments, certain types of attributes may be considered a strong indicator of a particular entity or event. For example, if two event data objects are associated with the same location and time, it may be inferred with a high level of confidence that the two event data objects are associated with the same event. On the other hand, certain types of attributes may be considered to be less indicative of a particular entity or event. For example, two data objects sharing the same source may not be considered a strong indicator that the objects should be associated with the same event.


In some embodiments, a confidence level indicating a probability that different objects may be considered to be associated with the same entity or event may be determined, based upon matching attribute data between the data objects. The data objects may be considered to be associated with the same entity or event for the purpose of calculating a score if the confidence level satisfies a threshold confidence level. For example, as discussed above, matching locations may be considered to satisfy the threshold confidence level, while matching sources may not.


In some embodiments, if a confidence level for two or more data objects satisfies the threshold confidence level, the data objects may not be automatically associated with the same entity or event. Instead, a notification may be presented to a user (e.g., at a user interface), prompting the user to confirm that the objects should be associated with the same entity or event.



FIG. 3 illustrates a flowchart of a process for associating a plurality of data objects with a single entity or event, in accordance with some embodiments. At block 302, a plurality of data objects is identified. In some embodiments, the data objects may be from different data sources (e.g., an entity object from an entity database and an event object from an events database). In some embodiments, the data objects may be from the same data source (e.g., multiple event objects from an events database).


As discussed above, each data object may be associated with one or more different attributes. At block 304, one or more shared attributes between the identified data objects are identified. For example, it may be determined that a first data object and a second data object are both associated with values for an address attribute. In some embodiments, a pair of data objects may share a plurality of different attributes. For example, in some embodiments, a field report object may be associated with attributes corresponding with information that a humanitarian aid worker is required to record when submitting a field report. As such, different field report objects will tend to be associated with the same attributes.


At block 306, a confidence level is determined, based at least in part upon the attribute types and values of the identified shared attributes. The shared attributes between the data objects may be analyzed to determine whether the values of the shared attributes match. The confidence level may be determined based upon matches and/or mismatches of the identified attribute values.


In addition, as discussed above, different types of attributes may be weighted differently when determining a confidence level. For example, a matching address may be a strong indicator that different data objects should be associated with the same entity, while matching sources, by itself, may be a weak indicator that the data objects should be associated with the same entity. Conversely, an address or source mismatch may both be strong indicators that the data objects are not associated with the same entity or event, while a mismatch in certain other attributes may be a weak indicator that the data objects are not associated with the same entity or event.


At block 308, a determination is made as to whether to confidence level satisfies a threshold level. If so, the process may proceed to block 310, where the identified data objects are associated with the same entity or event. In some embodiments, instead of automatically associating the data objects with the same entity or event in response to the confidence level satisfying a threshold, a notification may be displayed to a user at a user interface, prompting the user to manually determine whether the data objects should be associated with the same entity. In addition, a basis for determining the confidence level (e.g., matching attribute types) may be displayed.


In some embodiments, the data objects may be automatically associated with the same entity or event if the confidence level satisfies a first threshold level, while being presented to a user for confirmation if the confidence level satisfies a second threshold level lower than the first threshold level. For example, in some embodiments, data objects associated with a common value for a first attribute may be automatically determined to be associated with the same entity or event, while data objects associated with a common value for a second, different attribute may require manual confirmation by a user. In some embodiments, determining which data objects should be associated with the same entity/event allows for more complete information to be obtained for the entity/event, which may be used to calculate one or more scores associated with the entity or event (e.g., risk score, need score, and/or the like).


In some embodiments, because data objects may not always be associated with entities with complete certainty, probabilistic or “soft” resolution may be used to associate retrieved data objects with entities or events without modifying the underlying data objects or data sources. For example, while it may be acceptable for the data from two different data objects to associated with the same entity or event for the purpose of calculating a score for the entity or event, it may not be appropriate for the data objects to be combined in the source/original data store, due to uncertainty that the data objects actually correspond to the same entity or event.


Through the use of soft resolution, the functionality and usability of the system may be improved. By aggregating and associating different data objects with entities or events for certain applications (e.g., determining one or more scores, such as risk scores, need scores, and/or the like), a greater amount of data may be obtained, allowing for more accurate results (e.g., in determining scores for an entity or event). On the other hand, preserving the original data objects preserved for use in other applications where it may not be appropriate to aggregate the data objects.



FIG. 4 illustrates a block diagram of a system for performing soft resolution on data objects from a plurality of data sources, and assessing scores of the entity or event resolutions, in accordance with some embodiments. As illustrated in FIG. 4, a resolution assessment platform 402 may access a plurality of data sources 160. In some embodiments, the resolution assessment platform 402 is implemented by the resolution engine 120, metrics engine 122, or some combination thereof.


The data sources 160 may comprise entity data sources, event data sources, or other types of data sources. For example, in some embodiments, the data sources 160 may comprise one or more data sources as illustrated in FIG. 2 (e.g., entity database 202, aid database 204, natural disaster database 206, field reports database 208, and/or other databases 210).


Retrieved data objects from the plurality of data sources 160 may be processed using soft resolution, in order to determine which data objects should be associated with which entities or events for the purpose of assessing scores associated with the entities or events. In some embodiments, attribute types associated with the retrieved data objects may be compared to determining matching attribute types between pairs of data objects, from which a confidence level may be calculated. Pairs of data objects associated with a confidence value satisfying one or more confidence thresholds may be determined to be associated with the same entity or event (e.g., automatically or through manual user confirmation).


In some embodiments, soft resolution may comprise generating metadata to be associated with the retrieved data objects, wherein the metadata indicates which entities or events are associated with which data objects. As illustrated in FIG. 4, the soft resolution metadata 404 may be stored separately from the one or more data sources 160, such that the underlying data objects of the data sources 160 remain unchanged.


In some embodiments, the soft resolution metadata 404 may be used to form one or more aggregated entity or event data objects, wherein all available data associated with the entity or event is aggregated. In some embodiments, an aggregated data object may be constructed by accessing the soft resolution metadata 404 to determine which data objects are associated with the entity or event. The data associated with the data objects may be combined or aggregated to determine attribute values associated with the entity or event. For example, an entity may be associated with a first entity data object having values for birthday first set of attributes, and a second entity data object having values a second, different set of attributes, a first event data object corresponding to an event that the entity is associated with, and/or the like. In some embodiments, one or more aggregate attributes for the entity or event may be determined based upon the data objects associated with the entity or event. For example, a particular entity may be associated with a plurality of event data objects, which may be used to determine an event history for the entity, as well as one or more additional attributes based upon the determined event history (e.g., time between events, type of association with events, and/or the like). In another embodiment, a particular event may be associated with the plurality of entity data objects, which can be used to determine one or more aggregate attributes associated with the event (e.g., number of entities associated with event, types of entities associated with event, and/or the like).


In some embodiments, an entity may be associated with one or more affiliations (e.g., group affiliations). For example, an entity may be associated with an address or location associated with a particular group (e.g., a country, company, or other organization), may be identified in a field report data object as having been observed certain attributes associated with a group, may be associated with other entities believed to be affiliated with the group, and/or the like. In some embodiments, an affiliation score may be determined for an entity indicating a likelihood that the entity has a certain affiliation. In some embodiments, the calculated additional attributes for the entity may be stored as part of the soft entity resolution metadata 404 or in another data store (not shown).


In some embodiments, how additional attributes such as affiliation are determined may be updated by one or more users. For example, a fraud detection officer tasked with detecting and analyzing possible instances of money laundering may observe different types of indicators for identifying affiliations between two or more different entities. As a result, how affiliation is determined for the entities may be updated based upon the knowledge of the officer.


Scoring Model


In some embodiments, once an aggregated data object for an entity or event is created, a score may be calculated for the entity or event, indicating an applicability of the entity or event to a certain parameter. For example, in the context of fraud detection, a score for an entity may indicate a risk that the entity will commit fraud during a future time period. In the context of humanitarian aid, a score for an entity may indicate a level of need of the entity. In some embodiments, the score may be determined using a scoring model 406. In some embodiments, an entity may be associated with more than one score, each score being associated with a different type of parameter.


In some embodiments, the scoring model 406 extracts one or more attributes associated with an aggregated data object. The one or more attributes (also referred to as scoring factors) may correspond to attributes associated with an entity believed to be indicative of the parameter associated with the score (e.g., risk, need, etc.). These may include attributes related to a current status of the entity (e.g., geographic location of the entity, financial status of the entity, and/or the like), an event history associated with the entity (types of event associated with the entity, time since most recent event, time between events, and/or the like), and/or the like. The model may weigh the extracted attributes to make a determination of the score associated with the entity.


In some embodiments, one or more extracted attributes may be combined or aggregated to form an aggregated attribute to be used by the model. For example, in some embodiments, the scoring model may weigh a first attribute differently based upon a value of a second attribute.


In some embodiments, the scoring model 406 may utilize machine learning to determine scoring factors to extract and how the extracted scoring factors should be weighed in determining a score to be associated with the entity or event. In some embodiments, the models must first be trained before they can be used to determine scores to be associated with entities or events. For example, sample training data may be received by the model. The sample training data may comprise data associated with one or more entities or events from which one or more scoring factors may be extracted. The sample training data is processed by the model to determine scores for the one or more entities or events.


The results produced by the model can then be compared with one or more expected results, and a determination is made as to whether the model results are satisfactory. If the results are not satisfactory, the model may be updated based upon the comparison between the model results and the expected results. For example, in some embodiments, the weights afforded to certain extracted scoring factors may be altered, based upon how indicative they are of the parameter associated with the score.


In addition to training data, the model may continue to be updated during use. For example, as additional data is received for analyzed entities or events, the model may be further updated and improved, e.g., based upon the reported accuracy/inaccuracy of the model. For example, in the context of fraud detection, if it is determined that a particular entity has committed fraud, the attributes associated with the entity may be analyzed to determine one or more updates to the scoring model for calculating a score.


After scores have been calculated for one or more entities or events, a listing of entities or events may be displayed to a user at a user interface 408. The listing may be based upon the calculated scores (e.g., the listing may comprise entities having the highest scores). In some embodiments, a user may use the user interface 408 to modify the displayed listing of entities (e.g., add or remove entities from the listing), combine listed entities, bookmark listed entities, and/or other operations. Operations that may be performed using the user interface 408 will be described in greater detail below, in association with FIGS. 6-7.


Risk Determination Process



FIG. 5 illustrates a flowchart of a process for performing resolution and determining scores for aggregated data objects, in accordance with some embodiments. At block 502, a plurality of data sources is accessed. As discussed above, the plurality of data sources may include one or more entity data sources 502a, one or more event data sources 502b, and one or more other types of data sources 502c (e.g., data sources directed to other types of entities). At block 504, data objects are retrieved from the accessed data stores.


At block 506, attributes that may be used to identify an entity or event associated with the retrieved data objects are identified. For example, the attributes for identifying an event may comprise time and date attributes. In some embodiments, attributes for identifying an entity may comprise description attributes, associated location attributes, type attributes, title attributes, and/or the like.


At block 508, metadata is generated to associate retrieved data objects with entities or events, based at least in part upon the identified attributes. In some embodiments, identified attributes of different data objects may be compared and used to generate a confidence value that the data objects should be associated with the same entity or event. The attributes may be weighed based upon a type of attribute (e.g., certain types of attributes may be considered more indicative for identifying an entity or event) and whether the values of the attribute match between the different data objects. In some embodiments, if the confidence value meets a confidence threshold, the data objects may be automatically associated with the same entity or event.


In some embodiments, the generated metadata may be stored separately from the underlying data objects retrieved from the plurality of data stores. As such, the underlying data objects will remain unchanged.


At block 510, one or more additional attributes may be determined for each entity or event, based upon the data objects associated with the entity or event. For example, event data objects associated with an entity may be used to determine an event history associated with the entity, as well as one or more attribute values that may derived from the event history (e.g., number of events, time between events, time since most recent events, and/or the like). For example, in a fraud detection context, an entity associated with an account or account owner may be associated with multiple transaction events, which may be used to determine a transaction history associated with the entity, as well as one or more attributes associated with the transaction history (e.g., number of transactions, time between transactions, amount of transactions, and/or the like). In some embodiments, different attributes from different data objects may be used to determine a relationship of the entity or event. For example, a particular entity may be determined as being associated with a group or organization.


At block 512, a score is determined for each aggregated data object corresponding to a particular entity or event, based at least in part upon one or more scoring factors. The one or more scoring factors correspond to at least a portion of the attributes or combinations of attributes associated with the entity or event. In some embodiments, the score may be calculated using a scoring model, based upon the risk factors associated with the entity or event.


At block 514, a listing of one or more entities or events may be displayed to a user via a user interface at a user device. For example, a listing of entities or events having the highest scores may be displayed. In some embodiments, the entities displayed may be based at least in part upon one or more attributes of the user (e.g., user location, user department, user interests, and/or the like). The user may be able to modifying the listing by adding or removing entities or events from the list.


In some embodiments, a user may select a particular entity or event to view an aggregated data object for the entity or event. The aggregated data object may display data on attributes that would be useful to the user in dealing with the entity or event (e.g., photos, description, addresses). The aggregated data object may also display one or more attributes used in generating the score for the entity or event (e.g., most relevant scoring factors).


User Interface



FIGS. 6A and 6B illustrate example user interfaces that may be generated to display entities and/or events and their associated scores. In some embodiments, the user interface may display to a user a listing of entities 602 in descending order of scores. It is understood that although FIGS. 6A and 6B illustrate a listing of entities, in other embodiments, user interfaces may be generated to display events instead of entities, or some combination thereof. Displayed data for each entity may include title attributes of the entity such as an ID corresponding to the entity or a name of the entity, one or more additional attributes associated with the entity (e.g., a location associated with the entity, a most recent event associated with the entity), and the calculated score for the entity.


In some embodiments, the user may select a particular entity to view more detailed information on the selected entity. In some embodiments, the additional information for the entity may be retrieved using an aggregated data object associated with the entity. In some embodiments, the user may print out the additional information retrieved using the aggregated data object for the entity, for use when the user may not have access to a user device.


In some embodiments, the user may, when viewing the entity listing, believe that two or more entities in the listing actually refer to the same entity. As such, the user may select the desired entities and select the “Resolve” button 604 to combine the selected entities. In response, the data objects associated with each of the selected entities will be associated with a single aggregated data object. In addition, any additional attributes (e.g., history attributes, affiliation attributes, and/or the like) may be recalculated. In some embodiments, an updated score will also be calculated for the entity.


In some embodiments, the user may also modify the entities displayed in the listing. For example, the user may select the “Remove” button 606 to remove one or more selected entities from the listing. For example, the user may, using their own knowledge, determine that a particular entity is not relevant and thus select the entity for removal.


In some embodiments, the user may bookmark one or more of the listed entities (e.g., by selecting one or more entities and selecting the “Bookmark” button 608). As such, the user is able maintain a customized list of entities, even as the listing reflecting the entities with the highest scores changes. In some embodiments, the user may maintain multiple different lists of bookmarks (e.g., different lists associated with different locations, different types of events, and/or the like). A user interface element, such as a pull-down menu 610 may be used by the user to select a list for the selected entities to be added to.


In some embodiments, the user may be interested in certain entities that are not part of the displayed listing (e.g., due to their calculated scores being lower than the listed entities). As such, the user may desire to add one or more additional entities to the displayed listing (e.g., by selecting an “add entities” button 612).



FIG. 6B illustrates a user interface that may be used to add additional entities to the displayed entity list. As illustrated in FIG. 6B, an additional display panel allowing the user to search available entities is displayed (e.g., in response to the user selecting the “add” button 612). The user may search for one or more keywords relating to a desired data object (e.g., using search field 614) and receive a listing of search results 616. The user may select one or more entities 618 from the search results 616 to be added to the listing 602.


In some embodiments, the entities displayed to the user (e.g., listing 602) may be based at least in part upon one or more attributes of the user. For example, the user may be associated with a particular geographic area and/or a particular organization. As such, the listing 602 may be limited to entities associated with the geographic area, are relevant to the organization, etc. In some embodiments, the attributes of the user may be detected automatically. In some embodiments, the user may input one or more preferences (e.g., only list entities or events associated with certain locations, having certain attributes, and/or the like).



FIG. 7 illustrates a flowchart of a process for displaying a listing of entities to a user at a user interface, in accordance with some embodiments. At block 702, entity or event data and associated scores are identified. In some embodiments, the entity or event data may comprise all entity/event data, entity/event data corresponding to entities associated with a score above a threshold value, and/or the like.


At block 704, one or more attributes associated with the user are identified. These may include one or more attributes related to an identify of the user 704a, one or more attributes related to an organization associated with the user 704b, one or more attributes related to a location of the user 704c, and/or the like. In some embodiments, the one or more attributes may be determined automatically based upon a user profile of the user, one or more user-inputted credentials, and/or the like.


At block 706, a listing of entities/events is generated based upon the scores associated with the entities/events, and the identified user attributes. This may be done is order to present a listing of entities/events that will be most relevant to the user. For example, in some embodiments, a user may be associated with a location or region. Entities/events associated with locations that are in proximity to the location or region associated with the user may be more relevant to the user. As such, the listing of entities may be based upon the locations associated with the entities/events relative to the location associated with the user. For example, in some embodiments, entities/events having an associated location greater than a threshold distance from the location associated with the user may be excluded from the listing.


In some embodiments, the user may be associated with a particular organization. For example, a fraud detection officer may be associated with an organization that specializes in handling particular types of fraud (e.g., money laundering, insider trading, and/or the like). As such, entity listings generated for different users may be customized based upon the organization associated with the users (e.g., a listing generated for a user in the money laundering organization will contain entities or events associated with money laundering, while a listing generated for a user in the insider trading organization will contain entities or events associated with insider trading).


In some embodiments, other types of user attributes may be used to determine how the listing of entities or events is generated. In addition, in some embodiments the user may input one or more preferences for generating the listing. For example, the user may specify one or more additional criteria for which to generate the listing.


In some embodiments, one or more rules specifying which entities or events will be included in the listing may be generated. For example, in some embodiments, if a score for a particular entity is sufficiently high, the entity may be included in the listing for a particular user, even if the entity does not have attributes corresponding to the user (e.g., location, organization, and/or the like).


At block 708, user interface data is generated to display the listing of entities or events to the user at a user interface. As such, functionality of the system may be improved, as each user receives a listing of entities or events that is most relevant to them, based upon their attributes and/or preferences.


Event-Based Scoring Model


In some embodiments, events may change how a score is determined for certain entities, even if the attributes of the entities do not change. In some embodiments, the occurrence of a first event may increase the score of certain entities. In some embodiments, a first event may comprise an event involving a first set entities. In response, it may be expected that a second event will occur in the near-future. As a result, the scoring model (e.g., for calculating a score indicating a likelihood that the entity may be involved in an event) may be adjusted such that the first set of entities or other entities associated with the first set of entities will have a higher score. Thus, in some embodiments, a determined score associated with an entity may increase in response to an occurrence of an event, even if none of the attributes values associated with the entity have changed. In addition, the score associated with the entity may be adjusted even if the entity is not a participant of an event or otherwise directly associated with the event.


In another example, in the context of fraud monitoring, certain events may cause the risk scores associated with some entities to change. For example, an event corresponding to a large drop in the stock market may cause the scores for certain types of entities (e.g., entities with net assets below a certain value, entities with a certain amount of money invested in stocks, and/or the like) to increase or decrease, despite none of the attributes associated with the entities changing.



FIG. 8 illustrates a flowchart of a process for updating scores in response to events, in accordance with some embodiments. At block 802, an indication of an event may be received. In some embodiments, the indication of an event may be received as a new data object from an event data source (e.g., a field reports database, a calendar database, and/or the like). In some embodiments, the indication of the event may be received through another source. For example, an indication of a field report event may be received as field reports are received, instead of being retrieved from a field report data store.


At block 804, associations between the event and one or more scoring factors are identified. In some embodiments, the event may be associated with one or more different attributes (e.g., type of event, location of event, entities involved in the event, and/or the like) that may be associated with one or more scoring factors. In some embodiments, events of certain types may cause the weights of certain scoring factors to change. For example, an event of a certain type may cause the scores of entities associated with a location within a certain proximity of the event location to change (e.g., a natural disaster event may cause the need scores for entities associated with locations within a certain proximity to the event to rise).


At block 806, the one or more scoring factors are updated, based at least in part upon the identified associations. In some embodiments, weights associated with certain scoring factors may increase or decrease based upon attributes associated with the event. For example, as discussed above, a weight associated with an entity having certain attributes may be increased or decreased in response to the occurrence of the event, based upon a relationship or association with an entity associated with the event (e.g., a participant at the event), a location associated with the event, and/or the like. In some embodiments, the one or more scoring factors may comprise a combination of different scoring factors (e.g., scores may be increased for entities having a membership with certain organizations and associated with a location within a threshold distance of the location of the event, while scores for entities outside the threshold distance may remain unchanged, regardless of membership).


At block 808, scores for one or more entities are recalculated, based upon the one or more updated scoring factors. In addition, the updated scores may be propagated to a plurality of user devices, causing updates to one or more user interfaces to reflect the changed scores. For example, a user at a user device may receive an alert of the event and an updated entity listing based upon the event. As such, the user may receive updated scoring information in substantially real-time in response to events. By updating scores in substantially real-time in response to events, users can be kept up to date on the latest conditions, allowing them to allocate attention and resources accordingly. In some embodiments, entities having updated scores in the displayed listing may be highlighted or displayed with some other visual affect (e.g., up/down arrows indicating that a score for an entity has gone up or down), allowing the user to quickly determine which of the displayed entities were affected by the event.


In some embodiments, the user may also receive an indication of a basis for the updated score information. For example, an indication of a basis for an updated score due to an event may be displayed next to an updated score for a particular entity (e.g., an indication of the event, an indication of one or more scoring factors associated with the update, and/or the like). In some embodiments, the user may select a particular entity having an updated score due to the event, in order to view a display containing a basis for the updated score (e.g., a listing of the updated scoring factors used to update the score, a listing of the most relevant scoring factors used to update the score, and/or the like).


In some embodiments, scores for entities may continue to be updated in response to the occurrence of the event. For example, in some embodiments, the updated weights of certain factors may continue to be adjusted as time passes after the occurrence of the event. In some embodiments, changes to the weights of certain factors may be reverted after a threshold period of time has passed following the event.


Data Item-Centric Data Model


To provide a framework for the description of specific systems and methods provided above and below, an example database system 1210 using an ontology 1205 will now be described in reference to FIGS. 9A-9C. This description is provided for the purpose of providing an example and is not intended to limit the techniques to the example data model, the example database system, or the example database system's use of an ontology to represent information.


In one embodiment, a body of data is conceptually structured according to data item-centric data model represented by ontology 1205. The conceptual data model is independent of any particular database used for durably storing one or more database(s) 1209 based on the ontology 1205. For example, each object of the conceptual data model may correspond to one or more rows in a relational database or an entry in Lightweight Directory Access Protocol (LDAP) database, or any combination of one or more databases.



FIG. 9A illustrates data item-centric conceptual data model (which may also be referred to as an “object-centric conceptual data model”) according to an embodiment. An ontology 1205, as noted above, may include stored information providing a data model for storage of data in the database 1209. The ontology 1205 may be defined by one or more data item types (which may also be referred to as “object types”), which may each be associated with one or more property types. At the highest level of abstraction, data item 1201 (which may also be referred to as a “data object” or “object”) is a container for information representing things in the world. For example, data item 1201 can represent an entity such as a person, a place, an organization, a market instrument, or other noun. Data item 1201 can represent an event that happens at a point in time or for a duration. Data item 1201 can represent a document or other unstructured data source such as a file (for example, a malware file), an email message, a news report, or a written paper or article. Each data item 1201 is associated with a unique identifier that uniquely identifies the data item within the database system.


Different types of data items may have different property types. For example, a “file” data item (as described above) may have various property types as described above (for example, various hash property types, associated file property types, various external analysis property types, and/or the like), a “Person” data item may have an “Eye Color” property type, and an “Event” data item may have a “Date” property type. Each property 1203 as represented by data in the database system 1210 may have a property type defined by the ontology 1205 used by the database 1205. Properties of data items may, in an embodiment, themselves be data items and/or associated with data items. For example, file data items may be associated with various analysis information items, as described above. Analysis information items may comprise data items and/or properties associated with data items (for example, file data items).


Items may be instantiated in the database 1209 in accordance with the corresponding data item definition for the particular data item in the ontology 1205. For example, a specific monetary payment (e.g., an item of type “event”) of US$30.00 (e.g., a property of type “currency”) taking place on Mar. 27, 2009 (e.g., a property of type “date”) may be stored in the database 1209 as an event data item with associated currency and date properties as defined within the ontology 1205.


The data objects defined in the ontology 1205 may support property multiplicity. In particular, a data item 1201 may be allowed to have more than one property 1203 of the same property type. For example, a “Person” data item may have multiple “Address” properties or multiple “Name” properties.


Each link 1202 represents a connection between two data items 1201. In one embodiment, the connection is either through a relationship, an event, or through matching properties. A relationship connection may be asymmetrical or symmetrical. For example, “Person” data item A may be connected to “Person” data item B by a “Child Of” relationship (where “Person” data item B has an asymmetric “Parent Of” relationship to “Person” data item A), a “Kin Of” symmetric relationship to “Person” data item C, and an asymmetric “Member Of” relationship to “Organization” data item X. The type of relationship between two data items may vary depending on the types of the data items. For example, “Person” data item A may have an “Appears In” relationship with “Document” data item Y or have a “Participate In” relationship with “Event” data item E. As an example of an event connection, two “Person” data items may be connected by an “Airline Flight” data item representing a particular airline flight if they traveled together on that flight, or by a “Meeting” data item representing a particular meeting if they both attended that meeting. In one embodiment, when two data items are connected by an event, they are also connected by relationships, in which each data item has a specific relationship to the event, such as, for example, an “Appears In” relationship.


As an example of a matching properties connection, two “Person” data items representing a brother and a sister, may both have an “Address” property that indicates where they live. If the brother and the sister live in the same home, then their “Address” properties likely contain similar, if not identical property values. In one embodiment, a link between two data items may be established based on similar or matching properties (e.g., property types and/or property values) of the data items. These are just some examples of the types of connections that may be represented by a link and other types of connections may be represented; embodiments are not limited to any particular types of connections between data items. For example, a document might contain references to two different items. For example, a document may contain a reference to a payment (one item), and a person (a second item). A link between these two items may represent a connection between these two entities through their co-occurrence within the same document.


Each data item 1201 may have multiple links with another data item 1201 to form a link set 1204. For example, two “Person” data items representing a husband and a wife could be linked through a “Spouse Of” relationship, a matching “Address” property, and one or more matching “Event” properties (e.g., a wedding). Each link 1202 as represented by data in a database may have a link type defined by the database ontology used by the database.



FIG. 9B is a block diagram illustrating example components and data that may be used in identifying and storing data according to an ontology. In this example, the ontology may be configured, and data in the data model populated, by a system of parsers and ontology configuration tools. In the embodiment of FIG. 9B, input data 1300 is provided to parser 1302. The input data may comprise data from one or more sources. For example, an institution may have one or more databases with information on credit card transactions, rental cars, and people. The databases may contain a variety of related information and attributes about each type of data, such as a “date” for a credit card transaction, an address for a person, and a date for when a rental car is rented. The parser 1302 is able to read a variety of source input data types and determine which type of data it is reading.


In accordance with the discussion above, the example ontology 1205 comprises stored information providing the data model of data stored in database 1209, and the ontology is defined by one or more data item types 1310, one or more property types 1316, and one or more link types 1330. Based on information determined by the parser 1302 or other mapping of source input information to item type, one or more data items 1201 may be instantiated in the database 209 based on respective determined item types 1310, and each of the items 1201 has one or more properties 1203 that are instantiated based on property types 1316. Two data items 1201 may be connected by one or more links 1202 that may be instantiated based on link types 1330. The property types 1316 each may comprise one or more data types 1318, such as a string, number, etc. Property types 1316 may be instantiated based on a base property type 1320. For example, a base property type 1320 may be “Locations” and a property type 1316 may be “Home.”


In an embodiment, a user of the system uses an item type editor 1324 to create and/or modify the item types 1310 and define attributes of the item types. In an embodiment, a user of the system uses a property type editor 1326 to create and/or modify the property types 1316 and define attributes of the property types. In an embodiment, a user of the system uses link type editor 1328 to create the link types 1330. Alternatively, other programs, processes, or programmatic controls may be used to create link types and property types and define attributes, and using editors is not required.


In an embodiment, creating a property type 1316 using the property type editor 1326 involves defining at least one parser definition using a parser editor 1322. A parser definition comprises metadata that informs parser 1302 how to parse input data 1300 to determine whether values in the input data can be assigned to the property type 1316 that is associated with the parser definition. In an embodiment, each parser definition may comprise a regular expression parser 1304A or a code module parser 1304B. In other embodiments, other kinds of parser definitions may be provided using scripts or other programmatic elements. Once defined, both a regular expression parser 1304A and a code module parser 1304B can provide input to parser 1302 to control parsing of input data 1300.


Using the data types defined in the ontology, input data 1300 may be parsed by the parser 1302 determine which item type 1310 should receive data from a record created from the input data, and which property types 1316 should be assigned to data from individual field values in the input data. Based on the item/object-property mapping 1301, the parser 1302 selects one of the parser definitions that is associated with a property type in the input data. The parser parses an input data field using the selected parser definition, resulting in creating new or modified data 1303. The new or modified data 1303 is added to the database 1209 according to ontology 205 by storing values of the new or modified data in a property of the specified property type. As a result, input data 1300 having varying format or syntax can be created in database 1209. The ontology 1205 may be modified at any time using item/object type editor 1324, property type editor 1326, and link type editor 1328, or under program control without human use of an editor. Parser editor 1322 enables creating multiple parser definitions that can successfully parse input data 1300 having varying format or syntax and determine which property types should be used to transform input data 300 into new or modified input data 1303.


The properties, data items, and links (e.g. relationships) between the data items can be visualized using a graphical user interface (“GUI”). For example, FIG. 9C displays a user interface showing a graph representation 1403 of relationships (including relationships and/or links 1404, 1405, 1406, 1407, 1408, 1409, 1410, 1411, 1412, and 1413) between the data items (including data items 1421, 1422, 1423, 1424, 1425, 1426, 1427, 1428, and 1429) that are represented as nodes in the example of FIG. 9C. In this embodiment, the data items include person data items 1421, 1422, 1423, 1424, 1425, and 1426; a flight item 1427; a financial account 1428; and a computer data item 1429. In this example, each person node (associated with person data items), flight node (associated with flight data items), financial account node (associated with financial account data items), and computer node (associated with computer data items) may have relationships and/or links with any of the other nodes through, for example, other data items such as payment data items.


For example, in FIG. 9C, relationship 1404 is based on a payment associated with the individuals indicated in person data items 1421 and 1423. The link 1404 represents these shared payments (for example, the individual associated with data item 1421 may have paid the individual associated with data item 1423 on three occasions). The relationship is further indicated by the common relationship between person data items 1421 and 1423 and financial account data item 1428. For example, link 1411 indicates that person data item 1421 transferred money into financial account data item 1428, while person data item 1423 transferred money out of financial account data item 1428. In another example, the relationships between person data items 1424 and 1425 and flight data item 1427 are indicated by links 1406, 1409, and 1410. In this example, person data items 1424 and 1425 have a common address and were passengers on the same flight data item 1427. In an embodiment, further details related to the relationships between the various items may be displayed. For example, links 1411 and 1412 may, in some embodiments, indicate the timing of the respective money transfers. In another example, the time of the flight associated with the flight data item 1427 may be shown.


Relationships between data items may be stored as links, or in some embodiments, as properties, where a relationship may be detected between the properties. In some cases, as stated above, the links may be directional. For example, a payment link may have a direction associated with the payment, where one person data item is a receiver of a payment, and another person data item is the payer of payment.


In various embodiments, data items may further include malware analysis metadata and/or links. Such malware analysis metadata may be accessed by the data analysis system for displaying objects and features on the user interface (as described above).


In addition to visually showing relationships between the data items, the user interface may allow various other manipulations. For example, the data items within database 1108 may be searched using a search interface 1450 (e.g., text string matching of data item properties), inspected (e.g., properties and associated data viewed), filtered (e.g., narrowing the universe of data items into sets and subsets by properties or relationships), and statistically aggregated (e.g., numerically summarized based on summarization criteria), among other operations and visualizations.


Implementation Mechanisms


According to an embodiment, the data analysis system and other methods and techniques described herein are implemented by one or more special-purpose computing devices. For example, the server computing system 110 and/or client computer system 130, as illustrated in FIG. 1, may be implemented as special-purpose computing devices. The special-purpose computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the techniques, or may include one or more general purpose hardware processors programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the techniques. The special-purpose computing devices may be desktop computer systems, server computer systems, portable computer systems, handheld devices, networking devices or any other device or combination of devices that incorporate hard-wired and/or program logic to implement the techniques.


For example, FIG. 10 is a block diagram that illustrates a computer system 1000 upon which the various systems and methods discussed herein may be implemented. For example, in various embodiments, any of the computing systems illustrated in FIG. 1, such client computer system 130 and/or server computing system 110, may be implemented as computer systems 1000 and/or servers 1030 as illustrated in FIG. 10.


Computer system 1000 includes a bus 1002 or other communication mechanism for communicating information, and a hardware processor, or multiple processors, 1004 coupled with bus 1002 for processing information. Hardware processor(s) 1004 may be, for example, one or more general purpose microprocessors.


Computer system 1000 also includes a main memory 1006, such as a random access memory (RAM), cache and/or other dynamic storage devices (or computer readable storage medium), coupled to bus 1002 for storing information and instructions to be executed by processor 1004. Main memory 1006 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 1004. Such instructions, when stored in storage media (e.g., computer readable storage medium) accessible to processor 1004, render computer system 1000 into a special-purpose machine that is customized to perform the operations specified in the instructions.


Computer system 1000 further includes a read only memory (ROM) 10010 or other static storage device (e.g., computer readable storage medium) coupled to bus 1002 for storing static information and instructions for processor 1004. A storage device 1010, such as a magnetic disk, optical disk, or USB thumb drive (Flash drive), and/or any other suitable data store, is provided and coupled to bus 1002 for storing information (for example, file data items, analysis information data items, submission data items, and/or the like) and instructions.


Computer system 1000 may be coupled via bus 1002 to a display 1012, such as a cathode ray tube (CRT), LCD display, or touch screen display, for displaying information to a computer user and/or receiving input from the user. An input device 1014, including alphanumeric and other keys, is coupled to bus 1002 for communicating information and command selections to processor 1004. Another type of user input device is cursor control 1016, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 1004 and for controlling cursor movement on display 1012. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane. In some embodiments, the same direction information and command selections as cursor control may be implemented via receiving touches on a touch screen without a cursor.


Computing system 1000 may include a user interface module (e.g., implemented by workflow engine 124 as illustrated in FIG. 1), and/or various other types of modules to implement one or more graphical user interface of the data analysis system, as described above. The modules may be stored in a mass storage device (e.g., a computer readable storage medium) as executable software codes that are executed by the computing device(s).


As mentioned, computer system 1000 may implement the techniques described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the computer system causes or programs computer system 1000 to be a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 1000 in response to processor(s) 1004 executing one or more sequences of one or more modules and/or instructions contained in main memory 1006. Such instructions may be read into main memory 1006 from another storage medium, such as storage device 1010. Execution of the sequences of instructions contained in main memory 1006 causes processor(s) 1004 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.


Various forms of media (e.g., computer readable storage mediums) may be involved in carrying one or more sequences of one or more instructions to processor 1004 for execution. For example, the instructions may initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer can load the instructions and/or modules into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 1000 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 1002. Bus 1002 carries the data to main memory 1006, from which processor 1004 retrieves and executes the instructions. The instructions received by main memory 1006 may optionally be stored on storage device 1010 either before or after execution by processor 1004.


Computer system 1000 also includes a communication interface 10110 coupled to bus 1002. Communication interface 1018 provides a two-way data communication coupling to a network link 1020 that is connected to a local network 1022. For example, communication interface 1018 may be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 1018 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN (or WAN component to communicated with a WAN). Wireless links may also be implemented. In any such implementation, communication interface 1018 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.


Network link 1020 typically provides data communication through one or more networks to other data devices. For example, network link 1020 may provide a connection through local network 1022 to a host computer 1024 or to data equipment operated by an Internet Service Provider (ISP) 1026. ISP 1026 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 1028. Local network 1022 and Internet 1028 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 1020 and through communication interface 1018, which carry the digital data to and from computer system 1000, are example forms of transmission media. In some embodiments, network 150 (as illustrated in FIG. 1) may correspond to a local network 1022 and/or Internet 1028.


Computer system 1000 can send messages and receive data, including program code, through the network(s), network link 1020 and communication interface 1018. In the Internet example, a server 1030 might transmit a requested code for an application program through Internet 1028, ISP 1026, local network 1022 and communication interface 1018. For example, in an embodiment various aspects of the data analysis system may be implemented on one or more of the servers 1030 and may be transmitted to and from the computer system 1000. For example, submitted malware data items may be transmitted to one of the servers 1030, aspects of the basic analysis may be implemented on one or more of the servers 1030, and/or aspects of the external analysis may be implemented on one or more of the servers 1030. In an example, requests for external analyses of file data items may be transmitted to one or more third-party servers 1030 (from, for example, the computer system 1000 and/or another server 1030 of the system), and analysis data may then be transmitted back from third-party servers 1030.


In an embodiment, the data analysis system may be accessible by the user through a web-based viewer, such as a web browser. In this embodiment, the user interfaces of the system may be generated by a server (such as one of the servers 1030) and/or the computer system 1000 and transmitted to the web browser of the user. The user may then interact with the user interfaces through the web-browser. In an embodiment, the computer system 1000 may comprise a mobile electronic device, such as a cell phone, smartphone, and/or tablet. The system may be accessible by the user through such a mobile electronic device, among other types of electronic devices.


Additional Implementation Details and Embodiments


Various embodiments of the present disclosure may be a system, a method, and/or a computer program product at any possible technical detail level of integration. The computer program product may include a computer readable storage medium (or mediums) having computer readable program instructions thereon for causing a processor to carry out aspects of the present disclosure.


For example, the functionality described herein may be performed as software instructions are executed by, and/or in response to software instructions being executed by, one or more hardware processors and/or any other suitable computing devices. The software instructions and/or other executable code may be read from a computer readable storage medium (or mediums).


The computer readable storage medium can be a tangible device that can retain and store data and/or instructions for use by an instruction execution device. The computer readable storage medium may be, for example, but is not limited to, an electronic storage device (including any volatile and/or non-volatile electronic storage devices), a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. A non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a solid state drive, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing. A computer readable storage medium, as used herein, is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.


Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network. The network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.


Computer readable program instructions (as also referred to herein as, for example, “code,” “instructions,” “module,” “application,” “software application,” and/or the like) for carrying out operations of the present disclosure may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, configuration data for integrated circuitry, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++, or the like, and procedural programming languages, such as the “C” programming language or similar programming languages. Computer readable program instructions may be callable from other instructions or from itself, and/or may be invoked in response to detected events or interrupts. Computer readable program instructions configured for execution on computing devices may be provided on a computer readable storage medium, and/or as a digital download (and may be originally stored in a compressed or installable format that requires installation, decompression or decryption prior to execution) that may then be stored on a computer readable storage medium. Such computer readable program instructions may be stored, partially or fully, on a memory device (e.g., a computer readable storage medium) of the executing computing device, for execution by the computing device. The computer readable program instructions may execute entirely on a user's computer (e.g., the executing computing device), partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). In some embodiments, electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present disclosure.


Aspects of the present disclosure are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the disclosure. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.


These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart(s) and/or block diagram(s) block or blocks.


The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks. For example, the instructions may initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer may load the instructions and/or modules into its dynamic memory and send the instructions over a telephone, cable, or optical line using a modem. A modem local to a server computing system may receive the data on the telephone/cable/optical line and use a converter device including the appropriate circuitry to place the data on a bus. The bus may carry the data to a memory, from which a processor may retrieve and execute the instructions. The instructions received by the memory may optionally be stored on a storage device (e.g., a solid state drive) either before or after execution by the computer processor.


The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the blocks may occur out of the order noted in the Figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. In addition, certain blocks may be omitted in some implementations. The methods and processes described herein are also not limited to any particular sequence, and the blocks or states relating thereto can be performed in other sequences that are appropriate.


It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions. For example, any of the processes, methods, algorithms, elements, blocks, applications, or other functionality (or portions of functionality) described in the preceding sections may be embodied in, and/or fully or partially automated via, electronic hardware such application-specific processors (e.g., application-specific integrated circuits (ASICs)), programmable processors (e.g., field programmable gate arrays (FPGAs)), application-specific circuitry, and/or the like (any of which may also combine custom hard-wired logic, logic circuits, ASICs, FPGAs, etc. with custom programming/execution of software instructions to accomplish the techniques).


Any of the above-mentioned processors, and/or devices incorporating any of the above-mentioned processors, may be referred to herein as, for example, “computers,” “computer devices,” “computing devices,” “hardware computing devices,” “hardware processors,” “processing units,” and/or the like. Computing devices of the above-embodiments may generally (but not necessarily) be controlled and/or coordinated by operating system software, such as Mac OS, iOS, Android, Chrome OS, Windows OS (e.g., Windows XP, Windows Vista, Windows 7, Windows 8, Windows 10, Windows Server, etc.), Windows CE, Unix, Linux, SunOS, Solaris, Blackberry OS, VxWorks, or other suitable operating systems. In other embodiments, the computing devices may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface functionality, such as a graphical user interface (“GUI”), among other things.


As described above, in various embodiments certain functionality may be accessible by a user through a web-based viewer (such as a web browser), or other suitable software program). In such implementations, the user interface may be generated by a server computing system and transmitted to a web browser of the user (e.g., running on the user's computing system). Alternatively, data (e.g., user interface data) necessary for generating the user interface may be provided by the server computing system to the browser, where the user interface may be generated (e.g., the user interface data may be executed by a browser accessing a web service and may be configured to render the user interfaces based on the user interface data). The user may then interact with the user interface through the web-browser. User interfaces of certain implementations may be accessible through one or more dedicated software applications. In certain embodiments, one or more of the computing devices and/or systems of the disclosure may include mobile computing devices, and user interfaces may be accessible through such mobile computing devices (for example, smartphones and/or tablets).


Many variations and modifications may be made to the above-described embodiments, the elements of which are to be understood as being among other acceptable examples. All such modifications and variations are intended to be included herein within the scope of this disclosure. The foregoing description details certain embodiments. It will be appreciated, however, that no matter how detailed the foregoing appears in text, the systems and methods can be practiced in many ways. As is also stated above, it should be noted that the use of particular terminology when describing certain features or aspects of the systems and methods should not be taken to imply that the terminology is being re-defined herein to be restricted to including any specific characteristics of the features or aspects of the systems and methods with which that terminology is associated.


Conditional language, such as, among others, “can,” “could,” “might,” or “may,” unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements, and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.


The term “substantially” when used in conjunction with the term “real-time” forms a phrase that will be readily understood by a person of ordinary skill in the art. For example, it is readily understood that such language will include speeds in which no or little delay or waiting is discernible, or where such delay is sufficiently short so as not to be disruptive, irritating, or otherwise vexing to user.


Conjunctive language such as the phrase “at least one of X, Y, and Z,” or “at least one of X, Y, or Z,” unless specifically stated otherwise, is to be understood with the context as used in general to convey that an item, term, etc. may be either X, Y, or Z, or a combination thereof. For example, the term “or” is used in its inclusive sense (and not in its exclusive sense) so that when used, for example, to connect a list of elements, the term “or” means one, some, or all of the elements in the list. Thus, such conjunctive language is not generally intended to imply that certain embodiments require at least one of X, at least one of Y, and at least one of Z to each be present.


The term “a” as used herein should be given an inclusive rather than exclusive interpretation. For example, unless specifically noted, the term “a” should not be understood to mean “exactly one” or “one and only one”; instead, the term “a” means “one or more” or “at least one,” whether used in the claims or elsewhere in the specification and regardless of uses of quantifiers such as “at least one,” “one or more,” or “a plurality” elsewhere in the claims or specification.


The term “comprising” as used herein should be given an inclusive rather than exclusive interpretation. For example, a general purpose computer comprising one or more processors should not be interpreted as excluding other computer components, and may possibly include such components as memory, input/output devices, and/or network interfaces, among others.


While the above detailed description has shown, described, and pointed out novel features as applied to various embodiments, it may be understood that various omissions, substitutions, and changes in the form and details of the devices or processes illustrated may be made without departing from the spirit of the disclosure. As may be recognized, certain embodiments of the inventions described herein may be embodied within a form that does not provide all of the features and benefits set forth herein, as some features may be used or practiced separately from others. The scope of certain inventions disclosed herein is indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.

Claims
  • 1. A computer system, comprising: one or more computer readable storage mediums configured to store computer executable instructions; andone or more computer processors in communication with the one or more computer readable storage mediums and configured to execute the computer executable instructions in order to cause the computer system to:retrieve a plurality of data objects from a plurality of data stores, the data stores comprising at least a first data store containing data objects corresponding to entities, and a second data store containing data objects corresponding to events; andcreate aggregated data objects for one or more entities, wherein creating an aggregated data object for an entity comprises: determining a confidence value for an association between one or more of the retrieved data objects the entity based at least in part upon one or more attribute types and attribute values associated with the one or more data objects;generating metadata associating the one or more data objects with the entity based at least in part on the confidence value, wherein the metadata is stored separately from the one or more data objects, such that the one or more data objects are unchanged; anddetermining one or more additional attributes for the entity, based upon the data objects associated with the entity.
  • 2. The computer system of claim 1, wherein the one or more computer processors are further configured to: identify at least a portion of the attributes associated with the one or more entities as scoring factors;generate one or more scores for the one or more entities, based at least in part upon one or more values for the scoring factors for the entities determined using the data objects and additional attributes associated with the entities; andgenerate user interface information for displaying at least a portion of the entities, based at least in part upon the generated scores for the entities.
  • 3. The computer system of claim 2, wherein generating scores for the one or more entities comprises, for an entity of the one or more entities: processing the one or more scoring factors associated with the entity using a scoring model; andgenerating a score for the entity using the scoring model.
  • 4. The computer system of claim 3, wherein the scoring model comprises one or more weights associated with the one or more scoring factors.
  • 5. The computer system of claim 2, wherein the user interface information for displaying at least a portion of the entities is further based at least in part upon one or more user attributes.
  • 6. The computer system of claim 5, wherein the one or more user attributes comprises a user location.
  • 7. The computer system of claim 1, wherein the one or more additional attributes comprises a history attribute generated based at least in part upon one or more event data objects associated with the entity.
  • 8. The computer system of claim 1, wherein the data objects are associated with the entity if the confidence value exceeds a threshold confidence.
  • 9. The computer system of claim 1, wherein the one or more computer processors are further configured to: receive an indication of an event;associate the event with one or more scoring factors;in substantially real-time, update the one or more scoring factors based at least in part upon the event;re-generate scores for at least some of the one or more entities, based upon the updated scoring factors; andgenerate user interface information for displaying at least a portion of the entities, based at least in part upon the re-generated scores.
  • 10. The computer system of claim 9, wherein the one or more scoring factors are updated based upon a relationship between an entity attribute and an attribute value associated with the event.
  • 11. The computer system of claim 9, wherein the user interface information further comprises information for displaying a basis for a change in a score of a displayed entity, based upon the updated scoring factors.
  • 12. A computer-implemented method for performing probabilistic entity resolution, comprising: retrieving a plurality of data objects from a plurality of data stores, the data stores comprising at least a first data store containing data objects corresponding to entities, and a second data store containing data objects corresponding to events; andcreating aggregated data objects for one or more entities, wherein creating an aggregated data object for an entity comprises: determining a confidence value for an association between one or more of the retrieved data objects and the entity based at least in part upon one or more attribute types and attribute values associated with the one or more data objects;generating metadata associating the one or more data objects with the entity based at least in part on the confidence value, wherein the metadata is stored separately from the one or more data objects, such that the one or more data objects are unchanged; anddetermining one or more additional attributes for the entity, based upon the data objects associated with the entity.
  • 13. The computer-implemented method of claim 12, further comprising: identifying at least a portion of the attributes associated with the one or more entities as scoring factors;generating one or more scores for the one or more entities, based at least in part upon one or more values for the scoring factors for the entities determined using the data objects and additional attributes associated with the entities; andgenerating user interface information for displaying at least a portion of the entities, based at least in part upon the generated scores for the entities.
  • 14. The computer-implemented method of claim 13, wherein generating scores for the one or more entities comprises, for an entity of the one or more entities: processing the one or more scoring factors associated with the entity using a scoring model; andgenerating a score for the entity using the scoring model.
  • 15. The computer-implemented method of claim 13, wherein the user interface information for displaying at least a portion of the entities is further based at least in part upon one or more user attributes.
  • 16. The computer-implemented method of claim 12, wherein the one or more additional attributes comprises a history attribute generated based at least in part upon one or more event data objects associated with the entity.
  • 17. The computer-implemented method of claim 12, wherein the data objects are associated with the entity if the confidence value exceeds a threshold confidence.
  • 18. The computer-implemented method of claim 12, further comprising: receiving an indication of an event;associating the event with one or more scoring factors;in substantially real-time, updating the one or more scoring factors based at least in part upon the event;re-generating scores for at least some of the one or more entities, based upon the updated scoring factors; andgenerating user interface information for displaying at least a portion of the entities, based at least in part upon the re-generated scores.
  • 19. The computer-implemented method of claim 18, wherein the one or more scoring factors are updated based upon a relationship between an entity attribute and an attribute value associated with the event.
  • 20. The computer-implemented method of claim 18, wherein the user interface information further comprises information for displaying a basis for a change in a score of a displayed entity, based upon the updated scoring factors.
US Referenced Citations (559)
Number Name Date Kind
4899161 Morin et al. Feb 1990 A
4958305 Piazza Sep 1990 A
5109399 Thompson Apr 1992 A
5329108 Lamoure Jul 1994 A
5632009 Rao et al. May 1997 A
5670987 Doi et al. Sep 1997 A
5754182 Kobayashi May 1998 A
5781195 Marvin Jul 1998 A
5781704 Rossmo Jul 1998 A
5798769 Chiu et al. Aug 1998 A
5845300 Comer Dec 1998 A
6057757 Arrowsmith et al. May 2000 A
6091956 Hollenberg Jul 2000 A
6157747 Szeliski et al. Dec 2000 A
6161098 Wallman Dec 2000 A
6173067 Payton et al. Jan 2001 B1
6178432 Cook et al. Jan 2001 B1
6219053 Tachibana et al. Apr 2001 B1
6232971 Haynes May 2001 B1
6247019 Davies Jun 2001 B1
6279018 Kudrolli et al. Aug 2001 B1
6341310 Leshem et al. Jan 2002 B1
6366933 Ball et al. Apr 2002 B1
6369835 Lin Apr 2002 B1
6389289 Voce et al. May 2002 B1
6414683 Gueziec Jul 2002 B1
6456997 Shukla Sep 2002 B1
6483509 Rabenhorst Nov 2002 B1
6529900 Patterson et al. Mar 2003 B1
6549944 Weinberg et al. Apr 2003 B1
6560620 Ching May 2003 B1
6581068 Bensoussan et al. Jun 2003 B1
6594672 Lampson et al. Jul 2003 B1
6631496 Li et al. Oct 2003 B1
6642945 Sharpe Nov 2003 B1
6662103 Skolnick et al. Dec 2003 B1
6674434 Chojnacki et al. Jan 2004 B1
6714936 Nevin, III Mar 2004 B1
6757445 Knopp Jun 2004 B1
6775675 Nwabueze et al. Aug 2004 B1
6828920 Owen et al. Dec 2004 B2
6839745 Dingari et al. Jan 2005 B1
6877137 Rivette et al. Apr 2005 B1
6976210 Silva et al. Dec 2005 B1
6980984 Huffman et al. Dec 2005 B1
6985950 Hanson et al. Jan 2006 B1
7036085 Barros Apr 2006 B2
7043702 Chi et al. May 2006 B2
7055110 Kupka et al. May 2006 B2
7139800 Bellotti et al. Nov 2006 B2
7158878 Rasmussen et al. Jan 2007 B2
7162475 Ackerman Jan 2007 B2
7168039 Bertram Jan 2007 B2
7171427 Witowski et al. Jan 2007 B2
7269786 Malloy et al. Sep 2007 B1
7278105 Kitts Oct 2007 B1
7290698 Poslinski et al. Nov 2007 B2
7333998 Heckerman et al. Feb 2008 B2
7370047 Gorman May 2008 B2
7375732 Arcas May 2008 B2
7379811 Rasmussen et al. May 2008 B2
7379903 Caballero et al. May 2008 B2
7426654 Adams et al. Sep 2008 B2
7454466 Bellotti et al. Nov 2008 B2
7457706 Malero et al. Nov 2008 B2
7467375 Tondreau et al. Dec 2008 B2
7487139 Fraleigh et al. Feb 2009 B2
7502786 Liu et al. Mar 2009 B2
7519470 Brasche et al. Apr 2009 B2
7525422 Bishop et al. Apr 2009 B2
7529195 Gorman May 2009 B2
7529727 Arning et al. May 2009 B2
7529734 Dirisala May 2009 B2
7539666 Ashworth et al. May 2009 B2
7546245 Surpin et al. Jun 2009 B2
7558677 Jones Jun 2009 B2
7574409 Patinkin Aug 2009 B2
7574428 Leiserowitz et al. Aug 2009 B2
7579965 Bucholz Aug 2009 B2
7596285 Brown et al. Sep 2009 B2
7614006 Molander Nov 2009 B2
7617232 Gabbert et al. Nov 2009 B2
7620628 Kapur et al. Nov 2009 B2
7627812 Chamberlain et al. Dec 2009 B2
7634717 Chamberlain et al. Dec 2009 B2
7640173 Surpin et al. Dec 2009 B2
7663621 Allen et al. Feb 2010 B1
7703021 Flam Apr 2010 B1
7706817 Bamrah et al. Apr 2010 B2
7712049 Williams et al. May 2010 B2
7716067 Surpin et al. May 2010 B2
7716077 Mikurak May 2010 B1
7725530 Sah et al. May 2010 B2
7725547 Albertson et al. May 2010 B2
7730082 Sah et al. Jun 2010 B2
7730109 Rohrs et al. Jun 2010 B2
7770100 Chamberlain et al. Aug 2010 B2
7791616 Ioup et al. Sep 2010 B2
7805457 Viola et al. Sep 2010 B1
7809703 Balabhadrapatruni et al. Oct 2010 B2
7818658 Chen Oct 2010 B2
7870493 Pall et al. Jan 2011 B2
7872647 Mayer et al. Jan 2011 B2
7894984 Rasmussen et al. Feb 2011 B2
7899611 Downs et al. Mar 2011 B2
7917376 Bellin et al. Mar 2011 B2
7920963 Jouline et al. Apr 2011 B2
7933862 Chamberlain et al. Apr 2011 B2
7945852 Pilskains May 2011 B1
7962281 Rasmussen et al. Jun 2011 B2
7962495 Jain et al. Jun 2011 B2
7962848 Bertram Jun 2011 B2
7970240 Chao et al. Jun 2011 B1
7971150 Raskutti et al. Jun 2011 B2
7984374 Caro et al. Jun 2011 B2
8001465 Kudrolli et al. Aug 2011 B2
8001482 Bhattiprolu et al. Aug 2011 B2
8010545 Stefik et al. Aug 2011 B2
8010886 Gusmorino et al. Aug 2011 B2
8015487 Roy et al. Sep 2011 B2
8019709 Norton et al. Sep 2011 B2
8024778 Cash et al. Sep 2011 B2
8036632 Cona et al. Oct 2011 B1
8065080 Koch Nov 2011 B2
8082172 Chao et al. Dec 2011 B2
8085268 Carrino et al. Dec 2011 B2
8103543 Zwicky Jan 2012 B1
8134457 Velipasalar et al. Mar 2012 B2
8145703 Frishert et al. Mar 2012 B2
8185819 Sah et al. May 2012 B2
8214361 Sandler et al. Jul 2012 B1
8214764 Gemmell et al. Jul 2012 B2
8225201 Michael Jul 2012 B2
8229947 Fujinaga Jul 2012 B2
8230333 Decherd et al. Jul 2012 B2
8271461 Pike et al. Sep 2012 B2
8280880 Aymeloglu et al. Oct 2012 B1
8290926 Ozzie et al. Oct 2012 B2
8290942 Jones et al. Oct 2012 B2
8301464 Cave et al. Oct 2012 B1
8301904 Gryaznov Oct 2012 B1
8312367 Foster Nov 2012 B2
8312546 Alme Nov 2012 B2
8325178 Doyle, Jr. Dec 2012 B1
8352881 Champion et al. Jan 2013 B2
8368695 Howell et al. Feb 2013 B2
8397171 Klassen et al. Mar 2013 B2
8400448 Doyle, Jr. Mar 2013 B1
8407180 Ramesh et al. Mar 2013 B1
8411046 Kruzeniski et al. Apr 2013 B2
8412234 Gatmir-Motahari et al. Apr 2013 B1
8412707 Mianji Apr 2013 B1
8422825 Neophytou et al. Apr 2013 B1
8447722 Ahuja et al. May 2013 B1
8452790 Mianji May 2013 B1
8463036 Ramesh et al. Jun 2013 B1
8489331 Kopf et al. Jul 2013 B2
8489641 Seefeld et al. Jul 2013 B1
8498984 Hwang et al. Jul 2013 B1
8508533 Cervelli et al. Aug 2013 B2
8510743 Hackborn et al. Aug 2013 B2
8514082 Cova et al. Aug 2013 B2
8514229 Cervelli et al. Aug 2013 B2
8515207 Chau Aug 2013 B2
8554579 Tribble et al. Oct 2013 B2
8554653 Falkenborg et al. Oct 2013 B2
8554709 Goodson et al. Oct 2013 B2
8554840 Milgramm Oct 2013 B1
8564596 Carrino et al. Oct 2013 B2
8577911 Stepinski et al. Nov 2013 B1
8589273 Creeden et al. Nov 2013 B2
8595234 Siripuapu et al. Nov 2013 B2
8620641 Farnsworth et al. Dec 2013 B2
8639757 Zang et al. Jan 2014 B1
8646080 Williamson et al. Feb 2014 B2
8676597 Buehler et al. Mar 2014 B2
8676857 Adams et al. Mar 2014 B1
8689108 Duffield et al. Apr 2014 B1
8707185 Robinson et al. Apr 2014 B2
8713467 Goldenberg et al. Apr 2014 B1
8726379 Stiansen et al. May 2014 B1
8739278 Varghese May 2014 B2
8742934 Sarpy et al. Jun 2014 B1
8745516 Mason et al. Jun 2014 B2
8756244 Dassa et al. Jun 2014 B2
8768009 Smith Jul 2014 B1
8781169 Jackson et al. Jul 2014 B2
8787939 Papakipos et al. Jul 2014 B2
8799799 Cervelli et al. Aug 2014 B1
8799812 Parker Aug 2014 B2
8812960 Sun et al. Aug 2014 B1
8830322 Nerayoff et al. Sep 2014 B2
8832594 Thompson et al. Sep 2014 B1
8868486 Tamayo Oct 2014 B2
8868537 Colgrove et al. Oct 2014 B1
8917274 Ma et al. Dec 2014 B2
8924872 Bogomolov et al. Dec 2014 B1
8937619 Sharma et al. Jan 2015 B2
8938686 Erenrich et al. Jan 2015 B1
9009171 Grossman et al. Apr 2015 B1
9009827 Albertson et al. Apr 2015 B1
9021260 Falk et al. Apr 2015 B1
9021384 Beard et al. Apr 2015 B1
9043696 Meiklejohn et al. May 2015 B1
9043894 Dennison et al. May 2015 B1
10042524 Bogomolov et al. Aug 2018 B2
20010021936 Bertram Sep 2001 A1
20020003539 Abe Jan 2002 A1
20020033848 Sciammarella et al. Mar 2002 A1
20020065708 Senay et al. May 2002 A1
20020091707 Keller Jul 2002 A1
20020095658 Shulman Jul 2002 A1
20020116120 Ruiz et al. Aug 2002 A1
20020130867 Yang et al. Sep 2002 A1
20020130907 Chi et al. Sep 2002 A1
20020174201 Ramer et al. Nov 2002 A1
20020194119 Wright et al. Dec 2002 A1
20030028560 Kudrolli et al. Feb 2003 A1
20030039948 Donahue Feb 2003 A1
20030052896 Higgins et al. Mar 2003 A1
20030061200 Hubert Mar 2003 A1
20030103049 Kindratenko et al. Jun 2003 A1
20030140106 Raguseo Jul 2003 A1
20030144868 MacIntyre et al. Jul 2003 A1
20030163352 Surpin et al. Aug 2003 A1
20030200217 Ackerman Oct 2003 A1
20030225755 Iwayama et al. Dec 2003 A1
20030229848 Arend et al. Dec 2003 A1
20040030492 Fox et al. Feb 2004 A1
20040032432 Baynger Feb 2004 A1
20040039498 Ollis et al. Feb 2004 A1
20040064256 Barinek et al. Apr 2004 A1
20040085318 Hassler et al. May 2004 A1
20040095349 Bito et al. May 2004 A1
20040098236 Mayer et al. May 2004 A1
20040111410 Burgoon et al. Jun 2004 A1
20040126840 Cheng et al. Jul 2004 A1
20040143602 Ruiz et al. Jul 2004 A1
20040143796 Lerner et al. Jul 2004 A1
20040163039 McPherson et al. Aug 2004 A1
20040181554 Heckerman et al. Sep 2004 A1
20040193600 Kaasten et al. Sep 2004 A1
20040194549 Noel Oct 2004 A1
20040210847 Berson et al. Oct 2004 A1
20040221223 Yu et al. Nov 2004 A1
20040260702 Cragun et al. Dec 2004 A1
20040267746 Marcjan et al. Dec 2004 A1
20050027705 Sadri et al. Feb 2005 A1
20050028094 Allyn Feb 2005 A1
20050031197 Knopp Feb 2005 A1
20050034062 Bufkin et al. Feb 2005 A1
20050039119 Parks et al. Feb 2005 A1
20050065811 Chu et al. Mar 2005 A1
20050080769 Gemmell Apr 2005 A1
20050086207 Heuer et al. Apr 2005 A1
20050125715 Franco et al. Jun 2005 A1
20050162523 Darrell et al. Jul 2005 A1
20050166144 Gross Jul 2005 A1
20050180330 Shapiro Aug 2005 A1
20050182502 Iyengar Aug 2005 A1
20050182793 Keenan et al. Aug 2005 A1
20050183005 Denoue et al. Aug 2005 A1
20050210409 Jou Sep 2005 A1
20050246327 Yeung et al. Nov 2005 A1
20050251786 Citron et al. Nov 2005 A1
20050267652 Allstadt et al. Dec 2005 A1
20060026120 Carolan et al. Feb 2006 A1
20060026170 Kreitler et al. Feb 2006 A1
20060045470 Poslinski et al. Mar 2006 A1
20060059139 Robinson Mar 2006 A1
20060074866 Chamberlain et al. Apr 2006 A1
20060074881 Vembu et al. Apr 2006 A1
20060080619 Carlson et al. Apr 2006 A1
20060129746 Porter Jun 2006 A1
20060139375 Rasmussen et al. Jun 2006 A1
20060142949 Helt Jun 2006 A1
20060146050 Yamauchi Jul 2006 A1
20060149596 Surpin et al. Jul 2006 A1
20060203337 White Sep 2006 A1
20060218637 Thomas et al. Sep 2006 A1
20060235831 Adinolfi Oct 2006 A1
20060241974 Chao et al. Oct 2006 A1
20060242040 Rader et al. Oct 2006 A1
20060242630 Koike et al. Oct 2006 A1
20060251307 Florin et al. Nov 2006 A1
20060271277 Hu et al. Nov 2006 A1
20060279630 Aggarwal et al. Dec 2006 A1
20070011150 Frank Jan 2007 A1
20070016363 Huang et al. Jan 2007 A1
20070024620 Muller-Fischer et al. Feb 2007 A1
20070038962 Fuchs et al. Feb 2007 A1
20070057966 Ohno et al. Mar 2007 A1
20070078832 Ott et al. Apr 2007 A1
20070083541 Fraleigh et al. Apr 2007 A1
20070094389 Nussey et al. Apr 2007 A1
20070150369 Zivin Jun 2007 A1
20070174760 Chamberlain et al. Jul 2007 A1
20070188516 Loup et al. Aug 2007 A1
20070192265 Chopin et al. Aug 2007 A1
20070198571 Ferguson et al. Aug 2007 A1
20070208497 Downs et al. Sep 2007 A1
20070208498 Barker et al. Sep 2007 A1
20070208681 Bucholz Sep 2007 A1
20070208736 Tanigawa et al. Sep 2007 A1
20070240062 Christena et al. Oct 2007 A1
20070258642 Thota Nov 2007 A1
20070266336 Nojima et al. Nov 2007 A1
20070294643 Kyle Dec 2007 A1
20080010605 Frank Jan 2008 A1
20080016216 Worley et al. Jan 2008 A1
20080040684 Crump Feb 2008 A1
20080051989 Welsh Feb 2008 A1
20080052142 Bailey et al. Feb 2008 A1
20080077597 Butler Mar 2008 A1
20080077642 Carbone et al. Mar 2008 A1
20080082578 Hogue et al. Apr 2008 A1
20080098085 Krane et al. Apr 2008 A1
20080104019 Nath May 2008 A1
20080126951 Sood et al. May 2008 A1
20080140684 O'Reilly Jun 2008 A1
20080155440 Trevor et al. Jun 2008 A1
20080162616 Gross et al. Jul 2008 A1
20080163073 Becker et al. Jul 2008 A1
20080192053 Howell et al. Aug 2008 A1
20080195417 Surpin et al. Aug 2008 A1
20080195608 Clover Aug 2008 A1
20080222295 Robinson et al. Sep 2008 A1
20080223834 Griffiths et al. Sep 2008 A1
20080228512 Calkins et al. Sep 2008 A1
20080229056 Agarwal et al. Sep 2008 A1
20080255973 El Wade et al. Oct 2008 A1
20080263468 Cappione et al. Oct 2008 A1
20080267107 Rosenberg Oct 2008 A1
20080270468 Mao Oct 2008 A1
20080276167 Michael Nov 2008 A1
20080278311 Grange et al. Nov 2008 A1
20080288306 MacIntyre et al. Nov 2008 A1
20080294678 Gorman et al. Nov 2008 A1
20080301643 Appleton et al. Dec 2008 A1
20090002492 Velipasalar et al. Jan 2009 A1
20090027418 Maru et al. Jan 2009 A1
20090030915 Winter et al. Jan 2009 A1
20090055251 Shah et al. Feb 2009 A1
20090076845 Bellin et al. Mar 2009 A1
20090077081 Sarma Mar 2009 A1
20090088964 Schaaf et al. Apr 2009 A1
20090089273 Hicks Apr 2009 A1
20090100018 Roberts Apr 2009 A1
20090115786 Shmiasaki et al. May 2009 A1
20090119309 Gibson et al. May 2009 A1
20090125369 Kloosstra et al. May 2009 A1
20090125459 Norton et al. May 2009 A1
20090132921 Hwangbo et al. May 2009 A1
20090132953 Reed et al. May 2009 A1
20090143052 Bates et al. Jun 2009 A1
20090144262 White et al. Jun 2009 A1
20090144274 Fraleigh et al. Jun 2009 A1
20090158185 Lacevic et al. Jun 2009 A1
20090164934 Bhattiprolu et al. Jun 2009 A1
20090171939 Athsani et al. Jul 2009 A1
20090172511 Decherd et al. Jul 2009 A1
20090177962 Gusmorino et al. Jul 2009 A1
20090179892 Tsuda et al. Jul 2009 A1
20090187464 Bai et al. Jul 2009 A1
20090222400 Kupershmidt et al. Sep 2009 A1
20090222760 Halverson et al. Sep 2009 A1
20090234720 George et al. Sep 2009 A1
20090249244 Robinson et al. Oct 2009 A1
20090281839 Lynn et al. Nov 2009 A1
20090287470 Farnsworth et al. Nov 2009 A1
20090292626 Oxford Nov 2009 A1
20100011282 Dollard et al. Jan 2010 A1
20100042922 Bradateanu et al. Feb 2010 A1
20100057716 Stefik et al. Mar 2010 A1
20100063961 Guiheneuf et al. Mar 2010 A1
20100070523 Delgo et al. Mar 2010 A1
20100070842 Aymeloglu et al. Mar 2010 A1
20100070845 Facemire et al. Mar 2010 A1
20100070897 Aymeloglu et al. Mar 2010 A1
20100076968 Boyns et al. Mar 2010 A1
20100100963 Mahaffey Apr 2010 A1
20100103124 Kruzeniski et al. Apr 2010 A1
20100106420 Mattikalli et al. Apr 2010 A1
20100114887 Conway et al. May 2010 A1
20100122152 Chamberlain et al. May 2010 A1
20100131457 Heimendinger May 2010 A1
20100162176 Dunton Jun 2010 A1
20100191563 Schlaifer et al. Jul 2010 A1
20100198684 Eraker et al. Aug 2010 A1
20100199225 Coleman et al. Aug 2010 A1
20100228812 Uomini Sep 2010 A1
20100250412 Wagner Sep 2010 A1
20100280857 Liu et al. Nov 2010 A1
20100293174 Bennett et al. Nov 2010 A1
20100306713 Geisner et al. Dec 2010 A1
20100313119 Baldwin et al. Dec 2010 A1
20100318924 Frankel et al. Dec 2010 A1
20100321399 Ellren et al. Dec 2010 A1
20100325526 Ellis et al. Dec 2010 A1
20100325581 Finkelstein et al. Dec 2010 A1
20100330801 Rouh Dec 2010 A1
20110022312 McDonough et al. Jan 2011 A1
20110029526 Knight et al. Feb 2011 A1
20110047159 Baid et al. Feb 2011 A1
20110060753 Shaked et al. Mar 2011 A1
20110061013 Bilicki et al. Mar 2011 A1
20110074811 Hanson et al. Mar 2011 A1
20110078055 Faribault et al. Mar 2011 A1
20110078173 Seligmann et al. Mar 2011 A1
20110093327 Fordyce et al. Apr 2011 A1
20110106781 Pearson May 2011 A1
20110117878 Barash et al. May 2011 A1
20110119100 Ruhl et al. May 2011 A1
20110137766 Rasmussen et al. Jun 2011 A1
20110153384 Horne et al. Jun 2011 A1
20110161096 Buehler et al. Jun 2011 A1
20110167105 Ramakrishnan et al. Jul 2011 A1
20110167710 Ramakrishnan et al. Jul 2011 A1
20110170799 Carrino et al. Jul 2011 A1
20110173032 Payne et al. Jul 2011 A1
20110185316 Reid et al. Jul 2011 A1
20110208724 Jones et al. Aug 2011 A1
20110218934 Elser Sep 2011 A1
20110219450 McDougal et al. Sep 2011 A1
20110225198 Edwards et al. Sep 2011 A1
20110238553 Raj et al. Sep 2011 A1
20110258158 Resende et al. Oct 2011 A1
20110270705 Parker Nov 2011 A1
20110289397 Eastmond et al. Nov 2011 A1
20110289407 Naik et al. Nov 2011 A1
20110289420 Morioka et al. Nov 2011 A1
20110291851 Whisenant Dec 2011 A1
20110310005 Chen et al. Dec 2011 A1
20110314007 Dassa et al. Dec 2011 A1
20120019559 Siler et al. Jan 2012 A1
20120036013 Neuhaus et al. Feb 2012 A1
20120036434 Oberstein Feb 2012 A1
20120050293 Carlhian et al. Mar 2012 A1
20120066296 Appleton et al. Mar 2012 A1
20120072825 Sherkin et al. Mar 2012 A1
20120079363 Folting et al. Mar 2012 A1
20120084118 Bai et al. Apr 2012 A1
20120105632 Renkis May 2012 A1
20120106801 Jackson May 2012 A1
20120117082 Koperda et al. May 2012 A1
20120131512 Takeuchi et al. May 2012 A1
20120144335 Abeln et al. Jun 2012 A1
20120159307 Chung et al. Jun 2012 A1
20120159362 Brown et al. Jun 2012 A1
20120159363 DeBacker et al. Jun 2012 A1
20120159399 Bastide et al. Jun 2012 A1
20120170847 Tsukidate Jul 2012 A1
20120173985 Peppel Jul 2012 A1
20120196557 Reich et al. Aug 2012 A1
20120196558 Reich et al. Aug 2012 A1
20120203708 Psota et al. Aug 2012 A1
20120206469 Hulubei et al. Aug 2012 A1
20120208636 Feige Aug 2012 A1
20120221511 Gibson et al. Aug 2012 A1
20120221553 Wittmer et al. Aug 2012 A1
20120221580 Barney Aug 2012 A1
20120245976 Kumar et al. Sep 2012 A1
20120246148 Dror Sep 2012 A1
20120254129 Wheeler et al. Oct 2012 A1
20120284345 Costenaro et al. Nov 2012 A1
20120290879 Shibuya et al. Nov 2012 A1
20120296907 Long et al. Nov 2012 A1
20120311684 Paulsen et al. Dec 2012 A1
20120323888 Osann, Jr. Dec 2012 A1
20120330973 Ghuneim et al. Dec 2012 A1
20130006426 Healey et al. Jan 2013 A1
20130006725 Simanek et al. Jan 2013 A1
20130006916 McBride et al. Jan 2013 A1
20130018796 Kolhatkar et al. Jan 2013 A1
20130021445 Cossette-Pacheco et al. Jan 2013 A1
20130024202 Harris et al. Jan 2013 A1
20130046635 Grigg et al. Feb 2013 A1
20130046842 Muntz et al. Feb 2013 A1
20130060786 Serrano et al. Mar 2013 A1
20130061169 Pearcy et al. Mar 2013 A1
20130073377 Heath Mar 2013 A1
20130073454 Busch Mar 2013 A1
20130076732 Cervelli et al. Mar 2013 A1
20130078943 Biage et al. Mar 2013 A1
20130097482 Marantz et al. Apr 2013 A1
20130100134 Cervelli et al. Apr 2013 A1
20130110822 Ikeda et al. May 2013 A1
20130110877 Bonham et al. May 2013 A1
20130111320 Campbell et al. May 2013 A1
20130117651 Waldman et al. May 2013 A1
20130101159 Rosen Jun 2013 A1
20130150004 Rosen Jun 2013 A1
20130151148 Parundekar et al. Jun 2013 A1
20130151388 Falkenborg et al. Jun 2013 A1
20130157234 Gulli et al. Jun 2013 A1
20130166550 Buchmann et al. Jun 2013 A1
20130176321 Mitchell et al. Jul 2013 A1
20130179420 Park et al. Jul 2013 A1
20130224696 Wolfe et al. Aug 2013 A1
20130226953 Markovich et al. Aug 2013 A1
20130238616 Rose et al. Sep 2013 A1
20130246170 Gross et al. Sep 2013 A1
20130251233 Yang et al. Sep 2013 A1
20130262527 Hunter et al. Oct 2013 A1
20130263019 Castellanos et al. Oct 2013 A1
20130267207 Hao et al. Oct 2013 A1
20130268520 Fisher et al. Oct 2013 A1
20130279757 Kephart Oct 2013 A1
20130282696 John et al. Oct 2013 A1
20130282723 Petersen et al. Oct 2013 A1
20130290011 Lynn et al. Oct 2013 A1
20130290825 Arndt et al. Oct 2013 A1
20130297619 Chandarsekaran et al. Nov 2013 A1
20130311375 Priebatsch Nov 2013 A1
20140019936 Cohanoff Jan 2014 A1
20140032506 Hoey et al. Jan 2014 A1
20140033010 Richardt et al. Jan 2014 A1
20140040371 Gurevich et al. Feb 2014 A1
20140047357 Alfaro et al. Feb 2014 A1
20140059038 McPherson et al. Feb 2014 A1
20140067611 Adachi et al. Mar 2014 A1
20140068487 Steiger et al. Mar 2014 A1
20140095273 Tang et al. Apr 2014 A1
20140095509 Patton Apr 2014 A1
20140108068 Williams Apr 2014 A1
20140108380 Gotz et al. Apr 2014 A1
20140108985 Scott et al. Apr 2014 A1
20140129261 Bothwell et al. May 2014 A1
20140149436 Bahrami et al. May 2014 A1
20140156527 Grigg et al. Jun 2014 A1
20140157172 Peery et al. Jun 2014 A1
20140164502 Khodorenko et al. Jun 2014 A1
20140188847 Tang et al. Jul 2014 A1
20140189536 Lange et al. Jul 2014 A1
20140195515 Baker et al. Jul 2014 A1
20140195887 Ellis et al. Jul 2014 A1
20140267294 Ma Sep 2014 A1
20140267295 Sharma Sep 2014 A1
20140279824 Tamayo Sep 2014 A1
20140316911 Gross Oct 2014 A1
20140333651 Cervelli et al. Nov 2014 A1
20140337772 Cervelli et al. Nov 2014 A1
20140344230 Krause et al. Nov 2014 A1
20140361899 Layson Dec 2014 A1
20150019394 Unser et al. Jan 2015 A1
20150029176 Baxter et al. Jan 2015 A1
20150046870 Goldenberg et al. Feb 2015 A1
20150089424 Duffield et al. Mar 2015 A1
20150100897 Sun et al. Apr 2015 A1
20150100907 Erenrich et al. Apr 2015 A1
20150134666 Gattiker et al. May 2015 A1
20150169709 Kara et al. Jun 2015 A1
20150169726 Kara et al. Jun 2015 A1
20150170077 Kara et al. Jun 2015 A1
20150178877 Bogomolov et al. Jun 2015 A1
20150186821 Wang et al. Jul 2015 A1
20150187036 Wang et al. Jul 2015 A1
20160105334 Boe Apr 2016 A1
20180307391 Bogomolov et al. Oct 2018 A1
Foreign Referenced Citations (45)
Number Date Country
102014103482 Sep 2014 DE
102013222023 Jan 2015 DE
102014215621 Feb 2015 DE
0763201 Mar 1997 EP
1672527 Jun 2006 EP
2551799 Jan 2013 EP
2560134 Feb 2013 EP
2575107 Apr 2013 EP
2778913 Sep 2014 EP
2778977 Sep 2014 EP
2778983 Sep 2014 EP
2779082 Sep 2014 EP
2835745 Feb 2015 EP
2835770 Feb 2015 EP
2838039 Feb 2015 EP
2846241 Mar 2015 EP
2851852 Mar 2015 EP
2858014 Apr 2015 EP
2858018 Apr 2015 EP
2863326 Apr 2015 EP
2863346 Apr 2015 EP
2869211 May 2015 EP
2881868 Jun 2015 EP
2884439 Jun 2015 EP
2884440 Jun 2015 EP
2891992 Jul 2015 EP
2516155 Jan 2015 GB
2518745 Apr 2015 GB
2012778 Nov 2014 NL
2013306 Feb 2015 NL
624557 Dec 2014 NZ
WO 95032424 Nov 1995 WO
WO 00009529 Feb 2000 WO
WO 02065353 Aug 2002 WO
WO 2004057268 Jul 2004 WO
WO 2005013200 Feb 2005 WO
WO 2005104736 Nov 2005 WO
WO 2008064207 May 2008 WO
WO 2009061501 May 2009 WO
WO 2009123975 Oct 2009 WO
WO 2010000014 Jan 2010 WO
WO 2010030913 Mar 2010 WO
WO 2011058507 May 2011 WO
WO 2013010157 Jan 2013 WO
WO 2013102892 Jul 2013 WO
Non-Patent Literature Citations (213)
Entry
“Andy Turner's GISRUK 2012 Notes” <https://docs.google.com/document/d/1cTmxg7mVx5gd89lqblCYvCEnHA4QAivH4l4WpyPsqE4/edit?pli=1> printed Sep. 16, 2013 in 15 pages.
Barnes et al., “Viewshed Analysis”, GIS-ARC/INFO 2001, <www.evsc.virginia.edu/˜jhp7e/evsc466/student_pres/Rounds.pdf>.
Carver et al., “Real-Time Visibility Analysis and Rapid Viewshed Calculation Using a Voxel-Based Modelling Approach,” GISRUK 2012 Conference, Apr. 11-13, Lancaster UK, Apr. 13, 2012, pp. 6.
Ghosh, P., “A Solution of Polygon Containment, Spatial Planning, and Other Related Problems Using Minkowski Operations,” Computer Vision, Graphics, and Image Processing, 1990, vol. 49, pp. 1-35.
Gorr et al., “Crime Hot Spot Forecasting: Modeling and Comparative Evaluation”, Grant 98-IJ-CX-K005, May 6, 2002, 37 pages.
Haralick et al., “Image Analysis Using Mathematical Morphology,” Pattern Analysis and Machine Intelligence, IEEE Transactions, Jul. 1987, vol. PAMI-9, No. 4, pp. 532-550.
“HunchLab: Heat Map and Kernel Density Calculation for Crime Analysis,” Azavea Journal, printed from www.azavea.com/blogs/newsletter/v4i4/kernel-density-capabilities-added-to-hunchlab/ on Sep. 9, 2014, 2 pages.
Ipbucker, C., “Inverse Transformation for Several Pseudo-cylindrical Map Projections Using Jacobian Matrix,” ICCSA 2009, Part 1 LNCS 5592, pp. 553-564.
Levine, N., “Crime Mapping and the Crimestat Program,” Geographical Analysis, 2006, vol. 38, pp. 41-56.
Mandagere, Nagapramod, “Buffer Operations in GIS,” <http://www-users.cs.umn.edu/˜npramod/enc_pdf.pdf> retrieved Jan. 28, 2010, pp. 7.
Murray, C., Oracle Spatial Developer's Guide-6 Coordinate Systems (Spatial Reference Systems), <http://docs.oracle.com/cd/B28359_01/appdev.111/b28400.pdf>, Jun. 2009.
Open Street Map, “Amm's Diary:Unconnected ways and other data quality issues,” http://www.openstreetmap.org/user/amm/diary printed Jul. 23, 2012 in 3 pages.
POI Editor, “How to: Create Your Own Points of Interest,” <http://www.poieditor.com/articles/how_to_create_your_own_points_of_interest/> printed Jul. 22, 2012 in 4 pages.
Pozzi et al., “Vegetation and Population Density in Urban and Suburban Areas in the U.S.A.” Third International Symposium of Remote Sensing of Urban Areas Istanbul, Turkey, Jun. 2002, pp. 8.
Qiu, Fang, “3d Analysis and Surface Modeling”, <http://web.archive.org/web/20091202221925/http://www.utsa.edu/Irsg/Teaching/EES6513/08-3D.pdf> printed Sep. 16, 2013 in 26 pages.
Reddy et al., “Under the hood of GeoVRML 1.0,” SRI International, Proceedings of the fifth symposium on Vurtual Reality Modeling Language (Web3D-VRML), New York, NY, Feb. 2000, pp. 23-28. <http://pdf.aminer.org/000/648/038/under_the_hood_of_geovrml.pdf>.
Reibel et al., “Areal Interpolation of Population Counts Using Pre-classi_ed Land Cover Data,” Population Research and Policy Review, 2007, vol. 26, pp. 619-633.
Reibel, M., “Geographic Information Systems and Spatial Data Processing in Demography: a Review,” Population Research and Policy Review, 2007, vol. 26, pp. 601-618.
Rizzardi et al., “Interfacing U.S. Census Map Files with Statistical Graphics Software: Application and Use in Epidemiology,” Statistics in Medicine, Oct. 1993, vol. 12, No. 19-20, pp. 1953-1964.
Snyder, “Map Projections—A Working Manual,” U.S. Geological Survey Professional paper 1395, United States Government Printing Office, Washington: 1987, pp. 11-21 and 60-70.
Sonris, “Using the Area of Interest Tools,” <http://web.archive.org/web/20061001053327/http://sonris-www.dnr.state.la.us/gis/instruct_files/tutslide12> printed Jan. 3, 2013 in 1 page.
Tangelder et al., “Freeform Shape Matching Using Minkowski Operations,” The Netherlands, Jun. 1996, pp. 12.
Valentini et al., “Ensembles of Learning Machines”, M. Marinaro and R. Tagliaferri (Eds.): WIRN VIETRI 2002, LNCS 2486, pp. 3-20.
VB Forums, “Buffer a Polygon,” Internet Citation, <http://www.vbforums.com/showthread.php?198436-Buffer-a-Polygon>, Specifically Thread #1, #5 & #11 retrieved on May 2, 2013, pp. 8.
Vivid Solutions, “JTS Topology Suite: Technical Specifications,” <http://www.vividsolutions.com/jts/bin/JTS%20Technical%20Specs.pdf> Version 1.4, 2003, pp. 36.
Wikipedia, “Douglas Peucker-Algorithms,” <http://de.wikipedia.org/w/index.php?title=Douglas-Peucker-Algorithmus&oldid=91846042> printed Jul. 2011, pp. 2.
Wikipedia, “Ramer_Douglas_Peucker Algorithm,” <http://en.wikipedia.org/wiki/Ramer%E2%80%93Douglas%E2%80%93Peucker_algorithm> printed Jul. 2011, pp. 3.
Wongsuphasawat et al., “Visual Analytics for Transportation Incident Data Sets,” Transportation Research Record 2138, 2009, pp. 135-145.
Woodbridge, Stephen, “[gens-devel] Polygon simplification,” <http://lists.osgeo.org/pipermail/geos-devel/2011-May/005210.html> dated May 8, 2011, pp. 3.
International Search Report and Written Opinion in Application No. PCT/US2009/056703 dated Mar. 15, 2010.
Notice of Allowance for U.S. Appl. No. 13/948,859 dated Dec. 10, 2014.
Official Communication for Australian Patent Application No. 2010227081 dated Mar. 18, 2011.
Official Communication for Australian Patent Application No. 2010257305 dated Apr. 12, 2011.
Official Communication for Australian Patent Application No. 2010257305 dated Sep. 22, 2011.
Official Communication for European Patent Application No. 08839003.4 dated Jun. 12, 2013.
Official Communication for European Patent Application No. 08839003.4 dated Aug. 14, 2012.
Official Communication for European Patent Application No. 10195798.3 dated May 17, 2011.
Official Communication for European Patent Application No. 12186236.1 dated May 17, 2013.
Official Communication for European Patent Application No. 14189347.9 dated Oct. 13, 2017.
Official Communication for Great Britain Patent Application No. 1319225.7 dated May 2, 2014.
Official Communication for New Zealand Patent Application No. 616167 dated Oct. 10, 2013.
Official Communication for U.S. Appl. No. 12/840,673 dated Sep. 17, 2014.
Official Communication for U.S. Appl. No. 12/840,673 dated Jan. 2, 2015.
Official Communication for U.S. Appl. No. 13/728,879 dated Mar. 17, 2015.
Official Communication for U.S. Appl. No. 13/728,879 dated Jan. 27, 2015.
Official Communication for U.S. Appl. No. 14/135,289 dated Oct. 14, 2014.
Official Communication for U.S. Appl. No. 14/306,147 dated Feb. 19, 2015.
Official Communication for U.S. Appl. No. 14/319,161 dated Jan. 23, 2015.
Official Communication for U.S. Appl. No. 14/581,823 dated Sep. 1, 2017.
Official Communication for U.S. Appl. No. 14/581,823 dated Nov. 2, 2017.
Official Communication for U.S. Appl. No. 14/672,009 dated Jul. 14, 2017.
Official Communication for U.S. Appl. No. 14/672,009 dated May 26, 2017.
“A First Look: Predicting Market Demand for Food Retail using a Huff Analysis,” TRF Policy Solutions, Jul. 2012, pp. 30.
“A Quick Guide to UniProtKB Swiss-Prot & TrEMBL,” Sep. 2011, pp. 2.
“A Word About Banks and the Laundering of Drug Money,” Aug. 18, 2012, http://www.golemxiv.co.uk/2012/08/a-word-about-banks-and-the-laundering-of-drug-money/.
Acklen, Laura, “Absolute Beginner's Guide to Microsoft Word 2003,” Dec. 24, 2003, pp. 15-18, 34-41, 308-316.
Amnet, “5 Great Tools for Visualizing Your Twitter Followers,” posted Aug. 4, 2010, http://www.amnetblog.com/component/content/article/115-5-grate-tools-for-visualizing-your-twitter-followers.html.
Ananiev et al., “The New Modality API,” http://web.archive.org/web/20061211011958/http://java.sun.com/developer/technicalArticles/J2SE/Desktop/javase6/modality/ Jan. 21, 2006, pp. 8.
Bluttman et al., “Excel Formulas and Functions for Dummies,” 2005, Wiley Publishing, Inc., pp. 280, 284-286.
Boyce, Jim, “Microsoft Outlook 2010 Inside Out,” Aug. 1, 2010, retrieved from the internet https://capdtron.files.wordpress.com/2013/01/outlook-2010-inside_out.pdf.
Bugzilla@Mozilla, “Bug 18726—[feature] Long-click means of invoking contextual menus not supported,” http://bugzilla.mozilla.org/show_bug.cgi?id=18726 printed Jun. 13, 2013 in 11 pages.
Canese et al., “Chapter 2: PubMed: The Bibliographic Database,” The NCBI Handbook, Oct. 2002, pp. 1-10.
Celik, Tantek, “CSS Basic User Interface Module Level 3 (CSS3 UI),” Section 8 Resizing and Overflow, Jan. 17, 2012, retrieved from internet http://www.w3.org/TR/2012/WD-css3-ui-20120117/#resizing-amp-overflow retrieved on May 18, 2015.
Chen et al., “Bringing Order to the Web: Automatically Categorizing Search Results,” CHI 2000, Proceedings of the SIGCHI conference on Human Factors in Computing Systems, Apr. 1-6, 2000, The Hague, The Netherlands, pp. 145-152.
Chung, Chin-Wan, “Dataplex: An Access to Heterogeneous Distributed Databases,” Communications of the ACM, Association for Computing Machinery, Inc., vol. 33, No. 1, Jan. 1, 1990, pp. 70-80.
Conner, Nancy, “Google Apps: The Missing Manual,” May 1, 2008, pp. 15.
Definition “Identify”, downloaded Jan. 22, 2015, 1 page.
Definition “Overlay”, downloaded Jan. 22, 2015, 1 page.
Delcher et al., “Identifying Bacterial Genes and Endosymbiont DNA with Glimmer,” BioInformatics, vol. 23, No. 6, 2007, pp. 673-679.
Dramowicz, Ela, “Retail Trade Area Analysis Using the Huff Model,” Directions Magazine, Jul. 2, 2005 in 10 pages, http://www.directionsmag.com/articles/retail-trade-area-analysis-using-the-huff-model/123411.
GIS-NET 3 Public_Department of Regional Planning. Planning & Zoning Information for Unincorporated La County. Retrieved Oct. 2, 2013 from http://gis.planning.lacounty.gov/GIS-NET3_Public/Viewer.html.
Goswami, Gautam, “Quite Writly Said!,” One Brick at a Time, Aug. 21, 2005, pp. 7.
Griffith, Daniel A., “A Generalized Huff Model,” Geographical Analysis, Apr. 1982, vol. 14, No. 2, pp. 135-144.
Hansen et al., “Analyzing Social Media Networks with NodeXL: Insights from a Connected World”, Chapter 4, pp. 53-67 and Chapter 10, pp. 143-164, published Sep. 2010.
Hardesty, “Privacy Challenges: Analysis: Its Surprisingly Easy to Identify Individuals from Credit-Card Metadata,” MIT News on Campus and Around the World, MIT News Office, Jan. 29, 2015, 3 pages.
Hibbert et al., “Prediction of Shopping Behavior Using a Huff Model Within a GIS Framework,” Healthy Eating in Context, Mar. 18, 2011, pp. 16.
Hogue et al., “Thresher: Automating the Unwrapping of Semantic Content from the World Wide Web,” 14th International Conference on World Wide Web, WWW 2005: Chiba, Japan, May 10-14, 2005, pp. 86-95.
Huff et al., “Calibrating the Huff Model Using ArcGIS Business Analyst,” ESRI, Sep. 2008, pp. 33.
Huff, David L., “Parameter Estimation in the Huff Model,” ESRI, ArcUser, Oct.-Dec. 2003, pp. 34-36.
Kahan et al., “Annotea: an Open RDF Infrastructure for Shared Web Annotations”, Computer Networks, Elsevier Science Publishers B.V., vol. 39, No. 5, dated Aug. 5, 2002, pp. 589-608.
Keylines.com, “An Introduction to KeyLines and Network Visualization,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-White-Paper.pdf> downloaded May 12, 2014 in 8 pages.
Keylines.com, “KeyLines Datasheet,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-datasheet.pdf> downloaded May 12, 2014 in 2 pages.
Keylines.com, “Visualizing Threats: Improved Cyber Security Through Network Visualization,” Apr. 2014, <http://keylines.com/wp-content/uploads/2014/04/Visualizing-Threats1.pdf> downloaded May 12, 2014 in 10 pages.
Kitts, Paul, “Chapter 14: Genome Assembly and Annotation Process,” The NCBI Handbook, Oct. 2002, pp. 1-21.
Li et al., “Interactive Multimodal Visual Search on Mobile Device,” IEEE Transactions on Multimedia, vol. 15, No. 3, Apr. 1, 2013, pp. 594-607.
Liu, Tianshun, “Combining GIS and the Huff Model to Analyze Suitable Locations for a New Asian Supermarket in the Minneapolis and St. Paul, Minnesota USA,” Papers in Resource Analysis, 2012, vol. 14, pp. 8.
Madden, Tom, “Chapter 16: The BLAST Sequence Analysis Tool,” The NCBI Handbook, Oct. 2002, pp. 1-15.
Manno et al., “Introducing Collaboration in Single-user Applications through the Centralized Control Architecture,” 2010, pp. 10.
Manske, “File Saving Dialogs,” <http://www.mozilla.org/editor/ui_specs/FileSaveDialogs.html>, Jan. 20, 1999, pp. 7.
Map Builder, “Rapid Mashup Development Tool for Google and Yahoo Maps!” <http://web.archive.org/web/20090626224734/http://www.mapbuilder.net/> printed Jul. 20, 2012 in 2 pages.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.yahoo.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.bing.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.google.com.
Microsoft—Developer Network, “Getting Started with VBA in Word 2010,” Apr. 2010, <http://msdn.microsoft.com/en-us/library/ff604039%28v=office.14%29.aspx> as printed Apr. 4, 2014 in 17 pages.
Microsoft Office—Visio, “About connecting shapes,” <http://office.microsoft.com/en-us/visio-help/about-connecting-shapes-HP085050369.aspx> printed Aug. 4, 2011 in 6 pages.
Microsoft Office—Visio, “Add and glue connectors with the Connector tool,” <http://office.microsoft.com/en-us/visio-help/add-and-glue-connectors-with-the-connector-tool-HA010048532.aspx?CTT-1> printed Aug. 4, 2011 in 1 page.
Mizrachi, Ilene, “Chapter 1: GenBank: The Nuckeotide Sequence Database,” The NCBI Handbook, Oct. 2002, pp. 1-14.
Nierman, “Evaluating Structural Similarity in XML Documents”, 6 pages, 2002.
Olanoff, Drew, “Deep Dive with the New Google Maps for Desktop with Google Earth Integration, It's More than Just a Utility,” May 15, 2013, pp. 1-6, retrieved from the internet: http://web.archive.org/web/20130515230641/http://techcrunch.com/2013/05/15/deep-dive-with-the-new-google-maps-for-desktop-with-google-earth-integration-its-more-than-just-a-utility/.
Palmas et al. “An Edge-Bunding Layout for Interactive Parallel Coordinates” 2014 IEEE Pacific Visualization Symposium, pp. 57-64.
“Potential Money Laundering Warning Signs,” snapshot taken 2003, https://web.archive.org/web/20030816090055/http:/finsolinc.com/ANTI-MONEY%20LAUNDERING%20TRAINING%20GUIDES.pdf.
“Refresh CSS Ellipsis When Resizing Container—Stack Overflow,” Jul. 31, 2013, retrieved from internet http://stackoverflow.com/questions/17964681/refresh-css-ellipsis-when-resizing-container, retrieved on May 18, 2015.
Rouse, Margaret, “OLAP Cube,” <http://searchdatamanagement.techtarget.com/definition/OLAP-cube>, Apr. 28, 2012, pp. 16.
Sigrist, et al., “PROSITE, a Protein Domain Database for Functional Characterization and Annotation,” Nucleic Acids Research, 2010, vol. 38, pp. D161-D166.
Sirotkin et al., “Chapter 13: The Processing of Biological Sequence Data at NCBI,” The NCBI Handbook, Oct. 2002, pp. 1-11.
“The FASTA Program Package,” fasta-36.3.4, Mar. 25, 2011, pp. 29.
Thompson, Mick, “Getting Started with GEO,” Getting Started with GEO, Jul. 26, 2011.
Umagandhi et al., “Search Query Recommendations Using Hybrid User Profile with Query Logs,” International Journal of Computer Applications, vol. 80, No. 10, Oct. 1, 2013, pp. 7-18.
Wikipedia, “Federated Database System,” Sep. 7, 2013, retrieved from the internet on Jan. 27, 2015 http://en.wikipedia.org/w/index.php?title-Federated_database_system&oldid=571954221.
Yang et al., “HTML Page Analysis Based on Visual Cues”, A129, pp. 859-864, 2001.
Notice of Allowance for U.S. Appl. No. 14/102,394 dated Aug. 25, 2014.
Notice of Allowance for U.S. Appl. No. 14/108,187 dated Aug. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/135,289 dated Oct. 14, 2014.
Notice of Allowance for U.S. Appl. No. 14/192,767 dated Dec. 16, 2014.
Notice of Allowance for U.S. Appl. No. 14/225,084 dated May 4, 2015.
Notice of Allowance for U.S. Appl. No. 14/268,964 dated Dec. 3, 2014.
Notice of Allowance for U.S. Appl. No. 14/294,098 dated Dec. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/473,552 dated Jul. 24, 2015.
Notice of Allowance for U.S. Appl. No. 14/473,860 dated Jan. 5, 2015.
Notice of Allowance for U.S. Appl. No. 14/486,991 dated May 1, 2015.
Notice of Allowance for U.S. Appl. No. 14/504,103 dated May 18, 2015.
Notice of Allowance for U.S. Appl. No. 14/616,080 dated Apr. 2, 2015.
Official Communication for Australian Patent Application No. 2014201511 dated Feb. 27, 2015.
Official Communication for Australian Patent Application No. 2014202442 dated Mar. 19, 2015.
Official Communication for Australian Patent Application No. 2014210604 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014210614 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014213553 dated May 7, 2015.
Official Communication for Australian Patent Application No. 2014250678 dated Jun. 17, 2015.
Official Communication for European Patent Application No. 14158861.6 dated Jun. 16, 2014.
Official Communication for European Patent Application No. 14159464.8 dated Jul. 31, 2014.
Official Communication for European Patent Application No. 14180142.3 dated Feb. 6, 2015.
Official Communication for European Patent Application No. 14180281.9 dated Jan. 26, 2015.
Official Communication for European Patent Application No. 14180321.3 dated Apr. 17, 2015.
Official Communication for European Patent Application No. 14180432.8 dated Jun. 23, 2015.
Official Communication for European Patent Application No. 14186225.0 dated Feb. 13, 2015.
Official Communication for European Patent Application No. 14187739.9 dated Jul. 6, 2015.
Official Communication for European Patent Application No. 14187996.5 dated Feb. 12, 2015.
Official Communication for European Patent Application No. 14189344.6 dated Feb. 20, 2015.
Official Communication for European Patent Application No. 14189347.9 dated Mar. 4, 2015.
Official Communication for European Patent Application No. 14189802.3 dated May 11, 2015.
Official Communication for European Patent Application No. 14191540.5 dated May 27, 2015.
Official Communication for European Patent Application No. 14197879.1 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14197895.7 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14199182.8 dated Mar. 13, 2015.
Official Communication for Great Britain Patent Application No. 1404457.2 dated Aug. 14, 2014.
Official Communication for Great Britain Patent Application No. 1404574.4 dated Dec. 18, 2014.
Official Communication for Great Britain Patent Application No. 1408025.3 dated Nov. 6, 2014.
Official Communication for Great Britain Patent Application No. 1411984.6 dated Dec. 22, 2014.
Official Communication for Great Britain Patent Application No. 1413935.6 dated Jan. 27, 2015.
Official Communication for Netherlands Patent Application No. 2013306 dated Apr. 24, 2015.
Official Communication for New Zealand Patent Application No. 622513 dated Apr. 3, 2014.
Official Communication for New Zealand Patent Application No. 622517 dated Apr. 3, 2014.
Official Communication for New Zealand Patent Application No. 624557 dated May 14, 2014.
Official Communication for New Zealand Patent Application No. 627962 dated Aug. 5, 2014.
Official Communication for New Zealand Patent Application No. 628161 dated Aug. 25, 2014.
Official Communication for New Zealand Patent Application No. 628263 dated Aug. 12, 2014.
Official Communication for New Zealand Patent Application No. 628495 dated Aug. 19, 2014.
Official Communication for New Zealand Patent Application No. 628585 dated Aug. 26, 2014.
Official Communication for New Zealand Patent Application No. 628840 dated Aug. 28, 2014.
Official Communication for U.S. Appl. No. 12/556,318 dated Jul. 2, 2015.
Official Communication for U.S. Appl. No. 13/247,987 dated Apr. 2, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Mar. 4, 2015.
Official Communication for U.S. Appl. No. 13/835,688 dated Jun. 17, 2015.
Official Communication for U.S. Appl. No. 13/839,026 dated Aug. 4, 2015.
Official Communication for U.S. Appl. No. 14/148,568 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/148,568 dated Mar. 26, 2015.
Official Communication for U.S. Appl. No. 14/196,814 dated May 5, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 10, 2014.
Official Communication for U.S. Appl. No. 14/225,006 dated Feb. 27, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 2, 2014.
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 20, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Feb. 11, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Aug. 12, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated May 20, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/225,160 dated Jul. 29, 2014.
Official Communication for U.S. Appl. No. 14/268,964 dated Sep. 3, 2014.
Official Communication for U.S. Appl. No. 14/289,596 dated Jul. 18, 2014.
Official Communication for U.S. Appl. No. 14/289,596 dated Jan. 26, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Apr. 30, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated Jul. 22, 2014.
Official Communication for U.S. Appl. No. 14/289,599 dated May 29, 2015.
Official Communication for U.S. Appl. No. 14/294,098 dated Aug. 15, 2014.
Official Communication for U.S. Appl. No. 14/294,098 dated Nov. 6, 2014.
Official Communication for U.S. Appl. No. 14/306,138 dated Feb. 18, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 23, 2014.
Official Communication for U.S. Appl. No. 14/306,138 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Aug. 7, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 11, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated May 15, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Jul. 6, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Jun. 16, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Nov. 25, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 4, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Nov. 28, 2014.
Official Communication for U.S. Appl. No. 14/323,935 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Dec. 2, 2014.
Official Communication for U.S. Appl. No. 14/326,738 dated Jul. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/473,552 dated Feb. 24, 2015.
Official Communication for U.S. Appl. No. 14/486,991 dated Mar. 10, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Feb. 5, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated Aug. 19, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated May 18, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated Jul. 24, 2015.
Notice of Allowance for U.S. Appl. No. 14/581,823 dated Apr. 6, 2018.
Official Communication for European Patent Application No. 14189347.9 dated Jun. 8, 2018.
Official Communication for U.S. Appl. No. 14/672,009 dated Jan. 9, 2018.