Systems and interactive user interfaces for automatic generation of temporal representation of data objects

Information

  • Patent Grant
  • 9823818
  • Patent Number
    9,823,818
  • Date Filed
    Wednesday, April 6, 2016
    8 years ago
  • Date Issued
    Tuesday, November 21, 2017
    7 years ago
Abstract
Computer-implemented systems and methods are disclosed for automatically generating and displaying a chronology of events, where events may be represented by data objects in one or more databases. Events/data objects may be identified as relevant to an investigation or analysis based on specified criteria. A timeline may be generated based on the identified set of relevant events, and interactive user interfaces may be generated and displayed that present the events as a timeline and a list. Events may be selected from the timeline or the list, may be identified as key events in the chronology, and additional events related to a selected event may be determined and added to the chronology. Timelines may be compared to other data sets, including other timelines, other event lists, and other relevant data.
Description
TECHNICAL FIELD

The present disclosure relates to systems and techniques for data integration, analysis, and visualization. More specifically, the present disclosure relates to systems and techniques for integrating, analyzing, and visualizing data through the automatic generation of a chronology user interface.


BACKGROUND

Electronic record-keeping produces data sets with thousands or millions of records. The sheer quantity of information available for analysis may prevent meaningful conclusions from being drawn, or may prevent connections between events from even being discovered. Filtering data based on relationships, events, and common characteristics and visually presenting the data in concise and informative ways can help users to identify data relevant to a particular time, place, or sequence of events.


SUMMARY

The systems, methods, and devices described herein each have several aspects, no single one of which is solely responsible for its desirable attributes. Without limiting the scope of this disclosure, several non-limiting features will now be discussed briefly.


Embodiments of the present disclosure relate to automatic generation of interactive user interfaces for presentation of chronologies. A chronology user interface may be utilized to identify a set of relevant events. The events may be identified according to specified criteria, by comparing events and determining common attributes, or may be identified based on inputs, as described below. An interactive user interface may then be generated that displays the events, for example, as a timeline and a list. The timeline and list may be interactive. For example, an event on the timeline may be selected based on user input, and additional information or related events may be displayed. Modifiable attributes of events, such as annotations or associated keywords, may be modified via the user interface.


Chronology user interfaces may be used to compare timelines and to present and analyze various theories regarding a set of events and the relationship between them.


Additional embodiments of the disclosure are described below in reference to the appended claims, which may serve as an additional summary of the disclosure.


In various embodiments, computer systems are disclosed that comprise one or more hardware computer processors in communication with one or more non-transitory computer readable storage devices, wherein the one or more hardware computer processors are configured to execute the plurality of computer executable instructions in order to cause the computer system to operations comprising one or more aspects of the above-described embodiments (including one or more aspects of the appended claims).


In various embodiments, computer-implemented methods are disclosed in which, under control of one or more hardware computing devices configured with specific computer executable instructions, one or more aspects of the above-described embodiments (including one or more aspects of the appended claims) are implemented and/or performed.


In various embodiments, computer-readable storage mediums storing software instructions are disclosed, wherein, in response to execution by a computing system having one or more hardware processors, the software instructions configure the computing system to perform operations comprising one or more aspects of the above-described embodiments (including one or more aspects of the appended claims).


In an embodiment, a computer system is disclosed comprising a data store configured to store computer executable instructions, and a processor that, when executing the computer-executable instructions, is configured to: obtain a plurality of events, each of the plurality of events comprising a respective set of event attributes, the event attributes comprising at least an event timestamp; obtain one or more criteria associated with event attributes; determine, based at least in part on the one or more criteria, a set of relevant events within the plurality of events; generate for display a user interface based at least in part on the set of relevant events and the corresponding event attributes, the user interface comprising (1) a timeline interface for display of a timeline, wherein positions of individual relevant events within the timeline interface are determined based at least in part on respective event timestamps, and (2) a list interface for display of the relevant events; and cause display of the user interface.


In some embodiments, the event attributes further comprise at least one of an event source, an event type, an event icon, an event duration, an event title, a set of event participants, a set of event keywords, and a set of event annotations.


In some embodiments, the processor is further configured to: obtain input identifying a selected event displayed within the user interface; generate for display a first update to the user interface, the first update comprising (1) an indication of selection of the selected event within the timeline interface, and (2) an indication of selection of the selected event within the list interface; and cause display of the first update to the user interface.


In some embodiments, the processor is further configured to: determine that at least one of the plurality of events has been updated; generate an alert indicating that the at least one of the plurality of events has been updated; and cause transmission of the alert to a computing device, wherein transmission of the alert causes the computing device to activate an application for display of the alert, and wherein the alert enables display of an updated user interface, the updated user interface based at least in part on an updated event of the plurality of events.


In some embodiments, the processor is further configured to: determine, based at least in part on the selected event, one or more related events from within the plurality of events; generate for display an event selection interface for display of the one or more related events; and cause display of the event selection interface.


In some embodiments, the processor configured to determine the one or more related events is configured to determine the one or more related events based at least in part on one or more attributes of the selected event.


In some embodiments, the processor is further configured to obtain input identifying a related event within the one or more related events; generate for display a second update to the user interface, the second update comprising (1) display of the related event within the timeline interface and (2) display of the related event within the list interface; and cause display of the second update to the user interface.


In another embodiment, a computer-implemented method is disclosed comprising obtaining a plurality of events, each of the plurality of events comprising event attributes; determining, based at least in part on one or more criteria associated with event attributes, a set of relevant events within the plurality of events; generating for display a user interface based at least in part on the set of relevant events, the user interface comprising (1) a timeline interface for display of a timeline, wherein positions of individual relevant events within the timeline interface are determined based at least in part on respective event attributes, and (2) a list interface for display of the set of relevant events; and displaying the user interface.


In some embodiments, the timeline interface displays a subset of the set of relevant events as key events. In some embodiments, the event attributes include at least one configurable attribute, and in some embodiments the at least one configurable attribute comprises an event annotation, an event icon, an event description, or an event keyword.


In some embodiments, the computer-implemented method further comprises obtaining an input indicative of a modification of a configurable attribute for a relevant event; modifying the configurable attribute for the relevant event; generating for display a first update to the user interface, the first update comprising at least one of a modified configurable attribute within the timeline interface and a modified configurable attribute within the list interface; and causing display of the first update to the user interface.


In some embodiments, the computer-implemented method further comprises obtaining one or more updates to the plurality of events; determining, based at least in part on the one or more updates, an update to the set of relevant events; generating for display an alert message, the alert message indicating the update to the set of relevant alerts; and displaying the alert message.


In some embodiments, determining the update to the set of relevant events comprises at least one of determining an additional event to include in the set of relevant events, determining an event to remove from the set of relevant events, or determining a changed event attribute for an event in the set of relevant events.


In some embodiments, the computer-implemented method further comprises generating for display an updated user interface based at least in part on the update to the set of relevant events, the updated user interface comprising an updated timeline interface and an updated list interface; and displaying the updated user interface.


In another embodiment, a non-transitory computer-readable storage medium is disclosed, the non-transitory computer-readable storage medium including computer-executable instructions that, when executed by a processor, cause the processor to: determine, based at least in part on a first set of criteria, a first set of relevant events within a plurality of events; generate a first timeline corresponding to the first set of relevant events; generate for display a first user interface based at least in part on the first set of relevant events, the first user interface comprising (1) a timeline interface for display of at least a portion of the first timeline, wherein individual relevant events are positioned within the timeline interface based at least in part on respective attributes of the individual relevant events, and (2) a list interface for display of at least the portion of the first set of relevant events; and cause display of the first user interface.


In some embodiments, the computer-executable instructions further cause the processor to obtain the plurality of events. In some embodiments, the first set of criteria comprise one or more inputs, and each of the one or more inputs are indicative of a selection of a respective event. In some embodiments, the first set of criteria is associated with event attributes, and the first set of relevant events comprises events that satisfy the first set of criteria.


In some embodiments, the computer-executable instructions further cause the processor to determine, based at least in part on a second set of criteria, a second set of relevant events within the plurality of events; generate a second timeline corresponding to the second set of events; generate for display a second user interface based at least in part on the second set of relevant events; and cause display of the second user interface.


In some embodiments, the first set of relevant events and the second set of relevant events have at least one event in common. In some embodiments, the second user interface comprises a combined timeline interface for display of a combined timeline, the combined timeline comprising the first timeline and at least a portion of the second timeline; a first list interface for display of the first set of relevant events; and a second list interface for display the second set of relevant events. In some embodiments, the computer-executable instructions further cause the processor to generate the combined timeline based at least in part on the first timeline and the second timeline.





BRIEF DESCRIPTION OF THE DRAWINGS

Reference will now be made to the accompanying drawings, which illustrate exemplary embodiments of the present disclosure. In the drawings:



FIG. 1 is a block diagram of an example computer system consistent with embodiments of the present disclosure;



FIG. 2 illustrates one embodiment of a database system using an ontology;



FIG. 3 illustrates one embodiment of a system for creating data in a data store using a dynamic ontology;



FIG. 4 illustrates a sample user interface using relationships described in a data store using a dynamic ontology;



FIGS. 5A-5F are block diagrams of example user interfaces for presenting and interacting with chronologies in accordance with aspects of the present disclosure;



FIG. 6 is a flow diagram depicting an example routine for generating a chronology user interface in accordance with aspects of the present disclosure; and



FIG. 7 is a flow diagram depicting an example routing for identifying events related to the chronology and presenting an updated user interface in accordance with aspects of the present disclosure.





DETAILED DESCRIPTION

Reference will now be made in detail to example embodiments, the examples of which are illustrated in the accompanying drawings. Whenever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts.


Embodiments of the present disclosure relate to systems, methods, and computer-readable mediums for automatically generating and displaying a chronology. A chronology system obtains criteria for identifying a set of relevant events, and automatically generates for display an interactive user interface. The system processes input to designate key events, identify related events, and specify annotations or other modifiable attributes. Chronologies may be compared to each other, or to other data sets, to facilitate analysis and investigation.


According to some embodiments, the operations, techniques, and/or components described herein can be implemented by an electronic device, which can include one or more special-purpose computing devices. The special-purpose computing devices can be hard-wired to perform the operations, techniques, and/or components described herein, or can include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the operations, techniques and/or components described herein, or can include one or more hardware processors programmed to perform such features of the present disclosure pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices can also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the technique and other features of the present disclosure. The special-purpose computing devices can be desktop computer systems, portable computer systems, handheld devices, networking devices, or any other device that incorporates hard-wired and/or program logic to implement the techniques and other features of the present disclosure.


The one or more special-purpose computing devices can be generally controlled and coordinated by operating system software, such as iOS, Android, Blackberry, Chrome OS, Windows XP, Windows Vista, Windows 7, Windows 8, Windows Server, Windows CE, Unix, Linux, SunOS, Solaris, VxWorks, or other compatible operating systems. In other embodiments, the computing device can be controlled by a proprietary operating system. Operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface functionality, such as a graphical user interface (“GUI”), among other things.


In various embodiments of the present disclosure, large amounts of data are automatically and dynamically calculated interactively in response to user inputs, and the calculated data is efficiently and compactly presented to a user by the system. Thus, in some embodiments, the user interfaces described herein are more efficient as compared to previous user interfaces in which data is not dynamically updated and compactly and efficiently presented to the user in response to interactive inputs.


Further, as described herein, the system may be configured and/or designed to generate user interface data useable for rendering the various interactive user interfaces described. The user interface data may be used by the system, and/or another computer system, device, and/or software program (for example, a browser program), to render the interactive user interfaces. The interactive user interfaces may be displayed on, for example, electronic displays (including, for example, touch-enabled displays).


Additionally, it has been noted that design of computer user interfaces “that are useable and easily learned by humans is a non-trivial problem for software developers.” (Dillon, A. (2003) User Interface Design. MacMillan Encyclopedia of Cognitive Science, Vol. 4, London: MacMillan, 453-458.) The various embodiments of interactive and dynamic user interfaces of the present disclosure are the result of significant research, development, improvement, iteration, and testing. This non-trivial development has resulted in the user interfaces described herein which may provide significant cognitive and ergonomic efficiencies and advantages over previous systems. The interactive and dynamic user interfaces include improved human-computer interactions that may provide reduced mental workloads, improved decision-making, reduced work stress, and/or the like, for a user. For example, user interaction with the interactive user interfaces described herein may provide an optimized display of time-varying report-related information and may enable a user to more quickly access, navigate, assess, and digest such information than previous systems.


Further, the interactive and dynamic user interfaces described herein are enabled by innovations in efficient interactions between the user interfaces and underlying systems and components. For example, disclosed herein are improved methods of receiving user inputs, translation and delivery of those inputs to various system components, automatic and dynamic execution of complex processes in response to the input delivery, automatic interaction among various components and processes of the system, and automatic and dynamic updating of the user interfaces. The interactions and presentation of data via the interactive user interfaces described herein may accordingly provide cognitive and ergonomic efficiencies and advantages over previous systems.


Various embodiments of the present disclosure provide improvements to various technologies and technological fields. Additionally, various embodiments of the present disclosure are inextricably tied to computer technology. In particular, various embodiments rely on detection of user inputs via graphical user interfaces, calculation of updates to displayed electronic data based on those user inputs, automatic processing of related events, and presentation of the updates to displayed chronologies via interactive graphical user interfaces. Such features and others are intimately tied to, and enabled by, computer technology, and would not exist except for computer technology. For example, the interactions with displayed data described below in reference to various embodiments cannot reasonably be performed by humans alone, without the computer technology upon which they are implemented. Further, the implementation of the various embodiments of the present disclosure via computer technology enables many of the advantages described herein, including more efficient interaction with, and presentation of, various types of electronic image data.


Terms

In order to facilitate an understanding of the systems and methods discussed herein, a number of terms are defined below. The terms defined below, as well as other terms used herein, should be construed to include the provided definitions, the ordinary and customary meaning of the terms, and/or any other implied meaning for the respective terms. Thus, the definitions below do not limit the meaning of these terms, but only provide exemplary definitions.


Ontology: Stored information that provides a data model for storage of data in one or more databases. For example, the stored data may comprise definitions for object types and property types for data in a database, and how objects and properties may be related.


Database: A broad term for any data structure for storing and/or organizing data, including, but not limited to, relational databases (Oracle database, mySQL database, etc.), non-relational databases (for example, a NoSQL database), an in-memory database, spreadsheets, XML files, and text file, among others. The various terms “database,” “storage,” “data store,” and “data source” may be used interchangeably in the present disclosure.


Data Object or Object: A data container for information representing specific things in the world that have a number of definable properties. For example, a data object can represent an entity such as a person, a place, an organization, a market instrument, or other noun. A data object can represent an event that happens at a point in time or for a duration. A data object can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object may be associated with a unique identifier that uniquely identifies the data object. The object's attributes (e.g. metadata about the object) may be represented in one or more properties.


Object Type: Type of a data object (e.g., Person, Event, or Document). Object types may be defined by an ontology and may be modified or updated to include additional object types. An object definition (e.g., in an ontology) may include how the object is related to other objects, such as being a sub-object type of another object type (e.g. an agent may be a sub-object type of a person object type), and the properties the object type may have.


Properties: Attributes of a data object that represent individual data items. At a minimum, each property of a data object has a property type and a value or values. May also be referred to herein as “attributes.”


Property Type: The type of data a property is, such as a string, an integer, or a double. Property types may include complex property types, such as a series data values associated with timed ticks (e.g. a time series), etc. May also be referred to herein as “attribute types.”


Property Value: The value associated with a property, which is of the type indicated in the property type associated with the property. A property may have multiple values. May also be referred to herein as “attribute values.”


Link: A connection between two data objects, based on, for example, a relationship, an event, and/or matching properties. Links may be directional, such as one representing a payment from person A to B, or bidirectional.


Link Set: Set of multiple links that are shared between two or more data objects.


Example Devices/Systems


By way of example, FIG. 1 is a block diagram that illustrates an implementation of an electronic device 100, which, as described above, can comprise one or more electronic devices. Electronic device 100 includes a bus 102 or other communication mechanism for communicating information, and one or more hardware processors 104, coupled with bus 102 for processing information. One or more hardware processors 104 can be, for example, one or more microprocessors.


Electronic device 100 also includes a main memory 106, such as a random access memory (RAM) or other dynamic storage device, coupled to bus 102 for storing information and instructions to be executed by processor 104. Main memory 106 also can be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 104. Such instructions, when stored in non-transitory storage media accessible to one or more processors 104, render electronic device 100 into a special-purpose machine that is customized to perform the operations specified in the instructions.


Electronic device 100 further includes a read only memory (ROM) 108 or other static storage device coupled to bus 102 for storing static information and instructions for processor 104. A storage device 110, such as a magnetic disk, optical disk, or USB thumb drive (Flash drive), etc., is provided and coupled to bus 102 for storing information and instructions.


Electronic device 100 can be coupled via bus 102 to a display 112, such as a cathode ray tube (CRT), an LCD display, or a touchscreen, for displaying information to a computer user. An input device 114, including alphanumeric and other keys, is coupled to bus 102 for communicating information and command selections to processor 104. Another type of user input device is cursor control 116, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 104 and for controlling cursor movement on display 112. The input device typically has two degrees of freedom in two axes, a first axis (for example, x) and a second axis (for example, y), that allows the device to specify positions in a plane. In some embodiments, the same direction information and command selections as cursor control may be implemented via receiving touches on a touch screen without a cursor.


Electronic device 100 can include a user interface module to implement a GUI that may be stored in a mass storage device as executable software codes that are executed by the one or more computing devices. This and other modules may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.


In general, the word “module,” as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, Lua, C, and C++. A software module can be compiled and linked into an executable program, installed in a dynamic link library, or written in an interpreted programming language such as, for example, BASIC, Perl, Python, or Ruby. It will be appreciated that software modules can be callable from other modules or from themselves, and/or can be invoked in response to detected events or interrupts. Software modules configured for execution on computing devices can be provided on a computer readable medium, such as a compact disc, digital video disc, flash drive, magnetic disc, or any other tangible medium, or as a digital download (and can be originally stored in a compressed or installable format that requires installation, decompression, or decryption prior to execution). Such software code can be stored, partially or fully, on a memory device of the executing computing device, for execution by the computing device. Software instructions can be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules can be comprised of connected logic units, such as gates and flip-flops, and/or can be comprised of programmable units, such as programmable gate arrays or processors. The modules or computing device functionality described herein are preferably implemented as software modules, but can be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage.


Electronic device 100 can implement the techniques and other features described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the electronic device causes or programs electronic device 100 to be a special-purpose machine. According to some embodiments, the techniques and other features described herein are performed by electronic device 100 in response to one or more processors 104 executing one or more sequences of one or more instructions contained in main memory 106. Such instructions can be read into main memory 106 from another storage medium, such as storage device 150. Execution of the sequences of instructions contained in main memory 106 causes processor 104 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry can be used in place of or in combination with software instructions.


The term “non-transitory media” as used herein refers to any media storing data and/or instructions that cause a machine to operate in a specific fashion. Such non-transitory media can comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 110. Volatile media includes dynamic memory, such as main memory 106. Common forms of non-transitory media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge, and networked versions of the same.


Non-transitory media is distinct from, but can be used in conjunction with, transmission media. Transmission media participates in transferring information between storage media. For example, transmission media includes coaxial cables, copper wire, and fiber optics, including the wires that comprise bus 102. Transmission media can also take the form of acoustic or light waves, such as those generated during radio wave and infrared data communications.


Various forms of media can be involved in carrying one or more sequences of one or more instructions to processor 104 for execution. For example, the instructions can initially be carried on a magnetic disk or solid-state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to electronic device 100 can receive the data on the telephone line and use an infrared transmitter to convert the data to an infrared signal. An infrared detector can receive the data carried in the infrared signal and appropriate circuitry can place the data on bus 102. Bus 102 carries the data to main memory 106, from which processor 104 retrieves and executes the instructions. The instructions received by main memory 106 can optionally be stored on storage device 150 either before or after execution by processor 104.


Electronic device 100 also includes a communication interface 118 coupled to bus 102. Communication interface 118 provides a two-way data communication coupling to a network link 120 that is connected to a local network 122. For example, communication interface 118 can be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 118 can be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links can also be implemented. In any such implementation, communication interface 118 sends and receives electrical, electromagnetic, or optical signals that carry digital data streams representing various types of information.


Network link 120 typically provides data communication through one or more networks to other data devices. For example, network link 120 can provide a connection through local network 122 to a host computer 124 or to data equipment operated by an Internet Service Provider (ISP) 126. ISP 126 in turn provides data communication services through the worldwide packet data communication network now commonly referred to as the “Internet” 128. Local network 122 and Internet 128 both use electrical, electromagnetic, or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 120 and through communication interface 118, which carry the digital data to and from electronic device 100, are example forms of transmission media.


Electronic device 110 can send messages and receive data, including program code, through the network(s), network link 120 and communication interface 118. In the Internet example, a server 130 might transmit a requested code for an application program through Internet 128, ISP 126, local network 122, and communication interface 118. The received code can be executed by processor 104 as it is received, and/or stored in storage device 150 or other non-volatile storage for later execution. In some embodiments, server 130 can provide information for being displayed on a display, such as display 112.


Example Data Model


With reference now to FIG. 2, and to provide a framework for the following discussion of specific systems and methods described herein, an example database system 210 using an ontology 205 will now be described. This description is provided for the purpose of providing an example and is not intended to limit the techniques to the example data model, the example database system, or the example database system's use of an ontology to represent information.


In one embodiment, a body of data is conceptually structured according to an object-centric data model represented by ontology 205. The conceptual data model is independent of any particular database used for durably storing one or more database(s) 209 based on the ontology 205. For example, each object of the conceptual data model may correspond to one or more rows in a relational database or an entry in Lightweight Directory Access Protocol (LDAP) database, or any combination of one or more databases.



FIG. 2 illustrates an object-centric conceptual data model according to an embodiment. An ontology 205, as noted above, may include stored information providing a data model for storage of data in the database 209. The ontology 205 may be defined by one or more object types, which may each be associated with one or more property types. At the highest level of abstraction, data object 201 is a container for information representing things in the world. For example, data object 201 can represent an entity such as a person, a place, an organization, a market instrument, or other noun. Data object 201 can represent an event that happens at a point in time or for a duration. Data object 201 can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object 201 is associated with a unique identifier that uniquely identifies the data object within the database system.


Different types of data objects may have different property types. For example, a “Person” data object might have an “Eye Color” property type and an “Event” data object might have a “Date” property type. Each property 203 as represented by data in the database system 210 may have a property type defined by the ontology 205 used by the database 205.


Objects may be instantiated in the database 209 in accordance with the corresponding object definition for the particular object in the ontology 205. For example, a specific monetary payment (e.g., an object of type “event”) of US$30.00 (e.g., a property of type “currency”) taking place on Mar. 27, 2009 (e.g., a property of type “date”) may be stored in the database 209 as an event object with associated currency and date properties as defined within the ontology 205.


The data objects defined in the ontology 205 may support property multiplicity. In particular, a data object 201 may be allowed to have more than one property 203 of the same property type. For example, a “Person” data object might have multiple “Address” properties or multiple “Name” properties.


Each link 202 represents a connection between two data objects 201. In one embodiment, the connection is either through a relationship, an event, or through matching properties. A relationship connection may be asymmetrical or symmetrical. For example, “Person” data object A may be connected to “Person” data object B by a “Child Of” relationship (where “Person” data object B has an asymmetric “Parent Of” relationship to “Person” data object A), a “Kin Of” symmetric relationship to “Person” data object C, and an asymmetric “Member Of” relationship to “Organization” data object X. The type of relationship between two data objects may vary depending on the types of the data objects. For example, “Person” data object A may have an “Appears In” relationship with “Document” data object Y or have a “Participate In” relationship with “Event” data object E. As an example of an event connection, two “Person” data objects may be connected by an “Airline Flight” data object representing a particular airline flight if they traveled together on that flight, or by a “Meeting” data object representing a particular meeting if they both attended that meeting. In one embodiment, when two data objects are connected by an event, they are also connected by relationships, in which each data object has a specific relationship to the event, such as, for example, an “Appears In” relationship.


As an example of a matching properties connection, two “Person” data objects representing a brother and a sister, may both have an “Address” property that indicates where they live. If the brother and the sister live in the same home, then their “Address” properties likely contain similar, if not identical property values. In one embodiment, a link between two data objects may be established based on similar or matching properties (e.g., property types and/or property values) of the data objects. These are just some examples of the types of connections that may be represented by a link and other types of connections may be represented; embodiments are not limited to any particular types of connections between data objects. For example, a document might contain references to two different objects. For example, a document may contain a reference to a payment (one object), and a person (a second object). A link between these two objects may represent a connection between these two entities through their co-occurrence within the same document.


Each data object 201 can have multiple links with another data object 201 to form a link set 204. For example, two “Person” data objects representing a husband and a wife could be linked through a “Spouse Of” relationship, a matching “Address” property, and one or more matching “Event” properties (e.g., a wedding). Each link 202 as represented by data in a database may have a link type defined by the database ontology used by the database.



FIG. 3 is a block diagram illustrating exemplary components and data that may be used in identifying and storing data according to an ontology. In this example, the ontology may be configured, and data in the data model populated, by a system of parsers and ontology configuration tools. In the embodiment of FIG. 3, input data 300 is provided to parser 302. The input data may comprise data from one or more sources. For example, an institution may have one or more databases with information on credit card transactions, rental cars, and people. The databases may contain a variety of related information and attributes about each type of data, such as a “date” for a credit card transaction, an address for a person, and a date for when a rental car is rented. The parser 302 is able to read a variety of source input data types and determine which type of data it is reading.


In accordance with the discussion above, the example ontology 205 comprises stored information providing the data model of data stored in database 209, and the ontology is defined by one or more object types 310, one or more property types 316, and one or more link types 330. Based on information determined by the parser 302 or other mapping of source input information to object type, one or more data objects 201 may be instantiated in the database 209 based on respective determined object types 310, and each of the objects 201 has one or more properties 203 that are instantiated based on property types 316. Two data objects 201 may be connected by one or more links 202 that may be instantiated based on link types 330. The property types 316 each may comprise one or more data types 318, such as a string, number, etc. Property types 316 may be instantiated based on a base property type 320. For example, a base property type 320 may be “Locations” and a property type 316 may be “Home.”


In an embodiment, a user of the system uses an object type editor 324 to create and/or modify the object types 310 and define attributes of the object types. In an embodiment, a user of the system uses a property type editor 326 to create and/or modify the property types 316 and define attributes of the property types. In an embodiment, a user of the system uses link type editor 328 to create the link types 330. Alternatively, other programs, processes, or programmatic controls may be used to create link types and property types and define attributes, and using editors is not required.


In an embodiment, creating a property type 316 using the property type editor 326 involves defining at least one parser definition using a parser editor 322. A parser definition comprises metadata that informs parser 302 how to parse input data 300 to determine whether values in the input data can be assigned to the property type 316 that is associated with the parser definition. In an embodiment, each parser definition may comprise a regular expression parser 304A or a code module parser 304B. In other embodiments, other kinds of parser definitions may be provided using scripts or other programmatic elements. Once defined, both a regular expression parser 304A and a code module parser 304B can provide input to parser 302 to control parsing of input data 300.


Using the data types defined in the ontology, input data 300 may be parsed by the parser 302 determine which object type 310 should receive data from a record created from the input data, and which property types 316 should be assigned to data from individual field values in the input data. Based on the object-property mapping 301, the parser 302 selects one of the parser definitions that is associated with a property type in the input data. The parser parses an input data field using the selected parser definition, resulting in creating new or modified data 303. The new or modified data 303 is added to the database 209 according to ontology 205 by storing values of the new or modified data in a property of the specified property type. As a result, input data 300 having varying format or syntax can be created in database 209. The ontology 205 may be modified at any time using object type editor 324, property type editor 326, and link type editor 328, or under program control without human use of an editor. Parser editor 322 enables creating multiple parser definitions that can successfully parse input data 300 having varying format or syntax and determine which property types should be used to transform input data 300 into new or modified input data 303.


In an embodiment, the sources of input data 300 may themselves be stored or made available. For example, a user may provide a source document containing input data 300 that is processed by parser 302 as described above. The source document may be stored in the database 209 and associated with data objects 201 that are instantiated when parsing the source document. The source document may be made available when accessing data objects 201 associated with it, e.g., via the graphical user interface described below. In another embodiment, a user may provide a reference to a source document, such as a uniform resource locator (URL) or an Application Programming Interface (API) call that provides access to the source document. The reference may be utilized by the parser 302 to access and parse the source document, and may be associated with data objects 201 that are instantiated when parsing the source document. In still another embodiment, an API may be provided that allows sources of input data 300 to provide their data in a specified format.


The properties, objects, and links (e.g. relationships) between the objects can be visualized using a graphical user interface (GUI). For example, FIG. 4 displays a user interface showing a graph representation 403 of relationships (including relationships and/or links 404, 405, 406, 407, 408, 409, 410, 411, 412, and 413) between the data objects (including data objects 421, 422, 423, 424, 425, 426, 427, 428, and 429) that are represented as nodes in the example of FIG. 4. In this embodiment, the data objects include person objects 421, 422, 423, 424, 425, and 426; a flight object 427; a financial account 428; and a computer object 429. In this example, each person node (associated with person data objects), flight node (associated with flight data objects), financial account node (associated with financial account data objects), and computer node (associated with computer data objects) may have relationships and/or links with any of the other nodes through, for example, other objects such as payment objects.


For example, in FIG. 4, relationship 404 is based on a payment associated with the individuals indicated in person data objects 421 and 423. The link 404 represents these shared payments (for example, the individual associated with data object 421 may have paid the individual associated with data object 423 on three occasions). The relationship is further indicated by the common relationship between person data objects 421 and 423 and financial account data object 428. For example, link 411 indicates that person data object 421 transferred money into financial account data object 428, while person data object 423 transferred money out of financial account data object 428. In another example, the relationships between person data objects 424 and 425 and flight data object 427 are indicated by links 406, 409, and 410. In this example, person data objects 424 and 425 have a common address and were passengers on the same flight data object 427. In an embodiment, further details related to the relationships between the various objects may be displayed. For example, links 411 and 412 may, in some embodiments, indicate the timing of the respective money transfers. In another example, the time of the flight associated with the flight data object 427 may be shown.


Relationships between data objects may be stored as links, or in some embodiments, as properties, where a relationship may be detected between the properties. In some cases, as stated above, the links may be directional. For example, a payment link may have a direction associated with the payment, where one person object is a receiver of a payment, and another person object is the payer of payment.


In addition to visually showing relationships between the data objects, the user interface may allow various other manipulations. For example, the objects within database 108 may be searched using a search interface 450 (e.g., text string matching of object properties), inspected (e.g., properties and associated data viewed), filtered (e.g., narrowing the universe of objects into sets and subsets by properties or relationships), and statistically aggregated (e.g., numerically summarized based on summarization criteria), among other operations and visualizations. As a further example, the objects within database 108 may be associated with source documents, as described above, and the source documents may be inspected or exported when viewing associated objects. In some embodiments, selecting a source document for inspection may cause display of a separate user interface (e.g., in a separate window, document viewer, browser, or other application).


Example User Interfaces


With reference now to FIGS. 5A-5F, examples of user interfaces 500a-f for presentation of chronology information will be described. FIGS. 5A-5F depict simplified interfaces for purposes of illustration, and user interfaces 500a-f may have more or fewer components within the scope of the present disclosure. As non-limiting examples, user interfaces 500 may be implemented as a web page displayed via a web browser, or as an application executing on a tablet, mobile, or other computing device, such as the electronic device 110 and display 112 of FIG. 1.


Turning now to FIG. 5A, the illustrated user interface 500a includes a timeline interface 502. The timeline interface 502 includes timeline 504, which displays event icons 506-520 in chronological order. In some embodiments timeline 504 may be interactive, such that inputs received from a touchscreen or input device cause the timeline to scroll, zoom, or change the amount of detail displayed with regard to the event icons 506-520.


Event icons 506-520 correspond to a set of events (e.g., representing data objects of the type “event” and/or representing relationships among data objects, as described above) that are relevant to an investigation or analysis. In some embodiments, some or all of the relevant events may be obtained or determined based on criteria, such as events associated with a particular source, activity, participant, or other attribute. In other embodiments, the events displayed in the timeline interface 502 may be identified via inputs received from a touchscreen or input device. For example, input indicating that button 522 has been tapped or clicked may cause display of a separate interface for adding an event to the timeline. In further embodiments, timeline interface 502 or list interface 552 may be responsive to drag-and-drop input, such that dragging and dropping an event onto the timeline interface 502 or list interface 552 causes the event to be added to the set of relevant events and an additional event icon to be added to the timeline 504. For example, in some embodiments data objects may be dragged from one user interface (such as a user interface similar to the user interface of FIG. 4) to another user interface (such as the user interface of FIG. 5A) to add an event corresponding to the dragged data object to the set of events.


In the illustrated example, the events for which icons 506-520 are displayed relate to insider trading. Such events may be identified, for example, based on keyword criteria (e.g., a common company or stock symbol) or participant criteria (e.g., participants WC and DD) within a relevant timeframe. Although insider trading events are depicted for purposes of example, the present disclosure is not limited to a particular type of investigation or analysis. As non-limiting examples, a chronology user interface may be used to analyze and present events related to law enforcement (criminal investigations, prosecutions, defenses, anti-money laundering investigations, investigations into events leading up to use of a weapon (e.g., by a law enforcement officer), etc.), cybersecurity, finance (stock portfolios, loans, etc.), health issues (treatment plans, clinical drug trials, disease vectors, etc.), public safety, product marketing, anti-bribery workflows, cyber fraud detection, cyber breach detection, triage and analysis, investigations into consumer churn and retention (e.g., macro analysis (e.g., all customers), micro analysis (e.g., customer specific), analysis of events/trends that lead to a customer (or group) adopting or leaving products, buying more products, etc.), and/or other events that may be presented and analyzed chronologically.


Events 506-520 may correspond to various event categories or types, and may be displayed in the timeline interface 502 accordingly. As a non-limiting example, event 506 may correspond to transmission or receipt of an email message, and may accordingly be presented in the timeline interface 502 with a “mail” symbol or other designation. As further non-limiting examples, events 512, 514, and 520 may correspond to financial transactions (e.g., stock trades) occurring at the indicated times, events 508 and 512 may correspond to telephone calls, and event 516 may correspond to a document (e.g., an earnings report or press release) that was issued on the date and time indicated by the line connecting event 516 to the timeline 504. In various embodiments, some or all of events 506-520 may be labeled according to an event category, participant, source, keyword, annotation, or other attribute. In some embodiments, event labels or icons may be selected in response to input. For example, input may be received indicating that event participants should be displayed as icons. Although a particular embodiment is illustrated in FIGS. 5A-5D, one skilled in the art will understand that aspects of the present disclosure include various labels and representations. For example, event icons may be displayed in color, with the icon indicating the event type and the colors indicating participants, such that a blue “dollar sign” symbol indicates a stock trade performed by participant WC, and a blue-green “telephone” symbol indicates a phone call between participants WC and DD.


In FIG. 5A, event 508 is displayed in a minimized format that does not include an event icon. Illustratively, timeline interface 502 may display events in a minimized format to conserve space. In some embodiments, certain events in the timeline may be identified as key events, and the timeline interface 502 may display only the key events in the larger format. Key events may be identified, for example, according to specified criteria, via inputs obtained via the user interface 500, or combinations thereof. In further embodiments, a number of minimized formats may be utilized, and a particular minimized format may be selected based on, for example, the number of events on the timeline 504, the scope of the timeline 504, the size of the display, or other factors.


In some embodiments, the appearance or format of event icons 506-520 may be modified or customized. For example, events may be color-coded according to event type, participants, source, or other criteria. Event icons 506-520 may further be customized to display different icon shapes or symbols, or to display annotations or other attributes associated with some or all of the events.


The illustrated user interface 500a further includes a list interface 552. The list interface 552 displays the set of relevant events 506-520 in a list format. In the depicted embodiment, non-limiting examples of various event attributes are displayed in columns 554-564. For example, column 554 contains an event type or category, such as an email message, phone call, or stock trade. The event categories displayed in column 554 may correspond to event icons displayed in the timeline interface 502. As a further example, column 556 contains a date or date range associated with the event, which may correspond to the position of events in the timeline interface 502. The timeline interface 502 may not display corresponding information for all columns in the list interface 552. In FIG. 5A, the title column 558, participant column 560, keywords column 562, and notes column 564 are examples of event attributes that may not be displayed in the timeline interface. In some embodiments, columns of the list interface 552 may be selectable, and selection of a column may cause the selected attribute to be displayed or highlighted in the timeline interface 502.


The list interface 552 further includes event rows 566-574, which correspond to event icons 506-514 in the timeline interface 502. The list interface 552 may be scrollable, and input to the list interface 552 may cause, for example, rows corresponding to event icons 516-520 to scroll into view. In some embodiments, the list interface 552 may enable input or modification of event attributes that are modifiable, such as keywords, notes, or other attributes associated with the event. Other attributes, such as the date and participants in a phone call or other events, may not be modifiable.


With reference now to FIG. 5B, an example interface 500b for selecting displayed events and adding related events to the chronology will be described. Interface 500b may be displayed, for example, as a modification or update to interface 500a. Accordingly, only the elements of interface 500b that differ relative to interface 500a will be described.


In FIG. 5B, the timeline interface 502 has obtained a selection input. The selection input is illustrated as a hand cursor in the depicted embodiment, and indicates selection of the event associated with event icon 510 of FIG. 5A. The timeline interface 502 of FIG. 5B has accordingly updated to display event icon 510′, highlighting the selection. Because row 570 of the list interface 552 of FIG. 5A also corresponds to the selected event, the list interface 552 has updated to include row 570′, which highlights the selected event in the list. In some embodiments, events associated with audio or video, such as the telephone call event associated with event icon 512, may present the associated audio or video, or may present user interface controls for accessing the audio or video, when the corresponding event is selected in the timeline interface 502 or list interface 504.


Selected row 570′ further includes a list of related events 576. As described in more detail below, the chronology system may identify one or more events related to the selected event. Events may be identified as being related to the selected event, for example, by having common attributes. For example, a selected event may be a telephone call between parties X and Y at a particular date and time, and the related events may be other telephone calls between X and Y at other times. In FIG. 5B, the list of related events 576 includes rows 578a and 578b, each of which contains the attributes of a related event. Event rows 578a and 578b also include buttons 580a and 580b for viewing additional information regarding the respective event, and buttons 582a and 582b for adding the event to the chronology.


In some embodiments, interface 500b may include controls for modifying or customizing the appearance of a selected event icon, as described above. For example, the timeline interface 502 may display controls for annotating the selected event, and may display an annotation for the selected event icon 510′. As a further example, the list interface 552 may display controls for editing the modifiable attributes of a selected event.



FIG. 5C depicts an example interface 500c for comparing timelines. Interface 500c may be displayed, for example, as a modification or update to interface 500a. Accordingly, only the elements of interface 500c that differ relative to interface 500a will be described.


Interface 500c includes a second timeline interface 502′, which displays a second set of event icons 532, 514′, and 536 relative to the timeline 504 presented by timeline interface 502. For example, in the embodiment illustrated in FIG. 5C, the first timeline interface 502 is displayed above the timeline 504 and the second timeline interface 502′ is displayed below the timeline 504, which is common to both interfaces. In other embodiments (not depicted in FIG. 5C), the second timeline interface 502′ may overlap the first timeline interface 502. For example, the second set of event icons 532, 514′, and 536 may be displayed using a different color, shape, font, size, style, or otherwise distinguished from the first set of event icons 506-520. In further embodiments, the second timeline interface 502′ may include its own timeline 504′ rather than sharing a common timeline 504 with the first timeline interface 502.


Event icons 532, 514′, and 536 correspond to a second set of events. The second set of events may be relevant to the same investigation or analysis as the first set of events, or in some embodiments may represent an alternative theory of the case. In the embodiment illustrated in FIG. 5C, the first timeline interface 502 includes a series of events suggestive of insider trading, and the second timeline interface 502′ presents an alternate set of events that suggest trading based on publicly available information. For example, event icon 532, displaying a “television” symbol, corresponds to a television program airing at the time indicated by its position on the timeline 504. Event icon 514′ corresponds to a stock trade that occurred during the airing of the television program, and further corresponds to event icon 514 of the first timeline interface 502. That is, the stock trade associated with event icons 514 and 514′ is included in both the first set and the second set of events. Event icon 516 corresponds to a second stock trade during the airing of the television program.


In some embodiments, interface 500c includes a second list interface 552′, which includes columns 554′-560′ and event rows 582, 574′, and 586. Event rows 582, 574′ and 586 correspond to the events associated with icons 532, 514′, and 516 respectively, and row 574′ in the second list interface 552′ corresponds to row 574 in the first list interface 552. In other embodiments, a single list interface 552 may include events from both timelines 502 and 502′. Illustratively, a single list interface 552 may display the events of the first timeline 502 distinctly from the events of the second timeline 502′, using elements such as colors, icons, fonts, and the like to visually distinguish between the event sets.


Illustratively, interface 500c may be utilized to compare timelines created by different analysts, to assess whether events from a first and second timeline should be combined into a single timeline, to compare two potential timelines of events (e.g., compare two different theories of a case or present counter arguments to a theory), and/or the like. In some embodiments, interface 500c may include controls or other user interface elements to facilitate merging two or more timelines, and may further assign attributes to timeline events based on, for example, the original source or creator of the timeline. For example, interface 500c may generate a merged timeline that highlights each event according to whether the event was originally included in the first timeline, the second timeline, or both timelines.


With reference now to FIG. 5D, an example interface 500d for displaying historical data will be described. Interface 500d may be displayed, for example, as a modification or update to interface 500a. Accordingly, only the elements of interface 500d that differ relative to interface 500a will be described.


Interface 500d includes an historical interface 582 for displaying and interacting with historical data—that is, data associated with times displayed on the timeline 504. In the illustrated embodiment, the historical interface 582 displays historical stock prices 584 and trading volumes 586 for Acme, Inc. stock, which is the stock that was traded in events 510, 514, and 520. As other, non-limiting examples, the historical interface 582 may display data such as health statistics, account balances, sales volumes, or other data that varies over time and in relation to the set of relevant events.


In some embodiments, selection of an event in the timeline interface 502 or list interface 552 may cause an update to the historical interface 582. For example, selecting event icon 516 may highlight trading volumes or stock prices on the associated date. Selection of a particular date in the historical interface 582 may further cause display of an event selection interface, as described above with reference to FIG. 5B, that displays events within an interval of the selected date and time, and that provides controls for viewing these events or for adding these events to the set of relevant events.



FIG. 5E depicts an example interface 500e for displaying event participant Stimelines. Interface 500e may be displayed, for example, as a variation or alternative to interface 500a. Accordingly, only the elements of interface 500e that differ relative to interface 500a will be described.


Interface 500e includes participant timelines 524 and 526 within its timeline interface 502. Participant timeline 524 includes event icons 506, 508, 510, 512, all of which correspond to events that include “WC” as a participant. Participant timeline 526 includes event icons 506, 508, 512, and 514, all of which correspond to events that include “DD” as a participant. Events having multiple participants, such as the telephone call corresponding to event icon 508, display a connection between the participants. The timeline 504 further includes event 516, which does not have any participants and accordingly is not included on either of the participant timelines 524 and 526.


Interface 500e may display more or fewer participant timelines within the scope of the present disclosure. For example, interface 500e may display three participant timelines and include events with three participants, or may display a single participant timeline in response to, e.g., input selecting the participant. In some embodiments, participants may be added or removed from the timeline interface 502 based on received inputs. As a non-limiting example, the interface 500e may display a single participant timeline for X, which includes an email exchange between X, Y and Z. Selecting the email event may thus cause generation and/or display of participant timelines for Y and Z.



FIG. 5F depicts an example interface 500f that combines elements of the timeline interface 502 and the list interface 552. The timeline 504 of FIG. 5F is displayed vertically as a calendar rather than horizontally as a line, and entries in the list interface 552 are spaced according to their positions in the calendar. In some embodiments, the timeline interface of FIG. 502 may be interactive, such that “zooming in” on the calendar may cause display of a shorter time interval and allow, e.g., viewing of timelines at the daily or hourly level.


Chronology user interfaces 500a-f are non-limiting examples, and the present disclosure encompasses further variations that are not illustrated in the drawings. As an example, a timeline interface 502 may display a different set of events than the associated list interface 552, or one interface may display a subset of the events displayed in the other interface. As a further example, events corresponding to a date or time range may be displayed as blocks or ranges within the timeline interface. Further, various features of the various user interfaces 500a-f described above may be combined in various ways.


In various embodiments, one or more of the user interfaces 500a-f described above may include various additional features.


In some embodiments, events (as displayed in the example user interfaces described above) may be stored in an optimized database and/or table, which may be separate from the database 209, for example. For example, when a data object is added to a timeline/list of a chronology user interface, certain properties/attributes associated with that data object may be extracted and stored in the optimized database in association with the event that is added to the user interface. Advantageously, some properties associated with a data object may not be relevant to the display of the event in a chronology user interface, and thus may not be stored in the optimized database. Thus, when the chronology user interface is generated based on events stored in the optimized database, fewer properties of the events may need to be read (speeding up accessing and processing time), and the stored information may be less than with the full data objects were stored.


In some implementations, such events stored in an optimized database may include links back to the respective data objects. Thus, via a chronology user interface (such as one or more of the user interfaces 500a-f), the user may select to view the data object associated with the event, and the system may link to the data object by automatically executing an application associated with the data object (e.g., generate a user interface similar to the user interface of FIG. 4). In various implementations, the user may manually link an event to a data object, the user may manually enter an event (including its associated properties/attributes), the user may manually cause an event to be added via selection of a data object, the system many automatically add events associated with data objects, and/or the like.


In some implementations, multiple users may collaborate with one another via chronology user interfaces. For example, two users, via two different computing devices, may view a same chronology, and/or events associated with a same chronology. Either user may, for example, reorder or edit events and/or properties associated with events (e.g., a user may add comments or keywords to, or otherwise annotate, an event). In response the system may automatically propagate those changes to the user interface of the other user, such that the events viewed by either collaborating user may be updated automatically.


In some implementations, access to a chronology may be controlled. For example, a password may be required to access the chronology, or an access control list (ACL) may be used. An ACL is a set of one or more permissions that govern use (e.g., read access, write access, etc.) of the associated chronologies, or that govern use of individual rows or events within a chronology. For example, a permission may specify a set of one or more trustees (e.g., a user) and one or more use rights (e.g., read access, write access, etc.) of those trustees with respect to the associated chronology. As another example, a permission may specify a minimum security classification level (e.g., unclassified, secret, top secret, or a particular job title or position) such that only users that possess the minimum security classification level can use (e.g., read, write, etc.) rows or events in a chronology that involve a particular subject or that originate from a particular source.


In some implementations, the user interfaces may include one or more search user interface elements or search user interfaces by which the user may search or filter events (e.g., using keywords). For example, the user may specify certain terms, and any event that is associated with that term (e.g., includes the term in an annotation or other property) may be displayed to the user in a results list, or the user interface may be updated to only display events (in the list and/or timeline) that are associated with that term.


In some implementations, events displayed in a user interface may be linked to an original source (e.g., a source of the information associated with the event, a data object associated with the event, etc.) such that any changes to that original source may be automatically updated and reflected in the user interface. For example, new information may become available indicating a change to a time or date associated with an event. Accordingly, the system may determine that this change has been made, and automatically update the user interface to indicate the change. For example, the event may move on the timeline, or the list may be reordered. The user may be notified of the change via an indication in the user interface, and/or an alert may be sent to the user, as described below.


Additional examples of user interfaces and user interface features, and system for generating user interfaces, as described herein, may be found in U.S. patent application Ser. No. 14/841,338, filed Aug. 31, 2015, and titled “SYSTEM FOR PROVIDING DYNAMIC LINKED PANELS IN USER INTERFACE,” the entire disclosure of which is hereby made part of this specification as if set forth fully herein and incorporated by reference for all purposes, for all that it contains.


Further examples of user interfaces and user interface features, and systems for generating user interfaces with timelines and linking data objects may be found in U.S. patent application Ser. No. 14/670,089, filed Mar. 26, 2015, and titled “TEMPORAL REPRESENTATION OF STRUCTURED INFORMATION IN AN OBJECT MODEL,” the entire disclosure of which is hereby made part of this specification as if set forth fully herein and incorporated by reference for all purposes, for all that it contains.


Additional examples of Access Control Lists and ACL features, and systems for generating ACLs as described herein, may be found in U.S. Pat. No. 8,527,461 entitled “CROSS-ACL MULTI-MASTER REPLICATION” filed Nov. 27, 2012, U.S. Pat. No. 8,688,749 entitled “CROSS-ONTOLOGY MULTI-MASTER REPLICATION” filed Mar. 31, 2011, U.S. Pat. No. 9,081,975 entitled “SHARING INFORMATION BETWEEN NEXUSES THAT USE DIFFERENT CLASSIFICATION SCHEMES FOR INFORMATION ACCESS CONTROL” filed Oct. 22, 2012, and U.S. Pat. No. 8,838,538 entitled “TECHNIQUES FOR REPLICATING CHANGES TO ACCESS CONTROL LISTS ON INVESTIGATIVE ANALYSIS DATA” filed Jul. 31, 2013, the entire disclosure of each of which is hereby made part of this specification as if set forth fully herein and incorporated by reference for all purposes, for all that it contains.


Additional examples of user interfaces and user interface features, similar to those described herein, may be found in Appendix A of U.S. Provisional Application No. 62/272,526, the entire disclosure of which is hereby made part of this specification as if set forth fully herein and incorporated by reference for all purposes, for all that it contains.


In reference to Appendix A of U.S. Provisional Application No. 62/272,526, pages 2-12 illustrate examples of alternative designs of user interfaces similar to those described above. Pages 13-17 illustrate examples of alternative user interface designs and/or exported documents (e.g., exports of the chronology user interfaces) similar to the user interfaces described above. Pages 18-110 illustrate additional examples of alternative designs and functionality of user interfaces similar to those described above.


Example Methods/Routines


With reference now to FIG. 6, an example routine 600 for generation of a chronological interface will be described. At block 602, a set of events may be obtained. The set of events may illustratively include thousands or millions of events, only a portion of which may be relevant to an investigation or analysis. For example, the set of events may include all stock trades carried out within a particular timeframe, a log of all email messages sent or received, and so forth. Events may illustratively be obtained from a data store, such as the database 209 described above with reference to FIG. 2, and may correspond to data objects of a particular type (e.g., an “event” type) defined in an ontology 205. In some embodiments, a link may be maintained between the obtained events and their representations in the database 209. In other embodiments, one or more events may be obtained based on input data, and may further be linked to corresponding data objects in the database 209.


At block 604, criteria may be obtained for identifying a set of relevant events within the larger set of events. For example, the criteria may specify financial transactions involving certain participants, emails that include certain keywords, events on or about a particular date, or various other criteria. In some embodiments, the criteria may be inputs indicating selection of events. For example, the criteria may be a series of drag-and-drop inputs or other inputs indicating that an event should be added to the timeline. At block 606, a set of relevant events may be identified based on the criteria obtained at block 604.


At block 608, a timeline interface, such as the timeline interface 502 as described above, may be generated for display. Illustratively, the set of relevant events obtained at block 606 may be organized into a timeline according to their date and time attributes, and may be displayed as icons or symbols on the timeline as described above. At block 610 a list interface, such as the list interface 552, may be generated for display. One skilled in the art will appreciate that blocks 608 and 610 may be carried out independently of each other, and that the blocks may be carried out in parallel or in any order. At block 612, the timeline and list interfaces may be displayed.


In some embodiments, blocks 604-612 may be carried out repeatedly to generate updated timeline interfaces and list interfaces, or to generate additional timeline interfaces and list interfaces for display, as described above with reference to FIGS. 5B and 5C.


With reference now to FIG. 7, an example routine 700 for processing related events will be described. At block 702, an event may be obtained. For example, an event may be obtained based on input indicating a selection in the timeline interface 502 or the list interface 552.


At block 704, a set of related events may be determined. Illustratively, the set of related events may be determined based on further input, such as a request to display other telephone calls made within a specified timeframe of a selected telephone call event. In some embodiments, the set of related events may be determined automatically. For example, the set of relevant events may be analyzed to determine the relevance of a particular selected event, based on attributes or other criteria that the relevant events have in common. Events may be identified as potentially related if they are tagged with a certain keyword (such as “Acme, Inc.” in the examples above), involve certain participants, or occur relatively closely in time to events in the set of relevant events. The analysis may further consider key events in the timeline as indications of greater relevance. For example, stock trades that occur just before an earnings report may be identified as potentially related to the set of relevant events. In some embodiments, machine learning may be used to refine the automatic identification of related events based on previous manual selections of related events. For example, weighting factors may be applied to bias identification of related events in favor of events similar to those selected for inclusion in the timeline and list interfaces. In some embodiments, the identified events may be assigned a score or grade indicating a predicted likelihood of being related to the set of relevant events.


At block 706, a related event interface may be generated for display, as described above with reference to FIG. 5B. At block 708, the related event interface may be displayed.


Thereafter, at decision block 710, a determination may be made that one or more of the related events should be added to the set of relevant events. Illustratively, the determination may be made based on input selecting a related event for addition to the timeline, on scores or grades assigned to the related events, or based on other criteria. If the determination is that no related events should be added to the set, the routine 700 ends. Otherwise, the routine 700 branches to block 712, where an updated timeline interface may be generated that includes the newly added relevant event. At block 714, which may precede, follow, or be carried out in parallel with block 712, a list interface may be generated that includes the new relevant event. At block 716, the updated interfaces may be displayed, and thereafter the routine ends. In some embodiments, blocks 704-716 may be carried out iteratively to update the list of related events after a modification to the set of relevant events.


In some embodiments, an alert and/or notification may be automatically generated when data objects are updated, in response to changes to one or more events, in response to various changes in user interfaces, and/or the like, and the alert may be transmitted to the device operated by the entity associated with the alert and/or notification. The alert and/or notification can be transmitted at the time that the alert and/or notification is generated or at some determined time after generation of the alert and/or notification. When received by the device, the alert and/or notification can cause the device to display the alert and/or notification via the activation of an application on the device (e.g., a browser, a mobile application, etc.). For example, receipt of the alert and/or notification may automatically activate an application on the device, such as a messaging application (e.g., SMS or MMS messaging application), a standalone application (e.g., a chronology application), or a browser, for example, and display information included in the alert and/or notification. If the device is offline when the alert and/or notification is transmitted, the application may be automatically activated when the device is online such that the alert and/or notification is displayed. As another example, receipt of the alert and/or notification may cause a browser to open and be redirected to a login page generated by the chronology system so that the entity can log in to the chronology system and view the alert and/or notification. Alternatively, the alert and/or notification may include a URL of a webpage (or other online information) associated with the alert and/or notification, such that when the device (e.g., a mobile device) receives the alert, a browser (or other application) is automatically activated and the URL included in the alert and/or notification is accessed via the Internet. Alerts may include, for example, notifications to a user regarding a change to an event (e.g., by another user that is collaborating with the user), notifications to a user regarding new data objects that represent events related to events in a chronology, and/or the like.


Additional Embodiments

Embodiments of the present disclosure have been described herein with reference to numerous specific details that can vary from implementation to implementation. Certain adaptations and modifications of the described embodiments can be made. Other embodiments can be apparent to those skilled in the art from consideration of the specification and practice of the embodiments disclosed herein. It is intended that the specification and examples be considered as example only, with a true scope and spirit of the present disclosure being indicated by the following claims. It is also intended that the sequence of steps shown in figures are only for illustrative purposes and are not intended to be limited to any particular sequence of steps. As such, it is appreciated that these steps can be performed in a different order while implementing the example methods or processes disclosed herein. Further, the inventions illustratively disclosed herein suitably may be practiced in the absence of any element or aspect which is not specifically disclosed herein, such as without.


Depending on the embodiment, certain acts, events, or functions of any of the methods described herein can be performed in a different sequence, can be added, merged, or left out altogether (e.g., not all described acts or events are necessary for the practice of the algorithm). Moreover, in certain embodiments, acts or events can be performed concurrently, e.g., through multi-threaded processing, interrupt processing, or multiple processors or processor cores or on other parallel architectures, rather than sequentially.


The various illustrative logical blocks, modules and method elements described in connection with the embodiments disclosed herein can be implemented as electronic hardware, computer software or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. The described functionality can be implemented in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the disclosure.


The various illustrative logical blocks and modules described in connection with the embodiments disclosed herein can be implemented or performed by a machine, such as a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general purpose processor can be a microprocessor, but in the alternative, the processor can be a controller, microcontroller, or state machine, combinations of the same, or the like. A processor can also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.


The elements of a method, process, or algorithm described in connection with the embodiments disclosed herein can be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module can reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, a removable disk, a CD-ROM or any other form of computer-readable storage medium known in the art. A storage medium can be coupled to the processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium can be integral to the processor. The processor and the storage medium can reside in an ASIC. The ASIC can reside in a user terminal. In the alternative, the processor and the storage medium can reside as discrete components in a user terminal.


Conditional language used herein, such as, among others, “can,” “might,” “may,” “e.g.” and the like, unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or states. Thus, such conditional language is not generally intended to imply that features, elements and/or states are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without author input or prompting, whether these features, elements and/or states are included or are to be performed in any particular embodiment. The terms “comprising,” “including,” “having,” “involving” and the like are synonymous and are used inclusively, in an open-ended fashion, and do not exclude additional elements, features, acts, operations and so forth. For example, a general purpose computer comprising one or more processors should not be interpreted as excluding other computer components, and may possibly include such components as memory, input/output devices, and/or network interfaces, among others. Also, the term “or” is used in its inclusive sense (and not in its exclusive sense) so that when used, for example, to connect a list of elements, the term “or” means one, some or all of the elements in the list.


Disjunctive language such as the phrase “at least one of X, Y or Z,” unless specifically stated otherwise, is otherwise understood with the context as used in general to present that an item, term, etc., may be either X, Y or Z, or any combination thereof (e.g., X, Y and/or Z). Thus, such disjunctive language is not generally intended to, and should not, imply that certain embodiments require at least one of X, at least one of Y or at least one of Z to each be present.


Unless otherwise explicitly stated, articles such as “a” or “an” should generally be interpreted to include one or more described items. Accordingly, phrases such as “a device configured to” are intended to include one or more recited devices. Such one or more recited devices can also be collectively configured to carry out the stated recitations. For example, “a processor configured to carry out recitations A, B and C” can include a first processor configured to carry out recitation A working in conjunction with a second processor configured to carry out recitations B and C.


The term “comprising” as used herein should be given an inclusive rather than exclusive interpretation. For example, a general purpose computer comprising one or more processors should not be interpreted as excluding other computer components, and may possibly include such components as memory, input/output devices, and/or network interfaces, among others.


While the above detailed description has shown, described, and pointed out novel features as applied to various embodiments, it will be understood that various omissions, substitutions, and changes in the form and details of the devices or algorithms illustrated can be made without departing from the spirit of the disclosure. As will be recognized, certain embodiments described herein can be embodied within a form that does not provide all of the features and benefits set forth herein, as some features can be used or practiced separately from others. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.

Claims
  • 1. A system comprising: a data store configured to store computer-executable instructions; anda processor, wherein the computer-executable instructions, when executed, configure the processor to: obtain a plurality of events, each of the plurality of events comprising event attributes, the event attributes comprising at least an event timestamp;obtain one or more criteria associated with event attributes;determine, based at least in part on the one or more criteria, a set of relevant events within the plurality of events;determine, based at least in part on the one or more criteria, that one or more events within the set of relevant events are key events;generate for display a user interface based at least in part on the set of relevant events and the corresponding event attributes, the user interface comprising: a timeline interface for display of a timeline, wherein positions of individual relevant events within the timeline interface are determined based at least in part on respective event timestamps, and wherein the timeline interface displays the one or more events as key events; wherein the timeline is displayed as a horizontal line and time increments displayed in short vertical or hash indications intersecting the horizontal line,wherein key events are displayed as icons offset from the horizontal line and connected to the line by vertical lines intersecting the horizontal line at a position corresponding to the timestamps and at a position corresponding proportionally to the time increments;wherein the key event icons comprise indications of event type;a list interface for display of the set of relevant events, wherein the list interface is displayed in rows and columns with a single row corresponding to a single one of the plurality of events, and wherein additional information corresponding to the event is displayed in columns;cause display of the user interface; anddisplay, in response to a user selection of an event icon, a highlighted row corresponding to an event and comprising related events not displayed on the timeline, wherein each related event is separate from the highlighted event and has a different timestamp as the highlighted event, and wherein the related event is determined to be related based on having one or more event attributes in common with the highlighted event.
  • 2. The system of claim 1, wherein the event attributes further comprise at least one of an event source, an event type, an event icon, an event duration, an event title, a set of event participants, a set of event keywords, and a set of event annotations.
  • 3. The system of claim 1, wherein the processor is further configured to: obtain input identifying a selected event displayed within the user interface;generate for display a first update to the user interface, the first update comprising: an indication of selection of the selected event within the timeline interface; andan indication of selection of the selected event within the list interface; andcause display of the first update to the user interface.
  • 4. The system of claim 1, wherein the processor is further configured to: determine that at least one of the plurality of events has been updated;generate an alert indicating that the at least one of the plurality of events has been updated; andcause transmission of the alert to a computing device, wherein transmission of the alert causes the computing device to activate an application for display of the alert, and wherein the alert enables display of an updated user interface, the updated user interface based at least in part on an updated event of the plurality of events.
  • 5. The system of claim 3, wherein the processor is further configured to: determine, based at least in part on the selected event, one or more related events from within the plurality of events;generate for display an event selection interface for display of the one or more related events; andcause display of the event selection interface.
  • 6. The system of claim 5, wherein the processor configured to determine the one or more related events is configured to determine the one or more related events based at least in part on one or more attributes of the selected event.
  • 7. The system of claim 5, wherein the processor is further configured to: obtain input identifying a related event within the one or more related events, the input indicating that the related event should be added to the set of relevant events;generate for display a second update to the user interface, the second update comprising: display of the related event within the timeline interface; anddisplay of the related event within the list interface; andcause display of the second update to the user interface.
  • 8. A computer-implemented method comprising: obtaining a plurality of events, each of the plurality of events comprising event attributes;determining, based at least in part on one or more criteria associated with event attributes, a set of relevant events within the plurality of events;determining, based at least in part on the one or more criteria associated with event attributes, that a subset of the set of relevant events are key events;generating for display a user interface based at least in part on the set of relevant events, the user interface comprising: a timeline interface for display of a timeline, wherein positions of individual relevant events within the timeline interface are determined based at least in part on respective event attributes, wherein the timeline interface displays the subset of the set of relevant events as key events, wherein the timeline is displayed as a horizontal line and time increments displayed in short vertical or hash indications intersecting the horizontal line, wherein key events are displayed as icons offset from the horizontal line and connected to the line by vertical lines at a position corresponding to event timestamps and to the time increments, and wherein the key event icons comprise indications of event type; anda list interface for display of the set of relevant events, wherein the list interface is displayed in a table with a designated portion of the table corresponding to a single one of the plurality of events, and wherein additional information corresponding to the event is displayed in the table and not displayed on the timeline;displaying the user interface; anddisplaying, in response to a user selection of an event icon, a highlighted row corresponding to an event and comprising related events not displayed on the timeline, wherein each related event is separate from the highlighted event and has a different timestamp as the highlighted event, and wherein the related event is determined to be related based on having one or more event attributes in common with the highlighted event.
  • 9. The computer-implemented method of claim 8, wherein the event attributes include at least one configurable attribute.
  • 10. The computer-implemented method of claim 9, wherein the at least one configurable attribute comprises an event annotation, an event icon, an event description, or an event keyword.
  • 11. The computer-implemented method of claim 9 further comprising: obtaining an input indicative of a modification of a configurable attribute for a relevant event;modifying the configurable attribute for the relevant event;generating for display a first update to the user interface, the first update comprising at least one of a modified configurable attribute within the timeline interface and a modified configurable attribute within the list interface; andcausing display of the first update to the user interface.
  • 12. The computer-implemented method of claim 8 further comprising: obtaining one or more updates to the plurality of events;determining, based at least in part on the one or more updates, an update to the set of relevant events;generating for display an alert message, the alert message indicating the update to the set of relevant events; anddisplaying the alert message.
  • 13. The computer-implemented method of claim 12 further comprising: generating for display an updated user interface based at least in part on the update to the set of relevant events, the updated user interface comprising an updated timeline interface and an updated list interface; anddisplaying the updated user interface.
  • 14. A non-transitory computer-readable storage medium including computer-executable instructions that, when executed by a processor, cause the processor to: determine, based at least in part on a first set of criteria, a first set of relevant events within a plurality of events;determine, based at least in part on the first set of criteria, that a first subset of the first set of relevant events are key events;generate a first timeline corresponding to the first set of relevant events;generate for display a first user interface based at least in part on the first set of relevant events, the first user interface comprising: a timeline interface for display of at least a portion of the first timeline, wherein individual relevant events are positioned within the timeline interface based at least in part on respective attributes of the individual relevant events, wherein the timeline interface displays the first subset as key events, wherein the timeline is displayed as a horizontal line with time increments indicated on the horizontal line, wherein events are displayed as icons offset from the horizontal line and connected to the horizontal line by other lines intersecting the horizontal line at a position corresponding to event timestamps and to the time increments, and wherein the event icons comprise indications of event type; anda list interface for display of at least the portion of the first set of relevant events, wherein the list interface is displayed in rows with a single row corresponding to a single one of the plurality of events, and wherein additional information not displayed on the timeline corresponding to the event is displayed in the row;cause display of the first user interface; anddisplay, in response to a user selection of an event icon, a highlighted row corresponding to an event and comprising related events not displayed on the timeline, wherein each related event is separate from the highlighted event and has a different timestamp as the highlighted event, and wherein the related event is determined to be related based on having one or more event attributes in common with the highlighted event.
  • 15. The non-transitory computer-readable medium of claim 14, wherein the first set of criteria is associated with event attributes, and wherein the first set of relevant events comprises events that satisfy the first set of criteria.
  • 16. The non-transitory computer-readable medium of claim 14, wherein the computer-executable instructions further cause the processor to: determine, based at least in part on a second set of criteria, a second set of relevant events within the plurality of events;determine, based at least in part on the second set of criteria, that a second subset of the second set of relevant events are key events;generate a second timeline corresponding to the second set of events;generate for display a second user interface based at least in part on the second set of relevant events; andcause display of the second user interface.
  • 17. The non-transitory computer-readable medium of claim 16, wherein the first set of relevant events and the second set of relevant events have at least one event in common.
  • 18. The non-transitory computer-readable medium of claim 16, wherein the second user interface comprises: a combined timeline interface for display of a combined timeline, the combined timeline comprising the first timeline and at least a portion of the second timeline;a first list interface for display of the first set of relevant events; anda second list interface for display of the second set of relevant events.
  • 19. The non-transitory computer-readable medium of claim 18, wherein the computer-executable instructions further cause the processor to generate the combined timeline based at least in part on the first timeline and the second timeline.
  • 20. The non-transitory computer-readable medium of claim 18, wherein the second user interface further comprises: at least a portion of the first set of relevant events positioned above the combined timeline; andat least a portion of the second set of relevant events positioned below the combined timeline.
CROSS-REFERENCE TO RELATED APPLICATION

This application is a non-provisional of and claims priority to U.S. Provisional Application No. 62/272,526, which is hereby incorporated by reference in its entirety.

US Referenced Citations (585)
Number Name Date Kind
5109399 Thompson Apr 1992 A
5329108 Lamoure Jul 1994 A
5632009 Rao et al. May 1997 A
5670987 Doi et al. Sep 1997 A
5781704 Rossmo Jul 1998 A
5798769 Chiu et al. Aug 1998 A
5845300 Comer Dec 1998 A
6057757 Arrowsmith et al. May 2000 A
6091956 Hollenberg Jul 2000 A
6161098 Wallman Dec 2000 A
6167405 Rosensteel, Jr. et al. Dec 2000 A
6219053 Tachibana et al. Apr 2001 B1
6232971 Haynes May 2001 B1
6247019 Davies Jun 2001 B1
6279018 Kudrolli et al. Aug 2001 B1
6341310 Leshem et al. Jan 2002 B1
6366933 Ball et al. Apr 2002 B1
6369835 Lin Apr 2002 B1
6456997 Shukla Sep 2002 B1
6549944 Weinberg et al. Apr 2003 B1
6560620 Ching May 2003 B1
6581068 Bensoussan et al. Jun 2003 B1
6594672 Lampson et al. Jul 2003 B1
6631496 Li et al. Oct 2003 B1
6642945 Sharpe Nov 2003 B1
6674434 Chojnacki Jan 2004 B1
6714936 Nevin, III Mar 2004 B1
6775675 Nwabueze et al. Aug 2004 B1
6820135 Dingman Nov 2004 B1
6828920 Owen et al. Dec 2004 B2
6839745 Dingari et al. Jan 2005 B1
6877137 Rivette et al. Apr 2005 B1
6976210 Silva et al. Dec 2005 B1
6980984 Huffman et al. Dec 2005 B1
6985950 Hanson et al. Jan 2006 B1
7036085 Barros Apr 2006 B2
7043702 Chi et al. May 2006 B2
7055110 Kupka et al. May 2006 B2
7139800 Bellotti et al. Nov 2006 B2
7158878 Rasmussen et al. Jan 2007 B2
7162475 Ackerman Jan 2007 B2
7168039 Bertram Jan 2007 B2
7171427 Witkowski et al. Jan 2007 B2
7269786 Malloy et al. Sep 2007 B1
7278105 Kitts Oct 2007 B1
7290698 Poslinski et al. Nov 2007 B2
7333998 Heckerman et al. Feb 2008 B2
7370047 Gorman May 2008 B2
7379811 Rasmussen et al. May 2008 B2
7379903 Caballero et al. May 2008 B2
7426654 Adams et al. Sep 2008 B2
7451397 Weber et al. Nov 2008 B2
7454466 Bellotti et al. Nov 2008 B2
7467375 Tondreau et al. Dec 2008 B2
7487139 Fraleigh et al. Feb 2009 B2
7502786 Liu et al. Mar 2009 B2
7525422 Bishop et al. Apr 2009 B2
7529727 Arning et al. May 2009 B2
7529734 Dirisala May 2009 B2
7558677 Jones Jul 2009 B2
7574409 Patinkin Aug 2009 B2
7574428 Leiserowitz et al. Aug 2009 B2
7579965 Bucholz Aug 2009 B2
7596285 Brown et al. Sep 2009 B2
7614006 Molander Nov 2009 B2
7617232 Gabbert et al. Nov 2009 B2
7620628 Kapur et al. Nov 2009 B2
7627812 Chamberlain et al. Dec 2009 B2
7634717 Chamberlain et al. Dec 2009 B2
7703021 Flam Apr 2010 B1
7706817 Bamrah et al. Apr 2010 B2
7712049 Williams et al. May 2010 B2
7716077 Mikurak May 2010 B1
7725530 Sah et al. May 2010 B2
7725547 Albertson et al. May 2010 B2
7730082 Sah et al. Jun 2010 B2
7730109 Rohrs et al. Jun 2010 B2
7770100 Chamberlain et al. Aug 2010 B2
7805457 Viola et al. Sep 2010 B1
7809703 Balabhadrapatruni et al. Oct 2010 B2
7818658 Chen Oct 2010 B2
7870493 Pall et al. Jan 2011 B2
7894984 Rasmussen et al. Feb 2011 B2
7899611 Downs et al. Mar 2011 B2
7917376 Bellin et al. Mar 2011 B2
7920963 Jouline et al. Apr 2011 B2
7933862 Chamberlain et al. Apr 2011 B2
7941321 Greenstein et al. May 2011 B2
7962281 Rasmussen et al. Jun 2011 B2
7962495 Jain et al. Jun 2011 B2
7962848 Bertram Jun 2011 B2
7970240 Chao et al. Jun 2011 B1
7971150 Raskutti et al. Jun 2011 B2
7984374 Caro et al. Jul 2011 B2
8001465 Kudrolli et al. Aug 2011 B2
8001482 Bhattiprolu et al. Aug 2011 B2
8010545 Stefik et al. Aug 2011 B2
8015487 Roy et al. Sep 2011 B2
8024778 Cash et al. Sep 2011 B2
8036632 Cona et al. Oct 2011 B1
8103543 Zwicky Jan 2012 B1
8134457 Velipasalar et al. Mar 2012 B2
8145703 Frishert et al. Mar 2012 B2
8185819 Sah et al. May 2012 B2
8196184 Amirov et al. Jun 2012 B2
8214361 Sandler et al. Jul 2012 B1
8214764 Gemmell et al. Jul 2012 B2
8225201 Michael Jul 2012 B2
8229947 Fujinaga Jul 2012 B2
8230333 Decherd et al. Jul 2012 B2
8271461 Pike et al. Sep 2012 B2
8280880 Aymeloglu et al. Oct 2012 B1
8290926 Ozzie et al. Oct 2012 B2
8290942 Jones et al. Oct 2012 B2
8301464 Cave et al. Oct 2012 B1
8301904 Gryaznov Oct 2012 B1
8312367 Foster Nov 2012 B2
8312546 Alme Nov 2012 B2
8352881 Champion et al. Jan 2013 B2
8368695 Howell et al. Feb 2013 B2
8397171 Klassen et al. Mar 2013 B2
8412707 Mianji Apr 2013 B1
8447722 Ahuja et al. May 2013 B1
8452790 Mianji May 2013 B1
8463036 Ramesh et al. Jun 2013 B1
8489331 Kopf et al. Jul 2013 B2
8489623 Jain et al. Jul 2013 B2
8489641 Seefeld et al. Jul 2013 B1
8498984 Hwang et al. Jul 2013 B1
8510743 Hackborn et al. Aug 2013 B2
8514082 Cova et al. Aug 2013 B2
8515207 Chau Aug 2013 B2
8554579 Tribble et al. Oct 2013 B2
8554653 Falkenborg et al. Oct 2013 B2
8554709 Goodson et al. Oct 2013 B2
8560413 Quarterman Oct 2013 B1
8577911 Stepinski et al. Nov 2013 B1
8589273 Creeden et al. Nov 2013 B2
8595234 Siripurapu et al. Nov 2013 B2
8620641 Farnsworth et al. Dec 2013 B2
8639757 Zang et al. Jan 2014 B1
8646080 Williamson et al. Feb 2014 B2
8676857 Adams et al. Mar 2014 B1
8689108 Duffield et al. Apr 2014 B1
8713467 Goldenberg et al. Apr 2014 B1
8726379 Stiansen et al. May 2014 B1
8739278 Varghese May 2014 B2
8742934 Sarpy et al. Jun 2014 B1
8744890 Bernier Jun 2014 B1
8745516 Mason et al. Jun 2014 B2
8781169 Jackson et al. Jul 2014 B2
8787939 Papakipos et al. Jul 2014 B2
8788407 Singh et al. Jul 2014 B1
8799799 Cervelli et al. Aug 2014 B1
8812960 Sun et al. Aug 2014 B1
8830322 Nerayoff et al. Sep 2014 B2
8832594 Thompson et al. Sep 2014 B1
8868486 Tamayo Oct 2014 B2
8868537 Colgrove et al. Oct 2014 B1
8917274 Ma et al. Dec 2014 B2
8924872 Bogomolov et al. Dec 2014 B1
8937619 Sharma et al. Jan 2015 B2
8938686 Erenrich et al. Jan 2015 B1
9009171 Grossman et al. Apr 2015 B1
9009827 Albertson et al. Apr 2015 B1
9021260 Falk et al. Apr 2015 B1
9021384 Beard et al. Apr 2015 B1
9043696 Meiklejohn et al. May 2015 B1
9043894 Dennison et al. May 2015 B1
9069842 Melby Jun 2015 B2
9116975 Shankar et al. Aug 2015 B2
9146954 Boe et al. Sep 2015 B1
9202249 Cohen et al. Dec 2015 B1
9223773 Isaacson Dec 2015 B2
9229952 Meacham et al. Jan 2016 B1
9250759 Commons Feb 2016 B1
9256664 Chakerian et al. Feb 2016 B2
9298678 Chakerian et al. Mar 2016 B2
9319288 Somaiya et al. Apr 2016 B2
9367872 Visbal et al. Jun 2016 B1
9626088 Ma et al. Apr 2017 B2
9646396 Sharma et al. May 2017 B2
20010021936 Bertram Sep 2001 A1
20020033848 Sciammarella et al. Mar 2002 A1
20020065708 Senay et al. May 2002 A1
20020091707 Keller Jul 2002 A1
20020095658 Shulman et al. Jul 2002 A1
20020116120 Ruiz et al. Aug 2002 A1
20020130907 Chi et al. Sep 2002 A1
20020174201 Ramer et al. Nov 2002 A1
20020194119 Wright et al. Dec 2002 A1
20030028560 Kudrolli et al. Feb 2003 A1
20030036848 Sheha et al. Feb 2003 A1
20030039948 Donahue Feb 2003 A1
20030140106 Raguseo Jul 2003 A1
20030144868 MacIntyre et al. Jul 2003 A1
20030163352 Surpin et al. Aug 2003 A1
20030172014 Quackenbush et al. Sep 2003 A1
20030200217 Ackerman Oct 2003 A1
20030225755 Iwayama et al. Dec 2003 A1
20030229848 Arend et al. Dec 2003 A1
20040032432 Baynger Feb 2004 A1
20040064256 Barinek et al. Apr 2004 A1
20040085318 Hassler et al. May 2004 A1
20040095349 Bito et al. May 2004 A1
20040103124 Kupkova May 2004 A1
20040111410 Burgoon et al. Jun 2004 A1
20040126840 Cheng et al. Jul 2004 A1
20040143602 Ruiz et al. Jul 2004 A1
20040143796 Lerner et al. Jul 2004 A1
20040160309 Stilp Aug 2004 A1
20040163039 Gorman Aug 2004 A1
20040181554 Heckerman et al. Sep 2004 A1
20040193600 Kaasten et al. Sep 2004 A1
20040205524 Richter et al. Oct 2004 A1
20040221223 Yu et al. Nov 2004 A1
20040260702 Cragun et al. Dec 2004 A1
20040267746 Marcjan et al. Dec 2004 A1
20050027705 Sadri et al. Feb 2005 A1
20050028094 Allyn Feb 2005 A1
20050039119 Parks et al. Feb 2005 A1
20050065811 Chu et al. Mar 2005 A1
20050078858 Yao et al. Apr 2005 A1
20050080769 Gemmell Apr 2005 A1
20050086207 Heuer et al. Apr 2005 A1
20050125715 Franco et al. Jun 2005 A1
20050154628 Eckart et al. Jul 2005 A1
20050154769 Eckart et al. Jul 2005 A1
20050162523 Darrell et al. Jul 2005 A1
20050166144 Gross Jul 2005 A1
20050180330 Shapiro Aug 2005 A1
20050182793 Keenan et al. Aug 2005 A1
20050183005 Denoue et al. Aug 2005 A1
20050210409 Jou Sep 2005 A1
20050246327 Yeung et al. Nov 2005 A1
20050251786 Citron et al. Nov 2005 A1
20060026120 Carolan et al. Feb 2006 A1
20060026170 Kreitler et al. Feb 2006 A1
20060045470 Poslinski et al. Mar 2006 A1
20060059139 Robinson Mar 2006 A1
20060074866 Chamberlain et al. Apr 2006 A1
20060074881 Vembu et al. Apr 2006 A1
20060080619 Carlson et al. Apr 2006 A1
20060093222 Saffer et al. May 2006 A1
20060129746 Porter Jun 2006 A1
20060139375 Rasmussen et al. Jun 2006 A1
20060142949 Helt Jun 2006 A1
20060143034 Rothermel Jun 2006 A1
20060149596 Surpin et al. Jul 2006 A1
20060184889 Molander Aug 2006 A1
20060203337 White Sep 2006 A1
20060209085 Wong et al. Sep 2006 A1
20060218637 Thomas et al. Sep 2006 A1
20060241974 Chao et al. Oct 2006 A1
20060242040 Rader Oct 2006 A1
20060242630 Koike et al. Oct 2006 A1
20060271277 Hu et al. Nov 2006 A1
20060279630 Aggarwal et al. Dec 2006 A1
20070011150 Frank Jan 2007 A1
20070016363 Huang et al. Jan 2007 A1
20070038646 Thota Feb 2007 A1
20070038962 Fuchs et al. Feb 2007 A1
20070057966 Ohno et al. Mar 2007 A1
20070078832 Ott et al. Apr 2007 A1
20070083541 Fraleigh et al. Apr 2007 A1
20070094389 Nussey et al. Apr 2007 A1
20070150369 Zivin Jun 2007 A1
20070174760 Chamberlain et al. Jul 2007 A1
20070192265 Chopin et al. Aug 2007 A1
20070198571 Ferguson et al. Aug 2007 A1
20070208497 Downs et al. Sep 2007 A1
20070208498 Barker et al. Sep 2007 A1
20070208736 Tanigawa et al. Sep 2007 A1
20070233709 Abnous Oct 2007 A1
20070240062 Christena et al. Oct 2007 A1
20070250784 Riley Oct 2007 A1
20070266336 Nojima et al. Nov 2007 A1
20070294200 Au Dec 2007 A1
20070294643 Kyle Dec 2007 A1
20080016216 Worley et al. Jan 2008 A1
20080040275 Paulsen et al. Feb 2008 A1
20080040684 Crump Feb 2008 A1
20080051989 Welsh Feb 2008 A1
20080052142 Bailey et al. Feb 2008 A1
20080077597 Butler Mar 2008 A1
20080077642 Carbone et al. Mar 2008 A1
20080082486 Lermant et al. Apr 2008 A1
20080104019 Nath May 2008 A1
20080126951 Sood et al. May 2008 A1
20080148398 Mezack et al. Jun 2008 A1
20080155440 Trevor et al. Jun 2008 A1
20080195417 Surpin et al. Aug 2008 A1
20080195608 Clover Aug 2008 A1
20080222295 Robinson et al. Sep 2008 A1
20080243711 Aymeloglu et al. Oct 2008 A1
20080249983 Meisels et al. Oct 2008 A1
20080255973 El Wade et al. Oct 2008 A1
20080263468 Cappione et al. Oct 2008 A1
20080267107 Rosenberg Oct 2008 A1
20080276167 Michael Nov 2008 A1
20080278311 Grange et al. Nov 2008 A1
20080281819 Tenenbaum et al. Nov 2008 A1
20080288306 Maclntyre et al. Nov 2008 A1
20080288475 Kim et al. Nov 2008 A1
20080294994 Kruger et al. Nov 2008 A1
20080301643 Appleton et al. Dec 2008 A1
20090002492 Velipasalar et al. Jan 2009 A1
20090027418 Maru et al. Jan 2009 A1
20090030915 Winter et al. Jan 2009 A1
20090055251 Shah et al. Feb 2009 A1
20090070162 Leonelli et al. Mar 2009 A1
20090076845 Bellin et al. Mar 2009 A1
20090088964 Schaaf et al. Apr 2009 A1
20090094166 Aymeloglu et al. Apr 2009 A1
20090119309 Gibson et al. May 2009 A1
20090125359 Knapic May 2009 A1
20090125369 Kloostra et al. May 2009 A1
20090125459 Norton et al. May 2009 A1
20090132921 Hwangbo et al. May 2009 A1
20090132953 Reed et al. May 2009 A1
20090143052 Bates et al. Jun 2009 A1
20090144262 White et al. Jun 2009 A1
20090144274 Fraleigh et al. Jun 2009 A1
20090150854 Elaasar et al. Jun 2009 A1
20090164934 Bhattiprolu et al. Jun 2009 A1
20090171939 Athsani et al. Jul 2009 A1
20090172511 Decherd et al. Jul 2009 A1
20090172821 Daira et al. Jul 2009 A1
20090177962 Gusmorino et al. Jul 2009 A1
20090179892 Tsuda et al. Jul 2009 A1
20090187464 Bai et al. Jul 2009 A1
20090192957 Subramanian et al. Jul 2009 A1
20090222400 Kupershmidt et al. Sep 2009 A1
20090222759 Drieschner Sep 2009 A1
20090222760 Halverson et al. Sep 2009 A1
20090234720 George et al. Sep 2009 A1
20090249244 Robinson et al. Oct 2009 A1
20090254970 Agarwal et al. Oct 2009 A1
20090281839 Lynn et al. Nov 2009 A1
20090287470 Farnsworth et al. Nov 2009 A1
20090292626 Oxford Nov 2009 A1
20090327208 Bittner et al. Dec 2009 A1
20100011282 Dollard et al. Jan 2010 A1
20100030722 Goodson et al. Feb 2010 A1
20100042922 Bradateanu et al. Feb 2010 A1
20100057716 Stefik et al. Mar 2010 A1
20100070523 Delgo et al. Mar 2010 A1
20100070842 Aymeloglu et al. Mar 2010 A1
20100070845 Facemire et al. Mar 2010 A1
20100070897 Aymeloglu et al. Mar 2010 A1
20100100963 Mahaffey Apr 2010 A1
20100103124 Kruzeniski et al. Apr 2010 A1
20100106752 Eckardt et al. Apr 2010 A1
20100114887 Conway et al. May 2010 A1
20100122152 Chamberlain et al. May 2010 A1
20100131457 Heimendinger May 2010 A1
20100162176 Dunton Jun 2010 A1
20100185691 Irmak et al. Jul 2010 A1
20100191563 Schlaifer et al. Jul 2010 A1
20100198684 Eraker et al. Aug 2010 A1
20100199225 Coleman et al. Aug 2010 A1
20100228812 Uomini Sep 2010 A1
20100250412 Wagner Sep 2010 A1
20100280857 Liu et al. Nov 2010 A1
20100293174 Bennett et al. Nov 2010 A1
20100306029 Jolley Dec 2010 A1
20100306713 Geisner et al. Dec 2010 A1
20100313119 Baldwin et al. Dec 2010 A1
20100318924 Frankel et al. Dec 2010 A1
20100321399 Ellren et al. Dec 2010 A1
20100325526 Ellis et al. Dec 2010 A1
20100325581 Finkelstein et al. Dec 2010 A1
20100330801 Rouh Dec 2010 A1
20110004498 Readshaw Jan 2011 A1
20110029526 Knight et al. Feb 2011 A1
20110047159 Baid et al. Feb 2011 A1
20110060753 Shaked et al. Mar 2011 A1
20110061013 Bilicki et al. Mar 2011 A1
20110066933 Ludwig Mar 2011 A1
20110074811 Hanson et al. Mar 2011 A1
20110078055 Faribault et al. Mar 2011 A1
20110078173 Seligmann et al. Mar 2011 A1
20110093327 Fordyce, III et al. Apr 2011 A1
20110107196 Foster May 2011 A1
20110113348 Twiss May 2011 A1
20110117878 Barash et al. May 2011 A1
20110119100 Ruhl et al. May 2011 A1
20110131547 Elaasar Jun 2011 A1
20110137766 Rasmussen et al. Jun 2011 A1
20110153384 Horne et al. Jun 2011 A1
20110161096 Buehler et al. Jun 2011 A1
20110161137 Ubalde et al. Jun 2011 A1
20110167105 Ramakrishnan et al. Jul 2011 A1
20110170799 Carrino et al. Jul 2011 A1
20110173032 Payne et al. Jul 2011 A1
20110181598 O'Neall et al. Jul 2011 A1
20110185316 Reid et al. Jul 2011 A1
20110208724 Jones et al. Aug 2011 A1
20110213655 Henkin Sep 2011 A1
20110218934 Elser Sep 2011 A1
20110219321 Gonzalez et al. Sep 2011 A1
20110219450 McDougal et al. Sep 2011 A1
20110225198 Edwards et al. Sep 2011 A1
20110238495 Kang Sep 2011 A1
20110238553 Raj et al. Sep 2011 A1
20110251951 Kolkowitz Oct 2011 A1
20110252351 Sikora Oct 2011 A1
20110258158 Resende et al. Oct 2011 A1
20110270705 Parker Nov 2011 A1
20110289397 Eastmond et al. Nov 2011 A1
20110289407 Naik et al. Nov 2011 A1
20110289420 Morioka et al. Nov 2011 A1
20110291851 Whisenant Dec 2011 A1
20110310005 Chen et al. Dec 2011 A1
20110314007 Dassa et al. Dec 2011 A1
20120004904 Shin et al. Jan 2012 A1
20120019559 Siler et al. Jan 2012 A1
20120036013 Neuhaus et al. Feb 2012 A1
20120036434 Oberstein Feb 2012 A1
20120050293 Carlhian et al. Mar 2012 A1
20120066296 Appleton et al. Mar 2012 A1
20120072825 Sherkin et al. Mar 2012 A1
20120075324 Cardno et al. Mar 2012 A1
20120079363 Folting et al. Mar 2012 A1
20120084118 Bai et al. Apr 2012 A1
20120106801 Jackson May 2012 A1
20120116828 Shannon May 2012 A1
20120117082 Koperda et al. May 2012 A1
20120131512 Takeuchi et al. May 2012 A1
20120137235 TS et al. May 2012 A1
20120144335 Abeln et al. Jun 2012 A1
20120159307 Chung et al. Jun 2012 A1
20120159362 Brown et al. Jun 2012 A1
20120159399 Bastide et al. Jun 2012 A1
20120170847 Tsukidate Jul 2012 A1
20120173985 Peppel Jul 2012 A1
20120180002 Campbell et al. Jul 2012 A1
20120196557 Reich et al. Aug 2012 A1
20120196558 Reich et al. Aug 2012 A1
20120197651 Robinson et al. Aug 2012 A1
20120203708 Psota et al. Aug 2012 A1
20120208636 Feige Aug 2012 A1
20120221511 Gibson et al. Aug 2012 A1
20120221553 Wittmer et al. Aug 2012 A1
20120221580 Barney Aug 2012 A1
20120245976 Kumar et al. Sep 2012 A1
20120246148 Dror Sep 2012 A1
20120254129 Wheeler et al. Oct 2012 A1
20120284345 Costenaro et al. Nov 2012 A1
20120290879 Shibuya et al. Nov 2012 A1
20120296907 Long et al. Nov 2012 A1
20120311684 Paulsen et al. Dec 2012 A1
20120323888 Osann, Jr. Dec 2012 A1
20120330801 McDougal et al. Dec 2012 A1
20120330973 Ghuneim et al. Dec 2012 A1
20130006426 Healey et al. Jan 2013 A1
20130006725 Simanek et al. Jan 2013 A1
20130006916 McBride et al. Jan 2013 A1
20130018796 Kolhatkar et al. Jan 2013 A1
20130024268 Manickavelu Jan 2013 A1
20130046635 Grigg et al. Feb 2013 A1
20130046842 Muntz et al. Feb 2013 A1
20130050217 Armitage Feb 2013 A1
20130060786 Serrano et al. Mar 2013 A1
20130061169 Pearcy et al. Mar 2013 A1
20130073377 Heath Mar 2013 A1
20130073454 Busch Mar 2013 A1
20130078943 Biage et al. Mar 2013 A1
20130086482 Parsons Apr 2013 A1
20130097482 Marantz et al. Apr 2013 A1
20130101159 Chao et al. Apr 2013 A1
20130106860 De Pauw et al. May 2013 A1
20130110822 Ikeda et al. May 2013 A1
20130110877 Bonham et al. May 2013 A1
20130111320 Campbell et al. May 2013 A1
20130117011 Ahmed et al. May 2013 A1
20130117651 Waldman et al. May 2013 A1
20130150004 Rosen Jun 2013 A1
20130151148 Parundekar et al. Jun 2013 A1
20130151388 Falkenborg et al. Jun 2013 A1
20130157234 Gulli et al. Jun 2013 A1
20130166550 Buchmann et al. Jun 2013 A1
20130169666 Pacheco et al. Jul 2013 A1
20130176321 Mitchell et al. Jul 2013 A1
20130179420 Park et al. Jul 2013 A1
20130224696 Wolfe et al. Aug 2013 A1
20130225212 Khan Aug 2013 A1
20130226318 Procyk Aug 2013 A1
20130226953 Markovich et al. Aug 2013 A1
20130232045 Tai Sep 2013 A1
20130238616 Rose et al. Sep 2013 A1
20130246170 Gross et al. Sep 2013 A1
20130251233 Yang et al. Sep 2013 A1
20130262527 Hunter et al. Oct 2013 A1
20130263019 Castellanos et al. Oct 2013 A1
20130267207 Hao et al. Oct 2013 A1
20130268520 Fisher et al. Oct 2013 A1
20130276000 Neeman Oct 2013 A1
20130279757 Kephart Oct 2013 A1
20130282696 John et al. Oct 2013 A1
20130290011 Lynn et al. Oct 2013 A1
20130290825 Arndt et al. Oct 2013 A1
20130297619 Chandrasekaran et al. Nov 2013 A1
20130311375 Priebatsch Nov 2013 A1
20130335419 Bondesen et al. Dec 2013 A1
20140006938 Black Jan 2014 A1
20140019461 Bredenberg et al. Jan 2014 A1
20140019936 Cohanoff Jan 2014 A1
20140032506 Hoey et al. Jan 2014 A1
20140033010 Richardt et al. Jan 2014 A1
20140040371 Gurevich et al. Feb 2014 A1
20140046934 Zhou et al. Feb 2014 A1
20140047319 Eberlein Feb 2014 A1
20140047357 Alfaro et al. Feb 2014 A1
20140059038 McPherson et al. Feb 2014 A1
20140059498 McCormack et al. Feb 2014 A1
20140067611 Adachi et al. Mar 2014 A1
20140068487 Steiger et al. Mar 2014 A1
20140074855 Zhao et al. Mar 2014 A1
20140095273 Tang et al. Apr 2014 A1
20140095509 Patton Apr 2014 A1
20140108068 Williams Apr 2014 A1
20140108380 Gotz et al. Apr 2014 A1
20140108985 Scott et al. Apr 2014 A1
20140129261 Bothwell et al. May 2014 A1
20140149436 Bahrami et al. May 2014 A1
20140156527 Grigg et al. Jun 2014 A1
20140157172 Peery et al. Jun 2014 A1
20140164502 Khodorenko et al. Jun 2014 A1
20140189536 Lange et al. Jul 2014 A1
20140195515 Baker et al. Jul 2014 A1
20140195887 Ellis et al. Jul 2014 A1
20140214579 Shen et al. Jul 2014 A1
20140222521 Chait Aug 2014 A1
20140244388 Manouchehri et al. Aug 2014 A1
20140258246 Lo Faro et al. Sep 2014 A1
20140267294 Ma Sep 2014 A1
20140267295 Sharma Sep 2014 A1
20140279824 Tamayo Sep 2014 A1
20140282177 Wang et al. Sep 2014 A1
20140310266 Greenfield Oct 2014 A1
20140316911 Gross Oct 2014 A1
20140333651 Cervelli et al. Nov 2014 A1
20140337772 Cervelli et al. Nov 2014 A1
20140344230 Krause et al. Nov 2014 A1
20140351070 Christner et al. Nov 2014 A1
20150019394 Unser et al. Jan 2015 A1
20150046870 Goldenberg et al. Feb 2015 A1
20150073929 Psota et al. Mar 2015 A1
20150081370 Lo et al. Mar 2015 A1
20150089424 Duffield et al. Mar 2015 A1
20150100897 Sun et al. Apr 2015 A1
20150100907 Erenrich et al. Apr 2015 A1
20150134371 Shivakumar et al. May 2015 A1
20150134666 Gattiker et al. May 2015 A1
20150169709 Kara et al. Jun 2015 A1
20150169726 Kara et al. Jun 2015 A1
20150170077 Kara et al. Jun 2015 A1
20150178690 May Jun 2015 A1
20150178825 Huerta Jun 2015 A1
20150178877 Bogomolov et al. Jun 2015 A1
20150186821 Wang et al. Jul 2015 A1
20150187036 Wang et al. Jul 2015 A1
20150213631 Vander Broek Jul 2015 A1
20150227295 Meiklejohn et al. Aug 2015 A1
20150227847 Noel et al. Aug 2015 A1
20150229532 Somaiya et al. Aug 2015 A1
20150229546 Somaiya et al. Aug 2015 A1
20150242401 Liu Aug 2015 A1
20150254878 Sharma et al. Sep 2015 A1
20150309719 Ma et al. Oct 2015 A1
20150317342 Grossman et al. Nov 2015 A1
20150324868 Kaftan et al. Nov 2015 A1
20150341212 Hsiao Nov 2015 A1
20150347903 Saxena et al. Dec 2015 A1
20150363478 Haynes Dec 2015 A1
20150378996 Kesin et al. Dec 2015 A1
20160004667 Chakerian et al. Jan 2016 A1
20160006749 Cohen et al. Jan 2016 A1
20160034545 Shankar et al. Feb 2016 A1
20160098173 Slawinski et al. Apr 2016 A1
20160162497 Cho et al. Jun 2016 A1
20170109030 Mingione Apr 2017 A1
20170109910 Sharma et al. Apr 2017 A1
20170116294 Ma et al. Apr 2017 A1
Foreign Referenced Citations (42)
Number Date Country
2014250678 Feb 2016 AU
102014103482 Sep 2014 DE
102014215621 Feb 2015 DE
1191463 Mar 2002 EP
1672527 Jun 2006 EP
2551799 Jan 2013 EP
2560134 Feb 2013 EP
2778977 Sep 2014 EP
2778983 Sep 2014 EP
2779082 Sep 2014 EP
2835745 Feb 2015 EP
2835770 Feb 2015 EP
2838039 Feb 2015 EP
2846241 Mar 2015 EP
2851852 Mar 2015 EP
2858014 Apr 2015 EP
2858018 Apr 2015 EP
2863326 Apr 2015 EP
2863346 Apr 2015 EP
2869211 May 2015 EP
2884439 Jun 2015 EP
2884440 Jun 2015 EP
2891992 Jul 2015 EP
2911078 Aug 2015 EP
2911100 Aug 2015 EP
2940603 Nov 2015 EP
2940609 Nov 2015 EP
2516155 Jan 2015 GB
2518745 Apr 2015 GB
2012778 Nov 2014 NL
2013306 Feb 2015 NL
624557 Dec 2014 NZ
WO 0009529 Feb 2000 WO
WO 02065353 Aug 2002 WO
WO 2005104736 Nov 2005 WO
WO 2008064207 May 2008 WO
WO 2009061501 May 2009 WO
WO 2010000014 Jan 2010 WO
WO 2010030913 Mar 2010 WO
WO 2010098958 Sep 2010 WO
WO 2013010157 Jan 2013 WO
WO 2013102892 Jul 2013 WO
Non-Patent Literature Citations (256)
Entry
Timelines Everywhere 2012, Heather OCull.
About 80 Minutes, “Palantir in a Number of Parts—Part 6—Graph,” Mar. 21, 2013, pp. 1-6, retrieved from the internet http://about80minutes.blogspot.nl/2013/03/palantir-in-number-of-parts-part-6-graph.html retrieved on Aug. 18, 2015.
Bluttman et al., “Excel Formulas and Functions for Dummies,” 2005, Wiley Publishing, Inc., pp. 280, 284-286.
Chung, Chin-Wan, “Dataplex: An Access to Heterogeneous Distributed Databases,” Communications of the ACM, Association for Computing Machinery, Inc., vol. 33, No. 1, Jan. 1, 1990, pp. 70-80.
Gesher, Ari, “Palantir Screenshots in the Wild: Swing Sightings,” The Palantir Blog, Sep. 11, 2007, pp. 1-12, retrieved from the internet https://www.palantir.com/2007/09/palantir-screenshots/ retrieved on Aug. 18, 2015.
Hardesty, “Privacy Challenges: Analysis: It's Surprisingly Easy to Identify Individuals from Credit-Card Metadata,” MIT News on Campus and Around the World, MIT News Office, Jan. 29, 2015, 3 pages.
Keylines.com, “An Introduction to KeyLines and Network Visualization,” Mar. 2014, http://keylines.com/wp-content/uploads/2014/03/KeyLines-White-Paper.pdf downloaded May 12, 2014 in 8 pages.
Keylines.com, “KeyLines Datasheet,” Mar. 2014, http://keylines.com/wp-content/uploads/2014/03/KeyLines-datasheet.pdf downloaded May 12, 2014 in 2 pages.
Keylines.com, “Visualizing Threats: Improved Cyber Security Through Network Visualization,” Apr. 2014, http://keylines.com/wp-content/uploads/2014/04/Visualizing-Threats1.pdf downloaded May 12, 2014 in 10 pages.
Palantir Technolgies, “Palantir Labs—Timeline,” Oct. 1, 2010, retrieved from the internet https://www.youtube.com/watch?v=JCgDW5bru9M retrieved on Aug. 19, 2015.
Palmas et al', “An Edge-Bunding Layout for Interactive Parallel Coordinates” 2014 IEEE Pacific Visualization Symposium, pp. 57-64.
Wikipedia, “Federated Database System,” Sep. 7, 2013, retrieved from the internet on Jan. 27, 2015 http://en.wikipedia.org/w/index.php?title=Federated—database—system&oldid=571954221.
Wright et al., “Palantir Technologies VAST 2010 Challenge Text Records—Investigations into Arms Dealing,” Oct. 29, 2010, pp. 1-10, retrieved from the internet http://hcll2.cs.umd.edu/newvarepository/VAST%20Challenge%202010/challenges/MC1%20-%20Investigations%20into%20Arms%20Dealing/entries/Palantir%20Technoiogies/ retrieved on Aug. 20, 2015.
Vose et al., “Help File for ModelRisk Version 5,” 2007, Vose Software, pp. 349-353. [Uploaded in 2 Parts].
Notice of Allowance for U.S. Appl. No. 13/247,987 dated Mar. 17, 2016.
Notice of Allowance for U.S. Appl. No. 14/323,935 dated Oct. 1, 2015.
Notice of Allowance for U.S. Appl. No. 14/326,738 dated Nov. 18, 2015.
Notice of Allowance for U.S. Appl. No. 14/504,103 dated May 18, 2015.
Official Communication for European Patent Application No. 14189344.6 dated Feb. 20, 2015.
Official Communication for European Patent Application No. 14189344.6 dated Feb. 29, 2016.
Official Communication for European Patent Application No. 14197879.1 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14197895.7 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14197938.5 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 15165244.3 dated Aug. 27, 2015.
Official Communication for New Zealand Patent Application No. 622513 dated Apr. 3, 2014.
Official Communication for U.S. Appl. No. 13/247,987 dated Apr. 2, 2015.
Official Communication for U.S. Appl. No. 13/247,987 dated Sep. 22, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 14, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Mar. 17, 2016.
Official Communication for U.S. Appl. No. 14/306,138 dated Feb. 18, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 23, 2014.
Official Communication for U.S. Appl. No. 14/306,138 dated Dec. 24, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Dec. 3, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Feb. 19, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Dec. 24, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Aug. 7, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/306,154 dated Feb. 1, 2016.
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 11, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated May 15, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Nov. 16, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 17, 2016.
Official Communication for U.S. Appl. No. 14/306,154 dated Jul. 6, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 1, 2016.
Official Communication for U.S. Appl. No. 14/319,765 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Jun. 16, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Nov. 25, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 4, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Nov. 28, 2014.
Official Communication for U.S. Appl. No. 14/323,935 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Dec. 2, 2014.
Official Communication for U.S. Appl. No. 14/326,738 dated Jul. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Feb. 5, 2015.
Official Communication for U.S. Appl. No. 14/645,304 dated Jan. 25, 2016.
Official Communication for U.S. Appl. No. 14/874,690 dated Jun. 1, 2016.
Official Communication for U.S. Appl. No. 14/874,690 dated Dec. 21, 2015.
Official Communication for U.S. Appl. No. 14/948,009 dated Feb. 25, 2016.
“A First Look: Predicting Market Demand for Food Retail using a Huff Analysis,” TRF Policy Solutions, Jul. 2012, pp. 30.
“A Quick Guide to UniProtKB Swiss-Prot & TrEMBL,” Sep. 2011, pp. 2.
“A Word About Banks and the Laundering of Drug Money,” Aug. 18, 2012, http://www.golemxiv.co.uk/2012/08/a-word-about-banks-and-the-laundering-of-drug-money/.
“Money Laundering Risks and E-Gaming: A European Overview and Assessment,” 2009, http://www.cf.ac.uk/socsi/resources/Levi—Final—Money—Laundering—Risks—egaming.pdf.
“Potential Money Laundering Warning Signs,” snapshot taken 2003, https://web.archive.org/web/20030816090055/http:/finsolinc.com/ANTI-MONEY%20LAUNDERING%20TRAINING%20GUIDES.pdf.
“Refresh CSS Ellipsis When Resizing Container—Stack Overflow,” Jul. 31, 2013, retrieved from internet http://stackoverflow.com/questions/17964681/refresh-css-ellipsis-when-resizing-container, retrieved on May 18, 2015.
“The FASTA Program Package,” fasta-36.3.4, Mar. 25, 2011, pp. 29.
“Using Whois Based Geolocation and Google Maps API for Support Cybercrime Investigations,” http://wseas.us/e-library/conferences/2013/Dubrovnik/TELECIRC/TELECIRC-32.pdf.
Acklen, Laura, “Absolute Beginner's Guide to Microsoft Word 2003,” Dec. 24, 2003, pp. 15-18, 34-41, 308-316.
Alfred, Rayner “Summarizing Relational Data Using Semi-Supervised Genetic Algorithm-Based Clustering Techniques”, Journal of Computer Science, 2010, vol. 6, No. 7, pp. 775-784.
Alur et al., “Chapter 2: IBM InfoSphere DataStage Stages,” IBM InfoSphere DataStage Data Flow and Job Design, Jul. 1, 2008, pp. 35-137.
Amnet, “5 Great Tools for Visualizing Your Twitter Followers,” posted Aug. 4, 2010, http://www.amnetblog.com/component/content/article/115-5-grate-tools-for-visualizing-your-twitter-followers.html.
Ananiev et al., “The New Modality API,” http://web.archive.org/web/20061211011958/http://java.sun.com/developer/technicalArticles/J2SE/Desktop/javase6/modality/ Jan. 21, 2006, pp. 8.
Boyce, Jim, “Microsoft Outlook 2010 Inside Out,” Aug. 1, 2010, retrieved from the internet https://capdtron.files.wordpress.com/2013/01/outlook-2010-inside—out.pdf.
Bugzilla@Mozilla, “Bug 18726 —[feature] Long-click means of invoking contextual menus not supported,” http://bugzilla.mozilla.org/show—bug.cgi?id=18726 printed Jun. 13, 2013 in 11 pages.
Canese et al., “Chapter 2: PubMed: The Bibliographic Database,” The NCBI Handbook, Oct. 2002, pp. 1-10.
Celik, Tantek, “CSS Basic User Interface Module Level 3 (CSS3 UI),” Section 8 Resizing and Overflow, Jan. 17, 2012, retrieved from internet http://www.w3.org/TR/2012/WD-css3-ui-20120117/#resizing-amp-overflow retrieved on May 18, 2015.
Chen et al., “Bringing Order to the Web: Automatically Categorizing Search Results,” CHI 2000, Proceedings of the SIGCHI conference on Human Factors in Computing Systems, Apr. 1-6, 2000, The Hague, The Netherlands, pp. 145-152.
Conner, Nancy, “Google Apps: The Missing Manual,” May 1, 2008, pp. 15.
Definition “Identify”, downloaded Jan. 22, 2015, 1 page.
Definition “Overlay”, downloaded Jan. 22, 2015, 1 page.
Delcher et al., “Identifying Bacterial Genes and Endosymbiont DNA with Glimmer,” Biolnformatics, vol. 23, No. 6, 2007, pp. 673-679.
Dramowicz, Ela, “Retail Trade Area Analysis Using the Huff Model,” Directions Magazine, Jul. 2, 2005 in 10 pages, http://www.directionsmag.com/articles/retail-trade-area-analysis-using-the-huff-mode1/123411.
GIS-NET 3 Public —Department of Regional Planning. Planning & Zoning Information for Unincorporated LA County. Retrieved Oct. 2, 2013 from http://gis.planning.lacounty.gov/GIS-NET3—Public/Viewer.html.
Goswami, Gautam, “Quite Writly Said!,” One Brick at a Time, Aug. 21, 2005, pp. 7.
Griffith, Daniel A., “A Generalized Huff Model,” Geographical Analysis, Apr. 1982, vol. 14, No. 2, pp. 135-144.
Hansen et al., “Analyzing Social Media Networks with NodeXL: Insights from a Connected World”, Chapter 4, pp. 53-67 and Chapter 10, pp. 143-164, published Sep. 2010.
Hibbert et al., “Prediction of Shopping Behavior Using a Huff Model Within a GIS Framework,” Healthy Eating in Context, Mar. 18, 2011, pp. 16.
Hogue et al., “Thresher: Automating the Unwrapping of Semantic Content from the World Wide Web,” 14th International Conference on World Wide Web, WWW 2005: Chiba, Japan, May 10-14, 2005, pp. 86-95.
Huang et al., “Systematic and Integrative Analysis of Large Gene Lists Using DAVID Bioinformatics Resources,” Nature Protocols, 4.1, 2008, 44-57.
Huff et al., “Calibrating the Huff Model Using ArcGIS Business Analyst,” ESRI, Sep. 2008, pp. 33.
Huff, David L., “Parameter Estimation in the Huff Model,” ESRI, ArcUser, Oct.-Dec. 2003, pp. 34-36.
Janssen, Jan-Keno, “Wo bist'n du?—Googles Geodienst Latitude,” Jan. 17, 2011, pp. 86-88, retrieved from the internet on Jul. 30, 2015 http://www.heise.de/artikel-archiv/ct/2011/03/086/@00250@/ct.11.03.086-088.pdf.
Jelen, Bill, “Excel 2013 in Depth, Video Enhanced Edition,” Jan. 25, 2013.
Kahan et al., “Annotea: an Open RDF Infrastructure for Shared Web Annotations”, Computer Networks, Elsevier Science Publishers B.V., vol. 39, No. 5, dated Aug. 5, 2002, pp. 589-608.
Kitts, Paul, “Chapter 14: Genome Assembly and Annotation Process,” The NCBI Handbook, Oct. 2002, pp. 1-21.
Li et al., “Interactive Multimodal Visual Search on Mobile Device,” IEEE Transactions on Multimedia, vol. 15, No. 3, Apr. 1, 2013, pp. 594-607.
Liu, Tianshun, “Combining GIS and the Huff Model to Analyze Suitable Locations for a New Asian Supermarket in the Minneapolis and St. Paul, Minnesota USA,” Papers in Resource Analysis, 2012, vol. 14, pp. 8.
Madden, Tom, “Chapter 16: The BLAST Sequence Analysis Tool,” The NCBI Handbook, Oct. 2002, pp. 1-15.
Manno et al., “Introducing Collaboration in Single-user Applications through the Centralized Control Architecture,” 2010, pp. 10.
Manske, “File Saving Dialogs,” http://www.mozilla.org/editor/ui—specs/FileSaveDialogs.html, Jan. 20, 1999, pp. 7.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.yahoo.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.bing.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.google.com.
Microsoft—Developer Network, “Getting Started with VBA in Word 2010,” Apr. 2010, http://msdn.microsoft.com/en-us/library/ff604039%28v=office.14%29.aspx as printed Apr. 4, 2014 in 17 pages.
Microsoft Office—Visio, “About connecting shapes,” http://office.microsoft.com/en-us/visio-help/about-connecting-shapes-HP085050369.aspx printed Aug. 4, 2011 in 6 pages.
Microsoft Office—Visio, “Add and glue connectors with the Connector tool,” http://office.microsoft.com/en-us/visio-help/add-and-glue-connectors-with-the-connector-tool-HA010048532.aspx?CTT=1 printed Aug. 4, 2011 in 1 page.
Mizrachi, Ilene, “Chapter 1: GenBank: The Nuckeotide Sequence Database,” The NCBI Handbook, Oct. 2002, pp. 1-14.
Nierman, “Evaluating Structural Similarity in XML Documents”, 6 pages, 2002.
Nolan et al., “MCARTA: A Malicious Code Automated Run-Time Analysis Framework,” Homeland Security, 2012 IEEE Conference on Technologies for, Nov. 13, 2012, pp. 13-17.
Olanoff, Drew, “Deep Dive with the New Google Maps for Desktop with Google Earth Integration, It's More than Just a Utility,” May 15, 2013, pp. 1-6, retrieved from the internet: http://web.archive.org/web/20130515230641/http://techcrunch.com/2013/05/15/deep-dive-with-the-new-google-maps-for-desktop-with-google-earth-integration-its-more-than-just-a-utility/.
Perdisci et al., “Behavioral Clustering of HTTP-Based Malware and Signature Generation Using Malicious Network Traces,” USENIX, Mar. 18, 2010, pp. 1-14.
Psaltis, Andrew G., “Streaming Data—Designing the Real-Time Pipeline,” MEAP began Jul. 2014, Part 1, pp. 15.
Quest, “Toad for ORACLE 11.6—Guide to Using Toad,” Sep. 24, 2012, pp. 1-162.
Rouse, Margaret, “OLAP Cube,” http://searchdatamanagement.techtarget.com/definition/OLAP-cube, Apr. 28, 2012, pp. 16.
Shi et al., “A Scalable Implementation of Malware Detection Based on Network Connection Behaviors,” 2013 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, IEEE, Oct. 10, 2013, pp. 59-66.
Sigrist, et al., “PROSITE, a Protein Domain Database for Functional Characterization and Annotation,” Nucleic Acids Research, 2010, vol. 38, pp. D161-D166.
Sirotkin et al., “Chapter 13: The Processing of Biological Sequence Data at NCBI,” The NCBI Handbook, Oct. 2002, pp. 1-11.
Symantec Corporation, “E-Security Begins with Sound Security Policies,” Announcement Symantec, Jun. 14, 2001.
Thompson, Mick, “Getting Started with GEO,” Getting Started with GEO, Jul. 26, 2011.
Umagandhi et al., “Search Query Recommendations Using Hybrid User Profile with Query Logs,” International Journal of Computer Applications, vol. 80, No. 10, Oct. 1, 2013, pp. 7-18.
Wikipedia, “Mobile Web,” Jan. 23, 2015, retrieved from the internet on Mar. 15, 2016 https://en.wikipedia.org/w/index.php?title=Mobile—Web&oldid=643800164.
Windley, Phillip J., “The Live Web: Building Event-Based Connections in the Cloud,” Dec. 21, 2011, pp. 10, 216.
Yang et al., “HTML Page Analysis Based on Visual Cues”, A129, pp. 859-864, 2001.
International Search Report and Written Opinion in Application No. PCT/US2009/056703, dated Mar. 15, 2010.
Notice of Acceptance for Australian Patent Application No. 2014250678 dated Oct. 7, 2015.
Notice of Allowance for U.S. Appl. No. 12/556,318 dated Nov. 2, 2015.
Notice of Allowance for U.S. Appl. No. 14/102,394 dated Aug. 25, 2014.
Notice of Allowance for U.S. Appl. No. 14/108,187 dated Aug. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/135,289 dated Oct. 14, 2014.
Notice of Allowance for U.S. Appl. No. 14/148,568 dated Aug. 26, 2015.
Notice of Allowance for U.S. Appl. No. 14/192,767 dated Dec. 16, 2014.
Notice of Allowance for U.S. Appl. No. 14/192,767 dated Apr. 20, 2015.
Notice of Allowance for U.S. Appl. No. 14/225,084 dated May 4, 2015.
Notice of Allowance for U.S. Appl. No. 14/268,964 dated Dec. 3, 2014.
Notice of Allowance for U.S. Appl. No. 14/294,098 dated Dec. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/320,236 dated Jun. 29, 2016.
Notice of Allowance for U.S. Appl. No. 14/473,552 dated Jul. 24, 2015.
Notice of Allowance for U.S. Appl. No. 14/473,860 dated Feb. 27, 2015.
Notice of Allowance for U.S. Appl. No. 14/473,860 dated Jan. 5, 2015.
Notice of Allowance for U.S. Appl. No. 14/596,552 dated Dec. 23, 2016.
Notice of Allowance for U.S. Appl. No. 14/616,080 dated Apr. 2, 2015.
Notice of Allowance for U.S. Appl. No. 14/923,364 dated May 6, 2016.
Notice of Allowance for U.S. Appl. No. 14/948,009 dated May 6, 2016.
Official Communication for Australian Patent Application No. 2014201511 dated Feb. 27, 2015.
Official Communication for Australian Patent Application No. 2014202442 dated Mar. 19, 2015.
Official Communication for Australian Patent Application No. 2014210604 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014210614 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014213553 dated May 7, 2015.
Official Communication for Australian Patent Application No. 2014250678 dated Jun. 17, 2015.
Official Communication for European Patent Application No. 14158861.6 dated Jun. 16, 2014.
Official Communication for European Patent Application No. 14159464.8 dated Jul. 31, 2014.
Official Communication for European Patent Application No. 14180142.3 dated Feb. 6, 2015.
Official Communication for European Patent Application No. 14180281.9 dated Jan. 26, 2015.
Official Communication for European Patent Application No. 14180321.3 dated Apr. 17, 2015.
Official Communication for European Patent Application No. 14180432.8 dated Jun. 23, 2015.
Official Communication for European Patent Application No. 14186225.0 dated Feb. 13, 2015.
Official Communication for European Patent Application No. 14187739.9 dated Jul. 6, 2015.
Official Communication for European Patent Application No. 14187996.5 dated Feb. 12, 2015.
Official Communication for European Patent Application No. 14189347.9 dated Mar. 4, 2015.
Official Communication for European Patent Application No. 14189802.3 dated May 11, 2015.
Official Communication for European Patent Application No. 14191540.5 dated May 27, 2015.
Official Communication for European Patent Application No. 14199182.8 dated Mar. 13, 2015.
Official Communication for European Patent Application No. 15155845.9 dated Oct. 6, 2015.
Official Communication for European Patent Application No. 15155846.7 dated Jul. 8, 2015.
Official Communication for European Patent Application No. 15166137.8 dated Sep. 14, 2015.
Official Communication for European Patent Application No. 15175106.2 dated Nov. 5, 2015.
Official Communication for European Patent Application No. 15175151.8 dated Nov. 25, 2015.
Official Communication for European Patent Application No. 15183721.8 dated Nov. 23, 2015.
Official Communication for European Patent Application No. 16152984.7 dated Mar. 24, 2016.
Official Communication for Great Britain Patent Application No. 1404457.2 dated Aug. 14, 2014.
Official Communication for Great Britain Patent Application No. 1404574.4 dated Dec. 18, 2014.
Official Communication for Great Britain Patent Application No. 1408025.3 dated Nov. 6, 2014.
Official Communication for Great Britain Patent Application No. 1411984.6 dated Dec. 22, 2014.
Official Communication for Great Britain Patent Application No. 1413935.6 dated Jan. 27, 2015.
Official Communication for Netherlands Patent Application No. 2012437 dated Sep. 18, 2015.
Official Communication for Netherlands Patent Application No. 2013306 dated Apr. 24, 2015.
Official Communication for New Zealand Patent Application No. 622517 dated Apr. 3, 2014.
Official Communication for New Zealand Patent Application No. 624557 dated May 14, 2014.
Official Communication for New Zealand Patent Application No. 627962 dated Aug. 5, 2014.
Official Communication for New Zealand Patent Application No. 628161 dated Aug. 25, 2014.
Official Communication for New Zealand Patent Application No. 628263 dated Aug. 12, 2014.
Official Communication for New Zealand Patent Application No. 628495 dated Aug. 19, 2014.
Official Communication for New Zealand Patent Application No. 628585 dated Aug. 26, 2014.
Official Communication for New Zealand Patent Application No. 628840 dated Aug. 28, 2014.
Official Communication for U.S. Appl. No. 12/556,318 dated Jul. 2, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Mar. 4, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Aug. 6, 2015.
Official Communication for U.S. Appl. No. 13/835,688 dated Jun. 17, 2015.
Official Communication for U.S. Appl. No. 13/839,026 dated Aug. 4, 2015.
Official Communication for U.S. Appl. No. 14/102,394 dated Mar. 27, 2014.
Official Communication for U.S. Appl. No. 14/108,187 dated Apr. 17, 2014.
Official Communication for U.S. Appl. No. 14/108,187 dated Mar. 20, 2014.
Official Communication for U.S. Appl. No. 14/134,558 dated Oct. 7, 2015.
Official Communication for U.S. Appl. No. 14/135,289 dated Apr. 16, 2014.
Official Communication for U.S. Appl. No. 14/135,289 dated Jul. 7, 2014.
Official Communication for U.S. Appl. No. 14/148,568 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/148,568 dated Mar. 26, 2015.
Official Communication for U.S. Appl. No. 14/148,568 dated Mar. 27, 2014.
Official Communication for U.S. Appl. No. 14/192,767 dated Sep. 24, 2014.
Official Communication for U.S. Appl. No. 14/192,767 dated May 6, 2014.
Official Communication for U.S. Appl. No. 14/196,814 dated May 5, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 10, 2014.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 2, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Dec. 21, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Feb. 27, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 11, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 2, 2014.
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 20, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 26, 2016.
Official Communication for U.S. Appl. No. 14/225,084 dated Jan. 4, 2016.
Official Communication for U.S. Appl. No. 14/225,160 dated Feb. 11, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Aug. 12, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated May 20, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/225,160 dated Jan. 25, 2016.
Official Communication for U.S. Appl. No. 14/225,160 dated Jul. 29, 2014.
Official Communication for U.S. Appl. No. 14/268,964 dated Jul. 11, 2014.
Official Communication for U.S. Appl. No. 14/268,964 dated Sep. 3, 2014.
Official Communication for U.S. Appl. No. 14/289,596 dated Jul. 18, 2014.
Official Communication for U.S. Appl. No. 14/289,596 dated Jan. 26, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Apr. 30, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Aug. 5, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated Jul. 22, 2014.
Official Communication for U.S. Appl. No. 14/289,599 dated May 29, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated Sep. 4, 2015.
Official Communication for U.S. Appl. No. 14/294,098 dated Aug. 15, 2014.
Official Communication for U.S. Appl. No. 14/294,098 dated Nov. 6, 2014.
Official Communication for U.S. Appl. No. 14/306,147 dated Jun. 3, 2016.
Official Communication for U.S. Appl. No. 14/473,552 dated Feb. 24, 2015.
Official Communication for U.S. Appl. No. 14/473,860 dated Nov. 4, 2014.
Official Communication for U.S. Appl. No. 14/486,991 dated Mar. 10, 2015.
Official Communication for U.S. Appl. No. 14/490,612 dated Aug. 18, 2015.
Official Communication for U.S. Appl. No. 14/570,914 dated Sep. 16, 2016.
Official Communication for U.S. Appl. No. 14/570,914 dated Dec. 19, 2016.
Official Communication for U.S. Appl. No. 14/579,752 dated Aug. 19, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/596,552 dated Dec. 23, 2016.
Official Communication for U.S. Appl. No. 14/596,552 dated Sep. 23, 2016.
Official Communication for U.S. Appl. No. 14/596,552 dated Oct. 5, 2016.
Official Communication for U.S. Appl. No. 14/631,633 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated Oct. 16, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated May 18, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated Jul. 24, 2015.
Official Communication for U.S. Appl. No. 14/726,353 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/813,749 dated Sep. 28, 2015.
Restriction Requirement for U.S. Appl. No. 13/839,026 dated Apr. 2, 2015.
Map Builder, “Rapid Mashup Development Tool for Google and Yahoo Maps!” <http://web.archive.org/web/20090626224734/http://www.mapbuilder.net/> printed Jul. 20, 2012 in 2 pages.
Notice of Allowance for U.S. Appl. No. 14/570,914 dated Jan. 31, 2017.
Official Communication for U.S. Appl. No. 14/135,289 dated Oct. 14, 2014.
Official Communication for U.S. Appl. No. 14/148,559 dated Jun. 16, 2014.
Official Communication for U.S. Appl. No. 14/148,559 dated Apr. 2, 2014.
Official Communication for U.S. Appl. No. 15/392,624 dated Mar. 10, 2017.
Official Communication for U.S. Appl. No. 15/397,562 dated Mar. 14, 2017.
Provisional Applications (1)
Number Date Country
62272526 Dec 2015 US