The present disclosure relates to systems and techniques for data integration, analysis, and visualization. More specifically, the present disclosure relates to integration, analysis, and visualization of data objects in various contextual views.
Visualizations may enable faster and more thorough understandings of sets of data and information. Such visualizations of data and other information may be referred to as data visualizations. Data visualizations may, for example, visually transform and/or restructure data so as to provide new perspectives to a viewer of the visualization. A particular type of data visualization may be referred to as a contextual view. Examples of data visualizations include graphs, maps, tables, and/or lists, among others. Data visualizations may include displaying individual pieces of data in, for example, various arrangements, various sizes, various colors, and/or may include multi-dimensional aspects.
The systems, methods, and devices described herein each have several aspects, no single one of which is solely responsible for its desirable attributes. Without limiting the scope of this disclosure, several non-limiting features will now be discussed briefly.
A context-sensitive viewing system is disclosed in which various data visualizations, also referred to a contextual views, of a common set of data may be viewed by a user on an electronic device. Data in the context-sensitive viewing system may comprise data objects and associated properties and/or metadata. As a user of the system views and manipulates a first contextual view of a set of data objects, one or more other contextual views of the same set of data objects may be updated accordingly.
In various embodiments, a user of the context-sensitive viewing system may switch from a primary contextual view to a secondary contextual view, thereby making the switched-to contextual view the new primary contextual view. Data objects may be manipulated in any view, resulting in updates in the other views. Context switching may be accomplished through inputs from the user. For example, the user may click on a preview of a secondary view, and/or may scroll from one view to the next.
The context-sensitive viewing system advantageously enables a user to view a particular set of data objects in multiple visualization contexts. Previews of the set of data in other visualization may be quickly reviewed by the user to determine the most beneficial context for information extraction. Further, manipulations by the user in one context are propagated to the other contexts, allowing fast analysis of the impacts of changes to the set of data.
In an embodiment, a computer system is disclosed comprising one or more hardware processors in communication with a computer readable medium storing software modules including instructions that are executable by the one or more hardware processors, the software modules including at least: an electronic database configured to store a plurality of data objects and properties associated with each of the data objects; and a context viewing module configured to: generate a primary contextual view including a visualization of a set of data objects and associated properties; generate one or more secondary contextual views, each secondary contextual view including respective secondary visualizations of the set of data objects; receive a user input modifying the primary contextual view; and in response to receiving a user input modifying the primary contextual view, modify one or more of the secondary contextual views based at least in part on the user input.
According to an aspect, the context viewing module may be further configured to: in response to receiving a user input modifying the primary contextual view, determine which of the one or more of the secondary contextual views to modify.
According to an aspect, modifying one or more of the secondary contextual views may comprise modifying all of the one or more secondary contextual view.
According to an aspect, modifying one or more of the secondary contextual views may comprise modifying any of the one or more secondary contextual views that are currently viewable by a user.
According to an aspect, modifying one or more of the secondary contextual views may comprise modifying any of the one or more secondary contextual views that are immediately adjacent to the primary contextual view.
According to an aspect, each of the visualization and/or the secondary visualizations may include at least one of a graph, a map, a table, a timeline, a histogram, a list, a reader interface, or a postboard interface.
According to an aspect, the one or more secondary contextual views may comprise contextual previews.
According to an aspect, the one or more secondary contextual views may be viewable in a drawer or scrollbar on a user interface, and the one or more secondary contextual views may be selectable by a user.
According to an aspect, the one or more secondary contextual views may be substantially the same size as the primary contextual view.
According to an aspect, the one or more secondary contextual views may be configured to be accessible by a user through the use of a scrollbar.
According to an aspect, the scrollbar may include at least one of tick marks indicating the locations of the one or more secondary views or contextual previews accessible in pop up windows.
According to an aspect, the one or more secondary contextual views may be positioned laterally to the primary contextual view, and the secondary contextual views may be accessible by a user through a user input including at least one of a mouse cursor or a touch input.
According to an aspect, a user input modifying the primary contextual view may comprise at least one of adding data objects, removing data objects, modifying data objects, moving data objects, modifying properties associated with data objects, or modifying and/or manipulating links between data objects.
In an embodiment, a computer system is disclosed comprising one or more hardware processors in communication with a computer readable medium storing software modules including instructions that are executable by the one or more hardware processors, the software modules including at least: a context viewing module configured to: display a first data visualization of a set of data objects and properties associated with data objects of the set of data objects; provide one or more secondary data visualizations of the set of data objects; receive a user input modifying the first data visualization; and in response to receiving a user input modifying the first data visualization, implement modifications to at least some of the one or more secondary data visualizations based at least in part on the user input.
According to an aspect, the context viewing module may be further configured to: in response to receiving a user input modifying the modifying the first data visualization, determine which of the one or more of the secondary data visualizations are currently displayed to the user, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
According to an aspect, the context viewing module may be further configured to: in response to receiving a user input modifying the modifying the first data visualization, determine which of the one or more of the secondary data visualizations are adjacent to the first data visualization, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
In an embodiment, a computer-implemented method of updating multiple contextual views is disclosed, the method comprising: providing an electronic database configured to store a plurality of data objects and metadata associated with each of the plurality of data objects; generating, by a computing system having one or more computer processors, based at least in part on the plurality of data objects and associated metadata, a primary contextual view and one or more secondary contextual views; receiving, via an input device of the computing system, a user input modifying the primary contextual view; determining, by the computing system, based on the received user input, modifications of the one or more secondary contextual views that correspond to the modification of the primary contextual view; modifying at least some of the one or more secondary contextual views based on the determined modifications.
According to an aspect, the method may further comprise providing, on an electronic display of the computing system, the generated primary contextual view and one or more of the secondary contextual views in a preview drawer.
According to an aspect, the method may further comprise providing, on an electronic display of the computing system, the generated primary contextual view; and providing, on the electronic display of the computer system, a scrollbar that enables a user to scroll to any of the one or more of the secondary contextual views and view any of the one or more of the secondary contextual views on the electronic display.
According to an aspect, modifying at least some of the one or more secondary contextual views based on the determined modifications may comprise modifying any secondary contextual views that are immediately viewable by a user.
Overview
A context-sensitive viewing system is disclosed in which various data visualizations, also referred to a contextual views, of a common set of data may be viewed by a user on an electronic device. Data in the context-sensitive viewing system may comprise data objects and associated properties and/or metadata, and may be stored in one or more electronic data stores. As a user of the system views and manipulates a first contextual view (also referred to as the “primary contextual view”) of a set of data objects, one or more other contextual views (also referred to as “secondary contextual views”) of the same set of data objects may be updated accordingly.
Updates to the secondary contextual views may, in various embodiments, happen real-time or may happen upon the occurrence of a triggering event (for example, a user input). In various embodiments, the secondary contextual views may comprise previews and/or thumbnails. Further, the secondary contextual views may be visible to the user simultaneously with the primary contextual view. For example, the user of the context-sensitive viewing system may view a particular set of data objects in multiple visualization contexts. Further, as the user updates the set of data objects in one context, the set of data objects may automatically be updated in one or more secondary contexts.
For the sake of brevity, contextual views may be referred to herein simply as “views” or “contexts.” For example, a primary contextual view may be referred to as a “primary view.” Additionally, the terms “contextual view” and “data visualization” may be used interchangeably.
In various ways and in various embodiments, a user of the context-sensitive viewing system may switch from a primary contextual view to a secondary contextual view, thereby making the switched-to contextual view the new primary contextual view. Data objects may be manipulated in any view, resulting in updates in the other views. Context switching may be accomplished through inputs from the user. For example, the user may click on a preview of a secondary view, and/or may scroll from one view to the next.
Examples of contextual views (and/or data visualizations) of the context-sensitive viewing system include, but are not limited to graphs, maps, tables, timelines, histograms, and/or lists, among other types of data visualizations. In an embodiment, a contextual view comprises a graph of connected data objects as described below. In an embodiment, a contextual view comprises an interactive mapping application, an example of which is described in U.S. patent application Ser. No. 13/917,571 filed on Jun. 13, 2013, and titled “Interactive Geospatial Map,” which is hereby incorporated by reference herein in its entirety and for all purposes. In an embodiment, a contextual view comprises a reader interface that enables a user to review large amounts of notes and other textual information. An example of such a reader interface is described in U.S. Provisional Patent Application No. 61/863,792, filed on Aug. 8, 2013, and titled “Cable Reader Labeling,”, which is hereby incorporated by reference herein in its entirety and for all purposes. In an embodiment, a contextual view comprises a postboard view in which notes and textual clips may be listed, an example of which is described in U.S. Provisional Patent Application No. 61/863,814, filed on Aug. 8, 2013, and titled “Cable Reader Snippets and Postboard,” which is hereby incorporated by reference herein in its entirety and for all purposes. In an embodiment, a contextual view comprises a time series graph, timeline, and/or histogram, examples of which are described in U.S. Pat. No. 8,280,880, titled “Generating Dynamic Date Sets That Represent Market Conditions,” and U.S. Pat. No. 8,280,880, titled “Filter Chains With Associated Views For Exploring Large Data Sets,” each of which is hereby incorporated by reference herein in its entirety and for all purposes.
In order to facilitate an understanding of the systems and methods discussed herein, a number of terms are defined below. The terms defined below, as well as other terms used herein, should be construed to include the provided definitions, the ordinary and customary meaning of the terms, and/or any other implied meaning for the respective terms. The definitions below do not limit the meaning of these terms, but only provide exemplary definitions.
Ontology: Stored information that provides a data model for storage of data in one or more databases. For example, the stored data may comprise definitions for object types and property types for data in a database, and how objects and properties may be related.
Database: A broad term for any data structure for storing and/or organizing data, including, but not limited to, relational databases (Oracle database, mySQL database, etc.), spreadsheets, XML files, and text file, among others.
Data Object or Object: A data container for information representing specific things in the world that have a number of definable properties. For example, a data object can represent an entity such as a person, a place, an organization, a market instrument, or other noun. A data object can represent an event that happens at a point in time or for a duration. A data object can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object may be associated with a unique identifier that uniquely identifies the data object. The object's attributes (e.g. metadata about the object) may be represented in one or more properties.
Object Type: Type of a data object (e.g., Person, Event, or Document). Object types may be defined by an ontology and may be modified or updated to include additional object types. An object definition (e.g., in an ontology) may include how the object is related to other objects, such as being a sub-object type of another object type (e.g. an agent may be a sub-object type of a person object type), and the properties the object type may have.
Properties: Attributes of a data object that represent individual data items. At a minimum, each property of a data object has a property type and a value or values.
Property Type: The type of data a property is, such as a string, an integer, or a double. Property types may include complex property types, such as series data values associated with timed ticks (e.g. a time series), etc.
Property Value: The value associated with a property, which is of the type indicated in the property type associated with the property. A property may have multiple values.
Link: A connection between two data objects, based on, for example, a relationship, an event, and/or matching properties. Links may be directional, such as one representing a payment from person A to B, or bidirectional.
Link Set: Set of multiple links that are shared between two or more data objects.
Contextual view, context, view, data representation: A visual representation of data that may include various organizations, transformations, and/or restructuring of data so as to provide new perspectives to a viewer of the visualization. Examples of contexts include graphs, maps, tables, timelines, histograms, and/or lists, among others. Contextual views may include displaying individual pieces of data in, for example, various arrangements, various sizes, various colors, and/or may include multi-dimensional aspects. Contextual views may enable faster and more thorough understandings of sets of data and information.
Example User Interfaces
The example user interface 101 includes a particular graphical contextual view and/or data visualization 103 of various data objects and relationships between those data objects. In the example user interface 101 of
In this example contextual view 103, each person node (associated with person data objects), flight node (associated with flight data objects), financial account node (associated with financial account data objects), and computer node (associated with computer data objects) may have relationships and/or links with any of the other nodes through, for example, other objects such as payment objects. As is described in detail in reference to
Turning back to
Relationships between data objects may be stored as links, or in some embodiments, as properties, where a relationship may be detected between the properties. In some cases the links may be directional. For example, a payment link may have a direction associated with the payment, where one person object is a receiver of a payment, and another person object is the payer of payment.
In addition to visually showing graphical data visualization 103, the user interface 101 may allow various manipulations. For example, the various data objects of the context-sensitive viewing system may be searched using a search interface 102 (e.g., text string matching of object properties), inspected (e.g., properties and associated data viewed), filtered (e.g., narrowing the universe of objects into sets and subsets by properties or relationships), and statistically aggregated (e.g., numerically summarized based on summarization criteria), among other operations and visualizations. Further, the various data objects represented in the data visualization 103 may be moved, accessed, deleted from the interface, among other manipulations. Additional data objects and associated links may be added to the data visualization 103, and exiting data objects and links may be edited and/or otherwise altered.
The user interface 101 further includes a user-accessible drawer 130. The drawer 130 may be opened or closed by a user of the context-sensitive viewing system. The drawer 130 is shown in a closed state in
In the user interface 101 of
In an embodiment, the user may provide an input that causes the indicators 152, 154, 156, 158, 160 to be replaced with previews of the respective contexts (as shown in
As described below in reference to
In various embodiments other types of contextual previews may be provided. For example, a timeline context may be provided in which the various event associated with the data objects of graphical primary contextual view 103 may be mapped. In various embodiments, any other types of changes to the primary contextual view may be reflected in the secondary contextual views/previews. For example, removing data object, editing data objects or properties, and the like.
In various embodiments, the drawer 130 may appear at different locations on the user interface 101, and/or may be a different size. For example, the drawer 130 may appear at a top of the user interface, or on either side of the user interface. In an embodiment, the location, size, and/or other appearance of the drawer 130 may be user-configurable.
In the embodiment of
In
In
In an embodiment, multiple of the same type of contextual view may be accessible to the user. For example, timeline contextual views may be available.
Example Operations
At block 302, the context-sensitive viewing system receives user input at the first contextual view and/or data visualization user interface. In this present example, the first contextual view comprises the primary view with which the user is currently interacting. User inputs may include, for example, adding and/or deleting data objects, manipulating data objects, altering and/or editing data object properties and/or links, among other inputs.
At block 304, updates to the contextual view are determined by the context-sensitive viewing system based on the user input. For example, if the user provides an input to add a data object to the view, information associated with the data object to be added may be retrieved from a particular data store. The retrieved data object may then be displayed to the user and/or otherwise represented on the first/primary contextual view.
At block 306, the user may optionally select a different contextual view. Selecting a second contextual view may be accomplished in any of the ways described above in reference to the user interfaces of
At block 308, similar to block 304, updates to other contextual views may optionally be determined and displayed to the user. For example, when the user adds a data object to the first/primary contextual view, the same data object may be added to one or more other contextual views of the context-sensitive viewing system, as appropriate. In an example, when the user adds a person data object to a first graphical contextual view, the same person data object may be added to one or more other graphical contextual views. Further, the location(s) associated with that person data object may be added to one or more other map-based contextual views. Additionally, cables or other information, and/or user-generated snippets or notes associated with that person data object may be added to one or more other relevant contextual views.
The particular other contextual views that may be updated may depend on, for example, the particular implementation of the context-sensitive viewing system, user settings, and/or processing capability of the system. In an embodiment, at block 310, all other contextual views are updated simultaneously with, or very soon after, the updating of the first contextual. In another embodiment, at block 312, contextual views that are adjacent to the first view may be updated when the first view is updated. For example, in the embodiment of
In an embodiment, updating of other contextual views is determined based on processing capability available to the context-sensitive viewing system. For example, additional contextual views may be updated when more processing capability is available. In another example, updates to particular contextual views may be delayed until visible to the user so as to reduce power consumption and/or processing power.
Implementation Mechanisms
Turning to
The client device 402 may be any computing device capable of receiving input and providing output to a user. For example, the client device 402 may provide a contextual view of a data visualization to the user, among other functions. The client device 402 may also be capable of communicating over the network 408, for example, to request data objects, data visualization information, and/or contextual view information from the server device 404. In some embodiments, the client device 402 may include non-transitory computer-readable medium storage for storing data objects, data visualization information, and/or contextual view information. In an embodiment, the context-sensitive viewing system may include a plurality of client devices, each of which may communicate with each other, and with the network 408.
The network 408 may be any wired network, wireless network, or combination thereof. In addition, the network 408 may be a personal area network, local area network, wide area network, cable network, satellite network, cellular telephone network, or combination thereof. Protocols and components for communicating via the Internet or any of the other aforementioned types of communication networks are well known to those skilled in the art of computer communications and thus, need not be described in more detail herein.
The server device 404 is a computing device that may perform a variety of tasks to implement the contextual views and data visualizations of the context-sensitive viewing system. For example, the server device 404 may generate a user interface, including various contextual views, for display to the user via the client device 402. Alternatively, the server device 404 may receive requests for data and/or data objects from the client device 402, and may provide the requested data to the client device 402. The server device 404 may also generate requested data visualizations and/or contextual views that may be transmitted over the network 408, and provided to the user via the client device 402. Additional operations of the server device 404 and/or the client device 402 are described in further detail with respect to
The server device 404 may be in communication with the database 406. The database 406 may store one or more data objects, data visualization information, and/or contextual view information. The database 406 may be embodied in hard disk drives, solid state memories, and/or any other type of non-transitory, computer-readable storage medium remotely or locally accessible to the server device 404. The database 406 may also be distributed or partitioned across multiple storage devices as is known in the art without departing from the spirit and scope of the present disclosure.
According to various embodiments, the techniques described herein may be implemented by one or more special-purpose computing devices. The special-purpose computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the techniques, or may include one or more general purpose hardware processors programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the techniques. The special-purpose computing devices may be desktop computer systems, server computer systems, portable computer systems, handheld devices, networking devices or any other device or combination of devices that incorporate hard-wired and/or program logic to implement the techniques.
Computing device(s), such as the client device 402 and/or the server device 404, are generally controlled and coordinated by operating system software, such as iOS, Android, Chrome OS, Windows XP, Windows Vista, Windows 7, Windows 8, Windows Server, Windows CE, Unix, Linux, SunOS, Solaris, iOS, Blackberry OS, VxWorks, or other compatible operating systems. In other embodiments, a computing device may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface functionality, such as a graphical user interface (“GUI”), among other things.
For example,
Computer system 420 also includes a main memory 426, such as a random access memory (RAM), cache and/or other dynamic storage devices, coupled to bus 422 for storing information and instructions to be executed by processor 424. Main memory 426 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 424. Such instructions, when stored in storage media accessible to processor 424, render computer system 420 into a special-purpose machine that is customized to perform the operations specified in the instructions.
Computer system 420 further includes a read only memory (ROM) 428 or other static storage device coupled to bus 422 for storing static information and instructions for processor 424. A storage device 430, such as a magnetic disk, optical disk, or USB thumb drive (Flash drive), etc., is provided and coupled to bus 422 for storing information and instructions.
Computer system 420 may be coupled via bus 422 to a display 432, such as a cathode ray tube (CRT) or LCD display (or touch screen), for displaying information to a computer user. An input device 434, including alphanumeric and other keys, is coupled to bus 422 for communicating information and command selections to processor 424. Another type of user input device is cursor control 436, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 424 and for controlling cursor movement on display 432. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane. In some embodiments, the same direction information and command selections as cursor control may be implemented via receiving touches on a touch screen without a cursor.
Computer system 420 may also include one or more modules 452 that may, as described above and below, provide various functionality of the context-sensitive viewing system. For example, one module 452 may comprise the client-side context viewing module 410 of
In general, the word “module,” as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, Lua, C or C++. A software module may be compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language such as, for example, BASIC, Perl, or Python. It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts. Software modules configured for execution on computing devices may be provided on a computer readable medium, such as a compact disc, digital video disc, flash drive, magnetic disc, or any other tangible medium, or as a digital download (and may be originally stored in a compressed or installable format that requires installation, decompression or decryption prior to execution). Such software code may be stored, partially or fully, on a memory device of the executing computing device, for execution by the computing device. Software instructions may be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors. The modules or computing device functionality described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage
Computer system 420 may implement the techniques described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the computer system causes or programs computer system 420 to be a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 420 in response to processor(s) 424 executing one or more sequences of one or more instructions contained in main memory 426. Such instructions may be read into main memory 426 from another storage medium, such as storage device 430. Execution of the sequences of instructions contained in main memory 426 causes processor(s) 424 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.
The terms “non-transitory media,” “computer-readable media,” and similar terms, as used herein refers to any media that store data and/or instructions that cause a machine to operate in a specific fashion. Such non-transitory media may comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 430. Volatile media includes dynamic memory, such as main memory 426. Common forms of non-transitory media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge, and networked versions of the same.
Non-transitory media is distinct from but may be used in conjunction with transmission media. Transmission media participates in transferring information between non-transitory media. For example, transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 422. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
Various forms of media may be involved in carrying one or more sequences of one or more instructions to processor 424 for execution. For example, the instructions may initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 420 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 422. Bus 422 carries the data to main memory 426, from which processor 424 retrieves and executes the instructions. The instructions received by main memory 426 may retrieve and execute the instructions. The instructions received by main memory 426 may optionally be stored on storage device 430 either before or after execution by processor 424.
Computer system 420 also includes a communication interface 438 coupled to bus 422. Communication interface 438 provides a two-way data communication coupling to a network link 440 that is connected to a local network 442. For example, communication interface 438 may be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 438 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN (or WAN component to communicate with a WAN). Wireless links may also be implemented. In any such implementation, communication interface 438 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
Network link 440 typically provides data communication through one or more networks (for example, network 408 of
Computer system 420 can send messages and receive data, including program code, through the network(s), network link 440 and communication interface 438. In the Internet example, a server 450 might transmit a requested code for an application program through Internet 448, ISP 446, local network 442 and communication interface 438.
The received code may be executed by processor 424 as it is received, and/or stored in storage device 430, or other non-volatile storage for later execution.
In an embodiment, the context-sensitive viewing system is implemented by the computer system 420. For example, data objects, data visualization information, and/or contextual view information may be stored in the storage device 430, and/or in an external database accessible through the local network 442 (for example, database 406 of
The context-sensitive viewing system advantageously enables a user to view a particular set of data objects in multiple visualization contexts. Previews of the set of data in other visualization may be quickly reviewed by the user to determine the most beneficial context for information extraction. Further, manipulations by the user in one context are propagated to the other contexts, allowing fast analysis of the impacts of changes to the set of data.
Object Centric Data Model
In one embodiment, a body of data is conceptually structured according to an object-centric data model represented by ontology 505. The conceptual data model is independent of any particular database used for durably storing one or more database(s) 509 based on the ontology 505. For example, each object of the conceptual data model may correspond to one or more rows in a relational database or an entry in Lightweight Directory Access Protocol (LDAP) database, or any combination of one or more databases.
An ontology 505, as noted above, may include stored information providing a data model for storage of data in the database 509. The ontology 505 may be defined by one or more object types, which may each be associated with one or more property types. At the highest level of abstraction, data object 501 is a container for information representing things in the world. For example, data object 501 can represent an entity such as a person, a place, an organization, a market instrument, or other noun. Data object 501 can represent an event that happens at a point in time or for a duration. Data object 501 can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object 501 is associated with a unique identifier that uniquely identifies the data object within the database system.
Different types of data objects may have different property types. For example, a “Person” data object might have an “Eye Color” property type and an “Event” data object might have a “Date” property type. Each property 503 as represented by data in the database system 510 may have a property type defined by the ontology 505 used by the database 505.
Objects may be instantiated in the database 509 in accordance with the corresponding object definition for the particular object in the ontology 505. For example, a specific monetary payment (e.g., an object of type “event”) of US$30.00 (e.g., a property of type “currency”) taking place on Mar. 27, 2009 (e.g., a property of type “date”) may be stored in the database 509 as an event object with associated currency and date properties as defined within the ontology 505.
The data objects defined in the ontology 505 may support property multiplicity. In particular, a data object 501 may be allowed to have more than one property 503 of the same property type. For example, a “Person” data object might have multiple “Address” properties or multiple “Name” properties.
Each link 502 represents a connection between two data objects 501. In one embodiment, the connection is either through a relationship, an event, or through matching properties. A relationship connection may be asymmetrical or symmetrical. For example, “Person” data object A may be connected to “Person” data object B by a “Child Of” relationship (where “Person” data object B has an asymmetric “Parent Of” relationship to “Person” data object A), a “Kin Of” symmetric relationship to “Person” data object C, and an asymmetric “Member Of” relationship to “Organization” data object X. The type of relationship between two data objects may vary depending on the types of the data objects. For example, “Person” data object A may have an “Appears In” relationship with “Document” data object Y or have a “Participate In” relationship with “Event” data object E. As an example of an event connection, two “Person” data objects may be connected by an “Airline Flight” data object representing a particular airline flight if they traveled together on that flight, or by a “Meeting” data object representing a particular meeting if they both attended that meeting. In one embodiment, when two data objects are connected by an event, they are also connected by relationships, in which each data object has a specific relationship to the event, such as, for example, an “Appears In” relationship.
As an example of a matching properties connection, two “Person” data objects representing a brother and a sister, may both have an “Address” property that indicates where they live. If the brother and the sister live in the same home, then their “Address” properties likely contain similar, if not identical property values. In one embodiment, a link between two data objects may be established based on similar or matching properties (e.g., property types and/or property values) of the data objects. These are just some examples of the types of connections that may be represented by a link and other types of connections may be represented; embodiments are not limited to any particular types of connections between data objects. For example, a document might contain references to two different objects. For example, a document may contain a reference to a payment (one object), and a person (a second object). A link between these two objects may represent a connection between these two entities through their co-occurrence within the same document.
Each data object 501 can have multiple links with another data object 501 to form a link set 504. For example, two “Person” data objects representing a husband and a wife could be linked through a “Spouse Of” relationship, a matching “Address” property, and one or more matching “Event” properties (e.g., a wedding). Each link 502 as represented by data in a database may have a link type defined by the database ontology used by the database.
In accordance with the discussion above, the example ontology 505 comprises stored information providing the data model of data stored in database 509, and the ontology is defined by one or more object types 610, one or more property types 616, and one or more link types 630. Based on information determined by the parser 602 or other mapping of source input information to object type, one or more data objects 501 may be instantiated in the database 509 based on respective determined object types 610, and each of the objects 501 has one or more properties 503 that are instantiated based on property types 616. Two data objects 501 may be connected by one or more links 502 that may be instantiated based on link types 630. The property types 616 each may comprise one or more data types 618, such as a string, number, etc. Property types 616 may be instantiated based on a base property type 620. For example, a base property type 620 may be “Locations” and a property type 616 may be “Home.”
In an embodiment, a user of the system uses an object type editor 624 to create and/or modify the object types 610 and define attributes of the object types. In an embodiment, a user of the system uses a property type editor 626 to create and/or modify the property types 616 and define attributes of the property types. In an embodiment, a user of the system uses link type editor 628 to create the link types 630. Alternatively, other programs, processes, or programmatic controls may be used to create link types and property types and define attributes, and using editors is not required.
In an embodiment, creating a property type 616 using the property type editor 626 involves defining at least one parser definition using a parser editor 622. A parser definition comprises metadata that informs parser 602 how to parse input data 600 to determine whether values in the input data can be assigned to the property type 616 that is associated with the parser definition. In an embodiment, each parser definition may comprise a regular expression parser 604A or a code module parser 604B. In other embodiments, other kinds of parser definitions may be provided using scripts or other programmatic elements. Once defined, both a regular expression parser 604A and a code module parser 604B can provide input to parser 602 to control parsing of input data 600.
Using the data types defined in the ontology, input data 600 may be parsed by the parser 602 determine which object type 610 should receive data from a record created from the input data, and which property types 616 should be assigned to data from individual field values in the input data. Based on the object-property mapping 601, the parser 602 selects one of the parser definitions that is associated with a property type in the input data. The parser parses an input data field using the selected parser definition, resulting in creating new or modified data 603. The new or modified data 603 is added to the database 509 according to ontology 505 by storing values of the new or modified data in a property of the specified property type. As a result, input data 600 having varying format or syntax can be created in database 509. The ontology 505 may be modified at any time using object type editor 624, property type editor 626, and link type editor 628, or under program control without human use of an editor. Parser editor 622 enables creating multiple parser definitions that can successfully parse input data 600 having varying format or syntax and determine which property types should be used to transform input data 600 into new or modified input data 603.
The properties, objects, and links (e.g. relationships) between the objects can be visualized using a graphical user interface (GUI). For example, as described above,
Additional Implementation Details
Each of the processes, methods, and algorithms described in the preceding sections may be embodied in, and fully or partially automated by, code modules executed by one or more computer systems or computer processors comprising computer hardware. The processes and algorithms may be implemented partially or wholly in application-specific circuitry.
The various features and processes described above may be used independently of one another, or may be combined in various ways. All possible combinations and subcombinations are intended to fall within the scope of this disclosure. In addition, certain method or process blocks may be omitted in some implementations. The methods and processes described herein are also not limited to any particular sequence, and the blocks or states relating thereto can be performed in other sequences that are appropriate. For example, described blocks or states may be performed in an order other than that specifically disclosed, or multiple blocks or states may be combined in a single block or state. The example blocks or states may be performed in serial, in parallel, or in some other manner. Blocks or states may be added to or removed from the disclosed example embodiments. The example systems and components described herein may be configured differently than described. For example, elements may be added to, removed from, or rearranged compared to the disclosed example embodiments.
Conditional language, such as, among others, “can,” “could,” “might,” or “may,” unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.
The term “comprising” as used herein should be given an inclusive rather than exclusive interpretation. For example, a general purpose computer comprising one or more processors should not be interpreted as excluding other computer components, and may possibly include such components as memory, input/output devices, and/or network interfaces, among others.
Any process descriptions, elements, or blocks in the flow diagrams described herein and/or depicted in the attached figures should be understood as potentially representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps in the process. Alternate implementations are included within the scope of the embodiments described herein in which elements or functions may be deleted, executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those skilled in the art.
It should be emphasized that many variations and modifications may be made to the above-described embodiments, the elements of which are to be understood as being among other acceptable examples. All such modifications and variations are intended to be included herein within the scope of this disclosure. The foregoing description details certain embodiments of the invention. It will be appreciated, however, that no matter how detailed the foregoing appears in text, the invention can be practiced in many ways. As is also stated above, it should be noted that the use of particular terminology when describing certain features or aspects of the invention should not be taken to imply that the terminology is being re-defined herein to be restricted to including any specific characteristics of the features or aspects of the invention with which that terminology is associated. The scope of the invention should therefore be construed in accordance with the appended claims and any equivalents thereof.
This application is a continuation of U.S. application Ser. No. 14/242,559, filed Apr. 1, 2014, and titled “CONTEXT-SENSITIVE VIEWS,” which is a continuation of U.S. application Ser. No. 14/095,798, filed Dec. 3, 2013, and titled “CONTEXT-SENSITIVE VIEWS,” which application claims a priority benefit under 35 U.S.C. § 119 to U.S. Provisional Patent Application No. 61/864,048, filed on Aug. 9, 2013, and titled “CONTEXT-SENSITIVE VIEWS”. All of the above-identified applications are hereby incorporated by reference herein in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
5109399 | Thompson | Apr 1992 | A |
5329108 | Lamoure | Jul 1994 | A |
5670987 | Doi et al. | Sep 1997 | A |
5781704 | Rossmo | Jul 1998 | A |
5798769 | Chiu et al. | Aug 1998 | A |
5845300 | Comer | Dec 1998 | A |
6057757 | Arrowsmith et al. | May 2000 | A |
6091956 | Hollenberg | Jul 2000 | A |
6161098 | Wallman | Dec 2000 | A |
6219053 | Tachibana et al. | Apr 2001 | B1 |
6232971 | Haynes | May 2001 | B1 |
6247019 | Davies | Jun 2001 | B1 |
6279018 | Kudrolli et al. | Aug 2001 | B1 |
6341310 | Leshem et al. | Jan 2002 | B1 |
6366933 | Ball et al. | Apr 2002 | B1 |
6369835 | Lin | Apr 2002 | B1 |
6456997 | Shukla | Sep 2002 | B1 |
6549944 | Weinberg et al. | Apr 2003 | B1 |
6560620 | Ching | May 2003 | B1 |
6581068 | Bensoussan et al. | Jun 2003 | B1 |
6594672 | Lampson et al. | Jul 2003 | B1 |
6631496 | Li et al. | Oct 2003 | B1 |
6642945 | Sharpe | Nov 2003 | B1 |
6714936 | Nevin, III | Mar 2004 | B1 |
6775675 | Nwabueze et al. | Aug 2004 | B1 |
6828920 | Owen et al. | Dec 2004 | B2 |
6839745 | Dingari et al. | Jan 2005 | B1 |
6877137 | Rivette et al. | Apr 2005 | B1 |
6976210 | Silva et al. | Dec 2005 | B1 |
6980984 | Huffman et al. | Dec 2005 | B1 |
6985950 | Hanson et al. | Jan 2006 | B1 |
7036085 | Barros | Apr 2006 | B2 |
7043702 | Chi et al. | May 2006 | B2 |
7055110 | Kupka et al. | May 2006 | B2 |
7139800 | Bellotti et al. | Nov 2006 | B2 |
7158878 | Rasmussen et al. | Jan 2007 | B2 |
7162475 | Ackerman | Jan 2007 | B2 |
7168039 | Bertram | Jan 2007 | B2 |
7171427 | Witowski et al. | Jan 2007 | B2 |
7269786 | Malloy et al. | Sep 2007 | B1 |
7278105 | Kitts | Oct 2007 | B1 |
7290698 | Poslinski et al. | Nov 2007 | B2 |
7333998 | Heckerman et al. | Feb 2008 | B2 |
7370047 | Gorman | May 2008 | B2 |
7379811 | Rasmussen et al. | May 2008 | B2 |
7379903 | Caballero et al. | May 2008 | B2 |
7426654 | Adams et al. | Sep 2008 | B2 |
7454466 | Bellotti et al. | Nov 2008 | B2 |
7467375 | Tondreau et al. | Dec 2008 | B2 |
7487139 | Fraleigh et al. | Feb 2009 | B2 |
7502786 | Liu et al. | Mar 2009 | B2 |
7525422 | Bishop et al. | Apr 2009 | B2 |
7529727 | Arning et al. | May 2009 | B2 |
7558677 | Jones | Jul 2009 | B2 |
7574428 | Leiserowitz et al. | Aug 2009 | B2 |
7579965 | Bucholz | Aug 2009 | B2 |
7596285 | Brown et al. | Sep 2009 | B2 |
7614006 | Molander | Nov 2009 | B2 |
7617232 | Gabbert et al. | Nov 2009 | B2 |
7620628 | Kapur et al. | Nov 2009 | B2 |
7627812 | Chamberlain et al. | Dec 2009 | B2 |
7634717 | Chamberlain et al. | Dec 2009 | B2 |
7703021 | Flam | Apr 2010 | B1 |
7712049 | Williams et al. | May 2010 | B2 |
7716077 | Mikurak | May 2010 | B1 |
7725530 | Sah et al. | May 2010 | B2 |
7725547 | Albertson et al. | May 2010 | B2 |
7730082 | Sah et al. | Jun 2010 | B2 |
7730109 | Rohrs et al. | Jun 2010 | B2 |
7770100 | Chamberlain et al. | Aug 2010 | B2 |
7805457 | Viola et al. | Sep 2010 | B1 |
7809703 | Balabhadrapatruni et al. | Oct 2010 | B2 |
7818658 | Chen | Oct 2010 | B2 |
7870493 | Pall et al. | Jan 2011 | B2 |
7894984 | Rasmussen et al. | Feb 2011 | B2 |
7899611 | Downs et al. | Mar 2011 | B2 |
7917376 | Bellin et al. | Mar 2011 | B2 |
7920963 | Jouline et al. | Apr 2011 | B2 |
7933862 | Chamberlain et al. | Apr 2011 | B2 |
7962281 | Rasmussen et al. | Jun 2011 | B2 |
7962495 | Jain et al. | Jun 2011 | B2 |
7962848 | Bertram | Jun 2011 | B2 |
7970240 | Chao et al. | Jun 2011 | B1 |
7971150 | Raskutti et al. | Jun 2011 | B2 |
7984374 | Caro et al. | Jul 2011 | B2 |
8001465 | Kudrolli et al. | Aug 2011 | B2 |
8001482 | Bhattiprolu et al. | Aug 2011 | B2 |
8010545 | Stefik et al. | Aug 2011 | B2 |
8015487 | Roy et al. | Sep 2011 | B2 |
8024778 | Cash et al. | Sep 2011 | B2 |
8036632 | Cona et al. | Oct 2011 | B1 |
8103543 | Zwicky | Jan 2012 | B1 |
8134457 | Velipasalar et al. | Mar 2012 | B2 |
8145703 | Frishert et al. | Mar 2012 | B2 |
8185819 | Sah et al. | May 2012 | B2 |
8214361 | Sandler et al. | Jul 2012 | B1 |
8214764 | Gemmell et al. | Jul 2012 | B2 |
8225201 | Michael | Jul 2012 | B2 |
8229947 | Fujinaga | Jul 2012 | B2 |
8230333 | Decherd et al. | Jul 2012 | B2 |
8271461 | Pike et al. | Sep 2012 | B2 |
8280880 | Aymeloglu et al. | Oct 2012 | B1 |
8290942 | Jones et al. | Oct 2012 | B2 |
8301464 | Cave et al. | Oct 2012 | B1 |
8301904 | Gryaznov | Oct 2012 | B1 |
8312367 | Foster | Nov 2012 | B2 |
8312546 | Alme | Nov 2012 | B2 |
8352881 | Champion et al. | Jan 2013 | B2 |
8368695 | Howell et al. | Feb 2013 | B2 |
8397171 | Klassen et al. | Mar 2013 | B2 |
8397177 | Barros | Mar 2013 | B2 |
8412707 | Mianji | Apr 2013 | B1 |
8447722 | Ahuja et al. | May 2013 | B1 |
8452790 | Mianji | May 2013 | B1 |
8463036 | Ramesh et al. | Jun 2013 | B1 |
8489331 | Kopf et al. | Jul 2013 | B2 |
8489641 | Seefeld et al. | Jul 2013 | B1 |
8498984 | Hwang et al. | Jul 2013 | B1 |
8514082 | Cova et al. | Aug 2013 | B2 |
8515207 | Chau | Aug 2013 | B2 |
8554579 | Tribble et al. | Oct 2013 | B2 |
8577911 | Stepinski et al. | Nov 2013 | B1 |
8589273 | Creeden et al. | Nov 2013 | B2 |
8595234 | Siripuapu et al. | Nov 2013 | B2 |
8620641 | Farnsworth et al. | Dec 2013 | B2 |
8639757 | Zang et al. | Jan 2014 | B1 |
8646080 | Williamson et al. | Feb 2014 | B2 |
8676857 | Adams et al. | Mar 2014 | B1 |
8689108 | Duffield et al. | Apr 2014 | B1 |
8713467 | Goldenberg | Apr 2014 | B1 |
8726379 | Stiansen et al. | May 2014 | B1 |
8739278 | Varghese | May 2014 | B2 |
8742934 | Sarpy et al. | Jun 2014 | B1 |
8745516 | Mason et al. | Jun 2014 | B2 |
8781169 | Jackson et al. | Jul 2014 | B2 |
8787939 | Papakipos et al. | Jul 2014 | B2 |
8799799 | Cervelli et al. | Aug 2014 | B1 |
8812960 | Sun et al. | Aug 2014 | B1 |
8830322 | Nerayoff et al. | Sep 2014 | B2 |
8832594 | Thompson et al. | Sep 2014 | B1 |
8868537 | Colgrove et al. | Oct 2014 | B1 |
8909656 | Kumar | Dec 2014 | B2 |
8917274 | Ma et al. | Dec 2014 | B2 |
8924872 | Bogomolov et al. | Dec 2014 | B1 |
8937619 | Sharma et al. | Jan 2015 | B2 |
8938686 | Erenrich et al. | Jan 2015 | B1 |
9009171 | Grossman et al. | Apr 2015 | B1 |
9009827 | Albertson et al. | Apr 2015 | B1 |
9021260 | Falk et al. | Apr 2015 | B1 |
9021384 | Beard | Apr 2015 | B1 |
9043696 | Meiklejohn et al. | May 2015 | B1 |
9043894 | Dennison et al. | May 2015 | B1 |
9116975 | Shankar | Aug 2015 | B2 |
9335911 | Elliot | May 2016 | B1 |
9454281 | Ward | Sep 2016 | B2 |
9454785 | Hunter | Sep 2016 | B1 |
9557882 | Goldenberg et al. | Jan 2017 | B2 |
9646396 | Sharma | May 2017 | B2 |
9727560 | Chakerian | Aug 2017 | B2 |
9734217 | Kara | Aug 2017 | B2 |
9741138 | Friedlander | Aug 2017 | B2 |
9767172 | Fackler | Sep 2017 | B2 |
9779525 | Sharma | Oct 2017 | B2 |
9785317 | Duffield | Oct 2017 | B2 |
9785328 | Slawinski | Oct 2017 | B2 |
9785773 | Falk | Oct 2017 | B2 |
20020033848 | Sciammarella et al. | Mar 2002 | A1 |
20020091707 | Keller | Jul 2002 | A1 |
20020095658 | Shulman | Jul 2002 | A1 |
20020116120 | Ruiz et al. | Aug 2002 | A1 |
20020130907 | Chi et al. | Sep 2002 | A1 |
20020174201 | Ramer et al. | Nov 2002 | A1 |
20020194119 | Wright et al. | Dec 2002 | A1 |
20030028560 | Kudrolli et al. | Feb 2003 | A1 |
20030039948 | Donahue | Feb 2003 | A1 |
20030144868 | MacIntyre et al. | Jul 2003 | A1 |
20030163352 | Surpin et al. | Aug 2003 | A1 |
20030225755 | Iwayama et al. | Dec 2003 | A1 |
20030229848 | Arend et al. | Dec 2003 | A1 |
20040032432 | Baynger | Feb 2004 | A1 |
20040064256 | Barinek et al. | Apr 2004 | A1 |
20040085318 | Hassler et al. | May 2004 | A1 |
20040095349 | Bito et al. | May 2004 | A1 |
20040111410 | Burgoon et al. | Jun 2004 | A1 |
20040126840 | Cheng et al. | Jul 2004 | A1 |
20040143602 | Ruiz et al. | Jul 2004 | A1 |
20040143796 | Lerner et al. | Jul 2004 | A1 |
20040163039 | Gorman | Aug 2004 | A1 |
20040181554 | Heckerman et al. | Sep 2004 | A1 |
20040193600 | Kaasten et al. | Sep 2004 | A1 |
20040221223 | Yu et al. | Nov 2004 | A1 |
20040260702 | Cragun et al. | Dec 2004 | A1 |
20050027705 | Sadri et al. | Feb 2005 | A1 |
20050028094 | Allyn | Feb 2005 | A1 |
20050039119 | Parks et al. | Feb 2005 | A1 |
20050080769 | Gemmell | Apr 2005 | A1 |
20050086207 | Heuer et al. | Apr 2005 | A1 |
20050125715 | Di Franco et al. | Jun 2005 | A1 |
20050162523 | Darrell et al. | Jul 2005 | A1 |
20050180330 | Shapiro | Aug 2005 | A1 |
20050182793 | Keenan et al. | Aug 2005 | A1 |
20050183005 | Denoue et al. | Aug 2005 | A1 |
20050246327 | Yeung et al. | Nov 2005 | A1 |
20050251786 | Citron et al. | Nov 2005 | A1 |
20060026120 | Carolan et al. | Feb 2006 | A1 |
20060026170 | Kreitler et al. | Feb 2006 | A1 |
20060059139 | Robinson | Mar 2006 | A1 |
20060074881 | Vembu et al. | Apr 2006 | A1 |
20060080619 | Carlson et al. | Apr 2006 | A1 |
20060129746 | Porter | Jun 2006 | A1 |
20060139375 | Rasmussen et al. | Jun 2006 | A1 |
20060142949 | Helt | Jun 2006 | A1 |
20060149596 | Surpin et al. | Jul 2006 | A1 |
20060203337 | White | Sep 2006 | A1 |
20060218637 | Thomas et al. | Sep 2006 | A1 |
20060241974 | Chao et al. | Oct 2006 | A1 |
20060242040 | Rader et al. | Oct 2006 | A1 |
20060242630 | Koike et al. | Oct 2006 | A1 |
20060271277 | Hu et al. | Nov 2006 | A1 |
20060279630 | Aggarwal et al. | Dec 2006 | A1 |
20070011150 | Frank | Jan 2007 | A1 |
20070016363 | Huang et al. | Jan 2007 | A1 |
20070038962 | Fuchs et al. | Feb 2007 | A1 |
20070057966 | Ohno et al. | Mar 2007 | A1 |
20070078832 | Ott et al. | Apr 2007 | A1 |
20070083541 | Fraleigh et al. | Apr 2007 | A1 |
20070150369 | Zivin | Jun 2007 | A1 |
20070174760 | Chamberlain et al. | Jul 2007 | A1 |
20070192265 | Chopin et al. | Aug 2007 | A1 |
20070198571 | Ferguson et al. | Aug 2007 | A1 |
20070208497 | Downs et al. | Sep 2007 | A1 |
20070208498 | Barker et al. | Sep 2007 | A1 |
20070208736 | Tanigawa et al. | Sep 2007 | A1 |
20070240062 | Christena et al. | Oct 2007 | A1 |
20070266336 | Nojima et al. | Nov 2007 | A1 |
20070294643 | Kyle | Dec 2007 | A1 |
20080040684 | Crump | Feb 2008 | A1 |
20080051989 | Welsh | Feb 2008 | A1 |
20080052142 | Bailey et al. | Feb 2008 | A1 |
20080077597 | Butler | Mar 2008 | A1 |
20080077642 | Carbone et al. | Mar 2008 | A1 |
20080104019 | Nath | May 2008 | A1 |
20080126951 | Sood et al. | May 2008 | A1 |
20080155440 | Trevor et al. | Jun 2008 | A1 |
20080195417 | Surpin et al. | Aug 2008 | A1 |
20080195608 | Clover | Aug 2008 | A1 |
20080222295 | Robinson et al. | Sep 2008 | A1 |
20080263468 | Cappione et al. | Oct 2008 | A1 |
20080267107 | Rosenberg | Oct 2008 | A1 |
20080276167 | Michael | Nov 2008 | A1 |
20080278311 | Grange et al. | Nov 2008 | A1 |
20080288306 | MacIntyre et al. | Nov 2008 | A1 |
20080301643 | Appleton et al. | Dec 2008 | A1 |
20090002492 | Velipasalar et al. | Jan 2009 | A1 |
20090027418 | Maru et al. | Jan 2009 | A1 |
20090030915 | Winter et al. | Jan 2009 | A1 |
20090055251 | Shah et al. | Feb 2009 | A1 |
20090076845 | Bellin et al. | Mar 2009 | A1 |
20090088964 | Schaaf et al. | Apr 2009 | A1 |
20090119309 | Gibson et al. | May 2009 | A1 |
20090125369 | Kloosstra et al. | May 2009 | A1 |
20090125459 | Norton et al. | May 2009 | A1 |
20090132921 | Hwangbo et al. | May 2009 | A1 |
20090132953 | Reed et al. | May 2009 | A1 |
20090144262 | White et al. | Jun 2009 | A1 |
20090144274 | Fraleigh et al. | Jun 2009 | A1 |
20090164934 | Bhattiprolu et al. | Jun 2009 | A1 |
20090171939 | Athsani et al. | Jul 2009 | A1 |
20090172511 | Decherd et al. | Jul 2009 | A1 |
20090177962 | Gusmorino et al. | Jul 2009 | A1 |
20090179892 | Tsuda et al. | Jul 2009 | A1 |
20090222400 | Kupershmidt et al. | Sep 2009 | A1 |
20090222760 | Halverson et al. | Sep 2009 | A1 |
20090234720 | George et al. | Sep 2009 | A1 |
20090249244 | Robinson et al. | Oct 2009 | A1 |
20090281839 | Lynn et al. | Nov 2009 | A1 |
20090287470 | Farnsworth et al. | Nov 2009 | A1 |
20090292626 | Oxford | Nov 2009 | A1 |
20100011282 | Dollard et al. | Jan 2010 | A1 |
20100042922 | Bradateanu et al. | Feb 2010 | A1 |
20100057716 | Stefik et al. | Mar 2010 | A1 |
20100070523 | Delgo et al. | Mar 2010 | A1 |
20100070842 | Aymeloglu et al. | Mar 2010 | A1 |
20100070845 | Facemire et al. | Mar 2010 | A1 |
20100070897 | Aymeloglu et al. | Mar 2010 | A1 |
20100100963 | Mahaffey | Apr 2010 | A1 |
20100103124 | Kruzeniski et al. | Apr 2010 | A1 |
20100114887 | Conway et al. | May 2010 | A1 |
20100122152 | Chamberlain et al. | May 2010 | A1 |
20100131457 | Heimendinger | May 2010 | A1 |
20100162176 | Dunton | Jun 2010 | A1 |
20100191563 | Schlaifer et al. | Jul 2010 | A1 |
20100198684 | Eraker et al. | Aug 2010 | A1 |
20100199225 | Coleman et al. | Aug 2010 | A1 |
20100228812 | Uomini | Sep 2010 | A1 |
20100250412 | Wagner | Sep 2010 | A1 |
20100280857 | Liu et al. | Nov 2010 | A1 |
20100293174 | Bennett et al. | Nov 2010 | A1 |
20100306713 | Geisner et al. | Dec 2010 | A1 |
20100313119 | Baldwin et al. | Dec 2010 | A1 |
20100318924 | Frankel et al. | Dec 2010 | A1 |
20100321399 | Ellren et al. | Dec 2010 | A1 |
20100325526 | Ellis et al. | Dec 2010 | A1 |
20100325581 | Finkelstein et al. | Dec 2010 | A1 |
20100330801 | Rouh | Dec 2010 | A1 |
20110029526 | Knight et al. | Feb 2011 | A1 |
20110047159 | Baid et al. | Feb 2011 | A1 |
20110060753 | Shaked et al. | Mar 2011 | A1 |
20110061013 | Bilicki et al. | Mar 2011 | A1 |
20110074811 | Hanson et al. | Mar 2011 | A1 |
20110078055 | Faribault et al. | Mar 2011 | A1 |
20110078173 | Seligmann et al. | Mar 2011 | A1 |
20110117878 | Barash et al. | May 2011 | A1 |
20110119100 | Ruhl et al. | May 2011 | A1 |
20110137766 | Rasmussen et al. | Jun 2011 | A1 |
20110153384 | Horne et al. | Jun 2011 | A1 |
20110161096 | Buehler et al. | Jun 2011 | A1 |
20110167105 | Ramakrishnan et al. | Jul 2011 | A1 |
20110170799 | Carrino et al. | Jul 2011 | A1 |
20110173032 | Payne et al. | Jul 2011 | A1 |
20110185316 | Reid et al. | Jul 2011 | A1 |
20110208724 | Jones et al. | Aug 2011 | A1 |
20110218934 | Elser | Sep 2011 | A1 |
20110219450 | McDougal et al. | Sep 2011 | A1 |
20110225198 | Edwards et al. | Sep 2011 | A1 |
20110238553 | Raj et al. | Sep 2011 | A1 |
20110258158 | Resende et al. | Oct 2011 | A1 |
20110270705 | Parker | Nov 2011 | A1 |
20110289397 | Eastmond et al. | Nov 2011 | A1 |
20110289407 | Naik et al. | Nov 2011 | A1 |
20110291851 | Whisenant | Dec 2011 | A1 |
20110310005 | Chen et al. | Dec 2011 | A1 |
20110314007 | Dassa et al. | Dec 2011 | A1 |
20120019559 | Siler et al. | Jan 2012 | A1 |
20120036013 | Neuhaus et al. | Feb 2012 | A1 |
20120036434 | Oberstein | Feb 2012 | A1 |
20120050293 | Carlhian et al. | Mar 2012 | A1 |
20120066296 | Appleton et al. | Mar 2012 | A1 |
20120072825 | Sherkin et al. | Mar 2012 | A1 |
20120079363 | Folting et al. | Mar 2012 | A1 |
20120106801 | Jackson | May 2012 | A1 |
20120117082 | Koperda et al. | May 2012 | A1 |
20120131512 | Takeuchi et al. | May 2012 | A1 |
20120144335 | Abeln et al. | Jun 2012 | A1 |
20120159307 | Chung et al. | Jun 2012 | A1 |
20120159362 | Brown et al. | Jun 2012 | A1 |
20120159399 | Bastide et al. | Jun 2012 | A1 |
20120173985 | Peppel | Jul 2012 | A1 |
20120196557 | Reich et al. | Aug 2012 | A1 |
20120196558 | Reich et al. | Aug 2012 | A1 |
20120208636 | Feige | Aug 2012 | A1 |
20120221511 | Gibson et al. | Aug 2012 | A1 |
20120221553 | Wittmer et al. | Aug 2012 | A1 |
20120221580 | Barney | Aug 2012 | A1 |
20120245976 | Kumar et al. | Sep 2012 | A1 |
20120246148 | Dror | Sep 2012 | A1 |
20120254129 | Wheeler et al. | Oct 2012 | A1 |
20120290879 | Shibuya et al. | Nov 2012 | A1 |
20120296907 | Long et al. | Nov 2012 | A1 |
20120311684 | Paulsen et al. | Dec 2012 | A1 |
20120323888 | Osann, Jr. | Dec 2012 | A1 |
20120330973 | Ghuneim et al. | Dec 2012 | A1 |
20130006725 | Simanek et al. | Jan 2013 | A1 |
20130018796 | Kolhatkar et al. | Jan 2013 | A1 |
20130046842 | Muntz et al. | Feb 2013 | A1 |
20130060786 | Serrano et al. | Mar 2013 | A1 |
20130061169 | Pearcy et al. | Mar 2013 | A1 |
20130073377 | Heath | Mar 2013 | A1 |
20130073454 | Busch | Mar 2013 | A1 |
20130078943 | Biage et al. | Mar 2013 | A1 |
20130097482 | Marantz et al. | Apr 2013 | A1 |
20130101159 | Chao et al. | Apr 2013 | A1 |
20130110822 | Ikeda et al. | May 2013 | A1 |
20130111320 | Campbell et al. | May 2013 | A1 |
20130117651 | Waldman et al. | May 2013 | A1 |
20130150004 | Rosen | Jun 2013 | A1 |
20130151148 | Parundekar et al. | Jun 2013 | A1 |
20130157234 | Gulli et al. | Jun 2013 | A1 |
20130166550 | Buchmann et al. | Jun 2013 | A1 |
20130176321 | Mitchell et al. | Jul 2013 | A1 |
20130179420 | Park et al. | Jul 2013 | A1 |
20130224696 | Wolfe et al. | Aug 2013 | A1 |
20130226953 | Markovich et al. | Aug 2013 | A1 |
20130238616 | Rose et al. | Sep 2013 | A1 |
20130246170 | Gross et al. | Sep 2013 | A1 |
20130251233 | Yang et al. | Sep 2013 | A1 |
20130262527 | Hunter et al. | Oct 2013 | A1 |
20130263019 | Castellanos et al. | Oct 2013 | A1 |
20130267207 | Hao et al. | Oct 2013 | A1 |
20130268520 | Fisher et al. | Oct 2013 | A1 |
20130279757 | Kephart | Oct 2013 | A1 |
20130282696 | John et al. | Oct 2013 | A1 |
20130290011 | Lynn et al. | Oct 2013 | A1 |
20130290825 | Arndt et al. | Oct 2013 | A1 |
20130297619 | Chandarsekaran et al. | Nov 2013 | A1 |
20130311375 | Priebatsch | Nov 2013 | A1 |
20140019936 | Cohanoff | Jan 2014 | A1 |
20140032506 | Hoey et al. | Jan 2014 | A1 |
20140033010 | Richardt et al. | Jan 2014 | A1 |
20140040371 | Gurevich et al. | Feb 2014 | A1 |
20140047357 | Alfaro et al. | Feb 2014 | A1 |
20140059038 | McPherson et al. | Feb 2014 | A1 |
20140068487 | Steiger et al. | Mar 2014 | A1 |
20140095273 | Tang et al. | Apr 2014 | A1 |
20140095509 | Patton | Apr 2014 | A1 |
20140108068 | Williams | Apr 2014 | A1 |
20140108380 | Gotz et al. | Apr 2014 | A1 |
20140108985 | Scott et al. | Apr 2014 | A1 |
20140129261 | Bothwell et al. | May 2014 | A1 |
20140149436 | Bahrami et al. | May 2014 | A1 |
20140156527 | Grigg et al. | Jun 2014 | A1 |
20140157172 | Peery et al. | Jun 2014 | A1 |
20140164502 | Khodorenko et al. | Jun 2014 | A1 |
20140164966 | Kim | Jun 2014 | A1 |
20140189536 | Lange et al. | Jul 2014 | A1 |
20140195515 | Baker et al. | Jul 2014 | A1 |
20140195887 | Ellis et al. | Jul 2014 | A1 |
20140267294 | Ma | Sep 2014 | A1 |
20140267295 | Sharma | Sep 2014 | A1 |
20140279824 | Tamayo | Sep 2014 | A1 |
20140316911 | Gross | Oct 2014 | A1 |
20140333651 | Cervelli et al. | Nov 2014 | A1 |
20140337772 | Cervelli et al. | Nov 2014 | A1 |
20150019394 | Unser et al. | Jan 2015 | A1 |
20150046870 | Goldenberg et al. | Feb 2015 | A1 |
20150089424 | Duffield et al. | Mar 2015 | A1 |
20150100897 | Sun et al. | Apr 2015 | A1 |
20150100907 | Erenrich et al. | Apr 2015 | A1 |
20150134666 | Gattiker et al. | May 2015 | A1 |
20150169709 | Kara et al. | Jun 2015 | A1 |
20150169726 | Kara et al. | Jun 2015 | A1 |
20150170077 | Kara et al. | Jun 2015 | A1 |
20160188143 | Kohlmeier | Jun 2016 | A1 |
Number | Date | Country |
---|---|---|
102014103482 | Sep 2014 | DE |
102014215621 | Feb 2015 | DE |
1672527 | Jun 2006 | EP |
2551799 | Jan 2013 | EP |
2778977 | Sep 2014 | EP |
2835745 | Feb 2015 | EP |
2835770 | Feb 2015 | EP |
2838039 | Feb 2015 | EP |
2846241 | Mar 2015 | EP |
2851852 | Mar 2015 | EP |
2858014 | Apr 2015 | EP |
2858018 | Apr 2015 | EP |
2863326 | Apr 2015 | EP |
2863346 | Apr 2015 | EP |
2869211 | May 2015 | EP |
2884439 | Jun 2015 | EP |
2884440 | Jun 2015 | EP |
2516155 | Jan 2015 | GB |
2518745 | Apr 2015 | GB |
2012778 | Nov 2014 | NL |
2013306 | Feb 2015 | NL |
624557 | Dec 2014 | NZ |
WO 2000009529 | Feb 2000 | WO |
WO 2005104736 | Nov 2005 | WO |
WO 2009061501 | May 2009 | WO |
WO 2010000014 | Jan 2010 | WO |
WO 2010030913 | Mar 2010 | WO |
WO 2013010157 | Jan 2013 | WO |
WO 2013102892 | Jul 2013 | WO |
Entry |
---|
“A First Look: Predicting Market Demand for Food Retail using a Huff Analysis,” TRF Policy Solutions, Jul. 2012, pp. 30. |
“A Quick Guide to UniProtKB Swiss-Prot & TrEMBL,” Sep. 2011, pp. 2. |
“A Word About Banks and the Laundering of Drug Money,” Aug. 18, 2012, http://www.golemxiv.co.uk/2012/08/a-word-about-banks-and-the-laundering-of-drug-money/. |
Acklen, Laura, “Absolute Beginner's Guide to Microsoft Word 2003,” Dec. 24, 2003, pp. 15-18, 34-41, 308-316. |
Amnet, “5 Great Tools for Visualizing Your Twitter Followers,” posted Aug. 4, 2010, http://www.amnetblog.com/component/content/article/115-5-grate-tools-for-visualizing-your-twitter-followers.html. |
Ananiev et al., “The New Modality API,” http://web.archive.org/web/20061211011958/http://java.sun.com/developer/technicalArticles/J2SE/Desktop/javase6/modality/ Jan. 21, 2006, pp. 8. |
Bluttman et al., “Excel Formulas and Functions for Dummies,” 2005, Wiley Publishing, Inc., pp. 280, 284-286. |
Boyce, Jim, “Microsoft Outlook 2010 Inside Out,” Aug. 1, 2010, retrieved from the internet https://capdtron.files.wordpress.com/2013/01/outlook-2010-inside_out.pdf. |
Bugzilla@Mozilla, “Bug 18726—[feature] Long-click means of invoking contextual menus not supported,” http://bugzilla.mozilla.org/show_bug.cgi?id=18726 printed Jun. 13, 2013 in 11 pages. |
Canese et al., “Chapter 2: PubMed: The Bibliographic Database,” The NCBI Handbook, Oct. 2002, pp. 1-10. |
Celik, Tantek, “CSS Basic User Interface Module Level 3 (CSS3 UI),” Section 8 Resizing and Overflow, Jan. 17, 2012, retrieved from internet http://www.w3.org/TR/2012/WD-css3-ui-20120117/#resizing-amp-overflow retrieved on May 18, 2015. |
Chen et al., “Bringing Order to the Web: Automatically Categorizing Search Results,” CHI 2000, Proceedings of the SIGCHI conference on Human Factors in Computing Systems, Apr. 1-6, 2000, The Hague, The Netherlands, pp. 145-152. |
Chung, Chin-Wan, “Dataplex: An Access to Heterogeneous Distributed Databases,” Communications of the ACM, Association for Computing Machinery, Inc., vol. 33, No. 1, Jan. 1, 1990, pp. 70-80. |
Conner, Nancy, “Google Apps: The Missing Manual,” May 1, 2008, pp. 15. |
Definition “Identify”, downloaded Jan. 22, 2015, 1 page. |
Definition “Overlay”, downloaded Jan. 22, 2015, 1 page. |
Delcher et al., “Identifying Bacterial Genes and Endosymbiont DNA with Glimmer,” BioInformatics, vol. 23, No. 6, 2007, pp. 673-679. |
Dramowicz, Ela, “Retail Trade Area Analysis Using the Huff Model,” Directions Magazine, Jul. 2, 2005 in 10 pages, http://www.directionsmag.com/articles/retail-trade-area-analysis-using-the-huff-model/123411. |
“The FASTA Program Package,” fasta-36.3.4, Mar. 25, 2011, pp. 29. |
GIS-NET 3 Public_Department of Regional Planning. Planning & Zoning Information for Unincorporated LA County. Retrieved Oct. 2, 2013 from http://gis.planning.lacounty.gov/GIS-NET3_Public/Viewer.html. |
Goswami, Gautam, “Quite Writly Said!,” One Brick at a Time, Aug. 21, 2005, pp. 7. |
Griffith, Daniel A., “A Generalized Huff Model,” Geographical Analysis, Apr. 1982, vol. 14, No. 2, pp. 135-144. |
Hansen et al., “Analyzing Social Media Networks with NodeXL: Insights from a Connected World”, Chapter 4, pp. 53-67 and Chapter 10, pp. 143-164, published Sep. 2010. |
Hardesty, “Privacy Challenges: Analysis: It's Surprisingly Easy to Identify Individuals from Credit-Card Metadata,” MIT News on Campus and Around the World, MIT News Office, Jan. 29, 2015, 3 pages. |
Hibbert et al., “Prediction of Shopping Behavior Using a Huff Model Within a GIS Framework,” Healthy Eating in Context, Mar. 18, 2011, pp. 16. |
Hogue et al., “Thresher: Automating the Unwrapping of Semantic Content from the World Wide Web,” 14th International Conference on World Wide Web, WWW 2005: Chiba, Japan, May 10-14, 2005, pp. 86-95. |
Huff et al., “Calibrating the Huff Model Using ArcGIS Business Analyst,” ESRI, Sep. 2008, pp. 33. |
Huff, David L., “Parameter Estimation in the Huff Model,” ESRI, ArcUser, Oct.-Dec. 2003, pp. 34-36. |
Kahan et al., “Annotea: an Open RDF Infrastructure for Shared Web Annotations”, Computer Networks, Elsevier Science Publishers B.V., vol. 39, No. 5, dated Aug. 5, 2002, pp. 589-608. |
Keylines.com, “An Introduction to KeyLines and Network Visualization,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-White-Paper.pdf> downloaded May 12, 2014 in 8 pages. |
Keylines.com, “KeyLines Datasheet,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-datasheet.pdf> downloaded May 12, 2014 in 2 pages. |
Keylines.com, “Visualizing Threats: Improved Cyber Security Through Network Visualization,” Apr. 2014, <http://keylines.com/wp-content/uploads/2014/04/Visualizing-Threats1.pdf> downloaded May 12, 2014 in 10 pages. |
Kitts, Paul, “Chapter 14: Genome Assembly and Annotation Process,” The NCBI Handbook, Oct. 2002, pp. 1-21. |
Li et al., “Interactive Multimodal Visual Search on Mobile Device,” IEEE Transactions on Multimedia, vol. 15, No. 3, Apr. 1, 2013, pp. 594-607. |
Liu, Tianshun, “Combining GIS and the Huff Model to Analyze Suitable Locations for a New Asian Supermarket in the Minneapolis and St. Paul, Minnesota USA,” Papers in Resource Analysis, 2012, vol. 14, pp. 8. |
Madden, Tom, “Chapter 16: The BLAST Sequence Analysis Tool,” The NCBI Handbook, Oct. 2002, pp. 1-15. |
Manno et al., “Introducing Collaboration in Single-user Applications through the Centralized Control Architecture,” 2010, pp. 10. |
Manske, “File Saving Dialogs,” <http://www.mozilla.org/editor/ui_specs/FileSaveDialogs.html>, Jan. 20, 1999, pp. 7. |
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.yahoo.com. |
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.bing.com. |
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.google.com. |
Microsoft—Developer Network, “Getting Started with VBA in Word 2010,” Apr. 2010, <http://msdn.microsoft.com/en-us/library/ff604039%28v=office.14%29.aspx> as printed Apr. 4, 2014 in 17 pages. |
Microsoft Office—Visio, “About connecting shapes,” <http://office.microsoft.com/en-us/visio-help/about-connecting-shapes-HP085050369.aspx> printed Aug. 4, 2011 in 6 pages. |
Microsoft Office—Visio, “Add and glue connectors with the Connector tool,” <http://office.microsoft.com/en-us/visio-help/add-and-glue-connectors-with-the-connector-tool-HA010048532.aspx?CTT=1> printed Aug. 4, 2011 in 1 page. |
Mizrachi, Ilene, “Chapter 1: GenBank: The Nuckeotide Sequence Database,” The NCBI Handbook, Oct. 2002, pp. 1-14. |
Nierman, “Evaluating Structural Similarity in XML Documents”, 6 pages, 2002. |
Olanoff, Drew, “Deep Dive with the New Google Maps for Desktop with Google Earth Integration, It's More than Just a Utility,” May 15, 2013, pp. 1-6, retrieved from the internet: http://web.archive.org/web/20130515230641/http://techcrunch.com/2013/05/15/deep-dive-with-the-new-google-maps-for-desktop-with-google-earth-integration-its-more-than-just-a-utility/. |
Palmas et al., “An Edge-Bunding Layout for Interactive Parallel Coordinates” 2014 IEEE Pacific Visualization Symposium, pp. 57-64. |
“Potential Money Laundering Warning Signs,” snapshot taken 2003, https://web.archive.org/web/20030816090055/http:/finsolinc.com/ANTI-MONEY%20LAUNDERING%20TRAINING%20GUIDES.pdf. |
“Refresh CSS Ellipsis When Resizing Container—Stack Overflow,” Jul. 31, 2013, retrieved from internet http://stackoverflow.com/questions/17964681/refresh-css-ellipsis-when-resizing-container, retrieved on May 18, 2015. |
Rouse, Margaret, “OLAP Cube,” <http://searchdatamanagement.techtarget.com/definition/OLAP-cube>, Apr. 28, 2012, pp. 16. |
Sigrist, et al., “Prosite, a Protein Domain Database for Functional Characterization and Annotation,” Nucleic Acids Research, 2010, vol. 38, pp. D161-D166. |
Sirotkin et al., “Chapter 13: The Processing of Biological Sequence Data at NCBI,” The NCBI Handbook, Oct. 2002, pp. 1-11. |
Umagandhi et al., “Search Query Recommendations Using Hybrid User Profile with Query Logs,” International Journal of Computer Applications, vol. 80, No. 10, Oct. 1, 2013, pp. 7-18. |
Wikipedia, “Federated Database System,” Sep. 7, 2013, retrieved from the internet on Jan. 27, 2015 http://en.wikipedia.org/w/index.php?title=Federated_database_system&oldid=571954221. |
Yang et al., “HTML Page Analysis Based on Visual Cues”, A129, pp. 859-864, 2001. |
IBM—i2 Integrated Law Enforcement, https://www-03.ibm.com/software/products/en/integrated-law-enforcement, as printed Feb. 15, 2017 in 2 pages. |
IBM—i2 Analyze, https://www-03.ibm.com/software/products/en/i2-analyze, as printed Feb. 15, 2017 in 2 pages. |
IBM—Data analysis—i2 Analyst's Notebook, http://www-03.ibm.com/software/products/en/analysts-notebook, as printed Feb. 16, 2017 in 2 pages. |
Visual Analysis, “Overview of merging timeline charts and creating hybrid charts,” available at https://www.youtube.com/watch?v=dl6jzNtEVpA, as published on Mar. 9, 2015. |
IBM Analytics, “IBM i2 Intelligence Analysis Portfolio Overview,” available at https://www.youtube.com/watch?v=ElFu_oUiaBY, as published on Sep. 24, 2015. |
i2—An IBM Company, “IBM i2 Intelligent Law Enforcement Demo,” available at https://www.youtube.com/watch?v=_KCXZ2iTMXQ, as published on Dec. 3, 2012. |
Yair Shaked, “IBM i2 Enterprise Insight Analysis—cyber Demo,” available at https://www.youtube.com/watch?v=ZXmTWKqkfF4, as published on Nov. 19, 2015. |
Visual Analysis, “Overview of importing data and creating timelines,” available at https://www.youtube.com/watch?v=SovxKrvkZZs, as published on Mar. 9, 2015. |
IBM Corporation, “IBM i2 Analyst's Notebook,” Aug. 2015, in 4 pages. |
IBM Corporation, “IBM i2 Analyst's Notebook Connector for Esri,” May 2012, in 3 pages. |
IBM Corporation, “IBM i2 Enterprise Insight Analysis V2.0 delivers a modern contextual user interface and enhanced software operational warehouse support,” http://www-01.ibm.com/common/ssi/ShowDoc.wss?docURL=/common/ssi/rep_ca/2/897/ENUS215- 302/index.html&lang=en&request_locale=en, as published on Sep. 1, 2015. |
IBM Support, “Software lifecycle—i2 Analyst's Notebook Premium 9.0.0,” https://www-01.ibm.com/software/support/lifecycleapp/PLCDetail.wss?q45=I570331B72886X86, as printed Mar. 7, 2017 in 1 page. |
IBM Support, “Software lifecycle—i2 Enterprise Insight Analysis 2.0.0,” https://www-01.ibm.com/software/support/lifecycleapp/PLCDetail.wss?q45=E170786H45496153, as printed Mar. 7, 2017 in 1 page. |
i2 A ChoicePoint Company, “i2 Analyst's Notebook 7 User Guide: Creating Charts” Jun. 2007, 373 pages. |
Gatewaynews, “New Crime Fighting Tool ‘Coplink’” available at https://www.youtube.com/watch?v=GbU6E0grnTw, as published on Mar. 8, 2008. |
COPLINK, “Incident Analyzer User Guide,” created Nov. 5, 2010 (as indicated by the PDF file metadata), 14 pages. |
Notice of Allowance for U.S. Appl. No. 14/242,559 dated Oct. 3, 2016. |
Notice of Allowance for U.S. Appl. No. 14/102,394 dated Aug. 25, 2014. |
Notice of Allowance for U.S. Appl. No. 14/108,187 dated Aug. 29, 2014. |
Notice of Allowance for U.S. Appl. No. 14/135,289 dated Oct. 14, 2014. |
Notice of Allowance for U.S. Appl. No. 14/192,767 dated Dec. 16, 2014. |
Notice of Allowance for U.S. Appl. No. 14/225,084 dated May 4, 2015. |
Notice of Allowance for U.S. Appl. No. 14/268,964 dated Dec. 3, 2014. |
Notice of Allowance for U.S. Appl. No. 14/294,098 dated Dec. 29, 2014. |
Notice of Allowance for U.S. Appl. No. 14/473,860 dated Jan. 5, 2015. |
Notice of Allowance for U.S. Appl. No. 14/486,991 dated May 1, 2015. |
Notice of Allowance for U.S. Appl. No. 14/504,103 dated May 18, 2015. |
Notice of Allowance for U.S. Appl. No. 14/616,080 dated Apr. 2, 2015. |
Official Communication for Australian Patent Application No. 2014201511 dated Feb. 27, 2015. |
Official Communication for Australian Patent Application No. 2014202442 dated Mar. 19, 2015. |
Official Communication for Australian Patent Application No. 2014210604 dated Jun. 5, 2015. |
Official Communication for Australian Patent Application No. 2014210614 dated Jun. 5, 2015. |
Official Communication for Australian Patent Application No. 2014213553 dated May 7, 2015. |
Official Communication for Australian Patent Application No. 2014250678 dated Jun. 17, 2015. |
Official Communication for European Patent Application No. 14158861.6 dated Jun. 16, 2014. |
Official Communication for European Patent Application No. 14159464.8 dated Jul. 31, 2014. |
Official Communication for European Patent Application No. 14180142.3 dated Feb. 6, 2015. |
Official Communication for European Patent Application No. 14180281.9 dated Jan. 26, 2015. |
Official Communication for European Patent Application No. 14180321.3 dated Apr. 17, 2015. |
Official Communication for European Patent Application No. 14186225.0 dated Feb. 13, 2015. |
Official Communication for European Patent Application No. 14187996.5 dated Feb. 12, 2015. |
Official Communication for European Patent Application No. 14189344.6 dated Feb. 20, 2015. |
Official Communication for European Patent Application No. 14189347.9 dated Mar. 4, 2015. |
Official Communication for European Patent Application No. 14189802.3 dated May 11, 2015. |
Official Communication for European Patent Application No. 14191540.5 dated May 27, 2015. |
Official Communication for European Patent Application No. 14197879.1 dated Apr. 28, 2015. |
Official Communication for European Patent Application No. 14197895.7 dated Apr. 28, 2015. |
Official Communication for European Patent Application No. 14199182.8 dated Mar. 13, 2015. |
Official Communication for Great Britain Patent Application No. 1404457.2 dated Aug. 14, 2014. |
Official Communication for Great Britain Patent Application No. 1404574.4 dated Dec. 18, 2014. |
Official Communication for Great Britain Patent Application No. 1408025.3 dated Nov. 6, 2014. |
Official Communication for Great Britain Patent Application No. 1411984.6 dated Dec. 22, 2014. |
Official Communication for Great Britain Patent Application No. 1413935.6 dated Jan. 27, 2015. |
Official Communication for Netherlands Patent Application No. 2013306 dated Apr. 24, 2015. |
Official Communication for New Zealand Patent Application No. 622513 dated Apr. 3, 2014. |
Official Communication for New Zealand Patent Application No. 622517 dated Apr. 3, 2014. |
Official Communication for New Zealand Patent Application No. 624557 dated May 14, 2014. |
Official Communication for New Zealand Patent Application No. 627962 dated Aug. 5, 2014. |
Official Communication for New Zealand Patent Application No. 628161 dated Aug. 25, 2014. |
Official Communication for New Zealand Patent Application No. 628263 dated Aug. 12, 2014. |
Official Communication for New Zealand Patent Application No. 628495 dated Aug. 19, 2014. |
Official Communication for New Zealand Patent Application No. 628585 dated Aug. 26, 2014. |
Official Communication for New Zealand Patent Application No. 628840 dated Aug. 28, 2014. |
Official Communication for U.S. Appl. No. 13/247,987 dated Apr. 2, 2015. |
Official Communication for U.S. Appl. No. 13/831,791 dated Mar. 4, 2015. |
Official Communication for U.S. Appl. No. 13/835,688 dated Jun. 17, 2015. |
Official Communication for U.S. Appl. No. 14/148,568 dated Oct. 22, 2014. |
Official Communication for U.S. Appl. No. 14/148,568 dated Mar. 26, 2015. |
Official Communication for U.S. Appl. No. 14/196,814 dated May 5, 2015. |
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 10, 2014. |
Official Communication for U.S. Appl. No. 14/225,006 dated Feb. 27, 2015. |
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 2, 2014. |
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 20, 2015. |
Official Communication for U.S. Appl. No. 14/225,160 dated Feb. 11, 2015. |
Official Communication for U.S. Appl. No. 14/225,160 dated May 20, 2015. |
Official Communication for U.S. Appl. No. 14/225,160 dated Oct. 22, 2014. |
Official Communication for U.S. Appl. No. 14/225,160 dated Jul. 29, 2014. |
Official Communication for U.S. Appl. No. 14/268,964 dated Sep. 3, 2014. |
Official Communication for U.S. Appl. No. 14/289,596 dated Jul. 18, 2014. |
Official Communication for U.S. Appl. No. 14/289,596 dated Jan. 26, 2015. |
Official Communication for U.S. Appl. No. 14/289,596 dated Apr. 30, 2015. |
Official Communication for U.S. Appl. No. 14/289,599 dated Jul. 22, 2014. |
Official Communication for U.S. Appl. No. 14/289,599 dated May 29, 2015. |
Official Communication for U.S. Appl. No. 14/294,098 dated Aug. 15, 2014. |
Official Communication for U.S. Appl. No. 14/294,098 dated Nov. 6, 2014. |
Official Communication for U.S. Appl. No. 14/306,138 dated Feb. 18, 2015. |
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 23, 2014. |
Official Communication for U.S. Appl. No. 14/306,138 dated May 26, 2015. |
Official Communication for U.S. Appl. No. 14/306,147 dated Feb. 19, 2015. |
Official Communication for U.S. Appl. No. 14/306,147 dated Sep. 9, 2014. |
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 11, 2015. |
Official Communication for U.S. Appl. No. 14/306,154 dated May 15, 2015. |
Official Communication for U.S. Appl. No. 14/306,154 dated Sep. 9, 2014. |
Official Communication for U.S. Appl. No. 14/319,765 dated Jun. 16, 2015. |
Official Communication for U.S. Appl. No. 14/319,765 dated Nov. 25, 2014. |
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 4, 2015. |
Official Communication for U.S. Appl. No. 14/323,935 dated Jun. 22, 2015. |
Official Communication for U.S. Appl. No. 14/323,935 dated Nov. 28, 2014. |
Official Communication for U.S. Appl. No. 14/323,935 dated Mar. 31, 2015. |
Official Communication for U.S. Appl. No. 14/326,738 dated Dec. 2, 2014. |
Official Communication for U.S. Appl. No. 14/326,738 dated Mar. 31, 2015. |
Official Communication for U.S. Appl. No. 14/473,552 dated Feb. 24, 2015. |
Official Communication for U.S. Appl. No. 14/486,991 dated Mar. 10, 2015. |
Official Communication for U.S. Appl. No. 14/504,103 dated Mar. 31, 2015. |
Official Communication for U.S. Appl. No. 14/504,103 dated Feb. 5, 2015. |
Official Communication for U.S. Appl. No. 14/579,752 dated May 26, 2015. |
Official Communication for U.S. Appl. No. 14/639,606 dated May 18, 2015. |
Number | Date | Country | |
---|---|---|---|
20170115865 A1 | Apr 2017 | US |
Number | Date | Country | |
---|---|---|---|
61864048 | Aug 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14242559 | Apr 2014 | US |
Child | 15398092 | US | |
Parent | 14095798 | Dec 2013 | US |
Child | 14242559 | US |