Generating object time series from data objects

Information

  • Patent Grant
  • 9779525
  • Patent Number
    9,779,525
  • Date Filed
    Wednesday, December 28, 2016
    9 years ago
  • Date Issued
    Tuesday, October 3, 2017
    8 years ago
Abstract
Systems and methods are presented for representing non-numerical data objects in an object time series. An object time series of can be created by establishing one or more associations, each association including a mapping of at least one point in time with one or more objects that include properties and values. Visual representation of an object time series may include displaying non-numerical values associated with objects in the object time series in association with respective points in time.
Description
TECHNICAL FIELD

The present disclosure relates to systems and techniques for data integration, analysis, and visualization. More specifically, representing non-numerical data objects in a time series.


BACKGROUND

Traditionally, time series are created and used to store numerical data associated with a plurality of points in time. Non-numerical data are usually stored and used as metadata.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 illustrates one embodiment of a visualization of an object time series that may be generated and presented to a user.



FIG. 2 illustrates one embodiment of a database system using an ontology.



FIG. 3 illustrates one embodiment of a system for creating data in a data store using a dynamic ontology.



FIG. 4 illustrates a sample user interface using relationships described in a data store using a dynamic ontology.



FIG. 5 illustrates an example figure that may be generated and presented to a user to visualize a traditional time series with associating numerical values.



FIG. 6 illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series associated with a specific instance of a payment event object.



FIG. 7 illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series and the slicing of the Object Time Series.



FIG. 8 illustrates a computer system with which certain methods discussed herein may be implemented.



FIG. 9 illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series associated with multiple types of object and functions that may be used to access property values of the person object.



FIG. 9A illustrates an example figure that may be generated and presented to a user to visualize two Object Time Series associated with status objects to show duration.



FIG. 9B illustrates an example figure that may be generated and presented to a user to visualize two Object Time Series associated with status objects indicating progress over time.



FIG. 9C illustrates an example figure that may be generated and presented to a user to visualize two Object Time Series associated with events.



FIG. 9D illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series associated with a status object.



FIG. 9E illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series associated with a status object and transitions among the various statuses.



FIG. 10 illustrates a computer system with which certain methods discussed herein may be implemented.





DETAILED DESCRIPTION OF CERTAIN EMBODIMENTS
Definitions

In order to facilitate an understanding of the systems and methods discussed herein, a number of terms are defined below. The terms defined below, as well as other terms used herein, should be construed to include the provided definitions, the ordinary and customary meaning of the terms, and/or any other implied meaning for the respective terms. Thus, the definitions below do not limit the meaning of these terms, but only provide exemplary definitions.


Ontology: Stored information that provides a data model for storage of data in one or more databases. For example, the stored data may comprise definitions for object types and property types for data in a database, and how objects and properties may be related.


Database: A broad term for any data structure for storing and/or organizing data, including, but not limited to, relational databases (Oracle database, mySQL database, etc.), spreadsheets, XML files, and text file, among others.


Data Object or Object: A data container for information representing specific things in the world that have a number of definable properties. For example, a data object can represent an entity such as a person, a place, an organization, a market instrument, or other noun. A data object can represent an event that happens at a point in time or for a duration. A data object can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object may be associated with a unique identifier that uniquely identifies the data object. The object's attributes (e.g. metadata about the object) may be represented in one or more properties.


Object Type: Type of a data object (e.g., Person, Event, or Document). Object types may be defined by an ontology and may be modified or updated to include additional object types. An object definition (e.g., in an ontology) may include how the object is related to other objects, such as being a sub-object type of another object type (e.g. an agent may be a sub-object type of a person object type), and the properties the object type may have.


Properties: Attributes of a data object that represent individual data items. At a minimum, each property of a data object has a property type and a value or values.


Property Type: The type of data a property is, such as a string, an integer, or a double. Property types may include complex property types, such as a series data values associated with timed ticks (e.g. a time series), etc.


Property Value: The value associated with a property, which is of the type indicated in the property type associated with the property. A property may have multiple values.


Link: A connection between two data objects, based on, for example, a relationship, an event, and/or matching properties. Links may be directional, such as one representing a payment from person A to B, or bidirectional.


Link Set: Set of multiple links that are shared between two or more data objects.


Object Time Series: associating multiple points in time with one or more objects, for example, People, Event, and so forth. Objects included in Object Time Series may be any type of objects, including numerical objects and non-numerical objects.


Overview of Object Time Series


Traditional time series may store data in the format of TIME, VALUE, otherwise known as a tick. For example, the stock price of Microsoft (“MSFT”) may be represented by a specific time of any given business day, and a value of the stock at that specific time. However, using traditional time series data types to store generic values that are not numerical may be very difficult because data about objects 201 and about specific object types such as person, event, document, and so forth, when associated with a time series, have been traditionally implemented as metadata associated with values in time series.


For example, in the MSFT stock price use case, specific milestones such as quarterly earnings reports, board meetings, and dividends may be associated with a specific value of the MSFT stock price in a time series. However, such milestones (event) data may not be easily represented because traditional time series treat the milestones as metadata (e.g. not objects in themselves, but as descriptive text data associated with a numeric time series). Accessing such metadata may require several steps of referencing and/or mapping. Slicing and sampling traditional time series data based on metadata is also inefficient and difficult. Details regarding traditional time series are discussed further below.


An Object Time Series may instead directly associate time with any object. For example, instead of limiting the value associated with a point in time to numerical values, any object type may be used, such as Person, Event, and so forth.



FIG. 1 illustrates one embodiment of a visualization of an object time series that may be generated and presented to a user. In this example, an Object Time Series 100 is plotted. The Object Time Series 100 spans from Jan. 1, 2013 to Apr. 1, 2013. The data object associated with each date in the Object Time Series 100 is a Contagious Disease Report object. As a data object, the Contagious Disease Report object may contain properties such as: (1) type of contagious disease; (2) location; (3) State; (4) a patient object (age, gender, occupation); (5) hospital where the patient is or has been treated. As can be seen in this example, the object associated with an Object Time Series may itself contain other object types, which may involve more properties. Thus, Object Time Series may associate rich information and object types with time/dates.


In some embodiments, Object Time Series 100 data may be plotted similar to traditional time series, such as the example in FIG. 1. In some other embodiments, data involved in object time series may be used to identify links and/or relationships. For example, if the Object Time Series 100 is named a Disease OTS, then executing a function such as DiseaseOTS.location( ) may automatically generate a list of all locations that a disease has been reported, which may be the following list: (Aliso Viejo, Miami, Mission Viejo). In another example, executing a function such as DiseaseOTS.occupation( ) may generate the following list automatically: (Coast Elementary, Rancho High School, Mission Viejo Shopping Center, Sunny Restaurant). Moreover, in some embodiments, the list of locations generated may be automatically mapped to the respective points in time. Advantageously, this would allow a disease to be tracked by location in relation to time.


As shown in the example in FIG. 1, the objects stored in Object Time Series 100 may include not only numerical values, but also other data types involving multiple properties. For example, the hospitals in Object Time Series 100 may be associated property values such as hospital location, treating physician, and so forth.


In some embodiments, Object Time Series 100 may also enable fast visualization and association between stored objects at various points in time. For example, because the occupation information may be directly accessible and visualized in Object Time Series 100, an analyst reviewing the DiseaseOTS Object Time Series may easily find that West Nile Virus has been reported regarding two Coast Elementary School children, reported on Jan. 1, 2013 and Feb. 9, 2013, respectively. Moreover, it may be easy to show that both Mission Viejo and Aliso Viejo have been reported to have West Nile occurrences twice. More details regarding Object Time Series are discussed further herein.


Object Centric Data Model


To provide a framework for the following discussion of specific systems and methods described herein, an example database system 210 using an ontology 205 will now be described. This description is provided for the purpose of providing an example and is not intended to limit the techniques to the example data model, the example database system, or the example database system's use of an ontology to represent information.


In one embodiment, a body of data is conceptually structured according to an object-centric data model represented by ontology 205. The conceptual data model is independent of any particular database used for durably storing one or more database(s) 209 based on the ontology 205. For example, each object of the conceptual data model may correspond to one or more rows in a relational database or an entry in Lightweight Directory Access Protocol (LDAP) database, or any combination of one or more databases.



FIG. 2 illustrates an object-centric conceptual data model according to an embodiment. An ontology 205, as noted above, may include stored information providing a data model for storage of data in the database 209. The ontology 205 may be defined by one or more object types, which may each be associated with one or more property types. At the highest level of abstraction, data object 201 is a container for information representing things in the world. For example, data object 201 can represent an entity such as a person, a place, an organization, a market instrument, or other noun. Data object 201 can represent an event that happens at a point in time or for a duration. Data object 201 can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object 201 is associated with a unique identifier that uniquely identifies the data object within the database system.


Different types of data objects may have different property types. For example, a “Person” data object might have an “Eye Color” property type and an “Event” data object might have a “Date” property type. Each property 203 as represented by data in the database system 210 may have a property type defined by the ontology 205 used by the database 205.


Objects may be instantiated in the database 209 in accordance with the corresponding object definition for the particular object in the ontology 205. For example, a specific monetary payment (e.g., an object of type “event”) of US$30.00 (e.g., a property of type “currency”) taking place on Mar. 27, 2009 (e.g., a property of type “date”) may be stored in the database 209 as an event object with associated currency and date properties as defined within the ontology 205.


The data objects defined in the ontology 205 may support property multiplicity. In particular, a data object 201 may be allowed to have more than one property 203 of the same property type. For example, a “Person” data object might have multiple “Address” properties or multiple “Name” properties.


Each link 202 represents a connection between two data objects 201. In one embodiment, the connection is either through a relationship, an event, or through matching properties. A relationship connection may be asymmetrical or symmetrical. For example, “Person” data object A may be connected to “Person” data object B by a “Child Of” relationship (where “Person” data object B has an asymmetric “Parent Of” relationship to “Person” data object A), a “Kin Of” symmetric relationship to “Person” data object C, and an asymmetric “Member Of” relationship to “Organization” data object X. The type of relationship between two data objects may vary depending on the types of the data objects. For example, “Person” data object A may have an “Appears In” relationship with “Document” data object Y or have a “Participate In” relationship with “Event” data object E. As an example of an event connection, two “Person” data objects may be connected by an “Airline Flight” data object representing a particular airline flight if they traveled together on that flight, or by a “Meeting” data object representing a particular meeting if they both attended that meeting. In one embodiment, when two data objects are connected by an event, they are also connected by relationships, in which each data object has a specific relationship to the event, such as, for example, an “Appears In” relationship.


As an example of a matching properties connection, two “Person” data objects representing a brother and a sister, may both have an “Address” property that indicates where they live. If the brother and the sister live in the same home, then their “Address” properties likely contain similar, if not identical property values. In one embodiment, a link between two data objects may be established based on similar or matching properties (e.g., property types and/or property values) of the data objects. These are just some examples of the types of connections that may be represented by a link and other types of connections may be represented; embodiments are not limited to any particular types of connections between data objects. For example, a document might contain references to two different objects. For example, a document may contain a reference to a payment (one object), and a person (a second object). A link between these two objects may represent a connection between these two entities through their co-occurrence within the same document.


Each data object 201 can have multiple links with another data object 201 to form a link set 204. For example, two “Person” data objects representing a husband and a wife could be linked through a “Spouse Of” relationship, a matching “Address” property, and one or more matching “Event” properties (e.g., a wedding). Each link 202 as represented by data in a database may have a link type defined by the database ontology used by the database.



FIG. 3 is a block diagram illustrating exemplary components and data that may be used in identifying and storing data according to an ontology. In this example, the ontology may be configured, and data in the data model populated, by a system of parsers and ontology configuration tools. In the embodiment of FIG. 3, input data 300 is provided to parser 302. The input data may comprise data from one or more sources. For example, an institution may have one or more databases with information on credit card transactions, rental cars, and people. The databases may contain a variety of related information and attributes about each type of data, such as a “date” for a credit card transaction, an address for a person, and a date for when a rental car is rented. The parser 302 is able to read a variety of source input data types and determine which type of data it is reading.


In accordance with the discussion above, the example ontology 205 comprises stored information providing the data model of data stored in database 209, and the ontology is defined by one or more object types 310, one or more property types 316, and one or more link types 330. Based on information determined by the parser 302 or other mapping of source input information to object type, one or more data objects 201 may be instantiated in the database 209 based on respective determined object types 310, and each of the objects 201 has one or more properties 203 that are instantiated based on property types 316. Two data objects 201 may be connected by one or more links 202 that may be instantiated based on link types 330. The property types 316 each may comprise one or more data types 318, such as a string, number, etc. Property types 316 may be instantiated based on a base property type 320. For example, a base property type 320 may be “Locations” and a property type 316 may be “Home.”


In an embodiment, a user of the system uses an object type editor 324 to create and/or modify the object types 310 and define attributes of the object types. In an embodiment, a user of the system uses a property type editor 326 to create and/or modify the property types 316 and define attributes of the property types. In an embodiment, a user of the system uses link type editor 328 to create the link types 330. Alternatively, other programs, processes, or programmatic controls may be used to create link types and property types and define attributes, and using editors is not required.


In an embodiment, creating a property type 316 using the property type editor 326 involves defining at least one parser definition using a parser editor 322. A parser definition comprises metadata that informs parser 302 how to parse input data 300 to determine whether values in the input data can be assigned to the property type 316 that is associated with the parser definition. In an embodiment, each parser definition may comprise a regular expression parser 304A or a code module parser 304B. In other embodiments, other kinds of parser definitions may be provided using scripts or other programmatic elements. Once defined, both a regular expression parser 304A and a code module parser 304B can provide input to parser 302 to control parsing of input data 300.


Using the data types defined in the ontology, input data 300 may be parsed by the parser 302 determine which object type 310 should receive data from a record created from the input data, and which property types 316 should be assigned to data from individual field values in the input data. Based on the object-property mapping 301, the parser 302 selects one of the parser definitions that is associated with a property type in the input data. The parser parses an input data field using the selected parser definition, resulting in creating new or modified data 303. The new or modified data 303 is added to the database 209 according to ontology 205 by storing values of the new or modified data in a property of the specified property type. As a result, input data 300 having varying format or syntax can be created in database 209. The ontology 205 may be modified at any time using object type editor 324, property type editor 326, and link type editor 328, or under program control without human use of an editor. Parser editor 322 enables creating multiple parser definitions that can successfully parse input data 300 having varying format or syntax and determine which property types should be used to transform input data 300 into new or modified input data 303.


The properties, objects, and the links (e.g. relationships) between the objects can be visualized using a graphical user interface (GUI). For example, FIG. 4 displays a user interface showing a graph representation 403 of relationships (including relationships or links 404, 405, 406, 407, 408) between the data objects (including data objects 411, 412, 413, 414, 415, 416) that are represented as nodes in the example of FIG. 4. In this embodiment, the data objects are person objects. In this example, the person nodes (associated with person data objects) may have relationships to other person nodes, for example, through payment objects. For example, relationship 404 is based on a payment associated with the individuals indicated in person data objects 411 and 413. The link 404 represents these shared payments (for example, the individual associated with data object 411 may have paid the individual associated with data object 413 on three occasions). These relationships may be stored as links, or in some embodiments, as properties, where a relationship may be detected between the properties. In some cases, as stated above, the links may be directional. For example, a payment link may have a direction associated with the payment, where one person object is a receiver of a payment, and another person object is the payer of payment.


In addition to visually showing relationships between the data objects, the user interface may allow various other manipulations. For example, the objects within database 108 may be searched using a search interface 420 (e.g., text string matching of object properties), inspected (e.g., properties and associated data viewed), filtered (e.g., narrowing the universe of objects into sets and subsets by properties or relationships), and statistically aggregated (e.g., numerically summarized based on summarization criteria), among other operations and visualizations.


Object Time Series



FIG. 5 illustrates an example figure that may be generated and presented to a user to visualize a traditional time series 500 with associated numeric values. In time series 500, the time series bar represents five months, beginning from Jan. 1, 2013 and ending on May 1, 2013. At the beginning of each month, a numerical value is associated with the date. For example, the numerical value associated with Jan. 1, 2013 is $230,450. The numerical value associated with Feb. 1, 2013 is $231,400. The numerical value associated with Mar. 1, 2013 is $229,905. The numerical value associated with Apr. 1, 2013 is $230,090. The numerical value associated with May 1, 2013 is $232,500.


Traditional time series such as time series 500 stores a date/time and a numerical value. However, in order to find out other data associated with the numerical value, further steps may be needed. For example, in some embodiments, a traditional time series 500 may implement an array that stores (DATE, VALUE) pairs, for example:







[


(


Jan
.




1

,




2013
,

230


,


450


)

,





(


Feb
.




1

,




2013
,

231


,


400


)

,










(


May





1

,




2013
,

232


,


500


)


]

.




Another array may be used with the traditional time series 500 to store metadata associated with a date/time, such as whether the home price was above or below average for that time period, which may also be referred to as a “tick.” For example:







[


(


Jan
.




1

,
2013
,


above



)

,





(


Feb
.




1

,
2013
,


below



)

,










(


May





1

,
2013
,


above



)


]

.




In situations where metadata regarding categories such as “above” a value, “below” a value, on or after a certain date are accessed, a separate look up needs to be performed before the metadata may be interpreted. For example, knowing that the metadata associated with Jan. 1, 2013 is “above” may tell a user that the specific home price on Jan. 1, 2013 is above average. However, a separate lookup needs to be performed to find out what the average home value is. Accessing details regarding the value associated with the traditional time series and accessing metadata thus involves two separate steps accessing values stored in two separate data structures, which may be computationally costly, especially if searching is involved.


Therefore, as shown in this example, in a traditional time series 500, finding the metadata associated with a numerical value may require a lookup step or a mapping step that looks up a separate array. If there are more than one metadata fields, such lookup step may be computationally costly.


Moreover, in a traditional time series 500, it is also difficult to create a sample or slice of the time series that also includes the metadata. For example, metadata associated with the home price example discussed above may include state and/or city, population, average household income. Creating a sample or a slice of the traditional time series 500 would involve mapping steps that separately look up each array or other data structure that stores the metadata involving state, array or data structure that stores population, and so forth.


A traditional time series may also store metadata as a set of intervals. Each interval may contain a starting data, and ending date, and a status during the interval. For example, a time series may track a house's price over time. Each tick represents a date and an associated price for the house. However, other metadata may also be tracked in association with that time series, such as the status of the loan associated with that house over time (e.g. “loan up-to-date”, “loan payment behind 30 days”, “loan payment behind 60 days”, etc.). A traditional time series may not use an array of ticks to represent this status information. Instead, it may use a set of intervals. For example, one interval may be [4/30/2011, 12/29/2011, “loan up-to-date”], and another may be [12/30/2011, 2/06/2012, “loan payment behind 30 days], and yet another may be [2/7/2012, 12/29/2012, “loan up-to-date”].


One problem with representing statuses as intervals is that they cannot be easily sliced or subsetted. For example, data in the data structure itself may need to be duplicated to split such intervals. If a time series is sliced on Jan. 1, 2012 as a split date, the interval needs to be applied to both new time series, with the interval now split. Object time series does not have these limitations. In addition, intervals may overlap, whereas ticks in an object time series may not overlap when tracking metadata, making it easier to perform transformations and analysis on the metadata.



FIG. 6 illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series 600 associated with a specific instance of a payment event object. In this embodiment, the Object Time Series 600 includes a time series bar that displays a time period between Oct. 1, 2012 and Jul. 1, 2013.


In this example, a Payment object type is associated with the Object Time Series. Due dates for the Payment object is set to the first day of each month. Each Payment object may include properties such as timeliness with values indicating whether a payment is on time, late, or in default. Each Payment object may also include properties such as payment with values indicating whether a payment was successful or unsuccessful. In some instances, a payment may be unsuccessful because a payment check is bounced by a bank or there are not sufficient funds to complete a payment transaction. In some embodiments, all the properties of an object may be configured to be displayed or plotted. In some other embodiments, some properties of an object may be configured to be displayed whereas some other properties of an object may be configured not to be displayed.


In some embodiments, objects in an Object Time Series may be displayed next to the respective points in time. For example, in FIG. 6, dates on which payments are made are displayed. For example, on Oct. 1, 2012, a payment is made and it is on time. On Nov. 1, 2012, a payment is made and it is also on time. On Dec. 1, 2012, an on-time payment is also made.


In some embodiments, certain objects may be configured to be not shown directly in a visualization of an Object Time Series. For example, in Object Time Series 600, missed payments may be configured in such a way that they are not displayed next to the due date, which may be the first day of a month. In some other embodiments, missed payments may also be displayed.


On Jan. 1, 2013, a payment is missed. However, on Jan. 13, 2013, a late payment is made, and it is successful. No payment is made on either Feb. 1, 2013. On Mar. 1, 2013, there is also no payment. On Mar. 13, 2013, a late payment is made. However, this late payment is unsuccessful. Payment dates are also missed on April 1, May 1, and June 1st. By Jul. 1, 2013, the property value of the Payment object is set to default after 5 consecutive months without a successful payment.


As shown in the example in FIG. 6, unlike traditional time series that use metadata to associate non-numerical values with respective points in time, any type of Object and property values may be directly associated with respective points in time in an Objective Time Series.


In some embodiments, information regarding intervals between two or more points in time may be easily identified by comparing the values of the objects associated with an Object Time Series. For example, in FIG. 6, information regarding the interval between Dec. 1, 2012 and Jan. 13, 2013 may be determined by comparing the values of the property timeliness. The value of the property timeliness of the Payment object associated with Dec. 1, 2012 is “on time.” The value of the property timeliness of the Payment object associated with Jan. 13, 2013 is “late”. Therefore, during the interval between Dec. 1, 2012 and Jan. 13, 2013, payment timeliness has changed from “on time” to “late”.


In some embodiments, the values in each object associated with an Object Time Series may also contain other objects. For example, the Payment object involved in Object Time Series 600 may contain a property payment success, and this property may contain a value that is an object in itself. For example, the value may contain a link object. For example, on Mar. 13, 2013, a late and unsuccessful payment is made. The value of payment unsuccessful may contain a link object, which links the unsuccessful payment to a detailed description of the failed payment transaction. In some embodiments, such information may be configured to be displayed, while in other embodiments, such information may be configured in such a way that it is not displayed next to an Object Time Series.



FIG. 7 illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series 700 and the slicing of the object time series. In the example of FIG. 7, two slicing functions are performed on the Object Time Series 700.


The first slicing function is Slice (Q4 2012), which executes computer-implemented instructions that automatically performs a slicing function that may slice the fourth quarter of the year 2012 from the Object Time Series 700. This may result in the generation of Object Time Series 720, which includes the three months in the fourth quarter of the year 2012 and the objects associated with the three months.


In some embodiments, the generated Object Time Series 720 may be further sliced. For example, a splicing function may be performed on the Object Time Series 720 to create another Object Time Series that has an equal or less duration of time as the Object Time Series 720.


In some embodiments, all the objects involved with the original Object Time Series 700 are kept in the generated Object Times Series 720. In some other embodiments, choices of which objects, properties, and/or value types may be kept and/or excluded may be specified in a splicing function. For example, a user may specify that he or she does not want to see whether payments are successful.


The second slicing function is Slice (Q2, 2013), which may execute computer-implemented instructions that automatically performs a slicing function that may slice the second quarter of the year 2013 from the Object Time Series 700. This may result in the generation of Object Time Series 740, which includes the three months in the second quarter of the year 2013 and the objects associated with the three months.


Although quarterly slicing is used as an example in FIG. 7, in some other embodiments, slicing functions may also be based on other arbitrary time intervals. Slicing a traditional time series is difficult because metadata needs to be looked up separately for the arbitrary time intervals, If there are multiple intervals, looking up metadata stored in a different data structure may involve several separate look ups and access operations. After a new Object Time Series is created by slicing, it may be saved, viewed, accessed, and recalled as any Object Time Series. It may also be sliced to create other Object Time Series.



FIG. 8 illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series 850 and the sampling of the Object Time Series 850.


In this example, a sampling function Sample (Quarterly) may execute computer-implemented instructions that automatically perform a sampling function that may sample the Object Time Series on a quarter-to-quarter basis. In some embodiments, a sampling function of an Object Time Series may also allow a user to specify a starting point in time or an ending point in time of a sampling function. In some other embodiments, a sampling function of an Object Time Series may use the beginning of the Object Time Series involved as the default starting point in time for the sampling function.


In this particular example, the “Sample (Quarterly)” function samples the Object Time Series 850 on a quarterly basis (every three months). In this particular embodiment, sampling begins on Oct. 1, 2012. Therefore, the points in time that are sampled on a quarterly basis include: Oct. 1, 2012, Jan. 1, 2012, Apr. 1, 2013, and Jul. 1, 2013.


In some embodiments, the sampling function may generate an Object Time Series 870 based on the sampled points in time and the objects associated with the points in time. In the example in FIG. 8, the Object Time Series 870 includes four points in time, Oct. 1, 2012, Jan. 1, 2013, Apr. 1, 2013, and Jul. 1, 2013.


In some embodiments, the sampling an Object Time Series also includes all the objects associated with the original Object Time Series for each sampled point in time. In this example, including all objects associate with the original Object Time Series 850 would include the Payment object's properties such as timeliness and whether payment was successful.


In some other embodiments, the sampling function may be configured to include a portion of the objects, properties, and/or values associated with the original Object Time Series. For example, a user may choose to execute a sampling function that does not include the value in the Payment object indicating whether payments were successfully made.


Although quarterly sampling is used as an example in FIG. 8, in some other embodiments, sampling functions may also be based on other arbitrary time intervals. This would be difficult and costly to implement using a traditional time series because all the associated metadata needs to be separately looked up from individual arrays, or by searching a set of intervals. After a new Object Time Series is created by sampling, it may be saved, viewed, accessed as any Object Time Series. It may also be sampled to create other Object Time Series.



FIG. 9 illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series associated with multiple types of objects and functions that may be used to access property values of the person object. In FIG. 9, Object Time Series 900 is shown. Object Time Series 900 may include multiple types of objects, such as Person and Activity. Each object associated with Object Time Series 900 may include one or more properties that may have one or more values.


In FIG. 9, the Object Time Series 900 includes multiple points in time beginning from Feb. 1, 2013 and ending on Aug. 1, 2013. For example, on Feb. 1, 2013, a person named Karl K. is associated with the date. He was reportedly to be in the city of Kyoto Japan on this date. He is reported to have an occupation of consultant. The Object Time Series 900 does not include or does not have information regarding his phone number. The Object Time Series 900 includes the activity that Karl K. is associated with as meetings.


Karl K. is again recorded in the Object Time Series 900 on Feb. 24, 2013. On the particular day, he was reported to be in Guam. His phone number is also recorded in the Object Time Series 900.


On Apr. 23, 2013, a person named Frank J. is recorded in the Object Time Series 900 as being in the city of Moscow. His phone number is not recorded. The activity associated with the date is a bank wire, and an account number is included. The date in this Object Time Series 900 is also associated with a car, which is a blue Mini in this particular example.


On Jun. 7, 2013, a person named Alex B. is recorded in the Object Time Series 900 as being in the city of Sao Paolo. Alex B. is recorded in the Object Time Series 900 as being seen with a person named Antonio B. and an unknown woman. Because more information about Antonio B. is available, the Object Time Series 900 may display the name Antonio B. as a link so that more information about Antonio B. may be found by following the link. The activity involved on this date is recorded in the Object Time Series 900 as equipment purchase and firearm purchase.


On Jul. 22, 2013, Frank J. is recorded in the Object Time Series 900 as being in the city of Kuala Lumpur. The activities he was associated with on this date include a bank wire and plane purchase.


In some embodiments, an Object Time Series such as Object Time Series 900 may include multiple types of objects. For example, the Object Time Series 900 includes a Person object type, which may include properties such as name, location, phone number, and so forth. The Object Time Series 900 may also include an Event or Activity object type, which may include the activities recorded for a particular point in time, such as bank wires, purchases, meetings, and so forth.


Depending on embodiments, an Object Time Series may be created by using a function such as object TimeSeries status=createObjectTime Series (“String”). As noted in FIG. 9, expressions such as 910: OTS.ValueOn(4/25/2013).Car may be used to access the car object associated with the date Apr. 25, 2013. In this particular example, the value of the car object type is a Blue Mini. An expression such as 920: OTS.ValueOn(7/1/2013).City may be used to access the city object associated with the date Jul. 1, 2013. Using an Object Time Series, information, values and properties that would be stored otherwise as metadata as in traditional time series may be implemented and accessed easily. Moreover, information, values and properties, especially non-numerical values, may be visually displayed and associated with respective points in time. For example, for each person associated with a timeline, the person's picture, if available, may be displayed next to the timeline. Other objects in an Object Time Series that may involve properties and values that may be visually displayed may also be configured to be displayed as associated with respective points in time.


Depending on the embodiment, new object types may be introduced into an Object Time Series by applying filters, sampling, slicing, or other functions or criteria on an Object Time Series. For example, in FIG. 9, the city object in the Object Time Series 900 may be used and associated with data/information related to weather in each city. Thus, a new object type may be created that includes weather for each city. In some embodiments, a new Object Time Series may be created, which includes mapping between a plurality of points in time, the city objects/values, and weather objects/values for the cities. Additional objects may also be created based on the city object or any other object in an existing object time series.



FIG. 9A illustrates an example figure that may be generated and presented to a user to visualize two Object Time Series associated with status objects to show duration. As can be seen in the visualization 920 of the two Object Time Series, the duration of a status may be easily displayed and presented to users using Object Time Series. In some embodiments, horizontal bars with various colors and/or patterns may easily show the transition from one status to another. The length of each status bar in 920 may also represent duration of a status. For example, Status 1 Object Time Series 921 has a duration of approximately 180 days. The duration may be easily presented as the length of the horizontal status bar.



FIG. 9B illustrates an example figure that may be generated and presented to a user to visualize two Object Time Series associated with status objects indicating progress over time. As shown in this figure, in some embodiments, the visualization 930 of two Object Time Series may be implemented as two lines. In this example, status 1 is represented as the lowest status vertically. Status 3 is represented as the highest status vertically. Status 2 is in the middle between status 1 and 3. Object Time Series 931, for example, switched from status 2 to status 3 around May 2012.



FIG. 9C illustrates an example figure that may be generated and presented to a user to visualize two Object Time Series associated with various events. As shown in the visualization 940 of the two Object Time Series in FIG. 9C, the two Object Time Series are associated with 3 event types each. The events may be displayed in the same visualization by arranging event types related to the first Object Time Series on top, and event types related to the second Object Time Series below the first Object Time Series. Using this visualization technique, events and event types associated with the two Object Time Series may be compared to each other directly.



FIG. 9D illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series associated with a status object and a range of values associated with each status. In FIG. 9D, the Object Time Series 950 involves statuses that are related to various amounts. For example, as shown in FIG. 9D, Status 3 is associated with amounts ranging from 30K to approximately 150K. Using a visualization technique as shown in FIG. 9D, ranges of values associated with an object may also be displayed and associated directly with a plurality of points in time. In particular FIG. 9D represents one type of visualization that combines the renderings for status (e.g. the background color), events (e.g. the event dots), and progress (e.g. the line graph portion).



FIG. 9E illustrates an example figure that may be generated and presented to a user to visualize an Object Time Series associated with a status object, ranges of values involved with each status, and transitions among the various statuses. As shown in FIG. 9E, three different statuses are involved in the Object Time Series 960 in this example. Status 1 is associated with list prices above $225K, and the transition from status 1 to status 2 happened around October 2010. Status 2 is associated with list prices around $150K, and the transition from status 2 and status 3 took place between November 2010 and May 2011. Status 3 is associated with list prices above $25K but below $50K. Like with FIG. 9E, this figure represents one type of visualization that combines the renderings for status (e.g. the blue, green, and orange colors), events (e.g. the event dots), and progress (e.g. the line graph portion).


Implementation Mechanisms


According to one embodiment, the techniques described herein are implemented by one or more special-purpose computing devices. The special-purpose computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the techniques, or may include one or more general purpose hardware processors programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the techniques. The special-purpose computing devices may be desktop computer systems, server computer systems, portable computer systems, handheld devices, networking devices or any other device or combination of devices that incorporate hard-wired and/or program logic to implement the techniques.


Computing device(s) are generally controlled and coordinated by operating system software, such as iOS, Android, Chrome OS, Windows XP, Windows Vista, Windows 7, Windows 8, Windows Server, Windows CE, Unix, Linux, SunOS, Solaris, iOS, Blackberry OS, VxWorks, or other compatible operating systems. In other embodiments, the computing device may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface functionality, such as a graphical user interface (“GUI”), among other things.


For example, FIG. 10 is a block diagram that illustrates a computer system 800 upon which an embodiment may be implemented. Computer system 800 includes a bus 802 or other communication mechanism for communicating information, and a hardware processor, or multiple processors, 804 coupled with bus 802 for processing information. Hardware processor(s) 804 may be, for example, one or more general purpose microprocessors.


Computer system 800 also includes a main memory 806, such as a random access memory (RAM), cache and/or other dynamic storage devices, coupled to bus 802 for storing information and instructions to be executed by processor 804. Main memory 806 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 804. Such instructions, when stored in storage media accessible to processor 804, render computer system 800 into a special-purpose machine that is customized to perform the operations specified in the instructions.


Computer system 800 further includes a read only memory (ROM) 808 or other static storage device coupled to bus 802 for storing static information and instructions for processor 804. A storage device 810, such as a magnetic disk, optical disk, or USB thumb drive (Flash drive), etc., is provided and coupled to bus 802 for storing information and instructions.


Computer system 800 may be coupled via bus 802 to a display 812, such as a cathode ray tube (CRT) or LCD display (or touch screen), for displaying information to a computer user. An input device 814, including alphanumeric and other keys, is coupled to bus 802 for communicating information and command selections to processor 804. Another type of user input device is cursor control 816, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 804 and for controlling cursor movement on display 812. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane. In some embodiments, the same direction information and command selections as cursor control may be implemented via receiving touches on a touch screen without a cursor.


Computing system 800 may include a user interface module to implement a GUI that may be stored in a mass storage device as executable software codes that are executed by the computing device(s). This and other modules may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.


In general, the word “module,” as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, Lua, C or C++. A software module may be compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language such as, for example, BASIC, Perl, or Python. It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts. Software modules configured for execution on computing devices may be provided on a computer readable medium, such as a compact disc, digital video disc, flash drive, magnetic disc, or any other tangible medium, or as a digital download (and may be originally stored in a compressed or installable format that requires installation, decompression or decryption prior to execution). Such software code may be stored, partially or fully, on a memory device of the executing computing device, for execution by the computing device. Software instructions may be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors. The modules or computing device functionality described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage


Computer system 800 may implement the techniques described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the computer system causes or programs computer system 800 to be a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 800 in response to processor(s) 804 executing one or more sequences of one or more instructions contained in main memory 806. Such instructions may be read into main memory 806 from another storage medium, such as storage device 810. Execution of the sequences of instructions contained in main memory 806 causes processor(s) 804 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.


The term “non-transitory media,” and similar terms, as used herein refers to any media that store data and/or instructions that cause a machine to operate in a specific fashion. Such non-transitory media may comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 810. Volatile media includes dynamic memory, such as main memory 806. Common forms of non-transitory media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge, and networked versions of the same.


Non-transitory media is distinct from but may be used in conjunction with transmission media. Transmission media participates in transferring information between nontransitory media. For example, transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 802. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.


Various forms of media may be involved in carrying one or more sequences of one or more instructions to processor 804 for execution. For example, the instructions may initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 800 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 802. Bus 802 carries the data to main memory 806, from which processor 804 retrieves and executes the instructions. The instructions received by main memory 806 may retrieves and executes the instructions. The instructions received by main memory 806 may optionally be stored on storage device 810 either before or after execution by processor 804.


Computer system 800 also includes a communication interface 818 coupled to bus 802. Communication interface 818 provides a two-way data communication coupling to a network link 820 that is connected to a local network 822. For example, communication interface 818 may be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 818 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN (or WAN component to communicated with a WAN). Wireless links may also be implemented. In any such implementation, communication interface 818 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.


Network link 820 typically provides data communication through one or more networks to other data devices. For example, network link 820 may provide a connection through local network 822 to a host computer 824 or to data equipment operated by an Internet Service Provider (ISP) 826. ISP 826 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 828. Local network 822 and Internet 828 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 820 and through communication interface 818, which carry the digital data to and from computer system 800, are example forms of transmission media.


Computer system 800 can send messages and receive data, including program code, through the network(s), network link 820 and communication interface 818. In the Internet example, a server 830 might transmit a requested code for an application program through Internet 828, ISP 826, local network 822 and communication interface 818.


The received code may be executed by processor 804 as it is received, and/or stored in storage device 810, or other non-volatile storage for later execution.


Each of the processes, methods, and algorithms described in the preceding sections may be embodied in, and fully or partially automated by, code modules executed by one or more computer systems or computer processors comprising computer hardware. The processes and algorithms may be implemented partially or wholly in application-specific circuitry.


The various features and processes described above may be used independently of one another, or may be combined in various ways. All possible combinations and subcombinations are intended to fall within the scope of this disclosure. In addition, certain method or process blocks may be omitted in some implementations. The methods and processes described herein are also not limited to any particular sequence, and the blocks or states relating thereto can be performed in other sequences that are appropriate. For example, described blocks or states may be performed in an order other than that specifically disclosed, or multiple blocks or states may be combined in a single block or state. The example blocks or states may be performed in serial, in parallel, or in some other manner. Blocks or states may be added to or removed from the disclosed example embodiments. The example systems and components described herein may be configured differently than described. For example, elements may be added to, removed from, or rearranged compared to the disclosed example embodiments.


Conditional language, such as, among others, “can,” “could,” “might,” or “may,” unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.


Any process descriptions, elements, or blocks in the flow diagrams described herein and/or depicted in the attached figures should be understood as potentially representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps in the process. Alternate implementations are included within the scope of the embodiments described herein in which elements or functions may be deleted, executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those skilled in the art.


It should be emphasized that many variations and modifications may be made to the above-described embodiments, the elements of which are to be understood as being among other acceptable examples. All such modifications and variations are intended to be included herein within the scope of this disclosure. The foregoing description details certain embodiments of the invention. It will be appreciated, however, that no matter how detailed the foregoing appears in text, the invention can be practiced in many ways. As is also stated above, it should be noted that the use of particular terminology when describing certain features or aspects of the invention should not be taken to imply that the terminology is being re-defined herein to be restricted to including any specific characteristics of the features or aspects of the invention with which that terminology is associated. The scope of the invention should therefore be construed in accordance with the appended claims and any equivalents thereof.

Claims
  • 1. A computer implemented method comprising: by a computer system comprising one or more computer hardware processors and one or more storage devices,parsing, according to a first parser definition, first data in a first format to generate first modified data different from the first data, the first data from a first data source;generating, from at least some of the first modified data, an event object;parsing, according to a second parser definition, second data in a second format to generate a non-event object, the non-event object corresponding to a data container for properties, and wherein a property of the properties comprises a property type corresponding to a first property value which has a non-numerical value, the second data from a second data source;determining, from a second property of at least one of the event object or the non-event object, an association between the event object and the non-event object;accessing the first property value of the non-event object;generating an object time series based at least on a plurality of points in time, wherein a first point in time of the plurality of points in time is associated with at least the event object; andcausing presentation of a visual representation of the object time series including representations of the event object with the first point in time, the non-event object with at least one respective point in time, and the non-numerical value, and wherein the visual representation of the object time series further comprises each object of the event object and the non-event object as visually connected with at least one respective point in time on the visual representation.
  • 2. The computer implemented method of claim 1, wherein parsing the second data further comprises: identifying, from the second parser definition, the property type for at least some of the second data.
  • 3. The computer implemented method of claim 1, further comprising: receiving user input data comprising a user defined link type, wherein determining the association between the event object and the non-event object is based at least on the user defined link type.
  • 4. The computer implemented method of claim 1, further comprising: receiving user input data comprising a user defined object type and a user defined property type associated with the user defined object type, wherein at least one of the event object or the non-event object corresponds to the user defined object type, the at least one of the event object or the non-event object comprising the user defined property type.
  • 5. The computer implemented method of claim 1, further comprising: determining that the event object is further associated with another non-event object; andcausing presentation of a graph visual representation including a representation of the non-event object as linked with the another non-event object.
  • 6. A non-transitory computer storage medium storing computer executable instructions that when executed by a computer hardware processor perform operations comprising: parsing, according to a first parser definition, first data in a first format to generate first modified data different from the first data, the first data from a first data source;generating, from at least some of the first modified data, an event object;parsing, according to a second parser definition, second data in a second format to generate a non-event object, the non-event object corresponding to a data container for properties, and wherein a property of the properties comprises a property type corresponding to a first property value which has a non-numerical value, the second data from a second data source;determining, from a second property of at least one of the event object or the non-event object, an association between the event object and the non-event object;accessing the first property value of the non-event object;generating an object time series based at least on a plurality of points in time, wherein a first point in time of the plurality of points in time is associated with at least the event object; andcausing presentation of a visual representation of the object time series including representations of the event object with the first point in time, the non-event object with at least one respective point in time, and the non-numerical value, and wherein the visual representation of the object time series further comprises each object of the event object and the non-event object as visually connected with at least one respective point in time on the visual representation.
  • 7. The non-transitory computer storage medium of claim 6, wherein parsing the second data further comprises: identifying, from the second parser definition, the property type for at least some of the second data.
  • 8. The non-transitory computer storage medium of claim 6, wherein the operations further comprise: receiving user input data comprising a user defined link type, wherein determining the association between the event object and the non-event object is based at least on the user defined link type.
  • 9. The non-transitory computer storage medium of claim 6, wherein the operations further comprise: receiving user input data comprising a user defined object type and a user defined property type associated with the user defined object type, wherein at least one of the event object or the non-event object corresponds to the user defined object type, the at least one of the event object or the non-event object comprising the user defined property type.
  • 10. The non-transitory computer storage medium of claim 6, wherein the operations further comprise: determining that the event object is further associated with another non-event object; andgenerating a graph visual representation including representations of the non-event object as linked with the another non-event object.
  • 11. The non-transitory computer storage medium of claim 6, wherein a data type of the property comprises at least one of: a text type or a number type.
  • 12. A system comprising: a plurality of data sources comprising a first data source and a second data source, wherein first data of the first data source is stored in a first format, the first format different from a second format of second data of the second data source;at least one computer hardware processor; anddata storage comprising instructions executable by the at least one computer hardware processor to cause the system to: parse, according to a first parser definition, the first data in the first format to generate first modified data different from the first data;generate, from at least some of the first modified data, an event object;parse, according to a second parser definition, the second data in the second format to generate a non-event object, the non-event object corresponding to a data container for properties, and wherein a property of the properties comprises a property type corresponding to a first property value which has a non-numerical value;determine, from a second property of at least one of the event object or the non-event object, an association between the event object and the non-event object;access the first property value of the non-event object;generate an object time series based at least on a plurality of points in time, wherein a first point in time of the plurality of points in time is associated with at least the event object; andcause presentation of a visual representation of the object time series including representations of the event object with the first point in time, the non-event object with at least one respective point in time, and the non-numerical value, and wherein the visual representation of the object time series further comprises each object of the event object and the non-event object as visually connected with at least one respective point in time on the visual representation.
  • 13. The system of claim 12, wherein parsing the second data further comprises: identifying, from the second parser definition, the property type for at least some of the second data.
  • 14. The system of claim 12, wherein the instructions executable by the at least one computer hardware processor further cause the system to: receive user input data comprising a user defined link type, wherein determining the association between the event object and the non-event object is based at least on the user defined link type.
  • 15. The system of claim 12, wherein the instructions executable by the at least one computer hardware processor further cause the system to: receive user input data comprising a user defined object type and a user defined property type associated with the user defined object type, wherein at least one of the event object or the non-event object corresponds to the user defined object type, the at least one of the event object or the non-event object comprising the user defined property type.
  • 16. The system of claim 12, wherein the instructions executable by the at least one computer hardware processor further cause the system to: determine that the event object is further associated with another non-event object; andcause presentation of a graph visual representation including a representation of the non-event object as linked with the another non-event object.
  • 17. The system of claim 12, wherein a data type of the property comprises at least one of: a text type or a number type.
INCORPORATION BY REFERENCE TO ANY PRIORITY APPLICATIONS

Any and all applications for which a foreign or domestic priority claim is identified in the Application Data Sheet as filed with the present application are hereby incorporated by reference under 37 CFR 1.57. This application is a continuation of U.S. patent application Ser. No. 14/596,552 entitled “GENERATING OBJECT TIME SERIES AND DATA OBJECTS,” filed Jan. 14, 2015, which is a continuation of U.S. patent application Ser. No. 14/148,559 entitled “GENERATING AN OBJECT TIME SERIES FROM DATA OBJECTS,” filed Jan. 6, 2014, now U.S. Pat. No. 8,937,619, which claims benefit of U.S. Provisional Patent Application No. 61/799,691 entitled “OBJECT TIME SERIES,” filed Mar. 15, 2013. Each of these applications are hereby incorporated by reference herein in their entireties.

US Referenced Citations (581)
Number Name Date Kind
5109399 Thompson Apr 1992 A
5329108 Lamoure Jul 1994 A
5632009 Rao et al. May 1997 A
5670987 Doi et al. Sep 1997 A
5781704 Rossmo Jul 1998 A
5798769 Chiu et al. Aug 1998 A
5845300 Comer Dec 1998 A
6057757 Arrowsmith et al. May 2000 A
6091956 Hollenberg Jul 2000 A
6161098 Wallman Dec 2000 A
6167405 Rosensteel, Jr. et al. Dec 2000 A
6219053 Tachibana et al. Apr 2001 B1
6232971 Haynes May 2001 B1
6247019 Davies Jun 2001 B1
6279018 Kudrolli et al. Aug 2001 B1
6341310 Leshem et al. Jan 2002 B1
6366933 Ball et al. Apr 2002 B1
6369835 Lin Apr 2002 B1
6456997 Shukla Sep 2002 B1
6549944 Weinberg et al. Apr 2003 B1
6560620 Ching May 2003 B1
6581068 Bensoussan et al. Jun 2003 B1
6594672 Lampson et al. Jul 2003 B1
6631496 Li et al. Oct 2003 B1
6642945 Sharpe Nov 2003 B1
6674434 Chojnacki et al. Jan 2004 B1
6714936 Nevin, III Mar 2004 B1
6775675 Nwabueze et al. Aug 2004 B1
6820135 Dingman Nov 2004 B1
6828920 Owen et al. Dec 2004 B2
6839745 Dingari et al. Jan 2005 B1
6877137 Rivette et al. Apr 2005 B1
6976210 Silva et al. Dec 2005 B1
6980984 Huffman et al. Dec 2005 B1
6985950 Hanson et al. Jan 2006 B1
7036085 Barros Apr 2006 B2
7043702 Chi et al. May 2006 B2
7055110 Kupka et al. May 2006 B2
7139800 Bellotti et al. Nov 2006 B2
7158878 Rasmussen et al. Jan 2007 B2
7162475 Ackerman Jan 2007 B2
7168039 Bertram Jan 2007 B2
7171427 Witowski et al. Jan 2007 B2
7269786 Malloy et al. Sep 2007 B1
7278105 Kitts Oct 2007 B1
7290698 Poslinski et al. Nov 2007 B2
7333998 Heckerman et al. Feb 2008 B2
7370047 Gorman May 2008 B2
7379811 Rasmussen et al. May 2008 B2
7379903 Caballero et al. May 2008 B2
7426654 Adams et al. Sep 2008 B2
7451397 Weber et al. Nov 2008 B2
7454466 Bellotti et al. Nov 2008 B2
7467375 Tondreau et al. Dec 2008 B2
7487139 Fraleigh et al. Feb 2009 B2
7502786 Liu et al. Mar 2009 B2
7525422 Bishop et al. Apr 2009 B2
7529727 Arning et al. May 2009 B2
7529734 Dirisala May 2009 B2
7558677 Jones Jul 2009 B2
7574409 Patinkin Aug 2009 B2
7574428 Leiserowitz et al. Aug 2009 B2
7579965 Bucholz Aug 2009 B2
7596285 Brown et al. Sep 2009 B2
7614006 Molander Nov 2009 B2
7617232 Gabbert et al. Nov 2009 B2
7620628 Kapur et al. Nov 2009 B2
7627812 Chamberlain et al. Dec 2009 B2
7634717 Chamberlain et al. Dec 2009 B2
7703021 Flam Apr 2010 B1
7706817 Bamrah et al. Apr 2010 B2
7712049 Williams et al. May 2010 B2
7716077 Mikurak May 2010 B1
7725530 Sah et al. May 2010 B2
7725547 Albertson et al. May 2010 B2
7730082 Sah et al. Jun 2010 B2
7730109 Rohrs et al. Jun 2010 B2
7770100 Chamberlain et al. Aug 2010 B2
7805457 Viola et al. Sep 2010 B1
7809703 Balabhadrapatruni et al. Oct 2010 B2
7818658 Chen Oct 2010 B2
7870493 Pall et al. Jan 2011 B2
7894984 Rasmussen et al. Feb 2011 B2
7899611 Downs et al. Mar 2011 B2
7917376 Bellin et al. Mar 2011 B2
7920963 Jouline et al. Apr 2011 B2
7933862 Chamberlain et al. Apr 2011 B2
7941321 Greenstein et al. May 2011 B2
7962281 Rasmussen et al. Jun 2011 B2
7962495 Jain et al. Jun 2011 B2
7962848 Bertram Jun 2011 B2
7970240 Chao et al. Jun 2011 B1
7971150 Raskutti et al. Jun 2011 B2
7984374 Caro et al. Jul 2011 B2
8001465 Kudrolli et al. Aug 2011 B2
8001482 Bhattiprolu et al. Aug 2011 B2
8010545 Stefik et al. Aug 2011 B2
8015487 Roy et al. Sep 2011 B2
8024778 Cash et al. Sep 2011 B2
8036632 Cona et al. Oct 2011 B1
8103543 Zwicky Jan 2012 B1
8134457 Velipasalar et al. Mar 2012 B2
8145703 Frishert et al. Mar 2012 B2
8185819 Sah et al. May 2012 B2
8196184 Amirov et al. Jun 2012 B2
8214361 Sandler et al. Jul 2012 B1
8214764 Gemmell et al. Jul 2012 B2
8225201 Michael Jul 2012 B2
8229947 Fujinaga Jul 2012 B2
8230333 Decherd et al. Jul 2012 B2
8271461 Pike et al. Sep 2012 B2
8280880 Aymeloglu et al. Oct 2012 B1
8290926 Ozzie et al. Oct 2012 B2
8290942 Jones et al. Oct 2012 B2
8301464 Cave et al. Oct 2012 B1
8301904 Gryaznov Oct 2012 B1
8312367 Foster Nov 2012 B2
8312546 Alme Nov 2012 B2
8352881 Champion et al. Jan 2013 B2
8368695 Howell et al. Feb 2013 B2
8397171 Klassen et al. Mar 2013 B2
8412707 Mianji Apr 2013 B1
8447722 Ahuja et al. May 2013 B1
8452790 Mianji May 2013 B1
8463036 Ramesh et al. Jun 2013 B1
8489331 Kopf et al. Jul 2013 B2
8489623 Jain et al. Jul 2013 B2
8489641 Seefeld et al. Jul 2013 B1
8498984 Hwang et al. Jul 2013 B1
8510743 Hackborn et al. Aug 2013 B2
8514082 Cova et al. Aug 2013 B2
8515207 Chau Aug 2013 B2
8554579 Tribble et al. Oct 2013 B2
8554653 Falkenborg et al. Oct 2013 B2
8554709 Goodson et al. Oct 2013 B2
8560413 Quarterman Oct 2013 B1
8577911 Stepinski et al. Nov 2013 B1
8589273 Creeden et al. Nov 2013 B2
8595234 Siripurapu et al. Nov 2013 B2
8620641 Farnsworth et al. Dec 2013 B2
8639757 Zang et al. Jan 2014 B1
8646080 Williamson et al. Feb 2014 B2
8676857 Adams et al. Mar 2014 B1
8689108 Duffield et al. Apr 2014 B1
8713467 Goldenberg et al. Apr 2014 B1
8726379 Stiansen et al. May 2014 B1
8739278 Varghese May 2014 B2
8742934 Sarpy et al. Jun 2014 B1
8744890 Bernier Jun 2014 B1
8745516 Mason et al. Jun 2014 B2
8781169 Jackson et al. Jul 2014 B2
8787939 Papakipos et al. Jul 2014 B2
8788407 Singh et al. Jul 2014 B1
8799799 Cervelli et al. Aug 2014 B1
8812960 Sun et al. Aug 2014 B1
8830322 Nerayoff et al. Sep 2014 B2
8832594 Thompson et al. Sep 2014 B1
8868486 Tamayo Oct 2014 B2
8868537 Colgrove et al. Oct 2014 B1
8917274 Ma et al. Dec 2014 B2
8924872 Bogomolov et al. Dec 2014 B1
8937619 Sharma et al. Jan 2015 B2
8938686 Erenrich et al. Jan 2015 B1
9009171 Grossman et al. Apr 2015 B1
9009827 Albertson et al. Apr 2015 B1
9021260 Falk et al. Apr 2015 B1
9021384 Beard et al. Apr 2015 B1
9043696 Meiklejohn et al. May 2015 B1
9043894 Dennison et al. May 2015 B1
9069842 Melby Jun 2015 B2
9116975 Shankar et al. Aug 2015 B2
9146954 Boe et al. Sep 2015 B1
9202249 Cohen et al. Dec 2015 B1
9223773 Isaacson Dec 2015 B2
9229952 Meacham et al. Jan 2016 B1
9250759 Commons Feb 2016 B1
9256664 Chakerian et al. Feb 2016 B2
9298678 Chakerian et al. Mar 2016 B2
9319288 Somaiya et al. Apr 2016 B2
9367872 Visbal et al. Jun 2016 B1
9626088 Ma et al. Apr 2017 B2
9646396 Sharma et al. May 2017 B2
20010021936 Bertram Sep 2001 A1
20020033848 Sciammarella et al. Mar 2002 A1
20020065708 Senay et al. May 2002 A1
20020091707 Keller Jul 2002 A1
20020095658 Shulman Jul 2002 A1
20020116120 Ruiz et al. Aug 2002 A1
20020130907 Chi et al. Sep 2002 A1
20020174201 Ramer et al. Nov 2002 A1
20020194119 Wright et al. Dec 2002 A1
20030028560 Kudrolli et al. Feb 2003 A1
20030036848 Sheha et al. Feb 2003 A1
20030039948 Donahue Feb 2003 A1
20030140106 Raguseo Jul 2003 A1
20030144868 MacIntyre et al. Jul 2003 A1
20030163352 Surpin et al. Aug 2003 A1
20030172014 Quackenbush et al. Sep 2003 A1
20030200217 Ackerman Oct 2003 A1
20030225755 Iwayama et al. Dec 2003 A1
20030229848 Arend et al. Dec 2003 A1
20040032432 Baynger Feb 2004 A1
20040064256 Barinek et al. Apr 2004 A1
20040085318 Hassler et al. May 2004 A1
20040095349 Bito et al. May 2004 A1
20040103124 Kupkova May 2004 A1
20040111410 Burgoon et al. Jun 2004 A1
20040126840 Cheng et al. Jul 2004 A1
20040143602 Ruiz et al. Jul 2004 A1
20040143796 Lerner et al. Jul 2004 A1
20040160309 Stilp Aug 2004 A1
20040163039 Gorman Aug 2004 A1
20040181554 Heckerman et al. Sep 2004 A1
20040193600 Kaasten et al. Sep 2004 A1
20040205524 Richter et al. Oct 2004 A1
20040221223 Yu et al. Nov 2004 A1
20040260702 Cragun et al. Dec 2004 A1
20040267746 Marcjan et al. Dec 2004 A1
20050027705 Sadri et al. Feb 2005 A1
20050028094 Allyn Feb 2005 A1
20050039119 Parks et al. Feb 2005 A1
20050065811 Chu et al. Mar 2005 A1
20050078858 Yao et al. Apr 2005 A1
20050080769 Gemmell Apr 2005 A1
20050086207 Heuer et al. Apr 2005 A1
20050125715 Franco et al. Jun 2005 A1
20050154628 Eckart et al. Jul 2005 A1
20050154769 Eckart et al. Jul 2005 A1
20050162523 Darrell et al. Jul 2005 A1
20050166144 Gross Jul 2005 A1
20050180330 Shapiro Aug 2005 A1
20050182793 Keenan et al. Aug 2005 A1
20050183005 Denoue et al. Aug 2005 A1
20050210409 Jou Sep 2005 A1
20050246327 Yeung et al. Nov 2005 A1
20050251786 Citron et al. Nov 2005 A1
20060026120 Carolan et al. Feb 2006 A1
20060026170 Kreitler et al. Feb 2006 A1
20060045470 Poslinski et al. Mar 2006 A1
20060059139 Robinson Mar 2006 A1
20060074866 Chamberlain et al. Apr 2006 A1
20060074881 Vembu et al. Apr 2006 A1
20060080619 Carlson et al. Apr 2006 A1
20060093222 Saffer et al. May 2006 A1
20060129746 Porter Jun 2006 A1
20060139375 Rasmussen et al. Jun 2006 A1
20060142949 Helt Jun 2006 A1
20060143034 Rothermel Jun 2006 A1
20060149596 Surpin et al. Jul 2006 A1
20060184889 Molander Aug 2006 A1
20060203337 White Sep 2006 A1
20060209085 Wong et al. Sep 2006 A1
20060218637 Thomas et al. Sep 2006 A1
20060241974 Chao et al. Oct 2006 A1
20060242040 Rader et al. Oct 2006 A1
20060242630 Koike et al. Oct 2006 A1
20060271277 Hu et al. Nov 2006 A1
20060279630 Aggarwal et al. Dec 2006 A1
20070011150 Frank Jan 2007 A1
20070016363 Huang et al. Jan 2007 A1
20070038646 Thota Feb 2007 A1
20070038962 Fuchs et al. Feb 2007 A1
20070057966 Ohno et al. Mar 2007 A1
20070078832 Ott et al. Apr 2007 A1
20070083541 Fraleigh et al. Apr 2007 A1
20070094389 Nussey et al. Apr 2007 A1
20070150369 Zivin Jun 2007 A1
20070174760 Chamberlain et al. Jul 2007 A1
20070192265 Chopin et al. Aug 2007 A1
20070198571 Ferguson et al. Aug 2007 A1
20070208497 Downs et al. Sep 2007 A1
20070208498 Barker et al. Sep 2007 A1
20070208736 Tanigawa et al. Sep 2007 A1
20070233709 Abnous Oct 2007 A1
20070240062 Christena et al. Oct 2007 A1
20070250784 Riley Oct 2007 A1
20070266336 Nojima et al. Nov 2007 A1
20070294200 Au Dec 2007 A1
20070294643 Kyle Dec 2007 A1
20080016216 Worley et al. Jan 2008 A1
20080040275 Paulsen et al. Feb 2008 A1
20080040684 Crump Feb 2008 A1
20080051989 Welsh Feb 2008 A1
20080052142 Bailey et al. Feb 2008 A1
20080077597 Butler Mar 2008 A1
20080077642 Carbone et al. Mar 2008 A1
20080082486 Lermant et al. Apr 2008 A1
20080104019 Nath May 2008 A1
20080126951 Sood et al. May 2008 A1
20080148398 Mezack et al. Jun 2008 A1
20080155440 Trevor et al. Jun 2008 A1
20080162616 Gross et al. Jul 2008 A1
20080195417 Surpin et al. Aug 2008 A1
20080195608 Clover Aug 2008 A1
20080222295 Robinson et al. Sep 2008 A1
20080243711 Aymeloglu et al. Oct 2008 A1
20080249983 Meisels et al. Oct 2008 A1
20080255973 El Wade et al. Oct 2008 A1
20080263468 Cappione et al. Oct 2008 A1
20080267107 Rosenberg Oct 2008 A1
20080276167 Michael Nov 2008 A1
20080278311 Grange et al. Nov 2008 A1
20080281819 Tenenbaum et al. Nov 2008 A1
20080288306 MacIntyre et al. Nov 2008 A1
20080288475 Kim et al. Nov 2008 A1
20080294994 Kruger et al. Nov 2008 A1
20080301643 Appleton et al. Dec 2008 A1
20090002492 Velipasalar et al. Jan 2009 A1
20090027418 Maru et al. Jan 2009 A1
20090030915 Winter et al. Jan 2009 A1
20090055251 Shah et al. Feb 2009 A1
20090070162 Leonelli et al. Mar 2009 A1
20090076845 Bellin et al. Mar 2009 A1
20090088964 Schaaf et al. Apr 2009 A1
20090094166 Aymeloglu et al. Apr 2009 A1
20090119309 Gibson et al. May 2009 A1
20090125359 Knapic May 2009 A1
20090125369 Kloostra et al. May 2009 A1
20090125459 Norton et al. May 2009 A1
20090132921 Hwangbo et al. May 2009 A1
20090132953 Reed et al. May 2009 A1
20090143052 Bates et al. Jun 2009 A1
20090144262 White et al. Jun 2009 A1
20090144274 Fraleigh et al. Jun 2009 A1
20090150854 Elaasar et al. Jun 2009 A1
20090164934 Bhattiprolu et al. Jun 2009 A1
20090171939 Athsani et al. Jul 2009 A1
20090172511 Decherd et al. Jul 2009 A1
20090172821 Daira et al. Jul 2009 A1
20090177962 Gusmorino et al. Jul 2009 A1
20090179892 Tsuda et al. Jul 2009 A1
20090187464 Bai et al. Jul 2009 A1
20090192957 Subramanian et al. Jul 2009 A1
20090222400 Kupershmidt et al. Sep 2009 A1
20090222759 Drieschner Sep 2009 A1
20090222760 Halverson et al. Sep 2009 A1
20090234720 George et al. Sep 2009 A1
20090249244 Robinson et al. Oct 2009 A1
20090254970 Agarwal et al. Oct 2009 A1
20090281839 Lynn et al. Nov 2009 A1
20090287470 Farnsworth et al. Nov 2009 A1
20090292626 Oxford Nov 2009 A1
20090327208 Bittner et al. Dec 2009 A1
20100011282 Dollard et al. Jan 2010 A1
20100030722 Goodson et al. Feb 2010 A1
20100057716 Stefik et al. Mar 2010 A1
20100070523 Delgo et al. Mar 2010 A1
20100070842 Aymeloglu et al. Mar 2010 A1
20100070845 Facemire et al. Mar 2010 A1
20100070897 Aymeloglu et al. Mar 2010 A1
20100100963 Mahaffey Apr 2010 A1
20100103124 Kruzensiki et al. Apr 2010 A1
20100106752 Eckardt et al. Apr 2010 A1
20100114887 Conway et al. May 2010 A1
20100122152 Chamberlain et al. May 2010 A1
20100131457 Heimendinger May 2010 A1
20100162176 Dunton Jun 2010 A1
20100185691 Irmak et al. Jul 2010 A1
20100191563 Schlaifer et al. Jul 2010 A1
20100198684 Eraker et al. Aug 2010 A1
20100199225 Coleman et al. Aug 2010 A1
20100228812 Uomini Sep 2010 A1
20100250412 Wagner Sep 2010 A1
20100280857 Liu et al. Nov 2010 A1
20100293174 Bennett et al. Nov 2010 A1
20100306029 Jolley Dec 2010 A1
20100306713 Geisner et al. Dec 2010 A1
20100313119 Baldwin et al. Dec 2010 A1
20100318924 Frankel et al. Dec 2010 A1
20100321399 Ellren et al. Dec 2010 A1
20100325526 Ellis et al. Dec 2010 A1
20100325581 Finkelstein et al. Dec 2010 A1
20100330801 Rouh Dec 2010 A1
20110004498 Readshaw Jan 2011 A1
20110029526 Knight et al. Feb 2011 A1
20110047159 Baid et al. Feb 2011 A1
20110060753 Shaked et al. Mar 2011 A1
20110061013 Bilicki et al. Mar 2011 A1
20110066933 Ludwig Mar 2011 A1
20110074811 Hanson et al. Mar 2011 A1
20110078055 Faribault et al. Mar 2011 A1
20110078173 Seligmann et al. Mar 2011 A1
20110093327 Fordyce et al. Apr 2011 A1
20110107196 Foster May 2011 A1
20110113348 Twiss et al. May 2011 A1
20110117878 Barash et al. May 2011 A1
20110119100 Ruhl et al. May 2011 A1
20110131547 Elaasar Jun 2011 A1
20110137766 Rasmussen et al. Jun 2011 A1
20110153384 Horne et al. Jun 2011 A1
20110161096 Buehler et al. Jun 2011 A1
20110161137 Ubalde et al. Jun 2011 A1
20110167105 Ramakrishnan et al. Jul 2011 A1
20110170799 Carrino et al. Jul 2011 A1
20110173032 Payne et al. Jul 2011 A1
20110181598 O'Neall et al. Jul 2011 A1
20110185316 Reid et al. Jul 2011 A1
20110208724 Jones et al. Aug 2011 A1
20110213655 Henkin Sep 2011 A1
20110218934 Elser Sep 2011 A1
20110219321 Gonzalez et al. Sep 2011 A1
20110219450 McDougal et al. Sep 2011 A1
20110225198 Edwards et al. Sep 2011 A1
20110238495 Kang Sep 2011 A1
20110238553 Raj et al. Sep 2011 A1
20110251951 Kolkowtiz Oct 2011 A1
20110252351 Sikora et al. Oct 2011 A1
20110258158 Resende et al. Oct 2011 A1
20110270705 Parker Nov 2011 A1
20110289397 Eastmond et al. Nov 2011 A1
20110289407 Naik et al. Nov 2011 A1
20110289420 Morioka et al. Nov 2011 A1
20110291851 Whisenant Dec 2011 A1
20110310005 Chen et al. Dec 2011 A1
20110314007 Dassa et al. Dec 2011 A1
20120004904 Shin et al. Jan 2012 A1
20120019559 Siler et al. Jan 2012 A1
20120036013 Neuhaus et al. Feb 2012 A1
20120036434 Oberstein Feb 2012 A1
20120050293 Carlhian et al. Mar 2012 A1
20120066296 Appleton et al. Mar 2012 A1
20120072825 Sherkin et al. Mar 2012 A1
20120075324 Cardno et al. Mar 2012 A1
20120079363 Folting et al. Mar 2012 A1
20120084118 Bai et al. Apr 2012 A1
20120106801 Jackson May 2012 A1
20120116828 Shannon May 2012 A1
20120117082 Koperda et al. May 2012 A1
20120131512 Takeuchi et al. May 2012 A1
20120137235 TS et al. May 2012 A1
20120144335 Abeln et al. Jun 2012 A1
20120159307 Chung et al. Jun 2012 A1
20120159362 Brown et al. Jun 2012 A1
20120159399 Bastide et al. Jun 2012 A1
20120170847 Tsukidate Jul 2012 A1
20120173985 Peppel Jul 2012 A1
20120180002 Campbell et al. Jul 2012 A1
20120196557 Reich et al. Aug 2012 A1
20120196558 Reich et al. Aug 2012 A1
20120197651 Robinson et al. Aug 2012 A1
20120203708 Psota et al. Aug 2012 A1
20120208636 Feige Aug 2012 A1
20120221511 Gibson et al. Aug 2012 A1
20120221553 Wittmer et al. Aug 2012 A1
20120221580 Barney Aug 2012 A1
20120245976 Kumar et al. Sep 2012 A1
20120246148 Dror Sep 2012 A1
20120254129 Wheeler et al. Oct 2012 A1
20120284345 Costenaro et al. Nov 2012 A1
20120290879 Shibuya et al. Nov 2012 A1
20120296907 Long et al. Nov 2012 A1
20120311684 Paulsen et al. Dec 2012 A1
20120323888 Osann, Jr. Dec 2012 A1
20120330801 McDougal et al. Dec 2012 A1
20120330973 Ghuneim et al. Dec 2012 A1
20130006426 Healey et al. Jan 2013 A1
20130006725 Simanek et al. Jan 2013 A1
20130006916 McBride et al. Jan 2013 A1
20130018796 Kolhatkar et al. Jan 2013 A1
20130024268 Manickavelu Jan 2013 A1
20130046635 Grigg et al. Feb 2013 A1
20130046842 Muntz et al. Feb 2013 A1
20130050217 Armitage Feb 2013 A1
20130060786 Serrano et al. Mar 2013 A1
20130061169 Pearcy et al. Mar 2013 A1
20130073377 Heath Mar 2013 A1
20130073454 Busch Mar 2013 A1
20130078943 Biage et al. Mar 2013 A1
20130086482 Parsons Apr 2013 A1
20130097482 Marantz et al. Apr 2013 A1
20130101159 Chao et al. Apr 2013 A1
20130106860 De Pauw et al. May 2013 A1
20130110822 Ikeda et al. May 2013 A1
20130110877 Bonham et al. May 2013 A1
20130111320 Campbell et al. May 2013 A1
20130117011 Ahmed et al. May 2013 A1
20130117651 Waldman et al. May 2013 A1
20130150004 Rosen Jun 2013 A1
20130151148 Parundekar et al. Jun 2013 A1
20130151388 Falkenborg et al. Jun 2013 A1
20130157234 Gulli et al. Jun 2013 A1
20130166550 Buchmann et al. Jun 2013 A1
20130169666 Pacheco et al. Jul 2013 A1
20130176321 Mitchell et al. Jul 2013 A1
20130179420 Park et al. Jul 2013 A1
20130224696 Wolfe et al. Aug 2013 A1
20130225212 Khan Aug 2013 A1
20130226318 Procyk Aug 2013 A1
20130226953 Markovich et al. Aug 2013 A1
20130232045 Tai Sep 2013 A1
20130238616 Rose et al. Sep 2013 A1
20130246170 Gross et al. Sep 2013 A1
20130251233 Yang et al. Sep 2013 A1
20130262527 Hunter et al. Oct 2013 A1
20130263019 Castellanos et al. Oct 2013 A1
20130267207 Hao et al. Oct 2013 A1
20130268520 Fisher et al. Oct 2013 A1
20130279757 Kephart Oct 2013 A1
20130282696 John et al. Oct 2013 A1
20130290011 Lynn et al. Oct 2013 A1
20130290825 Arndt et al. Oct 2013 A1
20130297619 Chandrasekaran et al. Nov 2013 A1
20130311375 Priebatsch Nov 2013 A1
20130335419 Bondeson et al. Dec 2013 A1
20140019461 Bredenberg et al. Jan 2014 A1
20140019936 Cohanoff Jan 2014 A1
20140032506 Hoey et al. Jan 2014 A1
20140033010 Richardt et al. Jan 2014 A1
20140040371 Gurevich et al. Feb 2014 A1
20140046934 Zhou Feb 2014 A1
20140047319 Eberlein Feb 2014 A1
20140047357 Alfaro et al. Feb 2014 A1
20140059038 McPherson et al. Feb 2014 A1
20140059498 McCormack et al. Feb 2014 A1
20140067611 Adachi et al. Mar 2014 A1
20140068487 Steiger et al. Mar 2014 A1
20140074855 Zhao et al. Mar 2014 A1
20140095273 Tang et al. Apr 2014 A1
20140095509 Patton Apr 2014 A1
20140108068 Williams Apr 2014 A1
20140108380 Gotz et al. Apr 2014 A1
20140108985 Scott et al. Apr 2014 A1
20140129261 Bothwell et al. May 2014 A1
20140149436 Bahrami et al. May 2014 A1
20140156527 Grigg et al. Jun 2014 A1
20140157172 Peery et al. Jun 2014 A1
20140164502 Khodorenko et al. Jun 2014 A1
20140189536 Lange et al. Jul 2014 A1
20140195515 Baker et al. Jul 2014 A1
20140195887 Ellis et al. Jul 2014 A1
20140214579 Shen et al. Jul 2014 A1
20140222521 Chait Aug 2014 A1
20140244388 Manouchehri et al. Aug 2014 A1
20140258246 Lo Faro et al. Sep 2014 A1
20140267294 Ma Sep 2014 A1
20140267295 Sharma Sep 2014 A1
20140279824 Tamayo Sep 2014 A1
20140282177 Wang et al. Sep 2014 A1
20140310266 Greenfield Oct 2014 A1
20140316911 Gross Oct 2014 A1
20140333651 Cervelli et al. Nov 2014 A1
20140337772 Cervelli et al. Nov 2014 A1
20140344230 Krause et al. Nov 2014 A1
20140351070 Christner et al. Nov 2014 A1
20150019394 Unser et al. Jan 2015 A1
20150046870 Goldenberg et al. Feb 2015 A1
20150073929 Psota et al. Mar 2015 A1
20150081370 Lo et al. Mar 2015 A1
20150089424 Duffield et al. Mar 2015 A1
20150100897 Sun et al. Apr 2015 A1
20150100907 Erenrich et al. Apr 2015 A1
20150134371 Shivakumar et al. May 2015 A1
20150134666 Gattiker et al. May 2015 A1
20150169709 Kara et al. Jun 2015 A1
20150169726 Kara et al. Jun 2015 A1
20150170077 Kara et al. Jun 2015 A1
20150178690 May et al. Jun 2015 A1
20150178825 Huerta Jun 2015 A1
20150178877 Bogomolov et al. Jun 2015 A1
20150186821 Wang et al. Jul 2015 A1
20150187036 Wang et al. Jul 2015 A1
20150213631 Vander Broek Jul 2015 A1
20150227295 Meiklejohn et al. Aug 2015 A1
20150227847 Noel et al. Aug 2015 A1
20150229532 Somaiya et al. Aug 2015 A1
20150229546 Somaiya et al. Aug 2015 A1
20150242401 Liu Aug 2015 A1
20150254878 Sharma et al. Sep 2015 A1
20150309719 Ma et al. Oct 2015 A1
20150317342 Grossman et al. Nov 2015 A1
20150324868 Kaftan et al. Nov 2015 A1
20150347903 Saxena et al. Dec 2015 A1
20150363478 Haynes Dec 2015 A1
20150378996 Kesin et al. Dec 2015 A1
20160004667 Chakerian et al. Jan 2016 A1
20160006749 Cohen et al. Jan 2016 A1
20160034545 Shankar et al. Feb 2016 A1
20160098173 Slawinski et al. Apr 2016 A1
20160162497 Cho et al. Jun 2016 A1
20170109030 Mingione Apr 2017 A1
20170116294 Ma et al. Apr 2017 A1
Foreign Referenced Citations (42)
Number Date Country
2014250678 Feb 2016 AU
102014103482 Sep 2014 DE
102014215621 Feb 2015 DE
1191463 Mar 2002 EP
1672527 Jun 2006 EP
2551799 Jan 2013 EP
2560134 Feb 2013 EP
2778977 Sep 2014 EP
2778983 Sep 2014 EP
2779082 Sep 2014 EP
2835745 Feb 2015 EP
2835770 Feb 2015 EP
2838039 Feb 2015 EP
2846241 Mar 2015 EP
2851852 Mar 2015 EP
2858014 Apr 2015 EP
2858018 Apr 2015 EP
2863326 Apr 2015 EP
2863346 Apr 2015 EP
2869211 May 2015 EP
2884439 Jun 2015 EP
2884440 Jun 2015 EP
2891992 Jul 2015 EP
2911078 Aug 2015 EP
2911100 Aug 2015 EP
2940603 Nov 2015 EP
2940609 Nov 2015 EP
2516155 Jan 2015 GB
2518745 Apr 2015 GB
2012778 Nov 2014 NL
2013306 Feb 2015 NL
624557 Dec 2014 NZ
WO 0009529 Feb 2000 WO
WO 02065353 Aug 2002 WO
WO 2005104736 Nov 2005 WO
WO 2008064207 May 2008 WO
WO 2009061501 May 2009 WO
WO 2010000014 Jan 2010 WO
WO 2010030913 Mar 2010 WO
WO 2010098958 Sep 2010 WO
WO 2013010157 Jan 2013 WO
WO 2013102892 Jul 2013 WO
Non-Patent Literature Citations (268)
Entry
“A First Look: Predicting Market Demand for Food Retail using a Huff Analysis,” TRF Policy Solutions, Jul. 2012, pp. 30.
“A Quick Guide to UniProtKB Swiss-Prot & TrEMBL,” Sep. 2011, pp. 2.
“A Word About Banks and the Laundering of Drug Money,” Aug. 18, 2012, http://www.golemxiv.co.uk/2012/08/a-word-about-banks-and-the-laundering-of-drug-money/.
“Money Laundering Risks and E-Gaming: A European Overview and Assessment,” 2009, http://www.cf.ac.uk/socsi/resources/Levi—Final—Money—Laundering—Risks—egaming.pdf.
“Potential Money Laundering Warning Signs,” snapshot taken 2003, https://web.archive.org/web/20030816090055/http:/finsolinc.com/ANTI-MONEY%20LAUNDERING%20TRAINING%20GUIDES.pdf.
“Refresh CSS Ellipsis When Resizing Container—Stack Overflow,” Jul. 31, 2013, retrieved from internet http://stackoverflow.com/questions/17964681/refresh-css-ellipsis-when-resizing-container, retrieved on May 18, 2015.
“The FASTA Program Package,” fasta-36.3.4, Mar. 25, 2011, pp. 29.
“Using Whois Based Geolocation and Google Maps API for Support Cybercrime Investigations,” http://wseas.us/e-library/conferences/2013/Dubrovnik/TELECIRC/TELECIRC-32.pdf.
About 80 Minutes, “Palantir in a Number of Parts—Part 6—Graph,” Mar. 21, 2013, pp. 1-6, retrieved from the internet http://about80minutes.blogspot.nl/2013/03/palantir-in-number-of-parts-part-6-graph.html retrieved on Aug. 18, 2015.
Acklen, Laura, “Absolute Beginner's Guide to Microsoft Word 2003,” Dec. 24, 2003, pp. 15-18, 34-41, 308-316.
Alfred, Rayner “Summarizing Relational Data Using Semi-Supervised Genetic Algorithm-Based Clustering Techniques”, Journal of Computer Science, 2010, vol. 6, No. 7, pp. 775-784.
Alur et al., “Chapter 2: IBM InfoSphere DataStage Stages,” IBM InfoSphere DataStage Data Flow and Job Design, Jul. 1, 2008, pp. 35-137.
Amnet, “5 Great Tools for Visualizing Your Twitter Followers,” posted Aug. 4, 2010, http://www.amnetblog.com/component/content/article/115-5-grate-tools-for-visualizing-your-twitter-followers.html.
Ananiev et al., “The New Modality API,” http://web.archive.org/web/20061211011958/http://java.sun.com/developer/technicalArticles/J2SE/Desktop/javase6/modality/ Jan. 21, 2006, pp. 8.
Bluttman et al., “Excel Formulas and Functions for Dummies,” 2005, Wiley Publishing, Inc., pp. 280, 284-286.
Boyce, Jim, “Microsoft Outlook 2010 Inside Out,” Aug. 1, 2010, retrieved from the internet https://capdtron.files.wordpress.com/2013/01/outlook-2010-inside—out.pdf.
Bugzilla@Mozilla, “Bug 18726—[feature] Long-click means of invoking contextual menus not supported,” http://bugzilla.mozilla.org/show—bug.cgi?id=18726 printed Jun. 13, 2013 in 11 pages.
Canese et al., “Chapter 2: PubMed: The Bibliographic Database,” The NCBI Handbook, Oct. 2002, pp. 1-10.
Celik, Tantek, “CSS Basic User Interface Module Level 3 (CSS3 UI),” Section 8 Resizing and Overflow, Jan. 17, 2012, retrieved from internet http://www.w3.org/TR/2012/WD-css3-ui-20120117/#resizing-amp-overflow retrieved on May 18, 2015.
Chen et al., “Bringing Order to the Web: Automatically Categorizing Search Results,” CHI '00, Proceedings of the SIGCHI conference on Human Factors in Computing Systems, Apr. 1-6, 2000, The Hague, The Netherlands, pp. 145-152.
Chung, Chin-Wan, “Dataplex: An Access to Heterogeneous Distributed Databases,” Communications of the ACM, Association for Computing Machinery, Inc., vol. 33, No. 1, Jan. 1, 1990, pp. 70-80.
Conner, Nancy, “Google Apps: The Missing Manual,” Sharing and Collaborating on Documents, May 1, 2008, pp. 93-97, 106-113 & 120-121.
Definition “Identify”, downloaded Jan. 22, 2015, 1 page.
Definition “Overlay”, downloaded Jan. 22, 2015, 1 page.
Delcher et al., “Identifying Bacterial Genes and Endosymbiont DNA with Glimmer,” BioInformatics, vol. 23, No. 6, 2007, pp. 673-679.
Dramowicz, Ela, “Retail Trade Area Analysis Using the Huff Model,” Directions Magazine, Jul. 2, 2005 in 10 pages, http://www.directionsmag.com/articles/retail-trade-area-analysis-using-the-huff-model/123411.
Gesher, Ari, “Palantir Screenshots in the Wild: Swing Sightings,” The Palantir Blog, Sep. 11, 2007, pp. 1-12, retrieved from the internet https://www.palantir.com/2007/09/palantir-screenshots/ retrieved on Aug. 18, 2015.
GIS-NET 3 Public—Department of Regional Planning. Planning & Zoning Information for Unincorporated LA County. Retrieved Oct. 2, 2013 from http://gis.planning.lacounty.gov/GIS-NET3—Public/Viewer.html.
Goswami, Gautam, “Quite ‘Writely’ Said!” One Brick at a Time, Aug. 21, 2005, pp. 7.
Griffith, Daniel A., “A Generalized Huff Model,” Geographical Analysis, Apr. 1982, vol. 14, No. 2, pp. 135-144.
Hansen et al., “Analyzing Social Media Networks with NodeXL: Insights from a Connected World”, Elsevier Science, Sep. 2010, Ch. 4 & 10, pp. 53-67 & 143-164.
Hardesty, “Privacy Challenges: Analysis: It's Surprisingly Easy to Identify Individuals from Credit-Card Metadata,” MIT News on Campus and Around the World, MIT News Office, Jan. 29, 2015, 3 pages.
Hibbert et al., “Prediction of Shopping Behavior Using a Huff Model Within a GIS Framework,” Healthy Eating in Context Mar. 18, 2011, pp. 16.
Hogue et al., “Thresher: Automating the Unwrapping of Semantic Content from the World Wide Web,” 14th International Conference on World Wide Web, WWW 2005: Chiba, Japan, May 10-14, 2005, pp. 86-95.
Huang et al., “Systematic and Integrative Analysis of Large Gene Lists Using DAVID Bioinformatics Resources,” Nature Protocols, vol. 4, No. 1, 2009, 44-57.
Huff et al., “Calibrating the Huff Model Using ArcGIS Business Analyst,” ESRI, Sep. 2008, pp. 33.
Huff, David L., “Parameter Estimation in the Huff Model,” ESRI, ArcUser, Oct.-Dec. 2003, pp. 34-36.
International Search Report and Written Opinion in Application No. PCT/US2009/056703, dated Mar. 15, 2010.
Kahan et al., “Annotea: an Open RDF Infrastructure for Shared Web Annotations”, Computer Networks, Elsevier Science Publishers B.V., vol. 39, No. 5, dated Aug. 5, 2002, pp. 589-608.
Keylines.com, “An Introduction to KeyLines and Network Visualization,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-White-Paper.pdf> downloaded May 12, 2014 in 8 pages.
Keylines.com, “KeyLines Datasheet,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-datasheet.pdf> downloaded May 12, 2014 in 2 pages.
Keylines.com, “Visualizing Threats: Improved Cyber Security Through Network Visualization,” Apr. 2014, <http://keylines.com/wp-content/uploads/2014/04/Visualizing-Threats1.pdf> downloaded May 12, 2014 in 10 pages.
Kitts, Paul, “Chapter 14: Genome Assembly and Annotation Process,” The NCBI Handbook, Oct. 2002, pp. 1-21.
Li et al., “Interactive Multimodal Visual Search on Mobile Device,” IEEE Transactions on Multimedia, vol. 15, No. 3, Apr. 1, 2013, pp. 594-607.
Liu, Tianshun, “Combining GIS and the Huff Model to Analyze Suitable Locations for a New Asian Supermarket in the Minneapolis and St. Paul, Minnesota USA,” Papers in Resource Analysis, 2012, vol. 14, pp. 8.
Madden, Tom, “Chapter 16: The BLAST Sequence Analysis Tool,” The NCBI Handbook, Oct. 2002, pp. 1-15.
Manno et al., “Introducing Collaboration in Single-user Applications through the Centralized Control Architecture,” 2010, pp. 10.
Manske, “File Saving Dialogs,” http://www.mozilla.org/editor/ui—specs/FileSaveDialogs.html Jan. 20, 1999, pp. 7.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.yahoo.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.bing.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.google.com.
Microsoft—Developer Network, “Getting Started with VBA in Word 2010,” Apr. 2010, <http://msdn.microsoft.com/en-us/library/ff604039%28v=office.14%29.aspx> as printed Apr. 4, 2014 in 17 pages.
Microsoft Office—Visio, “About connecting shapes,” http://office.microsoft.com/en-us/visio-help/about-connecting-shapes-HP085050369.aspx printed Aug. 4, 2011 in 6 pages.
Microsoft Office—Visio, “Add and glue connectors with the Connector tool,” http://office.microsoft.com/en-us/visio-help/add-and-glue-connectors-with-the-connector-tool-HA010048532.aspx?CTT=1 printed Aug. 4, 2011 in 1 page.
Mizrachi, Ilene, “Chapter 1: GenBank: The Nuckeotide Sequence Database,” The NCBI Handbook, Oct. 2002, pp. 1-14.
Nierman, “Evaluating Structural Similarity in XML Documents”, 6 pages, 2002.
Nolan et al., “MCARTA: A Malicious Code Automated Run-Time Analysis Framework,” Homeland Security, 2012 IEEE Conference on Technologies for, Nov. 13, 2012, pp. 13-17.
Olanoff, Drew, “Deep Dive with the New Google Maps for Desktop with Google Earth Integration, It's More than Just a Utility,” May 15, 2013, pp. 1-6, retrieved from the internet: http://web.archive.org/web/20130515230641/http://techcrunch.com/2013/05/15/deep-dive-with-the-new-google-maps-for-desktop-with-google-earth-integration-its-more-than-just-a-utility/.
Palantir Technolgies, “Palantir Labs—Timeline,” Oct. 1, 2010, retrieved from the internet https://www.youtube.com/watch?v=JCgDW5bru9M retrieved on Aug. 19, 2015.
Palmas, et al., “An Edge-Bundling Layout for Interactive Parallel Coordinates,” Proceedings of the 2014 IEEE Pacific Visualization Symposium, Mar. 2014, pp. 57-64.
Perdisci et al., “Behavioral Clustering of HTTP-Based Malware and Signature Generation Using Malicious Network Traces,” USENIX, Mar. 18, 2010, pp. 1-14.
Quest, “Toad for Oracle 11.6—Guide to Using Toad,” Sep. 24, 2012, pp. 1-162.
Rouse, Margaret, “OLAP Cube,” http://searchdatamanagement.techtarget.com/definition/OLAP-cube, Apr. 28, 2012, pp. 16.
Shi et al., “A Scalable Implementation of Malware Detection Based on Network Connection Behaviors,” 2013 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, IEEE, Oct. 10, 2013, pp. 59-66.
Sigrist, et al., “PROSITE, a Protein Domain Database for Functional Characterization and Annotation,” Nucleic Acids Research, 2010, vol. 38, pp. D161-D166.
Sirotkin et al., “Chapter 13: The Processing of Biological Sequence Data at NCBI,” The NCBI Handbook, Oct. 2002, pp. 1-11.
Symantec Corporation, “E-Security Begins with Sound Security Policies,” Announcement Symantec, Jun. 14, 2001.
Thompson, Mick, “Getting Started with GEO,” Getting Started with GEO, Jul. 26, 2011.
Umagandhi et al., “Search Query Recommendations Using Hybrid User Profile with Query Logs,” International Journal of Computer Applications, vol. 80, No. 10, Oct. 1, 2013, pp. 7-18.
Vose et al., “Help File for ModelRisk Version 5,” 2007, Vose Software, pp. 349-353. [Uploaded in 2 Parts].
Wikipedia, “Federated Database System,” Sep. 7, 2013, retrieved from the internet on Jan. 27, 2015 http://en.wikipedia.org/w/index.php?title=Federated—database—system&oldid=571954221.
Wright et al., “Palantir Technologies VAST 2010 Challenge Text Records—Investigations into Arms Dealing,” Oct. 29, 2010, pp. 1-10, retrieved from the internet http://hcil2.cs.umd.edu/newvarepository/VAST%20Challenge%202010/challenges/MC1%20-%20Investigations%20into%20Arms%20Dealing/entries/Palantir%20Technologies/ retrieved on Aug. 20, 2015.
Yang et al., “HTML Page Analysis Based on Visual Cues”, A129, pp. 859-864, 2001.
Notice of Acceptance for Australian Patent Application No. 2014250678 dated Oct. 7, 2015.
Notice of Allowance for U.S. Appl. No. 12/556,318 dated Nov. 2, 2015.
Notice of Allowance for U.S. Appl. No. 13/247,987 dated Mar. 17, 2016.
Notice of Allowance for U.S. Appl. No. 14/102,394 dated Aug. 25, 2014.
Notice of Allowance for U.S. Appl. No. 14/108,187 dated Aug. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/135,289 dated Oct. 14, 2014.
Notice of Allowance for U.S. Appl. No. 14/148,568 dated Aug. 26, 2015.
Notice of Allowance for U.S. Appl. No. 14/192,767 dated Dec. 16, 2014.
Notice of Allowance for U.S. Appl. No. 14/192,767 dated Apr. 20, 2015.
Notice of Allowance for U.S. Appl. No. 14/225,084 dated May 4, 2015.
Notice of Allowance for U.S. Appl. No. 14/268,964 dated Dec. 3, 2014.
Notice of Allowance for U.S. Appl. No. 14/294,098 dated Dec. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/320,236 dated Jun. 29, 2016.
Notice of Allowance for U.S. Appl. No. 14/323,935 dated Oct. 1, 2015.
Notice of Allowance for U.S. Appl. No. 14/326,738 dated Nov. 18, 2015.
Notice of Allowance for U.S. Appl. No. 14/473,552 dated Jul. 24, 2015.
Notice of Allowance for U.S. Appl. No. 14/473,860 dated Feb. 27, 2015.
Notice of Allowance for U.S. Appl. No. 14/473,860 dated Jan. 5, 2015.
Notice of Allowance for U.S. Appl. No. 14/504,103 dated May 18, 2015.
Notice of Allowance for U.S. Appl. No. 14/616,080 dated Apr. 2, 2015.
Official Communication for Australian Patent Application No. 2014201511 dated Feb. 27, 2015.
Official Communication for Australian Patent Application No. 2014202442 dated Mar. 19, 2015.
Official Communication for Australian Patent Application No. 2014210604 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014210614 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014213553 dated May 7, 2015.
Official Communication for Australian Patent Application No. 2014250678 dated Jun. 17, 2015.
Official Communication for European Patent Application No. 14158861.6 dated Jun. 16, 2014.
Official Communication for European Patent Application No. 14159464.8 dated Jul. 31, 2014.
Official Communication for European Patent Application No. 14180142.3 dated Feb. 6, 2015.
Official Communication for European Patent Application No. 14180281.9 dated Jan. 26, 2015.
Official Communication for European Patent Application No. 14180321.3 dated Apr. 17, 2015.
Official Communication for European Patent Application No. 14180432.8 dated Jun. 23, 2015.
Official Communication for European Patent Application No. 14186225.0 dated Feb. 13, 2015.
Official Communication for European Patent Application No. 14187739.9 dated Jul. 6, 2015.
Official Communication for European Patent Application No. 14187996.5 dated Feb. 12, 2015.
Official Communication for European Patent Application No. 14189344.6 dated Feb. 20, 2015.
Official Communication for European Patent Application No. 14189347.9 dated Mar. 4, 2015.
Official Communication for European Patent Application No. 14189802.3 dated May 11, 2015.
Official Communication for European Patent Application No. 14191540.5 dated May 27, 2015.
Official Communication for European Patent Application No. 14197879.1 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14197895.7 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14197938.5 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14199182.8 dated Mar. 13, 2015.
Official Communication for European Patent Application No. 15155845.9 dated Oct. 6, 2015.
Official Communication for European Patent Application No. 15155846.7 dated Jul. 8, 2015.
Official Communication for European Patent Application No. 15165244.3 dated Aug. 27, 2015.
Official Communication for European Patent Application No. 15175106.2 dated Nov. 5, 2015.
Official Communication for European Patent Application No. 15175151.8 dated Nov. 25, 2015.
Official Communication for European Patent Application No. 15183721.8 dated Nov. 23, 2015.
Official Communication for Great Britain Patent Application No. 1404457.2 dated Aug. 14, 2014.
Official Communication for Great Britain Patent Application No. 1404574.4 dated Dec. 18, 2014.
Official Communication for Great Britain Patent Application No. 1408025.3 dated Nov. 6, 2014.
Official Communication for Great Britain Patent Application No. 1411984.6 dated Dec. 22, 2014.
Official Communication for Great Britain Patent Application No. 1413935.6 dated Jan. 27, 2015.
Official Communication for Netherlands Patent Application No. 2012437 dated Sep. 18, 2015.
Official Communication for Netherlands Patent Application No. 2013306 dated Apr. 24, 2015.
Official Communication for New Zealand Patent Application No. 622513 dated Apr. 3, 2014.
Official Communication for New Zealand Patent Application No. 622517 dated Apr. 3, 2014.
Official Communication for New Zealand Patent Application No. 624557 dated May 14, 2014.
Official Communication for New Zealand Patent Application No. 627962 dated Aug. 5, 2014.
Official Communication for New Zealand Patent Application No. 628161 dated Aug. 25, 2014.
Official Communication for New Zealand Patent Application No. 628263 dated Aug. 12, 2014.
Official Communication for New Zealand Patent Application No. 628495 dated Aug. 19, 2014.
Official Communication for New Zealand Patent Application No. 628585 dated Aug. 26, 2014.
Official Communication for New Zealand Patent Application No. 628840 dated Aug. 28, 2014.
Official Communication for U.S. Appl. No. 12/556,318 dated Jul. 2, 2015.
Official Communication for U.S. Appl. No. 13/247,987 dated Apr. 2, 2015.
Official Communication for U.S. Appl. No. 13/247,987 dated Sep. 22, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Mar. 4, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Aug. 6, 2015.
Official Communication for U.S. Appl. No. 13/835,688 dated Jun. 17, 2015.
Official Communication for U.S. Appl. No. 13/839,026 dated Aug. 4, 2015.
Official Communication for U.S. Appl. No. 14/102,394 dated Mar. 27, 2014.
Official Communication for U.S. Appl. No. 14/108,187 dated Apr. 17, 2014.
Official Communication for U.S. Appl. No. 14/108,187 dated Mar. 20, 2014.
Official Communication for U.S. Appl. No. 14/134,558 dated Oct. 7, 2015.
Official Communication for U.S. Appl. No. 14/135,289 dated Apr. 16, 2014.
Official Communication for U.S. Appl. No. 14/135,289 dated Jul. 7, 2014.
Official Communication for U.S. Appl. No. 14/148,568 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/148,568 dated Mar. 26, 2015.
Official Communication for U.S. Appl. No. 14/148,568 dated Mar. 27, 2014.
Official Communication for U.S. Appl. No. 14/192,767 dated Sep. 24, 2014.
Official Communication for U.S. Appl. No. 14/192,767 dated May 6, 2014.
Official Communication for U.S. Appl. No. 14/196,814 dated May 5, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 10, 2014.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 2, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Dec. 21, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Feb. 27, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 11, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 2, 2014.
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 20, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 26, 2016.
Official Communication for U.S. Appl. No. 14/225,084 dated Jan. 4, 2016.
Official Communication for U.S. Appl. No. 14/225,160 dated Feb. 11, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Aug. 12, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated May 20, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/225,160 dated Jan. 25, 2016.
Official Communication for U.S. Appl. No. 14/225,160 dated Jul. 29, 2014.
Official Communication for U.S. Appl. No. 14/268,964 dated Jul. 11, 2014.
Official Communication for U.S. Appl. No. 14/268,964 dated Sep. 3, 2014.
Official Communication for U.S. Appl. No. 14/289,596 dated Jul. 18, 2014.
Official Communication for U.S. Appl. No. 14/289,596 dated Jan. 26, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Apr. 30, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Aug. 5, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated Jul. 22, 2014.
Official Communication for U.S. Appl. No. 14/289,599 dated May 29, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated Sep. 4, 2015.
Official Communication for U.S. Appl. No. 14/294,098 dated Aug. 15, 2014.
Official Communication for U.S. Appl. No. 14/294,098 dated Nov. 6, 2014.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 14, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Mar. 17, 2016.
Official Communication for U.S. Appl. No. 14/306,138 dated Feb. 18, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 23, 2014.
Official Communication for U.S. Appl. No. 14/306,138 dated Dec. 24, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Dec. 3, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Feb. 19, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Dec. 24, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Aug. 7, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/306,154 dated Feb. 1, 2016.
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 11, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated May 15, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Nov. 16, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 17, 2016.
Official Communication for U.S. Appl. No. 14/306,154 dated Jul. 6, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 1, 2016.
Official Communication for U.S. Appl. No. 14/319,765 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Jun. 16, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Nov. 25, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 4, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Nov. 28, 2014.
Official Communication for U.S. Appl. No. 14/323,935 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Dec. 2, 2014.
Official Communication for U.S. Appl. No. 14/326,738 dated Jul. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/473,552 dated Feb. 24, 2015.
Official Communication for U.S. Appl. No. 14/473,860 dated Nov. 4, 2014.
Official Communication for U.S. Appl. No. 14/486,991 dated Mar. 10, 2015.
Official Communication for U.S. Appl. No. 14/490,612 dated Aug. 18, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Feb. 5, 2015.
Official Communication for U.S. Appl. No. 14/570,914 dated Sep. 16, 2016.
Official Communication for U.S. Appl. No. 14/570,914 dated Dec. 19, 2016.
Official Communication for U.S. Appl. No. 14/579,752 dated Aug. 19, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/631,633 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated Oct. 16, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated May 18, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated Jul. 24, 2015.
Official Communication for U.S. Appl. No. 14/645,304 dated Jan. 25, 2016.
Official Communication for U.S. Appl. No. 14/726,353 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/813,749 dated Sep. 28, 2015.
Official Communication for U.S. Appl. No. 14/874,690 dated Jun. 1, 2016.
Official Communication for U.S. Appl. No. 14/874,690 dated Dec. 21, 2015.
Official Communication for U.S. Appl. No. 14/948,009 dated Feb. 25, 2016.
Official Communication for U.S. Appl. No. 15/092,456 dated Nov. 4, 2016.
Restriction Requirement for U.S. Appl. No. 13/839,026 dated Apr. 2, 2015.
Janssen, Jan-Keno, “Wo bist'n du?—Googles Geodienst Latitude,” Jan. 17, 2011, pp. 86-88, retrieved from the internet on Jul. 30, 2015 http://www.heise.de/artikel-archiv/ct/2011/03/086/@00250@/ct.11.03.086-088.pdf.
Jelen, Bill, “Excel 2013 in Depth, Video Enhanced Edition,” Jan. 25, 2013.
Map Builder, “Rapid Mashup Development Tool for Google and Yahoo Maps!” <http://web.archive.org/web/20090626224734/http://www.mapbuilder.net/> printed Jul. 20, 2012 in 2 pages.
Psaltis, Andrew G., “Streaming Data—Designing the Real-Time Pipeline,” Jan. 16, 2015, vol. MEAP VO3, pp. 0-12.
Wikipedia, “Mobile Web,” Jan. 23, 2015, retrieved from the internet on Mar. 15, 2016 https://en.wikipedia.org/w/index.php?title=Mobile—Web&oldid=643800164.
Windley, Phillip J., “The Live Web: Building Event-Based Connections in the Cloud,” Dec. 21, 2011, pp. 10, 216.
Notice of Allowance for U.S. Appl. No. 14/570,914 dated Jan. 31, 2017.
Notice of Allowance for U.S. Appl. No. 14/596,552 dated Dec. 23, 2016.
Notice of Allowance for U.S. Appl. No. 14/923,364 dated May 6, 2016.
Official Communication for European Patent Application No. 15166137.8 dated Sep. 14, 2015.
Official Communication for European Patent Application No. 16152984.7 dated Mar. 24, 2016.
Official Communication for U.S. Appl. No. 14/148,559 dated Jun. 16, 2014.
Official Communication for U.S. Appl. No. 14/148,559 dated Apr. 2, 2014.
Official Communication for U.S. Appl. No. 14/306,147 dated Jun. 3, 2016.
Official Communication for U.S. Appl. No. 14/596,552 dated Sep. 23, 2016.
Official Communication for U.S. Appl. No. 14/596,552 dated Oct. 5, 2016.
Official Communication for U.S. Appl. No. 15/092,456 dated Mar. 21, 2017.
Official Communication for U.S. Appl. No. 15/397,562 dated Mar. 14, 2017.
IBM—i2 Integrated Law Enforcement, https://www-03.ibm.com/software/products/en/integrated-law-enforcement, as printed Feb. 15, 2017 in 2 pages.
IBM—i2 Analyze, https://www-03.ibm.com/software/products/en/i2-analyze, as printed Feb. 15, 2017 in 2 pages.
IBM—Data analysis—i2 Analyst's Notebook, http://www-03.ibm.com/software/products/en/analysts-notebook, as printed Feb. 16, 2017 in 2 pages.
Visual Analysis, “Overview of merging timeline charts and creating hybrid charts,” available at https://www.youtube.com/watch?v=dl6jzNtEVpA, as published on Mar. 9, 2015.
IBM Analytics, “IBM i2 Intelligence Analysis Portfolio Overview,” available at https://www.youtube.com/watch?v=EIFu—oUiaBY, as published on Sep. 24, 2015.
i2—An IBM Company, “IBM i2 Intelligent Law Enforcement Demo,” available at https://www.youtube.com/watch?v=—KCXZ2iTMXQ, as published on Dec. 3, 2012.
Yair Shaked, “IBM i2 Enterprise Insight Analysis—cyber Demo,” available at https://www.youtube.com/watch?v=ZXmTWKqkfF4, as published on Nov. 19, 2015.
Visual Analysis, “Overview of importing data and creating timelines,” available at https://www.youtube.com/watch?v=SovxKrvkZZs, as published on Mar. 9, 2015.
IBM Corporation, “IBM i2 Analyst's Notebook,” Aug. 2015, in 4 pages.
IBM Corporation, “IBM i2 Analyst's Notebook Connector for Esri,” May 2012, in 3 pages.
IBM Corporation, “IBM i2 Enterprise Insight Analysis V2.0 delivers a modern contextual user interface and enhanced software operational warehouse support,” http://www-01.ibm.com/common/ssi/ShowDoc.wss?docURL=/common/ssi/rep—ca/2/897/ENUS215-302/index.html&lang=en&request—locale=en, as published on Sep. 1, 2015.
IBM Support, “Software lifecycle—i2 Analyst's Notebook Premium 9.0.0,” https://www-01.ibm.com/software/support/lifecycleapp/PLCDetail.wss?q45=I570331B72886X86, as printed Mar. 7, 2017 in 1 page.
IBM Support, “Software lifecycle—i2 Enterprise Insight Analysis 2.0.0,” https://www-01.ibm.com/software/support/lifecycleapp/PLCDetail.wss?q45=E170786H45496I53, as printed Mar. 7, 2017 in 1 page.
i2 A ChoicePoint Company, “i2 Analyst's Notebook 7 User Guide: Creating Charts” Jun. 2007, 373 pages.
Gatewaynews, “New Crime Fighting Tool ‘Coplink’” available at https://www.youtube.com/watch?v=GbU6E0grnTw, as published on Mar. 8, 2008.
Coplink, “Incident Analyzer User Guide,” created Nov. 5, 2010 (as indicated by the PDF file metadata), 14 pages.
Related Publications (1)
Number Date Country
20170109910 A1 Apr 2017 US
Provisional Applications (1)
Number Date Country
61799691 Mar 2013 US
Continuations (2)
Number Date Country
Parent 14596552 Jan 2015 US
Child 15392624 US
Parent 14148559 Jan 2014 US
Child 14596552 US