Systems and methods for organizing and identifying documents via hierarchies and dimensions of tags

Information

  • Patent Grant
  • 9727560
  • Patent Number
    9,727,560
  • Date Filed
    Wednesday, February 25, 2015
    9 years ago
  • Date Issued
    Tuesday, August 8, 2017
    7 years ago
Abstract
Computer-implemented systems and methods are disclosed to interface with one or more storage devices storing a plurality of documents, wherein each of the plurality of documents is associated with one or more tags of one or more predefined hierarchies of tags, wherein the one or more hierarchies of tags include multiple dimensions. In accordance with some embodiments, a method is provided to identify one or more documents from the data storage devices. The method comprises acquiring, via an interface, a selection of one or more tags of the one or more predefined hierarchies of tags. The method further comprises identifying one or more documents from the data storage devices in response to the selection, the identified one or more documents having tags that have a relationship with the selected tags, and providing data corresponding to the identified documents for displaying in the interface.
Description
BACKGROUND

Data is commonly stored in computer-based systems in fixed, rigidly structured data stores. For example, one common type of data store is a “flat” file such as a spreadsheet, plain-text document, or XML document. Another common type of data store is a relational database comprising one or more tables. Other examples of data stores that comprise structured data include, without limitation, files systems, object collections, record collections, arrays, hierarchical trees, linked lists, stacks, and combinations thereof.


Often, the underlying structure of these types of data stores is poorly suited for data analysis. One approach for facilitating a more efficient analysis of data in such data stores is to reorganize that data according to an object model that defines object structures and relationships between the object structures. Tagging is a method used to create objects, properties, or links between objects and/or properties in structured or unstructured data. It can add structure to unstructured data or add further structure to structured data. An exemplary system and method for tagging is described in detail in U.S. application Ser. No. 14/025,653, filed on Sep. 12, 2013, and titled “Systems and Methods for Providing a Tagging Interface for External Content,” which is incorporated herein by reference in its entirety.





BRIEF DESCRIPTION OF THE DRAWINGS

Reference will now be made to the accompanying drawings showing example embodiments of the present application, and in which:



FIG. 1 is a block diagram of an exemplary computer system with which embodiments described herein can be implemented, consistent with embodiments of the present disclosure.



FIG. 2 is a block diagram depicting an exemplary internal database system, consistent with embodiments of the present disclosure.



FIG. 3 is a chart illustrating an exemplary hierarchical structure of tags, consistent with embodiments of the present disclosure.



FIGS. 4A-4B are charts illustrating an exemplary object model reflecting relationships between tags, consistent with embodiments of the present disclosure.



FIG. 5 is a chart illustrating an exemplary object model reflecting relationships between combinations of tags of the exemplary hierarchical structure of tags depicted in FIG. 3, consistent with embodiments of the present disclosure.



FIGS. 6A-6B are screenshots depicting an exemplary interface for selecting one or more tags to identify a document, consistent with embodiments of the present disclosure.



FIGS. 7A-7E are screenshots depicting exemplary interfaces for identifying and displaying documents based on tags, consistent with embodiments of the present disclosure.



FIGS. 8A-8B are screenshots depicting an exemplary interface for identifying and displaying documents based on tags from previously identified documents, consistent with embodiments of the present disclosure.



FIG. 9 is a flowchart representing an exemplary method performed by an electronic device for identifying documents based on tags, consistent with embodiments of the present disclosure.





DETAILED DESCRIPTION OF EXEMPLARY EMBODIMENTS

Reference will now be made in detail to the embodiments, the examples of which are illustrated in the accompanying drawings. Whenever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts.


Embodiments of the present disclosure provide a means to organize and access data structured with tag objects (e.g. by associating a portion or part or entirety of the data with tags) by providing a pre-defined hierarchy of tags. As an exemplary illustration, the hierarchy can include one or more dimensions, each dimension comprising a set of tags corresponding to that dimension. The pre-defined hierarchy of tags facilitates tag-based identification and retrieval of the data associated with one or more selected tags that are part of the hierarchy, which can allow a user of the system to navigate through a very large data sets to identify appropriate data or documents associated with or related to the one or more selected tags.


Embodiments of the present disclosure further provide an interface allowing the user to navigate through very large data sets to identify and display appropriate data or documents associated with or related to the one or more selected tags. Via the interface, a user can input a selection of tags and retrieve a document associated with the tags selection, as well as other documents that are related to the tags selection. The interface also updates the tags selection based on a document retrieved by the user, allowing the user to identify other related documents. The interface further facilitates a user's navigation through a very large data sets to identify appropriate data or documents associated with or related to the one or more selected tags.


The tag objects can include one or more attributes, and a relationship can be defined between the attributes of each tag object (or combinations thereof). As an exemplary illustration, the tag object can include attributes including a tag label, a tag type, and one or more properties. Moreover, based on these attributes, one or more relationships between tags can be defined.


After the one or more tags are selected in the interface, data associated with those tags can be acquired. Moreover, one or more other tags related to the selected tags can be identified, which can allow data associated with the one or more other tags to also be acquired. This further facilitates tag-based identification and retrieval of the data associated with tags that are part of the hierarchy by, for example, allowing the user to navigate within a huge universe of data structured with tags, guided by the pre-defined hierarchy of tags, as well as the pre-defined relationship between the tags in the hierarchy.


According to some embodiments, the operations, techniques, and/or components described herein can be implemented by an electronic device, which can include one or more special-purpose computing devices. The special-purpose computing devices can be hard-wired to perform the operations, techniques, and/or components described herein, or can include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the operations, techniques and/or components described herein, or can include one or more hardware processors programmed to perform such features of the present disclosure pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices can also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the technique and other features of the present disclosure. The special-purpose computing devices can be desktop computer systems, portable computer systems, handheld devices, networking devices, or any other device that incorporates hard-wired and/or program logic to implement the techniques and other features of the present disclosure.


The one or more special-purpose computing devices can be generally controlled and coordinated by operating system software, such as iOS, Android, Blackberry, Chrome OS, Windows XP, Windows Vista, Windows 7, Windows 8, Windows Server, Windows CE, Unix, Linux, SunOS, Solaris, VxWorks, or other compatible operating systems. In other embodiments, the computing device can be controlled by a proprietary operating system. Operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface functionality, such as a graphical user interface (“GUI”), among other things.



FIG. 1 is a block diagram of an exemplary computer system 100 with which embodiments described herein can be implemented, consistent with embodiments of the present disclosure. Computer system 100 includes a bus 102 or other communication mechanism for communicating information, and one or more hardware processors 104 (denoted as processor 104 for purposes of simplicity) coupled with bus 102 for processing information. Hardware processor 104 can be, for example, one or microprocessors.


Computer system 100 also includes a main memory 106, such as a random access memory (RAM) or other dynamic storage device, coupled to bus 102 for storing information and instructions to be executed by processor 104. Main memory 106 also can be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 104. Such instructions, after being stored in non-transitory storage media accessible to processor 104, render computer system 100 into a special-purpose machine that is customized to perform the operations specified in the instructions.


Computer system 100 further includes a read only memory (ROM) 108 or other static storage device coupled to bus 102 for storing static information and instructions for processor 104. A storage device 110, such as a magnetic disk, optical disk, or USB thumb drive (Flash drive), etc., is provided and coupled to bus 102 for storing information and instructions.


Computer system 100 can be coupled via bus 102 to a display 112, such as a cathode ray tube (CRT), an liquid crystal display (LCD), or a touch screen, for displaying information to a computer user. An input device 114, including alphanumeric and other keys, is coupled to bus 102 for communicating information and command selections to processor 104. Another type of user input device is cursor control 116, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 104 and for controlling cursor movement on display 112. The input device typically has two degrees of freedom in two axes, a first axis (for example, x) and a second axis (for example, y), that allows the device to specify positions in a plane. In some embodiments, the same direction information and command selections as cursor control may be implemented via receiving touches on a touch screen without a cursor.


Computing system 100 can include a user interface module to implement a graphical user interface (GUI) that can be stored in a mass storage device as executable software codes that are executed by the one or more computing devices. This and other modules can include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, fields, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.


In general, the word “module,” as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, Lua, C or C++. A software module can be compiled and linked into an executable program, installed in a dynamic link library, or written in an interpreted programming language such as, for example, BASIC, Perl, or Python. It will be appreciated that software modules can be callable from other modules or from themselves, and/or can be invoked in response to detected events or interrupts. Software modules configured for execution on computing devices can be provided on a computer readable medium, such as a compact disc, digital video disc, flash drive, magnetic disc, or any other tangible medium, or as a digital download (and can be originally stored in a compressed or installable format that requires installation, decompression, or decryption prior to execution). Such software code can be stored, partially or fully, on a memory device of the executing computing device, for execution by the computing device. Software instructions can be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules can be comprised of connected logic units, such as gates and flip-flops, and/or can be comprised of programmable units, such as programmable gate arrays or processors. The modules or computing device functionality described herein are preferably implemented as software modules, but can be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that can be combined with other modules or divided into sub-modules despite their physical organization or storage.


Computer system 100 can implement the techniques described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the computer system causes or programs computer system 100 to be a special-purpose machine. According to some embodiments, the operations, functionalities, and techniques and other features described herein are performed by computer system 100 in response to processor 104 executing one or more sequences of one or more instructions contained in main memory 106. Such instructions can be read into main memory 106 from another storage medium, such as storage device 110. Execution of the sequences of instructions contained in main memory 106 causes processor 104 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry can be used in place of or in combination with software instructions.


The term “non-transitory media” as used herein refers to any non-transitory media storing data and/or instructions that cause a machine to operate in a specific fashion. Such non-transitory media can comprise non-volatile media and/or volatile media. Non-volatile media can include, for example, optical or magnetic disks, such as storage device 110. Volatile media can include dynamic memory, such as main memory 106. Common forms of non-transitory media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge, and networked versions of the same.


Non-transitory media is distinct from, but can be used in conjunction with, transmission media. Transmission media can participate in transferring information between storage media. For example, transmission media can include coaxial cables, copper wire and fiber optics, including the wires that comprise bus 102. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.


Various forms of media can be involved in carrying one or more sequences of one or more instructions to processor 104 for execution. For example, the instructions can initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 100 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 102. Bus 102 carries the data to main memory 106, from which processor 104 retrieves and executes the instructions. The instructions received by main memory 106 can optionally be stored on storage device 110 either before or after execution by processor 104.


Computer system 100 can also include a communication interface 118 coupled to bus 102. Communication interface 118 can provide a two-way data communication coupling to a network link 120 that can be connected to a local network 122. For example, communication interface 118 can be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 118 can be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links can also be implemented. In any such implementation, communication interface 118 can send and receive electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.


Network link 120 can typically provide data communication through one or more networks to other data devices. For example, network link 120 can provide a connection through local network 122 to a host computer 124 or to data equipment operated by an Internet Service Provider (ISP) 126. ISP 126 in turn can provide data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 128. Local network 122 and Internet 128 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 120 and through communication interface 118, which carry the digital data to and from computer system 100, can be example forms of transmission media.


Computer system 100 can send messages and receive data, including program code, through the network(s), network link 120 and communication interface 118. In the Internet example, a server 130 can transmit a requested code for an application program through Internet 128, ISP 126, local network 122 and communication interface 118.


The received code can be executed by processor 104 as it is received, and/or stored in storage device 110, or other non-volatile storage for later execution. In some embodiments, server 130 can provide information for being displayed on a display.



FIG. 2 is a block diagram depicting an exemplary internal database system 200, consistent with embodiments of the present disclosure. Among other things, system 200 facilitates transformation of one or more data sources, such as data sources 230, into an object model 260, whose semantics are defined by an ontology 250. The transformation can be performed for a variety of reasons. For example, a database administrator can wish to import data from data sources 230 into a database 270 for persistently storing object model 260. As another example, a data presentation component (not depicted) can transform input data from data sources 230 “on the fly” into object model 260. Object model 260 can then be utilized, in conjunction with ontology 250, for analysis through graphs and/or other data visualization techniques.


System 200 comprises a definition component 210 and a transformation component 220, both implemented by one or more processors on one or more computing devices executing hardware and/or software-based logic for providing various functionality described herein. As will be appreciated from the present disclosure, system 200 can comprise fewer or additional components that provide various functionalities described herein. Such components are, for clarity, omitted from FIG. 1. Moreover, the component(s) of system 200 responsible for providing various functionalities can further vary from embodiment to embodiment.


Definition component 210 generates and/or modifies ontology 250 and a schema map 240. Exemplary embodiments for defining an ontology (such as ontology 250) is described in U.S. Pat. No. 7,962,495 (the '495 patent), issued Jun. 14, 2011, the entire contents of which are expressly incorporated herein by reference for all purposes. Among other things, the '495 patent describes embodiments that define a dynamic ontology for use in creating data in a database. For creating a database ontology, one or more object types are created where each object type can include one or more properties. The attributes of object types or property types of the ontology can be edited or modified at any time.


In some embodiments, each property type is declared to be representative of one or more object types. A property type is representative of an object type when the property type is intuitively associated with the object type. For example, a property type of “geographical location” may be representative of an object type “locale” but not representative of an object type “style.”


Schema map 240 can define how various elements of schemas 235 for data sources 230 map to various elements of ontology 250. Definition component 210 receives, calculates, extracts, or otherwise identifies schemas 235 for data sources 230. Schemas 235 define the structure of data sources 230—for example, the names and other characteristics of tables, files, columns, fields, properties, and so forth. Definition component 210 furthermore optionally identifies sample data 236 from data sources 230. Definition component 210 can further identify object type, relationship, and property definitions from ontology 250, if any already exist. Definition component 210 can further identify pre-existing mappings from schema map 240, if such mappings exist.


Transformation component 220 can be invoked after schema map 140 and ontology 250 have been defined or redefined. Transformation component 220 identifies schema map 240 and ontology 250. Transformation component 120 further reads data sources 230 and identifies schemas 235 for data sources 230. For each element of ontology 250 described in schema map 240, transformation component 220 iterates through some or all of the data items of data sources 230, generating elements of object model 260 in the manner specified by schema map 240. In some embodiments, transformation component 220 can store a representation of each generated element of object model 260 in a database 270. In some embodiments, transformation component 220 is further configured to synchronize changes in object model 160 back to data sources 230.


Data sources 230 can be one or more sources of data, including, without limitation, spreadsheet files, databases, email folders, document collections, media collections, contact directories, and so forth. Data sources 230 can include structured data (e.g., a database, a .csv file, or any tab delimited or fixed-width file), semi-structured data (e.g., an email, an email server, or forms such as a suspicious activity report or currency transaction report), or unstructured data (e.g., encoded files such as PDF, sound, and image files). Data sources 230 can include data structures stored persistently in non-volatile memory. Data sources 230 can also or instead include temporary data structures generated from underlying data sources via data extraction components, such as a result set returned from a database server executing an database query.


Schema map 240, ontology 250, and schemas 235 can be stored in any suitable data structures, such as XML files, database tables, and so forth. In some embodiments, ontology 250 is maintained persistently. Schema map 240 can or cannot be maintained persistently, depending on whether the transformation process is perpetual or a one-time event. Schemas 235 need not be maintained in persistent memory, but can be cached for optimization.


Object model 260 comprises collections of elements such as typed objects, properties, and relationships. The collections can be structured in any suitable manner. In some embodiments, a database 270 stores the elements of object model 260, or representations thereof. In some embodiments, the elements of object model 260 are stored within database 270 in a different underlying format, such as in a series of object, property, and relationship tables in a relational database


Based on the identified information, definition component 210 can generate a graphical interface 215. Graphical interface 215 can be presented to users of a computing device via any suitable output mechanism (e.g., a display screen, an image projection, etc.), and can further accept input from users of the computing device via any suitable input mechanism (e.g., a keyboard, a mouse, a touch screen interface). Graphical interface 215 may feature a visual workspace that visually depicts representations of the elements of ontology 250 for which mappings are defined in schema map 240. Graphical interface 215 can further utilize the sample data 236 to provide the user with a preview of object model 260 as the user defines schema map 240. In response to the input via the various controls of graphical interface 215, definition component 210 can generate and/or modify ontology 250 and schema map 240, and/or identify object models and sample data schemas 235 and data sources 230.


In some embodiments, graphical interface 215 also provides a user with the ability to add structure to an unstructured document stored in data sources 230 by tagging one or more portions (e.g., text) within the document. Defining tags and applying these tags to a portion of the document can create tag objects, properties, or links creating a relationship between one or more tag objects and/or properties. In some embodiments, graphical interface 215 allows a user to input one or more pre-defined tags to retrieve a tagged document, and/or a set of related documents that are associated with other pre-defined tags which are different from, but have a relationship to, the one or more input pre-defined tags. In some embodiments, graphical interface 215 also displays to the user the tags associated with those related documents, and the user can use those tags to identify another tagged documents, and/or another set of related documents, thereby allowing the user to “move” between the documents stored in data sources 230 guided by the relationships between the pre-defined tags.



FIG. 3 is a chart 300 illustrating an exemplary hierarchical structure of tags 310 (“tag hierarchy”), consistent with embodiments of the present disclosure. In some embodiments, the exemplary tag hierarchy in FIG. 3 can provide part of the structure of object model 260 stored within database 270 in FIG. 2. Tag hierarchy 310 can include dimensions 320. For example, as shown in tag hierarchy 310, these dimensions include a locale dimension 340, a subject matter dimension 350, a medium dimension 360, and a style dimension 370. Each dimension includes a set of tags, which includes one or more tags linked to that dimension. In some embodiments, tags are created as objects with attributes, and these links can be established based on the attributes of the tags. The tags constitute a group of tags 330.


As shown in FIG. 3, locale dimension 340 can include a set of tags having attributes related to regions and countries. As an exemplary illustration, under the locale dimension 340, there is a United States tag 341 and a Germany tag 342. There can also be a further subset of tags (e.g., California tag 343 and Texas tag 344) under United States tag 341, where California tag 343 and Texas tag 344 have attributes indicating that they are associated with United States (e.g. being a state of the United States), which can allow California tag 343 and Texas tag 344 to be linked to United States tag 341. Similarly, California tag 343 can also have a further subset of tags (e.g. Palo Alto tag 345).


The relationship between a tag and any corresponding subset of tags can be based on attributes in that tag and in the corresponding subset of tags. For example, Palo Alto tag 345 has attributes indicating that it is associated with California (e.g., a city of the state of California), which can allow Palo Alto tag 345 to be linked to California tag 343.


Under tag hierarchy 310, subject matter dimension 350 can include tags with attributes related to a classification based on content. As an exemplary illustration, under subject matter dimension 350, there are scenery tag 351 and living tag 352, where scenery tag 351 has attributes indicating that the content is related to scenery (e.g. depicting or describing a scene), while living tag 352 has attributes indicating that the content is related to a living thing (e.g. depicting or describing a living organism, such as human). Scenery tag 351 can have a further subset of tags (e.g. architecture tag 353), where architecture tag 353 has attributes indicating that the content is related to architecture (e.g. depicting or describing buildings), which can allow architecture tag 353 to be linked with scenery tag 351. Similarly, living tag 352 can have a further subset of tags (e.g. people tag 354), where people tag 354 has attributes indicating that the content is related to a human (e.g. a portrait), which can allow tag 354 to be linked with living tag 352.


Under tag hierarchy 310, medium dimension 360 can include tags with attributes related to a classification based on a medium on which the content is rendered. As an exemplary illustration, there are paper tag 361 and film tag 362 under medium dimension 360. Furthermore, style dimension 370 can also include tags with attributes related to a classification based on a style of rendering the content. As an exemplary illustration, there are classical tag 371 and modern tag 372 under the style dimension 370. A person with ordinary skill in the art will understand that the dimensions and tags depicted in FIG. 3 are for illustration purposes only, and there is no limitation on the number of dimensions, how dimensions are defined, and how the tags are organized under each dimension.



FIG. 4A shows, in a chart 400, an exemplary object model reflecting relationships between tags, consistent with embodiments of the present disclosure. In chart 400, each circle represents a cell, and each line represents a relationship between cells. In some embodiments, a cell within the object model can be associated with one or more pre-defined tags, and the relationship between the cells can be defined based on a relationship between the attributes of the tags associated with the cells. Each of the cells can also be associated with the tagged documents stored in data sources 230 of FIG. 2 via, for example, common tags or related tags associated with both the cells and the tagged documents. A document can also be associated with one or more of the cells, if the document is tagged with multiple sets of tags that are associated with multiple cells.


Chart 400 also includes a sub-chart 410 which includes an exemplary subset of cells and relationships of the object model. FIG. 4B shows a close-up view of sub-chart 410. Sub-chart 410 illustrates cells 420, 430, 440, 450, 460, and 470, as well as relationships 425, 435, 445, 455, 465, 475, and 485. As an exemplary illustration, cell 420 can be associated with United States tag 341, cell 430 can be associated with California tag 343, cell 440 can be associated with Palo Alto tag 345, cell 450 can be associated with Germany tag 342, cell 460 can be associated with a Japan tag (not shown in tag hierarchy 310 of FIG. 3), and cell 470 can be associated with a Tokyo tag (not shown in tag hierarchy 310 of FIG. 3). Among these cells, cell 420 (with the United States tag), cell 450 (with the Germany tag), and cell 460 (with the Japan tag) can have relationship 455, 465, and 475 between each other by virtue of, for example, that the United States tag, the Germany tag, and the Japan tag all have attributes related to an indication of a country with a developed economy.


Cell 430 (with the California tag) has relationship 425 with cell 420 (with the United States tag) by virtue of, for example, that the California tag has attributes that link it to United States tag 341 (e.g. California being a state of United States), the link to which can allow the California tag to be related to the United States tag. Furthermore, cell 440 (with the Palo Alto tag) can also have relationship 435 with cell 430 by virtue of, for example, that the Palo Alto tag includes attributes that link it to California tag 343 (e.g. Palo Alto being a city of California), the link to which can allow the Palo Alto tag to be related to the California tag. Palo Alto tag 345 can also include attributes that link it to United States tag 341 (e.g. Palo Alto being a city of United States), the link to which can allow cell 440 to also have the relationship 445 with cell 420.


On the other hand, cell 460 (with the Japan tag) can have a relationship 485 with cell 470 (with the Tokyo tag) by virtue of, for example, that the Tokyo tag has attributes that link it to the Japan tag (e.g. Tokyo being a city of Japan), the link to which can allow the Tokyo tag to be related to the Japan tag. But in this exemplary illustration, the Tokyo tag may have no relationship with the Germany tag, the United States tag, the California tag, or the Palo Alto tag, therefore cell 470 may have no relationship with cells 420, 430, 440, or 450 within sub-chart 410.



FIG. 5 is a chart 500 illustrating an exemplary object model reflecting relationships between combinations of tags of the exemplary hierarchical structure of tags depicted in FIG. 3, consistent with embodiments of the present disclosure. In some embodiments, the object model shown in chart 500 includes cells 510, 530, 550, 570, and 590, each of which can be, respectively, associated with tag combinations 512, 532, 552, 572, and 592. Each tag combination includes one or more tags for each of its dimensions, which include, for example, locale dimension 340, subject matter dimension 350, medium dimension 360, and style dimension 370 as depicted in FIG. 3. The tag combination can include tags of tag hierarchy 310 as depicted in FIG. 3, and can include one or more tags for each dimension as depicted in FIG. 3. Each of these cells can also be associated with the tagged documents stored in data sources 230 of FIG. 2 via the tags. The object model shown in chart 500 also includes relationships 520, 540, 545, 565, 568, and 585 between the cells. As to be illustrated below, these relationships can be determined based on the relationship between tags within one or more dimensions.


As an exemplary illustration, cell 510 is associated with tag combination 512, which includes United States tag 341 under the locale dimension and scenery tag 351 under the subject matter dimension. Cell 530 is associated with tag combination 532, which includes California tag 343 under the locale dimension, scenery tag 351 under the subject matter dimension, paper tag 361 under the medium dimension, and classical tag 371 under the style dimension. Cell 570 is associated with tag combination 572. Tag combination 572 is otherwise identical to tag combination 532 except that tag combination 572 has Palo Alto tag 345 instead of California tag 343 under the locale dimension. Moreover, cell 550 is associated with tag combination 552, which includes Texas tag 344 under the locale dimension, scenery tag 351 under the subject matter dimension, paper tag 361 under the medium dimension, and modern tag 372 under the style dimension. Lastly, cell 590 is associated with tag combination 592. Tag combination 592 is otherwise identical to tag combination 552, except that tag combination 592 has Germany tag 342 instead of Texas tag 344 under the locale dimension.


Relationship 520 between cell 510 and cell 530 can be determined based on, for example, a relationship between the United States tag (associated with cell 510) and the California tag (associated with cell 530) under the locale dimension. A relationship 540 between cell 510 and cell 550 can also be determined based on, for example, a relationship between the Texas tag (associated with cell 550) and the United States tag under the locale dimension. Furthermore, relationship 545 between cell 530 and cell 550 can also be determined based on, for example, both the relationship between the California tag and the Texas tag under the locale dimension, as well as the relationship between the paper tag (associated with cell 530) and the film tag (associated with cell 550) under the medium dimension. In this particular example, because both cells 510 and 530 have scenery tag 351 for the subject matter dimension, the subject matter dimension can be ignored in determining relationship 520. Also, because cell 510 does not have tags for the medium and style dimensions, these dimensions can also be ignored in determining relationships 520 and 540.


As discussed before, the relationship between tags can be determined based, for example, the attributes of the tags. In addition, relationship between tags can also be established in other ways. For example, tags can become related to each other when both tags are associated with a document, with documents that have related metadata, or with a cell. Furthermore, relationship between tags can also be created manually according to any pre-defined condition.


In some embodiments, each tag combination in FIG. 5 can be represented as a multi-dimensional vector, with each dimension of tag hierarchy 310 represented by a vector dimension, and a combination of one or more tags under a dimension of tag hierarchy 310 contributes to a magnitude of the vector along that vector dimension, based on the attributes of the tags. The relationship between tags can then be calculated as, for example, an imaginary distance between the multi-dimensional vectors representing the tag combinations. In some embodiments, such imaginary distance can be calculated by first projecting the multi-dimensional vectors representing the tag combinations onto a pre-defined plane, and then calculating a distance between the projections on the pre-defined plane. In some embodiments, when the calculated distance exceeds a certain threshold, it can be determined that no relationship exists between the tag combinations. In some embodiments, the relationship between cells (or between tag combinations associated with the cells) can also be added manually with or without considering the calculated distance.


Referring back to FIG. 5, cell 570 is associated with tag combination 572, which includes Palo Alto tag 345 under the locale dimension, and the locale dimension is the only dimension with different tags when compared with tag combination 532 associated with cell 530. Relationship 565 can then be determined based on, for example, the relationship between California tag 343 and Palo Alto tag 345 under the locale dimension alone. Similarly, cell 590 is associated with tag combination 592, which includes Germany tag 342 under the locale dimension, and the locale dimension is the only dimension with different tags when compared with the tag combination 552 associated with cell 550. Relationship 585 can then be determined based on, for example, the relationship between Germany tag 342 and Texas tag 344 under the locale dimension alone.


Relationship 568 between cells 570 and 590 can also be determined based on, for example, both the relationship between the Palo Alto tag (associated with cell 570) and the Germany tag (associated with cell 590) under the locale dimension, as well as the relationship between the classical tag (associated with cell 570) and the modern tag (associated with cell 590) under the medium dimension. In some embodiments, as discussed above, each of the cells in object model 500 can be associated with documents stored in data sources 230 that are tagged with the same tags associated with each cell, and relationship 568 can be established by, for example, that a document stored in data sources 230 describes a Germany film derived from a Palo Alto novel, and therefore is tagged with tags including, for example, Germany tag 342, Palo Alto tag 345, paper tag 361, and film tag 362, etc., notwithstanding any calculated distance between these tags.



FIG. 6A and FIG. 6B are screenshots depicting an exemplary interface 600 for selecting one or more tags to identify a document, consistent with embodiments of the present disclosure. In some embodiments, the exemplary interface can be provided by an application. The application can be a web browser such as, for example, Google™ Chrome™, Mozilla™ Firefox™, Microsoft™ Internet Explorer™, etc.


In some embodiments, a bookmarklet is installed in the web browser. A bookmarklet can be a bookmark that is stored in a web browser and can contain JavaScript™ commands to extend the web browser's functionality. That is, a bookmarklet can be a simple “one-click” tool that can add functionality to the web browser. For example, a bookmarklet can modify the appearance of a web page within the web browser by changing the font size or the background color of the text, and/or extract data from a web page.


In some embodiments, a plug-in, instead of a bookmarklet, can be installed. A plug-in can be implemented as a set of software components that adds specific abilities to a larger software application, like a web browser, to enable customizing the functionality of the software application. For example, a plug-in can be installed in a web browser to enable the web browser to play video.


In some embodiments, the exemplary interface can be provided by a client-side application. All the exemplary interfaces discussed below can take in any form, such as being displayed as a pop-up window.


Referring back to FIG. 6A, interface 600 includes a locale field 602 for the locale dimension, a subject matter field 604 for the subject matter dimension, a medium field 606 for the medium dimension, and a style field 608 for the style dimension. Each of these fields can receive one or more tags as input to identify one or more relevant documents through interface 600, and can also display one or more tags as output through interface 600.


Fields 602, 604, 606, and 608 can receive input via any means. For example, interface 600 can allow a user to type in the tags or, in some embodiments as shown in FIG. 6B, further provides a pull-down menu 622 from which the user can choose one or more tags. In some embodiments, the field can also receive an incomplete text input, and then provide a list of suggested tags for the user to choose from. The list of suggested tags may include pre-defined tags that closely match the incomplete text input. In some embodiments, instead of providing a field for each dimension, interface 600 can provide a single field for tag selection for all dimensions, and the user can either type in a combination of tags into the single field, or select the tags from a pull-down menu provided by the single field.


In some embodiments, a search field 610 is also provided, allowing the user to search for documents based on text, rather than tags. After receiving the tags or the search text input, user interface 600 corresponds with object model 260 and/or database 270 to search for or identify the documents. In some embodiments, the user is provided an option to select, by clicking on button 612, to explore the result presented in a graphical map similar to chart 400 of FIG. 4A, where the graphical map can show one or more icons with links between them. In some embodiments, each icon in the graphical map represents a document and is selectable, and a selection of the icon can trigger a selection and a display of the document represented by the icon, while the link represents relationships between the tags associated with the documents represented by the icons.


In some embodiments, the user is provided an option to select, by clicking on button 614, to list the search result. The listing of search result will be discussed later.



FIG. 7A is a screenshot depicting an exemplary interface 700 for identifying and displaying documents based on tags, consistent with embodiments of the present disclosure. Based on one or more tags received in, for example, interface 600 of FIG. 6A, one or more documents (in this case, document 702) can be identified and displayed by virtue of the fact that, for example, document 702 is associated with a cell that is associated with the received tags.


Interface 700 may include fields 602, 604, 606, and 608 of interface 600 to display the tags received. In this exemplary illustration, California tag 343 is input under the locale dimension with field 602, scenery tag 351 is input under the subject matter dimension with field 604, and modern tag 372 is input under the style dimension with field 608, while no tag is input for the medium dimension. Document 702 titled “California Impressionism” can then be identified and displayed through interface 700 in response to the California tag, the scenery tag, and the modern tag input by virtue of, for example, document 702 being associated with a cell that is associated with these tags.


In some embodiments, interface 700 can also provide a means to access other documents related to document 702 or related to the tags selected. As shown in FIG. 7A, interface 700 provides a related-overview button 704, a linked-documents button 706, and a related-documents button 708.


In an exemplary illustration, after clicking on the related-overviews button, a pull-down menu 710 can be displayed, which includes options including US art market, US film overview, and US photography overview. FIG. 7B is a screenshot depicting that a document 712 titled “United States Art Market” is identified and displayed when the “US Art Market option” of pull-down menu 710 is selected. As shown in FIG. 7B, document 712 is associated with United States tag 341, which is a hierarchical superset of California tag 343, and scenery tag 351. The related-overviews option can allow the user to identify documents that are relatively more closely related to document 702 of FIG. 7A. The closer relationship can be determined base on, for example, that document 712 is associated with a tag (United States tag 341) that is within the same dimension (locale dimension 340) as one of the tags associated with document 702 (California tag 343), or that a distance between documents 702 and 712 is below a certain threshold, as indicated by the fact that they are both associated with scenery tag 351.


In another exemplary illustration, as shown in FIG. 7C, after clicking on the related-documents button 708, a pull-down menu 714 can be displayed, which includes an option “Introduction to World Art.” In some embodiments, related-documents button 708 can also provide access to documents that are more broadly related to document 702 of FIG. 7A. For example, as shown in FIG. 7C, a document 716 titled “Introduction to World Art” is identified and displayed when the “Introduction to World Art” option of pull-down menu 714 is selected. As shown in FIG. 7C, document 716 is associated with scenery tag 351 and living tag 352 under the subject matter dimension, and is also associated with paper tag 361 and film tag 362 under the style dimension. Document 716 can be determined to be more broadly related to document 702 of FIG. 7A based on, for example, that while documents 716 and 702 are both associated with scenery tag 351 under the subject matter dimension, document 716 is associated with tags that are not associated with document 702 within the same dimension (e.g., living tag 352). The determination can also be based on that document 716 is associated with one or more tags of a specific dimension, while document 702 is not associated with any tag from that specific dimension (e.g., paper tag 361 and film tag 362 of the medium dimension). Therefore, related-documents button 708 allows a user to access documents across more dimensions and tags than related-overview button 704.


In another exemplary illustration, as shown in FIG. 7D, after clicking on the linked-documents button, a pull-down menu 718 can be displayed. In some embodiments, linked documents button 708 can provide access to documents associated with tags that have a lateral relationship with the tags selected. For example, referring to FIG. 3, the California tag 343 and the Germany tag 342 can be lateral to each other within tag hierarchy 310.


According to FIG. 7D, a document 720 titled “Architecture of Germany” is identified and displayed, when Architecture of Germany option of pull-down menu 718 is selected. As shown in FIG. 7D, document 720 is associated with Germany tag 342 and architecture tag 353. In this example, document 720 also has a California tag 343 because document 720 discusses about some of the landmarks in Germany are designed by architects from California, as shown in paragraph 722. The linked-documents option thus can also allow the user to identify documents associated with at least a tag (e.g. Germany tag 342) that has a lateral relationship with any one of the selected tags (e.g. California tag 343).



FIG. 7E is a screenshot depicting an exemplary interface 750 for identifying and displaying documents based on tags, consistent with embodiments of the present disclosure. Interface 750 includes a search interface 752 which can allow a search and display of one or more documents based on tags. In some embodiments, search interface 752 can be activated by, for example, clicking on button 614 of interface 600 as depicted in FIG. 6A to list the search result. In some embodiments, search interface 752 includes a locale field 754 for the locale dimension, a subject matter field 756 for the subject matter dimension, a medium field 758 for the medium dimension, and an style field 760 for the style dimension, which can allow the user to specify tags under each dimension for the search. In some embodiments, interface 750 may further include fields 602, 604, 606, and 608 of interface 600, and the fields 754, 756, 758, and 760 of the search interface 752 can be synchronized with, respectively, fields 602, 604, 606, and 608. In this exemplary illustration, United States tag 341 is input for the locale dimension, and scenery tag 351 is input for the subject matter dimension. Both fields 754 and 756 of search interface 752 can then display the same tags as, respectively, fields 602 and 604 of interface 750. A search for documents that are associated with a combination of tags input through fields 754, 756, 758, and 760 can then be performed, after clicking on the “search” button 762. The search interface 752 may also allow the user to provide additional search conditions, such as limiting to the search result to, for example, a start date and an end date provided through input fields 764 and 766. The user can clear the search conditions (e.g. tags and start/end date) by clicking on the “clear” button 768. After the search is performed, search result 770 is displayed. In this exemplary embodiment, search result 770 displays metadata such as the file type and the title of the documents found. The user can also select a document from the search result 770, which can lead to the displaying of document 772.



FIG. 8A is a screenshot depicting an exemplary interface 800 for identifying and displaying documents based on tags from previously identified documents, consistent with embodiments of the present disclosure. Interface 800 includes fields 602-608 of interface 600. In this example, United States tag 341 is input for the locale dimension with field 602, scenery tag 351 is input for the subject matter dimension with field 604, paper tag 361 is input for the medium dimension with field 606, and classical tag 371 is input for the style dimension with field 608. Thus, in this illustration, a tag combination identical to tag combination 532 of FIG. 5 is input for the search. Interface 800 also includes a search results interface 820, which displays search results 821-827. Search results 821-827 can show a list of, for example, documents that are found based on the selected tags, with metadata for each document, such as title 830, date 832, and author 834. Each of the documents in the search results can be selected, with additional information of the selected document displayed, such as classification 842, and tag combination 847.


In this exemplary illustration, a document titled “Exhibition of Expressionism in Germany and France at Houston Art Museum” is chosen, and is displayed as document 850. Selected document 850 is tagged with, for example, tag combination 847, which is identical to tag combination 552 of FIG. 5, and which includes Texas tag 344, scenery tag 351, paper tag 361, and modern tag 372. Referring to FIG. 5, selected document 850 can be associated with cell 550, by virtue of having tag combination 847 which is identical to tag combination 552, and cell 550 has relationship 545 with cell 530 that is associated with tag combination 532 which is identical to the tag combination input for this search. In some embodiments, interface 800 may also allow the user to add or modify the tags associated with the chosen document. For example, the user can remove classical tag 371 from the document, or tag the document with other tags under the style dimension.



FIG. 8B is another screenshot depicting exemplary interface 800. After the selection of document 850, which is tagged with tag combination 847 (which is identical to tag combination 552), fields 602-608 can be populated with the tags of tag combination 847. In this exemplary illustration, the locale dimension, which has California tag 343 when the prior search is performed, can be populated with Texas tag 344 from selected document 850. Moreover, the style dimension, which has classical tag 371 when the prior search is performed, can be populated with modern tag 372, also from selected document 850. The user can then perform a new search, and an updated search results 850 is shown, which includes search results 851-854. Referring to FIG. 5, some of the documents in search result 850 may be associated with a cell that is related to cell 550 associated with tag combination 552 (which is identical to tag combination 847), such as cell 590, cell 510, and cell 530, etc. This can allow the user to begin with an initial group of tags to identify one or more documents related to the initial set of tags, and then receive additional or new sets of tags from the identified documents. The additional or new sets of tags can then be used to refine the user's exploration in the universe of documents stored in data sources 230, and the refinement can be guided by the predefined relationship between the tags, which can determine the set of related documents provided for a given set of tags.



FIG. 9 is a flowchart representing an exemplary method 900 performed by an electronic device for identifying documents based on selected tags, consistent with embodiments of the present disclosure. The selected tags can be part of a predefined tag hierarchy (e.g., tag hierarchy 310 of FIG. 3).


In this exemplary illustration, the electronic device (e.g., a computer system 100) can interact with one or more other devices and/or storage components (e.g., data sources 230, object model 260, and database 270 of system 200 depicted in FIG. 2) for assisting with the identification of documents. While the flowchart discloses the following steps in a particular order, it will be appreciated that at least some of the steps can be moved, modified, or deleted where appropriate, consistent with the teachings of the present disclosure. And while the following steps are indicated as being performed by an electronic device, it is appreciated that the steps can be performed by more than one electronic device.


In step 902, the electronic device acquires a selection of one or more tags for at least one dimension defined under the tag hierarchy. The selection can be provided by a web-browser, or by a client-side application, after receiving the selection from a user.


In step 904, after acquiring the tag selection, the electronic device identifies one or more cells that are associated with the selected tags, and/or one or more cells associated with tags related to the selected tags. As indicated above, these identified cells can be provided by an object model (e.g., object model 260). In some embodiments, the relationship can be determined based on the attributes of the tags. For example, if a cell has attributes that match the selected tags, that cell can be identified.


In some embodiments, a combination of tags of one or more dimensions within tag hierarchy 310 can be represented as a multi-dimensional vector, with each dimension of tag hierarchy 310 represented by a vector dimension, and a combination of one or more tags under a dimension of tag hierarchy 310 contributes to a magnitude of the vector along that vector dimension, based on the attributes of the tags. The relationship between tags can then be calculated as, for example, an imaginary distance between the multi-dimensional vectors representing the tag combinations. In some embodiments, such imaginary distance can be calculated by first projecting the multi-dimensional vectors representing the tag combinations onto a pre-defined plane, and then calculating a distance between the projections on the pre-defined plane. In some embodiments, when the calculated distance exceeds a certain threshold, it can be determined that no relationship exists between the tag combinations. In some embodiments, the relationship between cells (or between tag combinations associated with the cells) can also be added manually with or without considering the calculated distance.


In step 906, the electronic device identifies documents associated with the one or more identified cells. As indicated above, tagged documents are associated with cells.


In step 908, the electronic device provides data corresponding to the identified documents for display. The identified documents can be represented as a list similar to search results 820 depicted in FIG. 8A, or similar to a graphical representation as depicted in FIG. 4A.


In step 910, the electronic device further provides data facilitating retrieval of documents with tags related to the selected tags. The data can be provided and displayed after, for example, the electronic device detects the clicking of at least one of related overview button 704, a linked documents button 706, and a related documents button 708 of interface 700. The data facilitating retrieval of documents can be displayed in the same interface as the data for the identified documents. For example, interface 700 can further include a pop-up window that includes requested information.


In the foregoing specification, embodiments have been described with reference to numerous specific details that can vary from implementation to implementation. Certain adaptations and modifications of the described embodiments can be made. Other embodiments can be apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the invention being indicated by the following claims. It is also intended that the sequence of steps shown in figures are only for illustrative purposes and are not intended to be limited to any particular sequence of steps. As such, those skilled in the art can appreciate that these steps can be performed in a different order while implementing the same method.

Claims
  • 1. An apparatus comprising: a memory device that stores a set of instructions; andat least one processor configured to execute the set of instructions to: communicate with one or more data storage devices storing a plurality of documents and an object model, wherein: the object model comprises a plurality of cells and relationships among the plurality of cells,each cell of the object model is associated with a respective combination of document tags, andeach of the documents is associated with at least one respective cell of the plurality of cells based on corresponding combinations of document tags associated with the respective documents;determine relationships among the plurality of cells in the object model by at least: representing each cell as a respective multi-dimensional vector based on the combination of document tags associated with the respective cells; anddetermining an imaginary distance between the multi-dimensional vectors associated with each pair of cells of the plurality of cells;receive, via a user interface, search criteria for searching the plurality of documents, wherein the search criteria include indications of one or more document tags;responsive to receiving the search criteria: determine a first cell of the plurality of cells of the object model matching the one or more document tags indicated by in the search criteria;determine a second cell of the plurality of cells that is related to the first cell based on the object model; andidentify, from the data storage devices, at least one or more documents associated with the second cell; andprovide, for displaying in the user interface, indications of the one or more documents associated with the second cell.
  • 2. A non-transitory computer readable medium that stores a set of instructions that are executable by at least one processor of an electronic device to cause the electronic device to perform a method comprising: communicating with one or more data storage devices storing a plurality of documents and an object model, wherein: the object model comprises a plurality of cells and relationships among the plurality of cells,each cell of the object model is associated with a respective combination of document tags, andeach of the documents is associated with at least one respective cell of the plurality of cells based on corresponding combinations of document tags associated with the respective documents;determining relationships among the plurality of cells in the object model by at least: representing each cell as a respective multi-dimensional vector based on the combination of document tags associated with the respective cells; anddetermining an imaginary distance between the multi-dimensional vectors associated with each pair of cells of the plurality of cells;receiving, via a user interface, search criteria for searching the plurality of documents, wherein the search criteria include indications of one or more document tags;responsive to receiving the search criteria: determining a first cell of the plurality of cells of the object model matching the one or more document tags indicated by in the search criteria;determining a second cell of the plurality of cells that is related to the first cell based on the object model; andidentifying, from the data storage devices, at least one or more documents associated with the second cell; andproviding, for displaying in the user interface, indication of the one or more documents associated with the second cell.
  • 3. A computer-implemented method comprising: communicating with one or more data storage devices storing a plurality of documents and an object model, wherein: the object model comprises a plurality of cells and relationships among the plurality of cells,each cell of the object model is associated with a respective combination of document tags, andeach of the documents is associated with at least one respective cell of the plurality of cells based on corresponding combinations of document tags associated with the respective documents;determining relationships among the plurality of cells in the object model by at least: representing each cell as a respective multi-dimensional vector based on the combination of document tags associated with the respective cells; anddetermining an imaginary distance between the multi-dimensional vectors associated with each pair of cells of the plurality of cells;receiving, via a user interface, search criteria for searching the plurality of documents, wherein the search criteria include indications of one or more document tags;responsive to receiving the search criteria: determining a first cell of the plurality of cells of the object model matching the one or more document tags indicated by in the search criteria;determining a second cell of the plurality of cells that is related to the first cell based on the object model; andidentifying, from the data storage devices, at least one or more documents associated with the second cell; andproviding, for displaying in the user interface, indication of the one or more documents associated with the second cell.
  • 4. The apparatus of claim 1, wherein the at least one processor is further configured to execute the set of instructions to: identify, from the data storage devices, at least one or more documents associated with the first cell; andprovide, for displaying in the user interface, at least one of the one or more documents associated with the first cell.
  • 5. The apparatus of claim 1, wherein the at least one processor is further configured to execute the set of instructions to: receive, via the user interface, a selection of a document from the one or more documents associated with the second cell; andresponsive to the selection of the document, update the search criteria to include indications of one or more document tags associated with the document.
  • 6. The apparatus of claim 1, wherein the document tags associated with the cells include multiple dimensions, and wherein the user interface provides an input selection field for each dimension or a single input selection field that can receive a combination of dimensions of document tags.
  • 7. The apparatus of claim 1, wherein the user interface provides a drop down menu to display one or more document tags for selection.
  • 8. The apparatus of claim 1, wherein the user interface displays a list of selectable items representing the one or more documents associated with the second cell, wherein each selectable item corresponds to an identified document and includes a title and metadata for the document, and wherein a selection of one of the selectable items causes the document corresponding to the selected selectable item to be selected.
  • 9. The apparatus of claim 1, wherein the at least one processor is further configured to execute the set of instructions to: receive, via the user interface, an input of one or more document tags, while displaying a document; andupdate one or more document tags associated with the document based on the input of the one or more document tags.
  • 10. The apparatus of claim 1, wherein the at least one processor is further configured to execute the set of instructions to: determine a relationship between the first cell and the second cell of the plurality of cells by at least: determining a first document tag type common to both the first cell and the second cell, wherein respective values of the first document tag type associated with the first cell and the second cell are different; anddetermining a relationship between the respective values of the first document tag type associated with the first cell and the second cell.
  • 11. The apparatus of claim 1, wherein the at least one processor is further configured to execute the set of instructions to: determine the relationships among the plurality of cells in the object model by at least: for each pair of cells: comparing the imaginary distance between the multi-dimensional vectors associated with the pair of cells to a threshold; andin response to the imaginary distance satisfying the threshold, determining that a relationship exists between the pair of cells.
  • 12. The non-transitory computer readable medium of claim 2, wherein the method further comprises: identifying, from the data storage devices, at least one or more documents associated with the first cell; andproviding, for displaying in the user interface, at least one of the one or more documents associated with the first cell.
  • 13. The non-transitory computer readable medium of claim 2, wherein the method further comprises: receiving, via the user interface, a selection of a document from the one or more documents associated with the second cell; andresponsive to the selection of the document, updating the search criteria to include indications of one or more document tags associated with the document.
  • 14. The non-transitory computer readable medium of claim 2, wherein the document tags associated with the cells include multiple dimensions, and wherein the user interface provides an input selection field for each dimension or a single input selection field that can receive a combination of dimensions of document tags.
  • 15. The non-transitory computer readable medium of claim 2, wherein the user interface provides a drop down menu to display one more document tags for selection.
  • 16. The non-transitory computer readable medium of claim 2, wherein the user interface displays a list of selectable items representing the one or more documents associated with the second cell, wherein each item corresponds to an identified document and includes a title and metadata for the document, and wherein a selection of one of the selectable items triggers the document corresponding to the selected selectable item to be selected.
  • 17. The non-transitory computer readable medium of claim 2, wherein the method further comprises: receiving, via the user interface, an input of one or more document tags, while displaying a document; andupdating one or more tags associated with the document based on the input of the one or more document tags.
  • 18. The non-transitory computer readable medium of claim 2, wherein the method further comprises: determining the relationships among the plurality of cells in the object model by at least: for each pair of cells: comparing the imaginary distance between the multi-dimensional vectors associated with the pair of cells to a threshold; andin response to the imaginary distance satisfying the threshold, determining that a relationship exists between the pair of cells.
  • 19. The computer-implemented method of claim 3 further comprising: identifying, from the data storage devices, at least one or more documents associated with the first cell; andproviding, for displaying in the user interface, at least one of the one or more documents associated with the first cell.
  • 20. The computer-implemented method of claim 3 further comprising: receiving, via the user interface, a selection of a document from the one or more documents associated with the second cell; andresponsive to the selection of the document, updating the search criteria to include indications of one or more document tags associated with the document.
  • 21. The computer-implemented method of claim 3 further comprising: determining the relationships among the plurality of cells in the object model by at least: for each pair of cells: comparing the imaginary distance between the multi-dimensional vectors associated with the pair of cells to a threshold; andin response to the imaginary distance satisfying the threshold, determining that a relationship exists between the pair of cells.
US Referenced Citations (662)
Number Name Date Kind
5109399 Thompson Apr 1992 A
5329108 Lamoure Jul 1994 A
5632009 Rao et al. May 1997 A
5670987 Doi et al. Sep 1997 A
5781704 Rossmo Jul 1998 A
5798769 Chiu et al. Aug 1998 A
5845300 Comer Dec 1998 A
6057757 Arrowsmith et al. May 2000 A
6091956 Hollenberg Jul 2000 A
6161098 Wallman Dec 2000 A
6219053 Tachibana et al. Apr 2001 B1
6232971 Haynes May 2001 B1
6247019 Davies Jun 2001 B1
6279018 Kudrolli et al. Aug 2001 B1
6289338 Stoffel et al. Sep 2001 B1
6341310 Leshem et al. Jan 2002 B1
6366933 Ball et al. Apr 2002 B1
6369835 Lin Apr 2002 B1
6430305 Decker Aug 2002 B1
6456997 Shukla Sep 2002 B1
6463404 Appleby Oct 2002 B1
6523172 Martinez-Guerra et al. Feb 2003 B1
6539538 Brewster et al. Mar 2003 B1
6549944 Weinberg et al. Apr 2003 B1
6560620 Ching May 2003 B1
6581068 Bensoussan et al. Jun 2003 B1
6594672 Lampson et al. Jul 2003 B1
6631496 Li et al. Oct 2003 B1
6640231 Andersen et al. Oct 2003 B1
6642945 Sharpe Nov 2003 B1
6674434 Chojnacki et al. Jan 2004 B1
6714936 Nevin, III Mar 2004 B1
6748481 Parry et al. Jun 2004 B1
6775675 Nwabueze et al. Aug 2004 B1
6820135 Dingman Nov 2004 B1
6828920 Owen et al. Dec 2004 B2
6839745 Dingari et al. Jan 2005 B1
6877137 Rivette et al. Apr 2005 B1
6976210 Silva et al. Dec 2005 B1
6978419 Kantrowitz Dec 2005 B1
6980984 Huffman et al. Dec 2005 B1
6985950 Hanson et al. Jan 2006 B1
7027974 Busch et al. Apr 2006 B1
7036085 Barros Apr 2006 B2
7043702 Chi et al. May 2006 B2
7055110 Kupka et al. May 2006 B2
7089541 Ungar Aug 2006 B2
7139800 Bellotti et al. Nov 2006 B2
7158878 Rasmussen et al. Jan 2007 B2
7162475 Ackerman Jan 2007 B2
7168039 Bertram Jan 2007 B2
7171427 Witkowski et al. Jan 2007 B2
7237192 Stephenson et al. Jun 2007 B1
7240330 Fairweather Jul 2007 B2
7269786 Malloy et al. Sep 2007 B1
7278105 Kitts Oct 2007 B1
7290698 Poslinski et al. Nov 2007 B2
7333998 Heckerman et al. Feb 2008 B2
7370047 Gorman May 2008 B2
7379811 Rasmussen et al. May 2008 B2
7379903 Caballero et al. May 2008 B2
7426654 Adams et al. Sep 2008 B2
7454466 Bellotti et al. Nov 2008 B2
7467375 Tondreau et al. Dec 2008 B2
7487139 Fraleigh et al. Feb 2009 B2
7502786 Liu et al. Mar 2009 B2
7525422 Bishop et al. Apr 2009 B2
7529727 Arning et al. May 2009 B2
7529734 Dirisala May 2009 B2
7533069 Fairweather May 2009 B2
7558677 Jones Jul 2009 B2
7574409 Patinkin Aug 2009 B2
7574428 Leiserowitz et al. Aug 2009 B2
7579965 Bucholz Aug 2009 B2
7596285 Brown et al. Sep 2009 B2
7614006 Molander Nov 2009 B2
7617232 Gabbert et al. Nov 2009 B2
7620628 Kapur et al. Nov 2009 B2
7627812 Chamberlain et al. Dec 2009 B2
7634717 Chamberlain et al. Dec 2009 B2
7685083 Fairweather Mar 2010 B2
7703021 Flam Apr 2010 B1
7706817 Bamrah et al. Apr 2010 B2
7712049 Williams et al. May 2010 B2
7716077 Mikurak May 2010 B1
7725530 Sah et al. May 2010 B2
7725547 Albertson et al. May 2010 B2
7730082 Sah et al. Jun 2010 B2
7730109 Rohrs et al. Jun 2010 B2
7756843 Palmer Jul 2010 B1
7770100 Chamberlain et al. Aug 2010 B2
7805457 Viola et al. Sep 2010 B1
7809703 Balabhadrapatruni et al. Oct 2010 B2
7818658 Chen Oct 2010 B2
7870493 Pall et al. Jan 2011 B2
7877421 Berger et al. Jan 2011 B2
7894984 Rasmussen et al. Feb 2011 B2
7899611 Downs et al. Mar 2011 B2
7899796 Borthwick et al. Mar 2011 B1
7917376 Bellin et al. Mar 2011 B2
7920963 Jouline et al. Apr 2011 B2
7933862 Chamberlain et al. Apr 2011 B2
7933929 McClendon et al. Apr 2011 B1
7941321 Greenstein et al. May 2011 B2
7962281 Rasmussen et al. Jun 2011 B2
7962495 Jain et al. Jun 2011 B2
7962848 Bertram Jun 2011 B2
7970240 Chao et al. Jun 2011 B1
7971150 Raskutti et al. Jun 2011 B2
7984374 Caro et al. Jul 2011 B2
8001465 Kudrolli et al. Aug 2011 B2
8001482 Bhattiprolu et al. Aug 2011 B2
8010545 Stefik et al. Aug 2011 B2
8015487 Roy et al. Sep 2011 B2
8024778 Cash et al. Sep 2011 B2
8036632 Cona et al. Oct 2011 B1
8036971 Aymeloglu et al. Oct 2011 B2
8046283 Burns Oct 2011 B2
8054756 Chand et al. Nov 2011 B2
8103543 Zwicky Jan 2012 B1
8117022 Linker Feb 2012 B2
8132149 Shenfield et al. Mar 2012 B2
8134457 Velipasalar et al. Mar 2012 B2
8145703 Frishert et al. Mar 2012 B2
8185819 Sah et al. May 2012 B2
8214361 Sandler et al. Jul 2012 B1
8214490 Vos et al. Jul 2012 B1
8214764 Gemmell et al. Jul 2012 B2
8225201 Michael Jul 2012 B2
8229902 Vishniac et al. Jul 2012 B2
8229947 Fujinaga Jul 2012 B2
8230333 Decherd et al. Jul 2012 B2
8271461 Pike et al. Sep 2012 B2
8271948 Talozi et al. Sep 2012 B2
8280880 Aymeloglu et al. Oct 2012 B1
8290838 Thakur et al. Oct 2012 B1
8290926 Ozzie et al. Oct 2012 B2
8290942 Jones et al. Oct 2012 B2
8301464 Cave et al. Oct 2012 B1
8301904 Gryaznov Oct 2012 B1
8302855 Ma et al. Nov 2012 B2
8312367 Foster Nov 2012 B2
8312546 Alme Nov 2012 B2
8352881 Champion et al. Jan 2013 B2
8368695 Howell et al. Feb 2013 B2
8397171 Klassen et al. Mar 2013 B2
8412707 Mianji Apr 2013 B1
8447722 Ahuja et al. May 2013 B1
8452790 Mianji May 2013 B1
8463036 Ramesh et al. Jun 2013 B1
8473454 Evanitsky et al. Jun 2013 B2
8484115 Aymeloglu et al. Jul 2013 B2
8489331 Kopf et al. Jul 2013 B2
8489623 Jain et al. Jul 2013 B2
8489641 Seefeld et al. Jul 2013 B1
8498984 Hwang et al. Jul 2013 B1
8510743 Hackborn et al. Aug 2013 B2
8514082 Cova et al. Aug 2013 B2
8515207 Chau Aug 2013 B2
8554579 Tribble et al. Oct 2013 B2
8554653 Falkenborg et al. Oct 2013 B2
8554709 Goodson et al. Oct 2013 B2
8560413 Quarterman Oct 2013 B1
8560494 Downing et al. Oct 2013 B1
8577911 Stepinski et al. Nov 2013 B1
8589273 Creeden et al. Nov 2013 B2
8595234 Siripurapu et al. Nov 2013 B2
8620641 Farnsworth et al. Dec 2013 B2
8639757 Zang et al. Jan 2014 B1
8646080 Williamson et al. Feb 2014 B2
8676857 Adams et al. Mar 2014 B1
8688573 Rukonic et al. Apr 2014 B1
8689108 Duffield et al. Apr 2014 B1
8689182 Leithead et al. Apr 2014 B2
8713467 Goldenberg et al. Apr 2014 B1
8726379 Stiansen et al. May 2014 B1
8739278 Varghese May 2014 B2
8742934 Sarpy et al. Jun 2014 B1
8744890 Bernier Jun 2014 B1
8745516 Mason et al. Jun 2014 B2
8781169 Jackson et al. Jul 2014 B2
8787939 Papakipos et al. Jul 2014 B2
8788407 Singh et al. Jul 2014 B1
8799799 Cervelli et al. Aug 2014 B1
8812960 Sun et al. Aug 2014 B1
8830322 Nerayoff et al. Sep 2014 B2
8832594 Thompson et al. Sep 2014 B1
8855999 Elliot Oct 2014 B1
8868537 Colgrove et al. Oct 2014 B1
8903717 Elliot Dec 2014 B2
8917274 Ma et al. Dec 2014 B2
8924388 Elliot et al. Dec 2014 B2
8924389 Elliot et al. Dec 2014 B2
8924872 Bogomolov et al. Dec 2014 B1
8930897 Nassar Jan 2015 B2
8937619 Sharma et al. Jan 2015 B2
8938686 Erenrich et al. Jan 2015 B1
8949164 Mohler Feb 2015 B1
9009171 Grossman et al. Apr 2015 B1
9009827 Albertson et al. Apr 2015 B1
9021260 Falk et al. Apr 2015 B1
9021384 Beard et al. Apr 2015 B1
9043696 Meiklejohn et al. May 2015 B1
9043894 Dennison et al. May 2015 B1
9100428 Visbal Aug 2015 B1
9116975 Shankar et al. Aug 2015 B2
9129219 Robertson et al. Sep 2015 B1
9201920 Jain et al. Dec 2015 B2
9223773 Isaacson Dec 2015 B2
20010021936 Bertram Sep 2001 A1
20020033848 Sciammarella et al. Mar 2002 A1
20020065708 Senay et al. May 2002 A1
20020091707 Keller Jul 2002 A1
20020095360 Joao Jul 2002 A1
20020095658 Shulman Jul 2002 A1
20020103705 Brady Aug 2002 A1
20020116120 Ruiz et al. Aug 2002 A1
20020130907 Chi et al. Sep 2002 A1
20020147805 Leshem et al. Oct 2002 A1
20020174201 Ramer et al. Nov 2002 A1
20020194119 Wright et al. Dec 2002 A1
20030028560 Kudrolli et al. Feb 2003 A1
20030039948 Donahue Feb 2003 A1
20030126102 Borthwick Jul 2003 A1
20030140106 Raguseo Jul 2003 A1
20030144868 MacIntyre et al. Jul 2003 A1
20030163352 Surpin et al. Aug 2003 A1
20030172053 Fairweather Sep 2003 A1
20030177112 Gardner Sep 2003 A1
20030200217 Ackerman Oct 2003 A1
20030225755 Iwayama et al. Dec 2003 A1
20030227392 Ebert et al. Dec 2003 A1
20030229848 Arend et al. Dec 2003 A1
20040032432 Baynger Feb 2004 A1
20040034570 Davis Feb 2004 A1
20040044992 Muller et al. Mar 2004 A1
20040064256 Barinek et al. Apr 2004 A1
20040083466 Dapp et al. Apr 2004 A1
20040085318 Hassler et al. May 2004 A1
20040095349 Bito et al. May 2004 A1
20040111410 Burgoon et al. Jun 2004 A1
20040111480 Yue Jun 2004 A1
20040126840 Cheng et al. Jul 2004 A1
20040143602 Ruiz et al. Jul 2004 A1
20040143796 Lerner et al. Jul 2004 A1
20040153418 Hanweck Aug 2004 A1
20040163039 Gorman Aug 2004 A1
20040181554 Heckerman et al. Sep 2004 A1
20040193600 Kaasten et al. Sep 2004 A1
20040221223 Yu et al. Nov 2004 A1
20040236688 Bozeman Nov 2004 A1
20040260702 Cragun et al. Dec 2004 A1
20040267746 Marcjan et al. Dec 2004 A1
20050010472 Quatse et al. Jan 2005 A1
20050027705 Sadri et al. Feb 2005 A1
20050028094 Allyn Feb 2005 A1
20050039119 Parks et al. Feb 2005 A1
20050065811 Chu et al. Mar 2005 A1
20050080769 Gemmell et al. Apr 2005 A1
20050086207 Heuer et al. Apr 2005 A1
20050091420 Snover et al. Apr 2005 A1
20050125715 Franco et al. Jun 2005 A1
20050154628 Eckart et al. Jul 2005 A1
20050154769 Eckart et al. Jul 2005 A1
20050162523 Darrell et al. Jul 2005 A1
20050166144 Gross Jul 2005 A1
20050180330 Shapiro Aug 2005 A1
20050182793 Keenan et al. Aug 2005 A1
20050183005 Denoue et al. Aug 2005 A1
20050210409 Jou Sep 2005 A1
20050246327 Yeung et al. Nov 2005 A1
20050251786 Citron et al. Nov 2005 A1
20060026120 Carolan et al. Feb 2006 A1
20060026170 Kreitler et al. Feb 2006 A1
20060045470 Poslinski et al. Mar 2006 A1
20060059139 Robinson Mar 2006 A1
20060074866 Chamberlain et al. Apr 2006 A1
20060074881 Vembu et al. Apr 2006 A1
20060080619 Carlson et al. Apr 2006 A1
20060093222 Saffer et al. May 2006 A1
20060129746 Porter Jun 2006 A1
20060139375 Rasmussen et al. Jun 2006 A1
20060142949 Helt Jun 2006 A1
20060143034 Rothermel Jun 2006 A1
20060143075 Carr et al. Jun 2006 A1
20060143079 Basak et al. Jun 2006 A1
20060149596 Surpin et al. Jul 2006 A1
20060203337 White Sep 2006 A1
20060218637 Thomas et al. Sep 2006 A1
20060241974 Chao et al. Oct 2006 A1
20060242040 Rader Oct 2006 A1
20060242630 Koike et al. Oct 2006 A1
20060271277 Hu et al. Nov 2006 A1
20060279630 Aggarwal et al. Dec 2006 A1
20070000999 Kubo et al. Jan 2007 A1
20070011150 Frank Jan 2007 A1
20070011304 Error Jan 2007 A1
20070016363 Huang et al. Jan 2007 A1
20070038646 Thota Feb 2007 A1
20070038962 Fuchs et al. Feb 2007 A1
20070057966 Ohno et al. Mar 2007 A1
20070074169 Chess et al. Mar 2007 A1
20070078832 Ott et al. Apr 2007 A1
20070078872 Cohen Apr 2007 A1
20070083541 Fraleigh et al. Apr 2007 A1
20070094389 Nussey et al. Apr 2007 A1
20070112714 Fairweather May 2007 A1
20070150369 Zivin Jun 2007 A1
20070150801 Chidlovskii et al. Jun 2007 A1
20070156673 Maga Jul 2007 A1
20070174760 Chamberlain et al. Jul 2007 A1
20070185867 Maga Aug 2007 A1
20070192265 Chopin et al. Aug 2007 A1
20070198571 Ferguson et al. Aug 2007 A1
20070208497 Downs et al. Sep 2007 A1
20070208498 Barker et al. Sep 2007 A1
20070208736 Tanigawa et al. Sep 2007 A1
20070233709 Abnous Oct 2007 A1
20070240062 Christena et al. Oct 2007 A1
20070266336 Nojima et al. Nov 2007 A1
20070284433 Domenica et al. Dec 2007 A1
20070294643 Kyle Dec 2007 A1
20080016216 Worley et al. Jan 2008 A1
20080040684 Crump Feb 2008 A1
20080051989 Welsh Feb 2008 A1
20080052142 Bailey et al. Feb 2008 A1
20080069081 Chand et al. Mar 2008 A1
20080077597 Butler Mar 2008 A1
20080077642 Carbone et al. Mar 2008 A1
20080082486 Lermant et al. Apr 2008 A1
20080103996 Forman et al. May 2008 A1
20080104019 Nath May 2008 A1
20080126951 Sood et al. May 2008 A1
20080140387 Linker Jun 2008 A1
20080148398 Mezack et al. Jun 2008 A1
20080155440 Trevor et al. Jun 2008 A1
20080162616 Gross et al. Jul 2008 A1
20080195417 Surpin et al. Aug 2008 A1
20080195608 Clover Aug 2008 A1
20080222295 Robinson et al. Sep 2008 A1
20080228467 Womack et al. Sep 2008 A1
20080243711 Aymeloglu et al. Oct 2008 A1
20080244510 Cote Oct 2008 A1
20080255973 El Wade et al. Oct 2008 A1
20080263468 Cappione et al. Oct 2008 A1
20080267107 Rosenberg Oct 2008 A1
20080276167 Michael Nov 2008 A1
20080278311 Grange et al. Nov 2008 A1
20080281580 Zabokritski Nov 2008 A1
20080288306 MacIntyre et al. Nov 2008 A1
20080301643 Appleton et al. Dec 2008 A1
20080313132 Hao et al. Dec 2008 A1
20090002492 Velipasalar et al. Jan 2009 A1
20090027418 Maru et al. Jan 2009 A1
20090030915 Winter et al. Jan 2009 A1
20090055251 Shah et al. Feb 2009 A1
20090076845 Bellin et al. Mar 2009 A1
20090077062 Spivack Mar 2009 A1
20090088964 Schaaf et al. Apr 2009 A1
20090094166 Aymeloglu et al. Apr 2009 A1
20090106178 Chu Apr 2009 A1
20090112745 Stefanescu Apr 2009 A1
20090119309 Gibson et al. May 2009 A1
20090125359 Knapic May 2009 A1
20090125369 Kloostra et al. May 2009 A1
20090125459 Norton et al. May 2009 A1
20090132921 Hwangbo et al. May 2009 A1
20090132953 Reed et al. May 2009 A1
20090143052 Bates et al. Jun 2009 A1
20090144262 White et al. Jun 2009 A1
20090144274 Fraleigh et al. Jun 2009 A1
20090164934 Bhattiprolu et al. Jun 2009 A1
20090171939 Athsani et al. Jul 2009 A1
20090172511 Decherd et al. Jul 2009 A1
20090172821 Daira et al. Jul 2009 A1
20090177962 Gusmorino et al. Jul 2009 A1
20090179892 Tsuda et al. Jul 2009 A1
20090187464 Bai et al. Jul 2009 A1
20090187546 Whyte et al. Jul 2009 A1
20090187548 Ji et al. Jul 2009 A1
20090222400 Kupershmidt et al. Sep 2009 A1
20090222759 Drieschner Sep 2009 A1
20090222760 Halverson et al. Sep 2009 A1
20090234720 George et al. Sep 2009 A1
20090249244 Robinson et al. Oct 2009 A1
20090254970 Agarwal et al. Oct 2009 A1
20090271343 Vaiciulis et al. Oct 2009 A1
20090281839 Lynn et al. Nov 2009 A1
20090287470 Farnsworth et al. Nov 2009 A1
20090292626 Oxford Nov 2009 A1
20090307049 Elliott et al. Dec 2009 A1
20090313463 Pang et al. Dec 2009 A1
20090319418 Herz Dec 2009 A1
20090319891 MacKinlay Dec 2009 A1
20100011282 Dollard et al. Jan 2010 A1
20100030722 Goodson et al. Feb 2010 A1
20100031141 Summers et al. Feb 2010 A1
20100042922 Bradateanu et al. Feb 2010 A1
20100057622 Faith et al. Mar 2010 A1
20100057716 Stefik et al. Mar 2010 A1
20100070523 Delgo et al. Mar 2010 A1
20100070842 Aymeloglu et al. Mar 2010 A1
20100070845 Facemire et al. Mar 2010 A1
20100070897 Aymeloglu et al. Mar 2010 A1
20100098318 Anderson Apr 2010 A1
20100100963 Mahaffey Apr 2010 A1
20100103124 Kruzeniski et al. Apr 2010 A1
20100114887 Conway et al. May 2010 A1
20100122152 Chamberlain et al. May 2010 A1
20100131457 Heimendinger May 2010 A1
20100131502 Fordham May 2010 A1
20100161735 Sharma Jun 2010 A1
20100162176 Dunton Jun 2010 A1
20100191563 Schlaifer et al. Jul 2010 A1
20100198684 Eraker et al. Aug 2010 A1
20100199225 Coleman et al. Aug 2010 A1
20100204983 Chung et al. Aug 2010 A1
20100228812 Uomini Sep 2010 A1
20100235915 Memon et al. Sep 2010 A1
20100250412 Wagner Sep 2010 A1
20100262688 Hussain et al. Oct 2010 A1
20100280857 Liu et al. Nov 2010 A1
20100293174 Bennett et al. Nov 2010 A1
20100306285 Shah et al. Dec 2010 A1
20100306713 Geisner et al. Dec 2010 A1
20100312837 Bodapati et al. Dec 2010 A1
20100313119 Baldwin et al. Dec 2010 A1
20100318924 Frankel et al. Dec 2010 A1
20100321399 Ellren et al. Dec 2010 A1
20100325526 Ellis et al. Dec 2010 A1
20100325581 Finkelstein et al. Dec 2010 A1
20100330801 Rouh Dec 2010 A1
20110029526 Knight et al. Feb 2011 A1
20110047159 Baid et al. Feb 2011 A1
20110060753 Shaked et al. Mar 2011 A1
20110061013 Bilicki et al. Mar 2011 A1
20110066933 Ludwig Mar 2011 A1
20110074811 Hanson et al. Mar 2011 A1
20110078055 Faribault et al. Mar 2011 A1
20110078173 Seligmann et al. Mar 2011 A1
20110093327 Fordyce et al. Apr 2011 A1
20110099133 Chang et al. Apr 2011 A1
20110117878 Barash et al. May 2011 A1
20110119100 Ruhl et al. May 2011 A1
20110137766 Rasmussen et al. Jun 2011 A1
20110153384 Horne et al. Jun 2011 A1
20110161096 Buehler et al. Jun 2011 A1
20110167105 Ramakrishnan et al. Jul 2011 A1
20110170799 Carrino et al. Jul 2011 A1
20110173032 Payne et al. Jul 2011 A1
20110173093 Psota et al. Jul 2011 A1
20110185316 Reid et al. Jul 2011 A1
20110208565 Ross et al. Aug 2011 A1
20110208724 Jones et al. Aug 2011 A1
20110213655 Henkin Sep 2011 A1
20110218934 Elser Sep 2011 A1
20110218955 Tang Sep 2011 A1
20110219450 McDougal et al. Sep 2011 A1
20110225198 Edwards et al. Sep 2011 A1
20110238495 Kang Sep 2011 A1
20110238553 Raj et al. Sep 2011 A1
20110251951 Kolkowitz Oct 2011 A1
20110258158 Resende et al. Oct 2011 A1
20110258216 Supakkul et al. Oct 2011 A1
20110270604 Qi et al. Nov 2011 A1
20110270705 Parker Nov 2011 A1
20110270834 Sokolan et al. Nov 2011 A1
20110289397 Eastmond et al. Nov 2011 A1
20110289407 Naik et al. Nov 2011 A1
20110289420 Morioka et al. Nov 2011 A1
20110291851 Whisenant Dec 2011 A1
20110295649 Fine Dec 2011 A1
20110310005 Chen et al. Dec 2011 A1
20110314007 Dassa et al. Dec 2011 A1
20110314024 Chang et al. Dec 2011 A1
20120011238 Rathod Jan 2012 A1
20120011245 Gillette et al. Jan 2012 A1
20120019559 Siler et al. Jan 2012 A1
20120022945 Falkenborg et al. Jan 2012 A1
20120036013 Neuhaus et al. Feb 2012 A1
20120036434 Oberstein Feb 2012 A1
20120050293 Carlhian et al. Mar 2012 A1
20120054284 Rakshit Mar 2012 A1
20120059853 Jagota Mar 2012 A1
20120066166 Curbera et al. Mar 2012 A1
20120066210 Mukerjee et al. Mar 2012 A1
20120066296 Appleton et al. Mar 2012 A1
20120072825 Sherkin et al. Mar 2012 A1
20120079363 Folting et al. Mar 2012 A1
20120084117 Tavares et al. Apr 2012 A1
20120084118 Bai et al. Apr 2012 A1
20120084287 Lakshminarayan et al. Apr 2012 A1
20120106801 Jackson May 2012 A1
20120110515 Abramoff May 2012 A1
20120117082 Koperda et al. May 2012 A1
20120131512 Takeuchi et al. May 2012 A1
20120137235 Ts et al. May 2012 A1
20120144335 Abeln et al. Jun 2012 A1
20120159307 Chung et al. Jun 2012 A1
20120159362 Brown et al. Jun 2012 A1
20120159399 Bastide et al. Jun 2012 A1
20120170847 Tsukidate Jul 2012 A1
20120173381 Smith Jul 2012 A1
20120173985 Peppel Jul 2012 A1
20120180002 Campbell et al. Jul 2012 A1
20120191446 Binsztok et al. Jul 2012 A1
20120196557 Reich et al. Aug 2012 A1
20120196558 Reich et al. Aug 2012 A1
20120197651 Robinson et al. Aug 2012 A1
20120203708 Psota et al. Aug 2012 A1
20120208636 Feige Aug 2012 A1
20120215784 King et al. Aug 2012 A1
20120221511 Gibson et al. Aug 2012 A1
20120221553 Wittmer et al. Aug 2012 A1
20120221580 Barney Aug 2012 A1
20120226523 Weiss Sep 2012 A1
20120245976 Kumar et al. Sep 2012 A1
20120246148 Dror Sep 2012 A1
20120254129 Wheeler et al. Oct 2012 A1
20120284345 Costenaro et al. Nov 2012 A1
20120290879 Shibuya et al. Nov 2012 A1
20120296907 Long et al. Nov 2012 A1
20120304150 Leithead et al. Nov 2012 A1
20120311684 Paulsen et al. Dec 2012 A1
20120323888 Osann, Jr. Dec 2012 A1
20120330801 McDougal et al. Dec 2012 A1
20120330973 Ghuneim et al. Dec 2012 A1
20130006426 Healey et al. Jan 2013 A1
20130006725 Simanek et al. Jan 2013 A1
20130006916 McBride et al. Jan 2013 A1
20130016106 Yip et al. Jan 2013 A1
20130018796 Kolhatkar et al. Jan 2013 A1
20130024268 Manickavelu Jan 2013 A1
20130046635 Grigg et al. Feb 2013 A1
20130046842 Muntz et al. Feb 2013 A1
20130054306 Bhalla Feb 2013 A1
20130057551 Ebert et al. Mar 2013 A1
20130060786 Serrano et al. Mar 2013 A1
20130061169 Pearcy et al. Mar 2013 A1
20130073377 Heath Mar 2013 A1
20130073454 Busch Mar 2013 A1
20130078943 Biage et al. Mar 2013 A1
20130086482 Parsons Apr 2013 A1
20130091084 Lee Apr 2013 A1
20130096988 Grossman et al. Apr 2013 A1
20130097482 Marantz et al. Apr 2013 A1
20130110746 Ahn May 2013 A1
20130110822 Ikeda et al. May 2013 A1
20130110877 Bonham et al. May 2013 A1
20130111320 Campbell et al. May 2013 A1
20130117651 Waldman et al. May 2013 A1
20130124193 Holmberg May 2013 A1
20130150004 Rosen Jun 2013 A1
20130151148 Parundekar et al. Jun 2013 A1
20130151388 Falkenborg et al. Jun 2013 A1
20130151453 Bhanot et al. Jun 2013 A1
20130157234 Gulli et al. Jun 2013 A1
20130166348 Scotto Jun 2013 A1
20130166480 Popescu et al. Jun 2013 A1
20130166550 Buchmann et al. Jun 2013 A1
20130176321 Mitchell et al. Jul 2013 A1
20130179420 Park et al. Jul 2013 A1
20130185245 Anderson Jul 2013 A1
20130185307 El-Yaniv et al. Jul 2013 A1
20130212081 Shenoy et al. Aug 2013 A1
20130224696 Wolfe et al. Aug 2013 A1
20130225212 Khan Aug 2013 A1
20130226318 Procyk Aug 2013 A1
20130226953 Markovich et al. Aug 2013 A1
20130232045 Tai Sep 2013 A1
20130238616 Rose et al. Sep 2013 A1
20130246170 Gross et al. Sep 2013 A1
20130246537 Gaddala Sep 2013 A1
20130246597 Iizawa et al. Sep 2013 A1
20130251233 Yang et al. Sep 2013 A1
20130262527 Hunter et al. Oct 2013 A1
20130263019 Castellanos et al. Oct 2013 A1
20130267207 Hao et al. Oct 2013 A1
20130268520 Fisher et al. Oct 2013 A1
20130275446 Jain et al. Oct 2013 A1
20130279757 Kephart Oct 2013 A1
20130282696 John et al. Oct 2013 A1
20130290011 Lynn et al. Oct 2013 A1
20130290825 Arndt et al. Oct 2013 A1
20130297619 Chandrasekaran et al. Nov 2013 A1
20130304770 Boero et al. Nov 2013 A1
20130311375 Priebatsch Nov 2013 A1
20140012796 Petersen et al. Jan 2014 A1
20140019936 Cohanoff Jan 2014 A1
20140032506 Hoey et al. Jan 2014 A1
20140033010 Richardt et al. Jan 2014 A1
20140040371 Gurevich et al. Feb 2014 A1
20140047319 Eberlein Feb 2014 A1
20140047357 Alfaro et al. Feb 2014 A1
20140058914 Song et al. Feb 2014 A1
20140059038 McPherson et al. Feb 2014 A1
20140067611 Adachi et al. Mar 2014 A1
20140068487 Steiger et al. Mar 2014 A1
20140074855 Zhao et al. Mar 2014 A1
20140095273 Tang et al. Apr 2014 A1
20140095509 Patton Apr 2014 A1
20140108068 Williams Apr 2014 A1
20140108380 Gotz et al. Apr 2014 A1
20140108985 Scott et al. Apr 2014 A1
20140123279 Bishop et al. May 2014 A1
20140129261 Bothwell et al. May 2014 A1
20140136285 Carvalho May 2014 A1
20140143009 Brice et al. May 2014 A1
20140149436 Bahrami et al. May 2014 A1
20140156527 Grigg et al. Jun 2014 A1
20140157172 Peery et al. Jun 2014 A1
20140164502 Khodorenko et al. Jun 2014 A1
20140189536 Lange et al. Jul 2014 A1
20140195515 Baker et al. Jul 2014 A1
20140195887 Ellis et al. Jul 2014 A1
20140214579 Shen et al. Jul 2014 A1
20140222521 Chait Aug 2014 A1
20140222793 Sadkin et al. Aug 2014 A1
20140229554 Grunin et al. Aug 2014 A1
20140244388 Manouchehri et al. Aug 2014 A1
20140267294 Ma Sep 2014 A1
20140267295 Sharma Sep 2014 A1
20140279824 Tamayo Sep 2014 A1
20140281850 Prakash et al. Sep 2014 A1
20140282121 Sun et al. Sep 2014 A1
20140310266 Greenfield Oct 2014 A1
20140316911 Gross Oct 2014 A1
20140333651 Cervelli et al. Nov 2014 A1
20140337772 Cervelli et al. Nov 2014 A1
20140344230 Krause et al. Nov 2014 A1
20140358829 Hurwitz Dec 2014 A1
20140366132 Stiansen et al. Dec 2014 A1
20150019394 Unser et al. Jan 2015 A1
20150046481 Elliot Feb 2015 A1
20150046870 Goldenberg et al. Feb 2015 A1
20150073929 Psota et al. Mar 2015 A1
20150073954 Braff Mar 2015 A1
20150089424 Duffield et al. Mar 2015 A1
20150095773 Gonsalves et al. Apr 2015 A1
20150100559 Nassar Apr 2015 A1
20150100897 Sun et al. Apr 2015 A1
20150100907 Erenrich et al. Apr 2015 A1
20150106379 Elliot et al. Apr 2015 A1
20150134666 Gattiker et al. May 2015 A1
20150135256 Hoy et al. May 2015 A1
20150142766 Jain et al. May 2015 A1
20150169709 Kara et al. Jun 2015 A1
20150169726 Kara et al. Jun 2015 A1
20150170077 Kara et al. Jun 2015 A1
20150178825 Huerta Jun 2015 A1
20150178877 Bogomolov et al. Jun 2015 A1
20150186821 Wang et al. Jul 2015 A1
20150187036 Wang et al. Jul 2015 A1
20150188872 White Jul 2015 A1
20150227295 Meiklejohn et al. Aug 2015 A1
20150261847 Ducott, III et al. Sep 2015 A1
20150309719 Ma et al. Oct 2015 A1
20150317342 Grossman et al. Nov 2015 A1
20150324868 Kaftan et al. Nov 2015 A1
20150338233 Cervelli et al. Nov 2015 A1
20150379413 Robertson et al. Dec 2015 A1
20160004764 Chakerian et al. Jan 2016 A1
Foreign Referenced Citations (61)
Number Date Country
2666364 Jan 2015 CA
102546446 Jul 2012 CN
103167093 Jun 2013 CN
102054015 May 2014 CN
102014103482 Sep 2014 DE
102014204827 Sep 2014 DE
102014204830 Sep 2014 DE
102014204834 Sep 2014 DE
102014204840 Sep 2014 DE
102014215621 Feb 2015 DE
1672527 Jun 2006 EP
2487610 Aug 2012 EP
2551799 Jan 2013 EP
2560134 Feb 2013 EP
2778913 Sep 2014 EP
2778914 Sep 2014 EP
2778977 Sep 2014 EP
2778986 Sep 2014 EP
2835745 Feb 2015 EP
2835770 Feb 2015 EP
2838039 Feb 2015 EP
2846241 Mar 2015 EP
2851852 Mar 2015 EP
2858014 Apr 2015 EP
2858018 Apr 2015 EP
2863326 Apr 2015 EP
2863346 Apr 2015 EP
2869211 May 2015 EP
2884439 Jun 2015 EP
2884440 Jun 2015 EP
2889814 Jul 2015 EP
2891992 Jul 2015 EP
2892197 Jul 2015 EP
2911078 Aug 2015 EP
2911100 Aug 2015 EP
2940603 Nov 2015 EP
2940609 Nov 2015 EP
2963595 Jan 2016 EP
3062244 Aug 2016 EP
2366498 Mar 2002 GB
2513007 Oct 2014 GB
2516155 Jan 2015 GB
2518745 Apr 2015 GB
2012778 Nov 2014 NL
2013306 Feb 2015 NL
624557 Dec 2014 NZ
WO 99033000 Jul 1999 WO
WO 99033004 Jul 1999 WO
WO 0009529 Feb 2000 WO
WO 02065353 Aug 2002 WO
WO 03060751 Jul 2003 WO
WO 2005104736 Nov 2005 WO
WO 2005116851 Dec 2005 WO
WO 2008064207 May 2008 WO
WO 2009061501 May 2009 WO
WO 2010000014 Jan 2010 WO
WO 2010030913 Mar 2010 WO
WO 2011071833 Jun 2011 WO
WO 2013010157 Jan 2013 WO
WO 2013102892 Jul 2013 WO
WO 2013115775 Aug 2013 WO
Non-Patent Literature Citations (329)
Entry
“A First Look: Predicting Market Demand for Food Retail using a Huff Analysis,” TRF Policy Solutions, Jul. 2012, pp. 30.
“A Quick Guide to UniProtKB Swiss-Prot & TrEMBL,” Sep. 2011, pp. 2.
Acklen, Laura, “Absolute Beginner's Guide to Microsoft Word 2003,” Dec. 24, 2003, pp. 15-18, 34-41, 308-316.
Ananiev et al., “The New Modality API,” http://web.archive.org/web/20061211011958/http://java.sun.com/developer/technicalArticles/J2SE/Desktop/javase6/modality/ Jan. 21, 2006, pp. 8.
Appacts, “Smart Thinking for Super Apps,” <http://www.appacts.com> Printed Jul. 18, 2013 in 4 pages.
Apsalar, “Data Powered Mobile Advertising,” “Free Mobile App Analytics” and various analytics related screen shots <http://apsalar.com> Printed Jul. 18, 2013 in 8 pages.
Bluttman et al., “Excel Formulas and Functions for Dummies,” 2005, Wiley Publishing, Inc., pp. 280, 284-286.
Bugzilla@Mozilla, “Bug 18726—[feature] Long-click means of invoking contextual menus not supported,” http://bugzilla.mozilla.org/show—bug.cgi?id=18726 printed Jun. 13, 2013 in 11 pages.
Canese et al., “Chapter 2: PubMed: The Bibliographic Database,” The NCBI Handbook, Oct. 2002, pp. 1-10.
Capptain—Pilot Your Apps, <http://www.capptain.com> Printed Jul. 18, 2013 in 6 pages.
Chen et al., “Bringing Order to the Web: Automatically Categorizing Search Results,” CHI 2000, Proceedings of the SIGCHI conference on Human Factors in Computing Systems, Apr. 1-6, 2000, The Hague, The Netherlands, pp. 145-152.
Chung, Chin-Wan, “Dataplex: An Access to Heterogeneous Distributed Databases,” Communications of the ACM, Association for Computing Machinery, Inc., vol. 33, No. 1, Jan. 1, 1990, pp. 70-80.
Conner, Nancy, “Google Apps: The Missing Manual,” May 1, 2008, pp. 15.
Countly Mobile Analytics, <http://count.ly/> Printed Jul. 18, 2013 in 9 pages.
Definition “Identify” downloaded Jan. 22, 2015, 1 page.
Definition “Overlay” downloaded Jan. 22, 2015, 1 page.
Delcher et al., “Identifying Bacterial Genes and Endosymbiont DNA with Glimmer,” BioInformatics, vol. 23, No. 6, 2007, pp. 673-679.
Distimo—App Analytics, <http://www.distimo.com/app-analytics> Printed Jul. 18, 2013 in 5 pages.
Dramowicz, Ela, “Retail Trade Area Analysis Using the Huff Model,” Directions Magazine, Jul. 2, 2005, http://www.directionsmag.com/articles/retail-trade-area-analysis-using-the-huff-model/123411 (10 pages).
Flurry Analytics, <http://www.flurry.com/> Printed Jul. 18, 2013 in 14 pages.
GIS-NET 3 Public—Department of Regional Planning. Planning & Zoning Information for Unincorporated LA County. Retrieved Oct. 2, 2013 from http://gis.planning.lacounty.gov/GIS-NET3—Public/Viewer.html.
Google Analytics Official Website—Web Analytics & Reporting, <http://www.google.com/analytics.index.html> Printed Jul. 18, 2013 in 22 pages.
Gorr et al., “Crime Hot Spot Forecasting: Modeling and Comparative Evaluation,” Grant 98-IJ-CX-K005, May 6, 2002, 37 pages.
Goswami, Gautam, “Quite Writly Said!,” One Brick at a Time, Aug. 21, 2005, pp. 7.
Griffith, Daniel A., “A Generalized Huff Model,” Geographical Analysis, Apr. 1982, vol. 14, No. 2, pp. 135-144.
Hansen et al. “Analyzing Social Media Networks with NodeXL: Insights from a Connected World”, Chapter 4, pp. 53-67 and Chapter 10, pp. 143-164, published Sep. 2010.
Hardesty, “Privacy Challenges: Analysis: Its Surprisingly Easy to Identify Individuals from Credit-Card Metadata,” MIT News on Campus and Around the World, MIT News Office, Jan. 29, 2015, 3 pages.
Hibbert et al., “Prediction of Shopping Behavior Using a Huff Model Within a GIS Framework,” Healthy Eating in Context, Mar. 18, 2011, pp. 16.
Hua et al., “A Multi-attribute Data Structure with Parallel Bloom Filters for Network Services” HiPC 2006, LNCS 4297, pp. 277-288, 2006.
Huff et al., “Calibrating the Huff Model Using ArcGIS Business Analyst,” ESRI, Sep. 2008, pp. 33.
Huff, David L., “Parameter Estimation in the Huff Model,” ESRI, ArcUser, Oct.-Dec. 2003, pp. 34-36.
“HunchLab: Heat Map and Kernel Density Calculation for Crime Analysis,” Azavea Journal, printed from wwvv.azavea.com/blogs/newsletter/v4i4/kernel-density-capabilities-added-to-hunchlab/ on Sep. 9, 2014, 2 pages.
Kahan et al., “Annotea: an open RDF infrastructure for shared Web annotations”, Computer Networks 39, pp. 589-608, 2002.
Keylines.com, “An Introduction to KeyLines and Network Visualization,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-White-Paper.pdf> downloaded May 12, 2014 in 8 pages.
Keylines.com, “KeyLines Datasheet,” Mar. 2014, <http://keylines.com/wp-content/uploads/2014/03/KeyLines-datasheet.pdf> downloaded May 12, 2014 in 2 pages.
Keylines.com, “Visualizing Threats: Improved Cyber Security Through Network Visualization,” Apr. 2014, <http://keylines.com/wp-content/uploads/2014/04/Visualizing-Threats1.pdf> downloaded May 12, 2014 in 10 pages.
Kitts, Paul, “Chapter 14: Genome Assembly and Annotation Process,” The NCBI Handbook, Oct. 2002, pp. 1-21.
Kontagent Mobile Analytics, <http://www.kontagent.com/> Printed Jul. 18, 2013 in 9 pages.
Liu, Tianshun, “Combining GIS and the Huff Model to Analyze Suitable Locations for a New Asian Supermarket in the Minneapolis and St. Paul, Minnesota USA,” Papers in Resource Analysis, 2012, vol. 14, pp. 8.
Madden, Tom, “Chapter 16: The BLAST Sequence Analysis Tool,” The NCBI Handbook, Oct. 2002, pp. 1-15.
Manno et al., “Introducing Collaboration in Single-user Applications through the Centralized Control Architecture,” 2010, pp. 10.
Manske, “File Saving Dialogs,” <http://www.mozilla.org/editor/ui—specs/FileSaveDialogs.html>, Jan. 20, 1999, pp. 7.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.bing.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.google.com.
Map of San Jose, CA. Retrieved Oct. 2, 2013 from http://maps.yahoo.com.
Microsoft—Developer Network, “Getting Started with VBA in Word 2010,” Apr. 2010, <http://msdn.microsoft.com/en-us/library/ff604039%28v=office.14%29.aspx> as printed Apr. 4, 2014 in 17 pages.
Microsoft Office—Visio, “About connecting shapes,” <http://office.microsoft.com/en-us/visio-help/about-connecting-shapes-HP085050369.aspx> printed Aug. 4, 2011 in 6 pages.
Microsoft Office—Visio, “Add and glue connectors with the Connector tool,” <http://office.microsoft.com/en-us/visio-help/add-and-glue-connectors-with-theconnector-tool-HA010048532.aspx?CTT=1> printed Aug. 4, 2011 in 1 page.
Mixpanel—Mobile Analytics, <https://mixpanel.com/> Printed Jul. 18, 2013 in 13 pages.
Mizrachi, Ilene, “Chapter 1: GenBank: The Nuckeotide Sequence Database,” The NCBI Handbook, Oct. 2002, pp. 1-14.
Nierman, “Evaluating Structural Similarity in XML Documents,” 2002, 6 pages.
Olanoff, Drew, “Deep Dive with the New Google Maps for Desktop with Google Earth Integration, It's More than Just a Utility,” May 15, 2013, pp. 1-6, retrieved from the internet: http://web.archive.org/web/20130515230641/http://techcrunch.com/2013/05/15/deep-dive-with-the-new-google-maps-for-desktop-with-google-earth-integration-its-more-than-just-a-utility/.
Open Web Analytics (OWA), <http://www.openwebanalytics.com/> Printed Jul. 19, 2013 in 5 pages.
Palmas et al., “An Edge-Bunding Layout for Interactive Parallel Coordinates” 2014 IEEE Pacific Visualization Symposium, pp. 57-64.
Piwik—Free Web Analytics Software. <http://piwik.org/> Printed Jul. 19, 2013 in18 pages.
Rouse, Margaret, “OLAP Cube,” <http://searchdatamanagement.techtarget.com/definition/OLAP-cube>, Apr. 28, 2012, pp. 16.
Sigrist, et al., “PROSITE, a Protein Domain Database for Functional Characterization and Annotation,” Nucleic Acids Research, 2010, vol. 38, pp. D161-D166.
Sirotkin et al., “Chapter 13: The Processing of Biological Sequence Data at NCBI,” The NCBI Handbook, Oct. 2002, pp. 1-11.
StatCounter—Free Invisible Web Tracker, Hit Counter and Web Stats, <http://statcounter.com/> Printed Jul. 19, 2013 in 17 pages.
TestFlight—Beta Testing on the Fly, <http://testflightapp.com/> Printed Jul. 18, 2013 in 3 pages.
“The FASTA Program Package,” fasta-36.3.4, Mar. 25, 2011, pp. 29.
trak.io, <http://trak.io/> printed Jul. 18, 2013 in 3 pages.
UserMetrix, <http://usermetrix.com/android-analytics> printed Jul. 18, 2013 in 3 pages.
Valentini et al., “Ensembles of Learning Machines,” M. Marinaro and R. Tagliaferri (Eds.): WIRN VIETRI 2002, LNCS 2486, pp. 3-20.
Vose et al., “Help File for ModelRisk Version 5,” 2007, Vose Software, pp. 349-353. [Uploaded in 2 Parts].
Wikipedia, “Federated Database System,” Sep. 7, 2013, retrieved from the internet on Jan. 27, 2015 http://en.wikipedia.org/w/index.php?title=Federated—database—system&oldid=571954221.
Yang et al., “HTML page Analysis Based on Visual Cues,” 2001, pp. 859-864.
Issue Notification for U.S. Appl. No. 13/917,571 dated Aug. 5, 2014.
Notice of Allowance for U.S. Appl. No. 14/102,394 dated Aug. 25, 2014.
Notice of Allowance for U.S. Appl. No. 14/108,187 dated Aug. 29, 2014.
Notice of Allowance for U.S. Appl. No. 14/135,289 dated Oct. 14, 2014.
Notice of Allowance for U.S. Appl. No. 14/268,964 dated Dec. 3, 2014.
Notice of Allowance for U.S. Appl. No. 14/473,860 dated Jan. 5, 2015.
Notice of Allowance for U.S. Appl. No. 14/294,098 dated Dec. 29, 2014.
Official Communication for New Zealand Patent Application No. 624557 dated May 14, 2014.
Official Communication for New Zealand Patent Application No. 628585 dated Aug. 26, 2014.
Official Communication for European Patent Application No. 14158861.6 dated Jun. 16, 2014.
Official Communication for New Zealand Patent Application No. 622517 dated Apr. 3, 2014.
Official Communication for New Zealand Patent Application No. 628263 dated Aug. 12, 2014.
Official Communication for Great Britain Patent Application No. 1404457.2 dated Aug. 14, 2014.
Official Communication for New Zealand Patent Application No. 627962 dated Aug. 5, 2014.
Official Communication for European Patent Application No. 14159464.8 dated Jul. 31, 2014.
Official Communication for European Patent Application No. 14159464.8 dated Aug. 20, 2014.
Official Communication for European Patent Application No. 14159464.8 dated Sep. 22, 2014.
Official Communication for New Zealand Patent Application No. 628840 dated Aug. 28, 2014.
Official Communication in New Zealand Patent Application No. 628495 dated Aug. 19, 2014.
Official Communication for Great Britain Patent Application No. 1408025.3 dated Nov. 6, 2014.
Official Communication for New Zealand Patent Application No. 628161 dated Aug. 25, 2014.
Official Communication for Great Britain Patent Application No. 1404574.4 dated Dec. 18, 2014.
Official Communication for Great Britain Patent Application No. 1411984.6 dated Dec. 22, 2014.
Official Communication for European Patent Application No. 14180281.9 dated Jan. 26, 2015.
Official Communication for European Patent Application No. 14180142.3 dated Feb. 6, 2015.
Official Communication for European Patent Application No. 14186225.0 dated Feb. 13, 2015.
Official Communication for European Patent Application No. 14189344.6 dated Feb. 20, 2015.
Official Communication for Great Britain Patent Application No. 1404499.4 dated Sep. 29, 2014.
Official Communication for Great Britain Patent Application No. 1404489.5 dated Oct. 6, 2014.
Official Communication for Great Britain Patent Application No. 1404486.1 dated Aug. 27, 2014.
Official Communication for Great Britain Patent Application No. 1404499.4 dated Aug. 20, 2014.
Official Communication for Great Britain Patent Application No. 1404489.5 dated Aug. 27, 2014.
Official Communication for New Zealand Patent Application No. 622473 dated Mar. 27, 2014.
Official Communication for New Zealand Patent Application No. 622473 dated Jun. 19, 2014.
Official Communication for U.S. Appl. No. 14/289,596 dated Jul. 18, 2014.
Official Communication for U.S. Appl. No. 14/289,599 dated Jul. 22, 2014.
Official Communication for U.S. Appl. No. 14/225,160 dated Jul. 29, 2014.
Official Communication for U.S. Appl. No. 14/268,964 dated Sep. 3, 2014.
Official Communication for U.S. Appl. No. 14/294,098 dated Aug. 15, 2014.
Official Communication for U.S. Appl. No. 14/294,098 dated Nov. 6, 2014.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 23, 2014.
Official Communication for U.S. Appl. No. 14/306,154 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/306,147 dated Sep. 9, 2014.
Official Communication for U.S. Appl. No. 14/319,765 dated Nov. 25, 2014.
Official Communication for U.S. Appl. No. 14/323,935 dated Nov. 28, 2014.
Official Communication for U.S. Appl. No. 14/326,738 dated Dec. 2, 2014.
Official Communication for U.S. Appl. No. 14/225,160 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/504,103 dated Feb. 5, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Jan. 26, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Feb. 4, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Feb. 18, 2015.
Official Communication for U.S. Appl. No. 14/451,221 dated Oct. 21, 2014.
Official Communication for U.S. Appl. No. 14/463,615 dated Nov. 13, 2014.
Official Communication for U.S. Appl. No. 13/827,491 dated Dec. 1, 2014.
Official Communication for U.S. Appl. No. 14/479,863 dated Dec. 26, 2014.
Official Communication for U.S. Appl. No. 14/319,161 dated Jan. 23, 2015.
Official Communication for U.S. Appl. No. 14/483,527 dated Jan. 28, 2015.
Official Communication for U.S. Appl. No. 14/463,615 dated Jan. 28, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Feb. 20, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Feb. 27, 2015.
“A Word About Banks and the Laundering of Drug Money,” Aug. 18, 2012, http://www.golemxiv.co.uk/2012/08/a-word-about-banks-and-the-laundering-of-drug-money/.
“Money Laundering Risks and E-Gaming: A European Overview and Assessment,” 2009, http://www.cf.ac.uk/socsi/resources/Levi—Final—Money—Laundering—Risks—egaming.pdf.
“Potential Money Laundering Warning Signs,” snapshot taken 2003, https://web.archive.org/web/20030816090055/http:/finsolinc.com/ANTI-MONEY%20LAUNDERING%20TRAINING%20GUIDES.pdf.
“Refresh CSS Ellipsis When Resizing Container—Stack Overflow,” Jul. 31, 2013, retrieved from internet http://stackoverflow.com/questions/17964681/refresh-css-ellipsis-when-resizing-container, retrieved on May 18, 2015.
“Using Whois Based Geolocation and Google Maps API for Support Cybercrime Investigations,” http://wseas.us/e-library/conferences/2013/Dubrovnik/TELECIRC/TELECIRC-32.pdf.
About 80 Minutes, “Palantir in a Number of Parts—Part 6—Graph,” Mar. 21, 2013, pp. 1-6.
Alur et al., “Chapter 2: IBM InfoSphere DataStage Stages,” IBM InfoSphere DataStage Data Flow and Job Design, Jul. 1, 2008, pp. 35-137.
Amnet, “5 Great Tools for Visualizing Your Twitter Followers,” posted Aug. 4, 2010, http://www.amnetblog.com/component/content/article/115-5-grate-tools-for-visualizing-your-twitter-followers.html.
Boyce, Jim, “Microsoft Outlook 2010 Inside Out,” Aug. 1, 2010, retrieved from the internet https://capdtron.files.wordpress.com/2013/01/outlook-2010-inside—out.pdf.
Celik, Tantek, “CSS Basic User Interface Module Level 3 (CSS3 UI),” Section 8 Resizing and Overflow, Jan. 17, 2012, retrieved from internet http://www.w3.org/TR/2012/WD-css3-ui-20120117/#resizing-amp-overflow retrieved on May 18, 2015.
Chaudhuri et al., “An Overview of Business Intelligence Technology,” Communications of the ACM, Aug. 2011, vol. 54, No. 8.
Cohn, et al., “Semi-supervised clustering with user feedback,” Constrained Clustering: Advances in Algorithms, Theory, and Applications 4.1 (2003): 17-32.
Gesher, Ari, “Palantir Screenshots in the Wild: Swing Sightings,” The Palantir Blog, Sep. 11, 2007, pp. 1-12.
Gu et al., “Record Linkage: Current Practice and Future Directions,” Jan. 15, 2004, pp. 32.
Hogue et al., “Thresher: Automating the Unwrapping of Semantic Content from the World Wide Web,” 14th International Conference on World Wide Web, WWW 2005: Chiba, Japan, May 10-14, 2005, pp. 86-95.
Huang et al., “Systematic and Integrative Analysis of Large Gene Lists Using DAVID Bioinformatics Resources,” Nature Protocols, 4.1, 2008, 44-57.
Li et al., “Interactive Multimodal Visual Search on Mobile Device,” IEEE Transactions on Multimedia, vol. 15, No. 3, Apr. 1, 2013, pp. 594-607.
Localytics—Mobile App Marketing & Analytics, http://www.localytics.com/ Printed Jul. 18, 2013 in 12 pages.
Map Builder, “Rapid Mashup Development Tool for Google and Yahoo Maps!” http://web.archive.org/web/20090626224734/http://www.mapbuilder.net/ printed Jul. 20, 2012 in 2 pages.
Nolan et al., “MCARTA: A Malicious Code Automated Run-Time Analysis Framework,” Homeland Security, 2012 IEEE Conference on Technologies for, Nov. 13, 2012, pp. 13-17.
Palantir Technologies, “Palantir Labs—Timeline,” Oct. 1, 2010, retrieved from the internet https://www.youtube.com/watch?v=JCgDW5bru9M.
Perdisci et al., “Behavioral Clustering of HTTP-Based Malware and Signature Generation Using Malicious Network Traces,” USENIX, Mar. 18, 2010, pp. 1-14.
QUEST, “Toad for Oracle 11.6—Guide to Using Toad,” Sep. 24, 2012, pp. 1-162.
Shi et al., “A Scalable Implementation of Malware Detection Based on Network Connection Behaviors,” 2013 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, IEEE, Oct. 10, 2013, pp. 59-66.
Symantec Corporation, “E-Security Begins with Sound Security Policies,” Announcement Symantec, Jun. 14, 2001.
Thompson, Mick, “Getting Started with GEO,” Getting Started with GEO, Jul. 26, 2011.
Umagandhi et al., “Search Query Recommendations Using Hybrid User Profile with Query Logs,” International Journal of Computer Applications, vol. 80, No. 10, Oct. 1, 2013, pp. 7-18.
Wang et al., “Research on a Clustering Data De-Duplication Mechanism Based on Bloom Filter,” IEEE 2010, 5 pages.
Wikipedia, “Multimap,” Jan. 1, 2013, https://en.wikipedia.org/w/index.php?title-Multimap&oldid=530800748.
Wright et al., “Palantir Technologies VAST 2010 Challenge Text Records—Investigations into Arms Dealing,” Oct. 29, 2010, pp. 1-10.
Notice of Acceptance for Australian Patent Application No. 2014250678 dated Oct. 7, 2015.
Notice of Allowance for U.S. Appl. No. 12/556,318 dated Nov. 2, 2015.
Notice of Allowance for U.S. Appl. No. 14/148,568 dated Aug. 26, 2015.
Notice of Allowance for U.S. Appl. No. 14/192,767 dated Dec. 16, 2014.
Notice of Allowance for U.S. Appl. No. 14/225,084 dated May 4, 2015.
Notice of Allowance for U.S. Appl. No. 14/319,161 dated May 4, 2015.
Notice of Allowance for U.S. Appl. No. 14/323,935 dated Oct. 1, 2015.
Notice of Allowance for U.S. Appl. No. 14/326,738 dated Nov. 18, 2015.
Notice of Allowance for U.S. Appl. No. 14/473,552 dated Jul. 24, 2015.
Notice of Allowance for U.S. Appl. No. 14/479,863 dated Mar. 31, 2015.
Notice of Allowance for U.S. Appl. No. 14/486,991 dated May 1, 2015.
Notice of Allowance for U.S. Appl. No. 14/504,103 dated May 18, 2015.
Notice of Allowance for U.S. Appl. No. 14/552,336 dated Nov. 3, 2015.
Notice of Allowance for U.S. Appl. No. 14/616,080 dated Apr. 2, 2015.
Notice of Allowance for U.S. Appl. No. 14/746,671 dated Jan. 21, 2016.
Official Communication for Australian Patent Application No. 2014201511 dated Feb. 27, 2015.
Official Communication for Australian Patent Application No. 2014202442 dated Mar. 19, 2015.
Official Communication for Australian Patent Application No. 2014210604 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014210614 dated Jun. 5, 2015.
Official Communication for Australian Patent Application No. 2014213553 dated May 7, 2015.
Official Communication for Australian Patent Application No. 2014250678 dated Jun. 17, 2015.
Official Communication for European Patent Application No. 14180321.3 dated Apr. 17, 2015.
Official Communication for European Patent Application No. 14180432.8 dated Jun. 23, 2015.
Official Communication for European Patent Application No. 14187739.9 dated Jul. 6, 2015.
Official Communication for European Patent Application No. 14187996.5 dated Feb. 12, 2015.
Official Communication for European Patent Application No. 14189347.9 dated Mar. 4, 2015.
Official Communication for European Patent Application No. 14189802.3 dated May 11, 2015.
Official Communication for European Patent Application No. 14191540.5 dated May 27, 2015.
Official Communication for European Patent Application No. 14197879.1 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14197895.7 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14197938.5 dated Apr. 28, 2015.
Official Communication for European Patent Application No. 14199182.8 dated Mar. 13, 2015.
Official Communication for European Patent Application No. 14200246.8 dated May 29, 2015.
Official Communication for European Patent Application No. 14200298.9 dated May 13, 2015.
Official Communication for European Patent Application No. 15155845.9 dated Oct. 6, 2015.
Official Communication for European Patent Application No. 15155846.7 dated Jul. 8, 2015.
Official Communication for European Patent Application No. 15165244.3 dated Aug. 27, 2015.
Official Communication for European Patent Application No. 15175106.2 dated Nov. 5, 2015.
Official Communication for European Patent Application No. 15175151.8 dated Nov. 25, 2015.
Official Communication for European Patent Application No. 15181419.1 dated Sep. 29, 2015.
Official Communication for European Patent Application No. 15183721.8 dated Nov. 23, 2015.
Official Communication for European Patent Application No. 15184764.7 dated Dec. 14, 2015.
Official Communication for European Patent Application No. 16157173.2 dated Jul. 5, 2016.
Official Communication for Great Britain Patent Application No. 1404486.1 dated May 21, 2015.
Official Communication for Great Britain Patent Application No. 1404489.5 dated May 21, 2015.
Official Communication for Great Britain Patent Application No. 1404499.4 dated Jun. 11, 2015.
Official Communication for Netherlands Patent Application No. 2012417 dated Sep. 18, 2015.
Official Communication for Netherlands Patent Application No. 2012421 dated Sep. 18, 2015.
Official Communication for Netherlands Patent Application No. 2012437 dated Sep. 18, 2015.
Official Communication for Netherlands Patent Application No. 2012438 dated Sep. 21, 2015.
Official Communication for Netherlands Patent Application No. 2013306 dated Apr. 24, 2015.
Official Communication for New Zealand Patent Application No. 622513 dated Apr. 3, 2014.
Official Communication for U.S. Appl. No. 12/556,318 dated Jul. 2, 2015.
Official Communication for U.S. Appl. No. 13/247,987 dated Apr. 2, 2015.
Official Communication for U.S. Appl. No. 13/247,987 dated Sep. 22, 2015.
Official Communication for U.S. Appl. No. 13/827,491 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 13/827,491 dated Oct. 9, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Mar. 4, 2015.
Official Communication for U.S. Appl. No. 13/831,791 dated Aug. 6, 2015.
Official Communication for U.S. Appl. No. 13/835,688 dated Jun. 17, 2015.
Official Communication for U.S. Appl. No. 13/839,026 dated Aug. 4, 2015.
Official Communication for U.S. Appl. No. 14/134,558 dated Oct. 7, 2015.
Official Communication for U.S. Appl. No. 14/141,252 dated Oct. 8, 2015.
Official Communication for U.S. Appl. No. 14/148,568 dated Mar. 26, 2015.
Official Communication for U.S. Appl. No. 14/196,814 dated May 5, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 10, 2014.
Official Communication for U.S. Appl. No. 14/225,006 dated Sep. 2, 2015.
Official Communication for U.S. Appl. No. 14/225,006 dated Dec. 21, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 11, 2015.
Official Communication for U.S. Appl. No. 14/225,084 dated Sep. 2, 2014.
Official Communication for U.S. Appl. No. 14/225,084 dated Jan. 4, 2016.
Official Communication for U.S. Appl. No. 14/225,160 dated Feb. 11, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated Aug. 12, 2015.
Official Communication for U.S. Appl. No. 14/225,160 dated May 20, 2015.
Official Communication for U.S. Appl. No. 14/289,596 dated Apr. 30, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated May 29, 2015.
Official Communication for U.S. Appl. No. 14/289,599 dated Sep. 4, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Sep. 14, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Dec. 24, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/306,138 dated Dec. 3, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Feb. 19, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Dec. 24, 2015.
Official Communication for U.S. Appl. No. 14/306,147 dated Aug. 7, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Mar. 11, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated May 15, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Nov. 16, 2015.
Official Communication for U.S. Appl. No. 14/306,154 dated Jul. 6, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/319,765 dated Jun. 16, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 14/323,935 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Jul. 31, 2015.
Official Communication for U.S. Appl. No. 14/326,738 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/463,615 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/463,615 dated May 21, 2015.
Official Communication for U.S. Appl. No. 14/463,615 dated Dec. 9, 2015.
Official Communication for U.S. Appl. No. 14/473,552 dated Feb. 24, 2015.
Official Communication for U.S. Appl. No. 14/483,527 dated Jun. 22, 2015.
Official Communication for U.S. Appl. No. 14/483,527 dated Oct. 28, 2015.
Official Communication for U.S. Appl. No. 14/486,991 dated Mar. 10, 2015.
Official Communication for U.S. Appl. No. 14/490,612 dated Aug. 18, 2015.
Official Communication for U.S. Appl. No. 14/504,103 dated Mar. 31, 2015.
Official Communication for U.S. Appl. No. 14/552,336 dated Jul. 20, 2015.
Official Communication for U.S. Appl. No. 14/562,524 dated Nov. 10, 2015.
Official Communication for U.S. Appl. No. 14/562,524 dated Sep. 14, 2015.
Official Communication for U.S. Appl. No. 14/571,098 dated Nov. 10, 2015.
Official Communication for U.S. Appl. No. 14/571,098 dated Mar. 11, 2015.
Official Communication for U.S. Appl. No. 14/571,098 dated Aug. 24, 2015.
Official Communication for U.S. Appl. No. 14/571,098 dated Aug. 5, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated Aug. 19, 2015.
Official Communication for U.S. Appl. No. 14/579,752 dated May 26, 2015.
Official Communication for U.S. Appl. No. 14/631,633 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated Oct. 16, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated May 18, 2015.
Official Communication for U.S. Appl. No. 14/639,606 dated Jul. 24, 2015.
Official Communication for U.S. Appl. No. 14/676,621 dated Oct. 29, 2015.
Official Communication for U.S. Appl. No. 14/676,621 dated Jul. 30, 2015.
Official Communication for U.S. Appl. No. 14/726,353 dated Sep. 10, 2015.
Official Communication for U.S. Appl. No. 14/746,671 dated Nov. 12, 2015.
Official Communication for U.S. Appl. No. 14/800,447 dated Dec. 10, 2015.
Official Communication for U.S. Appl. No. 14/813,749 dated Sep. 28, 2015.
Official Communication for U.S. Appl. No. 14/842,734 dated Nov. 19, 2015.
Restriction Requirement for U.S. Appl. No. 13/839,026 dated Apr. 2, 2015.
“A Tour of Pinboard,” <http://pinboard.in/tour> as printed May 15, 2014 in 6 pages.
“BackTult—JD Edwards One World Version Control System,” printed Jul. 23, 2007 in 1 page.
Delicious, <http://delicious.com/> as printed May 15, 2014 in 1 page.
Geiger, Jonathan G., “Data Quality Management, The Most Critical Initiative You Can Implement,” Data Warehousing, Management and Quality, Paper 098-29, SUGI 29, Intelligent Solutions, Inc., Bounder, CO, pp. 14, accessed Oct. 3, 2013.
Kahan et al., “Annotea: an Open RDF Infastructure for Shared Web Annotations”, Computer Networks, Elsevier Science Publishers B.V., vol. 39, No. 5, dated Aug. 5, 2002, pp. 589-608.
Klemmer et al., “Where Do Web Sites Come From? Capturing and Interacting with Design History,” Association for Computing Machinery, CHI 2002, Apr. 20-25, 2002, Minneapolis, MN, pp. 8.
Kokossi et al., “D7-Dynamic Ontoloty Management System (Design),” Information Societies Technology Programme, Jan. 10, 2002, pp. 1-27.
Miklau et al., “Securing History: Privacy and Accountability in Database Systems,” 3rd Biennial Conference on Innovative Data Systems Research (CIDR), Jan. 7-10, 2007, Asilomar, California, pp. 387-396.
Morrison et al., “Converting Users to Testers: An Alternative Approach to Load Test Script Creation, Parameterization and Data Corellation,” CCSC: Southeastern Conference, JCSC 28, Dec. 2, 2012, pp. 188-196.
Niepert et al., “A Dynamic Ontology for a Dynamic Reference Work”, Joint Conference on Digital Libraries, Jun. 17-22, 2007, Vancouver, British Columbia, Canada, pp. 1-10.
Nivas, Tuli, “Test Harness and Script Design Principles for Automated Testing of non-GUI or Web Based Applications,” Performance Lab, Jun. 2011, pp. 30-37.
Palantir, “Extracting and Transforming Data with Kite,” Palantir Technologies, Inc., Copyright 2010, pp. 38.
Palantir, “Kite,” https://docs.palantir.com/gotham/3.11.1.0/adminreference/datasources.11 printed Aug. 30, 2013 in 2 pages.
Palantir, “Kite Data-Integration Process Overview,” Palantir Technologies, Inc., Copyright 2010, pp. 48.
Palantir, “Kite Operations,” Palantir Technologies, Inc., Copyright 2010, p. 1.
Palantir, “The Repository Element,” https://docs.palantir.com/gotham/3.11.1.0/dataguide/kite—config—file.04 printed Aug. 30, 2013 in 2 pages.
Palantir, “Write a Kite Configuration File in Eclipse,” Palantir Technologies, Inc., Copyright 2010, pp. 2.
Palantir, https://docs.palantir.com/gotham/3.11.1.0/dataguide/baggage/KiteSchema.xsd printed Apr. 4, 2014 in 4 pages.
Palermo, Christopher J., “Memorandum,” [Disclosure relating to U.S. Appl. No. 13/916,447, filed Jun. 12, 2013, and related applications], Jan. 31, 2014 in 3 pages.
Wollrath et al., “A Distributed Object Model for the Java System,” Conference on Object-Oriented Technologies and Systems, Jun. 17-21, 1996, pp. 219-231.
Notice of Allowance for U.S. Appl. No. 14/044,800 dated Sep. 2, 2014.
Notice of Allowance for U.S. Appl. No. 14/508,696 dated Jul. 27, 2015.
Notice of Allowance for U.S. Appl. No. 14/533,433 dated Sep. 1, 2015.
Official Communication for U.S. Appl. No. 14/148,568 dated Oct. 22, 2014.
Official Communication for U.S. Appl. No. 14/508,696 dated Mar. 2, 2015.
Official Communication for U.S. Appl. No. 14/025,653 dated Oct. 6, 2015.
Official Communication for U.S. Appl. No. 14/526,066 dated Jan. 21, 2016.
Official Communication for U.S. Appl. No. 14/025,653 dated Mar. 3, 2016.
Official Communication for Australian Patent Application No. 2014201506 dated Feb. 27, 2015.
Official Communication for Australian Patent Application No. 2014201507 dated Feb. 27, 2015.
Official Communication for Australian Patent Application No. 2014201580 dated Feb. 27, 2015.
Official Communication for Canadian Patent Application No. 2666364 dated Jun. 4, 2012.
Official Communication for European Patent Application No. 14158958.0 dated Jun. 3, 2014.
Official Communication for European Patent Application No. 14158977.0 dated Jun. 10, 2014.
Official Communication for European Patent Application No. 14159629.6 dated Jul. 31, 2014.
Official Communication for European Patent Application No. 14158977.0 dated Apr. 16, 2015.
Official Communication for European Patent Application No. 14158958.0 dated Apr. 16, 2015.
Official Communication for Great Britain Patent Application No. 1404479.6 dated Aug. 12, 2014.
Official Communication for Great Britain Patent Application No. 1413935.6 dated Jan. 27, 2015.
Official Communication for Great Britain Patent Application No. 1404479.6 dated Jul. 9, 2015.
Official Communication for Great Britain Patent Application No. 1413935.6 dated Dec. 21, 2015.
Official Communication for Israel Patent Application No. 198253 dated Nov. 24, 2014.
Official Communication for New Zealand Patent Application No. 622389 dated Mar. 20, 2014.
Official Communication for New Zealand Patent Application No. 622404 dated Mar. 20, 2014.
Official Communication for New Zealand Patent Application No. 622414 dated Mar. 24, 2014.
Official Communication for New Zealand Patent Application No. 622497 dated Mar. 26, 2014.
Official Communication for New Zealand Patent Application No. 622484 dated Apr. 2, 2014.
Official Communication for New Zealand Patent Application No. 622497 dated Jun. 19, 2014.
Related Publications (1)
Number Date Country
20160246784 A1 Aug 2016 US