System, method and computer program for creating and manipulating data structures using an interactive graphical interface

Information

  • Patent Grant
  • 10146843
  • Patent Number
    10,146,843
  • Date Filed
    Tuesday, January 5, 2016
    8 years ago
  • Date Issued
    Tuesday, December 4, 2018
    5 years ago
Abstract
A computer network-implemented method for displaying a tag cloud representing a data structure. The data structure includes data entities and representations of relationships between the data entities. The method enables one or more users to interact with the data structure, and via the data structure, to interact with one another. In an example embodiment, the method comprises, via a computer processor, populating a display of a user interface showing a tag cloud populated with images of data entities from the data structure, such images, and in response to user input to the interface, associating one or more visual properties of the user interface with one or more of said relationships between data entities in the data structure.
Description
FIELD OF THE INVENTION

The invention relates generally to an interface for creating and manipulating data structures. Aspects of the invention relate more specifically to a system, method and computer program product for creating, visualizing, and manipulating (e.g., inputting to or changing) a data structure using an intuitive and interactive graphical interface.


BACKGROUND OF THE INVENTION

Knowledge can be represented using various types of data structures, including graphs. One such graph is known as a semantic network. A semantic network is a directed graph consisting of vertices, which represent concepts, and edges, which represent semantic relations between the concepts.


A semantic network can become unwieldy as it grows.


Tags can be used to represent concepts in the semantic network. Sets of tags, in a visualization known as “tag clouds”, can be used to represent relationships between concepts. Tag clouds are a familiar data visualization device on the Internet. Tag clouds are commonly used to represent tags in a meaningful way, for example to describe to a user the prevalence of tags in blogs and other Internet resources. Properties of words, such as a size, weight or colour, may represent properties of the underlying data. A tag cloud may be generated either manually or using computerized means.



FIGS. 1A to 1C illustrate examples of tag clouds. As shown in FIG. 1A, for example, a cloud may comprise several differently sized tags wherein larger tags in the cloud represent a greater (e.g., higher volume) use of the tag in a blog. As shown in FIGS. 1B and 1C, for example, the sizes of the tags may signify the importance of concepts relative to a context, importance being determined by a measure such as frequency of occurrence.


Information may be encoded in tag properties (such as size, weight or colour) but absolute and relative position is virtually meaningless in a tag cloud. The tags are typically single words, which may be ordered alphabetically or otherwise. The words may be aligned on a baseline or arranged in some other way, but this is typically done to conserve space or to obtain different visual effects.


However, these tag clouds are not interactive as they merely represent information graphically without any means for feedback from a user. With a tag cloud there is no means to manipulate the relationships between the tags.


The prior art does not discuss ways in which to present a simple visual representation of a hierarchical or polyhierarchical data structure (such as a taxonomy of terms or a semantic network) so as to enable the average computer user to create, visualize or manipulate the data structure. While manual entry of new concepts and relationships has been contemplated to a limited extent in the prior art, what has not been disclosed in the prior art is the use of a tag cloud to create (i.e., record) concepts and automatically infer relationships to existing concepts represented by the tags. What has also not been disclosed is a convenient way in which to visualize and manipulate relationships between the concepts represented by the tags. In other words, the prior art does not teach using the tag cloud as an input device to create and alter the represented semantic network or other data structure.


U.S. patent application Ser. No. 11/548,894 to Lewis et al. discloses a tag cloud that is presented to a user where each tag can lead to n-layers of relevant information.


U.S. patent application Ser. No. 11/540,628 to Hoskinson discloses a tag cloud that is computer generated in response to a search query. The tags, containing subject representation or labels, are linked to associated websites from where the information for generating the cloud is initially collected.


U.S. patent application Ser. No. 11/533,058 to Blanchard et al. discloses customizing a display of a presented tag cloud. These clouds are customizable in terms of their attributes such as font color, font size, borders, 3D-depth, shadowing, and so on. While changes in all these attributes contribute to visual display of the tags in the tag cloud, there is no corresponding material affect on the information represented by the tag cloud.


None of the above applications discuss ways in which to present a data structure to a user so as to visually represent relationships that may exist between concepts represented by the tags and enable the manipulation of the data structure by the user using the tags.


PCT/US2007/025873 to Lindermann, et al. discloses enabling a user to input, store and output in a graphical user interface concepts expressed as a word or combination of words and relationships between these concepts. The user provides the concepts to a thought engine and specifies the type or nature of relationship between concepts. A user builds and shares the generated semantic network.


Lindermann et al. is directed to enabling users that do not understand structured data to insert the data into the structure. The user inserts the data and the relationships explicitly and, therefore, must learn how the relationships are made. There is no automation provided for establishing relationships based on ways in which the user views the data. While Lindermann et al. discuss a user classifying the types of new relationships, there is no discussion of simple ways in which to establish the relationship with minimal user input. There is also no discussion at all of ways in which to easily manipulate existing relationships between concepts in a semantic network.


Therefore, what is required is a means and method by which to enable the average computer user to create, visualize or manipulate a data structure (particularly a hierarchical, polyhierarchical, associative or equivalence) data structure, using a tag cloud.


Also lacking heretofore is a way for users to interact with each other so that a community can create, visualize and/or manipulate such a data structure using a tag cloud, especially in a way that tracks changes made to the data structure over time, who made each change and when the change was made. A means and method which allows such user interaction would be very useful.


SUMMARY

A user interface is described for use in a computer network implemented method and apparatus for displaying a data structure and enabling one or more users to interact with the data structure by means of the user interface, the data structure including data entities and relationships between the data entities. One or more user context properties are associable with the data entities such that, for example, the data structure may be reflective of a user context. In this regard, the user context may, but will not necessarily, result in relationships between data entities being hierarchical, polyhierarchical, associative or equivalence relationships. The method comprises: populating the user interface with data entities from the data structure by enabling the one or more users to associate one or more visual properties of the user interface with the user context associable properties of the data structure, thereby enabling the one or more users to interact with the data structure by means of the user interface, using one or more computer processors.


The data entities are displayed by means of the user interface, enabling one or more users to interact with the data structure. The user interface enables a user to define the relationships between new and existing data entities without having to understand the nature of the relationships in the underlying data structure and without explicitly specifying the desired changes to the relationships.


Some embodiments discussed herein also provide a system configured to execute a computer-implemented method for displaying a data structure including data entities and relationships between the data entities, wherein one or more user context properties are associable with the data entities, and enabling one or more users to interact with the data structure including by creating the data structure and/or editing the data structure, the system comprising one or more computer devices including or being linked to: (a) at least one display, and (b) a user interface utility operable to: (i) present to one or more users a user interface, by means of the at least one display; and (b) populate the user interface with data entities from the data structure by enabling the one or more users to associate one or more visual properties of the user interface with the user context associable properties of the data structure, thereby enabling the one or more users to interact with the data structure by means of the user interface.


Some embodiments further provide a computer program product containing executable computer program instructions which, when executed by one or more computers having a display, presents a user interface corresponding to a data structure and enables one or more users to interact with the data structure including by creating the data structure, the data structure including data entities and relationships between the data entities, wherein one or more user context properties are associable with the data entities, the interaction with the data structure on a user interface including: populating the user interface with data entities from the data structure by enabling the one or more users to associate one or more visual properties of the user interface with the user context associable properties of the data structure, thereby enabling the one or more users to interact with the data structure by means of the user interface, using one or more computer processors.


According to some embodiments (whether method, system or computer program product), the data entities are displayed by means of the user interface, enabling one or more users to interact with the data structure. The user interface, in turn, enables a user to define the relationships between new and existing data entities without having to understand the nature of the relationships in the underlying data structure and without explicitly specifying the desired changes to the relationships.


While manual input to the user interface suffices in many situations, once a mass of processed data related to a user is archived, it is desirable to interactively link new inputs by the user and data processed by other users. Just setting aside vast amounts of curated data for a user does not take advantage of the great productive and useful aspects of a body of networked data that can enable a user(s) to connect his or her current activity with what has already been curated by that user and others, as well. Therefore, in some embodiments, in order to facilitate achievement of this potential, a navigation means is provided to permit a user to move smoothly from a global view of the thought cloud to a more local view.


In these respects, before explaining at least one embodiment of the invention in detail, it is to be understood that the invention is not limited in its application to the details of construction and to the arrangements of the components set forth in the following description or illustrated in the drawings. The invention is capable of other embodiments and of being practiced and carried out in various ways. Also, it is to be understood that the phraseology and terminology employed herein are for the purpose of description and should not be regarded as limiting.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A illustrates a tag cloud of the prior art.



FIG. 1B illustrates another tag cloud of the prior art.



FIG. 1C illustrates another tag cloud of the prior art.



FIG. 2 illustrates an example of a thought cloud as intended herein.



FIG. 3A illustrates inserting a new concept to the thought cloud.



FIG. 3B illustrates changing a label corresponding to a new concept of the thought cloud.



FIG. 3C illustrates a changed label corresponding to a new concept of the thought cloud.



FIG. 4A illustrates a label corresponding to an existing concept of the thought cloud.



FIG. 4B illustrates a means for resizing the label's font in the thought cloud.



FIG. 4C illustrates a resized font of the label.



FIG. 5A illustrates a label of the thought cloud to be repositioned.



FIG. 5B illustrates repositioning the label.



FIG. 5C illustrates a repositioned label.



FIG. 6A illustrates a means for changing content related to a concept.



FIG. 6B illustrates content related to a concept.



FIG. 7A illustrates a thought cloud.



FIG. 7B illustrates suggested concepts shown in the thought cloud.



FIG. 8 illustrates an example of a hierarchical semantic network operable with some embodiments implementing aspects of the system and method disclosed herein.



FIG. 9 illustrates an example of a user interface (UI) such as might be used to implement some aspects of the invention, in some embodiments, to provide a way for users to navigate between global and local views of a thought cloud and to manipulate, and explore relationships among, data in an underling semantic network's data structure,



FIG. 10 is a flow chart showing an example of how the UI of FIG. 9 may be used to implement some of the methodology discussed herein.



FIG. 11 is a flow chart showing an example of how the UI of FIG. 9 may be used to support collaboration via a shared global workspace view of a thought (tag) cloud.



FIG. 12 is a flow chart showing an example of how the UI of FIG. 9 may be used by multiple users to jointly curate concepts relating to a root concept.





DETAILED DESCRIPTION
Definitions

Data structure includes a means of organizing and accessing an information store comprised of knowledge representations based on relational links between data entities in the data structure.


Data entity means a knowledge representation present with a data structure.


Relationship includes hierarchical, polyhierarchical, associative, equivalence or other relationship between data entities in a data structure.


User context property means any quantitative or qualitative aspects of a concept that a user wants to portray in a data structure using an interface as taught herein, including for example relative topical importance, time, chronology, and/or physical attributes such as temperature, pressure, etc.


Concept, in the context of a semantic network, includes an embodiment of a data entity wherein each concept may represent a thought of a user.


Active concept means a concept under consideration.


Depth, in the generation of a semantic network, means a particular number of relationships distanced from the active concept within which to generate or display the semantic network.


Overview

Embodiments described herein provide a user interface to create, visualize and manipulate a data structure (e.g., for a semantic network). The user interface provides an input means to the data structure and, to a degree, a mechanism for viewing the data structure or a portion thereof. The data structure may comprise data entities and relationships between the data entities. One or more user context properties may be associable with the data entities such that, for example, the data structure may be reflective of a user context. In this regard, the user context may, but not necessarily, result in relationships between data entities being hierarchical, polyhierarchical, associative or equivalence.


The data entities are displayed by means of the user interface, enabling one or more users to interact with the data structure. The user interface enables a user to define the relationships between new and existing data entities without having to understand the nature of the relationships in the underlying data structure and without explicitly specifying the desired changes to the relationships.


The data structure may be a semantic network, wherein the data entities may be concepts. The semantic network may be provided by a semantic engine, for example as provided in PCT/CA2009/000567 to Sweeney et al. The semantic network may be variously structured, including hierarchical, polyhierarchical, cyclic, or acyclic data structures.


The user interface is designed for enabling a user to easily create, visualize and manipulate the data structure. Specifically, the data structure is represented in an easy to understand and intuitive format and enables the user to seamlessly manipulate relationships in the semantic network. The user does not have to understand the nature of the relationships in the underlying data structure or even have knowledge of the existence of relationships between any of its concepts. Rather, the relationships are defined based on the way in which the user thinks about the concepts in relation to one another.


For example, where the data structure is a semantic network, the user interface enables the creation and management of a contextual or graphical representation of the semantic network. The user interface is populated with interface components that are matched or correspond to the concepts and relationships. The interface components may, for example, be labels corresponding to representations of the concepts.


The user interface enables a layperson computer user to interact with a semantic engine for generating and refining a semantic network (which for example might include managing the relationships between concepts at one or more relationship levels in the semantic network and adding, editing or removing concepts).


The relationships are managed implicitly through the user's interaction with the user interface rather than requiring the user to explicitly define the changes to relationships in the semantic network. The labels or other interface components populated on the user interface are mapped to the relationships. The relationships are inferred from the mapping of the concepts on the user interface as managed by the user, which may reflect the way in which the user thinks about the concepts in relation to one another. The user can also create, edit or delete concepts from the semantic network by manipulating the interface components. By managing this contextual or graphical representation of the semantic network, the user is actually causing the invention to automatically create or manipulate the relationships of the semantic network.


The representation may include a plurality of dimensions in visual space for mapping one or more various properties or attributes relating to the data entities. The one or more properties may, for example, be visual properties including position, size, colour, distance, typeface, underlining, outlining, weight, gradient, time, and any other visually meaningful attribute.


The user interface may be understood as a thought cloud, that is, a cloud of tags in which a user's thoughts are manifested. The thought cloud is operable to display to a user a plurality of labels that reflect concepts that are related in the underlying semantic network, and are arranged as labels in the thought cloud based on their relationships. In this way, various user context properties of the underlying data structure and the relationships between entities in the data structure may be represented in the visualization.


A user may interact with the user interface to further create, visualize and manipulate the semantic network, for example to manage the semantic network in such a way that it reflects the way in which the user thinks about the concepts in relation to one another. The user may be provided with tools by the user interface including adding concepts to the semantic network and manipulating the relationships in the semantic network. In this way, a user is not burdened with managing, or even having an understanding of, the complexities of the corresponding data transformation formalisms that are being processed in accordance with the user's interactions with the semantic network. There is no requirement that the user interface explicitly display to the user the relationships among the concepts comprising the semantic network, however it may.


Embodiments disclosed herein also comprise a user interface for a semantic network in which users are provided with a plurality of suggested labels representing concepts in the semantic network that are potentially related to one or more labels input by the user to the user interface. The semantic network may comprise concepts and relationships between concepts. The user can optionally approve the suggested labels, whose mapping implicitly results in inferred relationships to other concepts in the semantic network, enabling the creation of new relationships from the existing concepts to the concept represented by the suggested label.


In the description that follows, a semantic network is used as an example data structure, and a thought cloud is used as an example user interface for the semantic network. However, it should be understood that the data structure could be any data structure comprising data entities and relationships between the data entities, wherein user context properties are associable with the data entities. Furthermore, it should be understood that the term “thought cloud” is used for convenience only.


Thought Cloud


Embodiments disclosed herein also comprise a user interface to create, visualize and manipulate a data structure that may, for example, be a semantic network comprising concepts and hierarchical or polyhierarchical relationships between the concepts. The semantic network, for example, may be provided by a semantic engine, for example as provided in PCT/CA2009/000567 to Sweeney et al., which is incorporated by reference, and may be initiated from an active concept. The user interface provides an input means to the semantic network that enables a user to manipulate relationships between new and existing concepts without having to understand the nature of the relationships in the underlying semantic network and without explicitly specifying the desired changes to the relationships.


The user interface enables the creation of a contextual or graphical representation of the semantic network. The user interface may be a thought cloud. The thought cloud displays to a user one or more interface components, that may for example be labels, corresponding to concepts that are hierarchically or polyhierarchically related in the underlying semantic network. The labels may have one or more properties wherein each property can be contextually or graphically represented. These properties enable each of the concepts to be represented by the user interface in an organized manner. This enables the user interface to represent the relationships without explicitly displaying the relationships to the user, so as to prevent confusion of users that do not have an understanding of the nature of existence of the relationships.


Whereas tag clouds are not interactive, a thought cloud is. Although tag clouds have been used as output devices, a thought cloud can also be used as an input device for enabling a user to create, visualize and manipulate the semantic network. The user may be provided with tools by the user interface including adding concepts to or modifying concepts in the semantic network and means by which to manipulate the hierarchical or polyhierarchical relationships in the semantic network without requiring the user to know it is doing so. In this way, a user is not burdened with managing, or even having knowledge of, the complexities of the corresponding data transformation formalisms that are being processed in accordance with the user's interactions with the semantic network.


A user of the thought cloud can, for example, move the labels relative to each other and change properties of the labels, which results in changes to the underlying semantic network without the user being exposed to the associated transformations. Thus a thought cloud enables a user to capture its thinking on a computer by creating, arranging and manipulating labels representing concepts in whatever way makes sense to the user. For example, there is no need for any particular alignment or arrangement of labels in a thought cloud, as the labels can be arranged and rearranged as the user desires.


Example Workflow

A user may use embodiments taught herein to create, visualize and manipulate a data structure that may, for example, be a semantic network. For example, the following workflow is one particular example usage of the present teachings. However, it should be understood that other uses are possible, for example where the data structure is a data structure other than a semantic network, or where the semantic network is fully generated by the user rather than a semantic engine.


The user may initially access the user interface and be provided by the user interface with means for inputting an active concept and optionally a depth. The user may input the active concept, which is obtained by a semantic engine. The semantic engine, in accordance with its implementation, generates a semantic network. The user interface displays to the user a thought cloud representing the semantic network. The thought cloud is optionally constrained to those concepts within the depth of the active concept, if the depth was specified by the user. It should be understood that the semantic engine may use the depth to limit the generation of the semantic network from the active concept.


The user then can visualize the semantic network by the thought cloud. The user may navigate the thought cloud, for example by panning across the thought cloud or zooming in and out of the thought cloud. The user may manipulate the labels representing the concepts as displayed in the thought cloud. For example, the user may move one or more of the concepts on the user interface, or may manipulate one or more properties of one or more concepts. Each manipulation is processed by the invention, which automatically alters the semantic network to reflect the user's manipulations of the concepts in accordance with translation rules.


The user can also add or delete concepts in the thought cloud. When adding a concept, the user may define properties for the concept. The concept and its properties can be processed by the invention, which automatically alters the semantic network to reflect the addition of the concept based on its particular properties in accordance with translation rules.


The user interface can provide the user with a means by which to save the semantic network it has created and manipulated. The user may save a plurality of its created and manipulated semantic networks, for example representing the semantic network at different historical points or dates. The user interface can also provide the user with a means to restore any of the saved semantic networks.


Visualizing a Thought Cloud


The user interface may be a screen or a part of a screen that serves as boundaries for visualizing a thought cloud.


A user interaction may be used for determining a starting point for building a thought cloud initially, in order to prevent an unwieldy number of concepts appearing in the thought cloud. For example, a semantic engine for generating a semantic network may be provided for receiving user interaction information (a context) from which an active concept may be derived. The semantic engine may be, for example, that described in PCT/CA2009/000567. The active concept could be considered the starting point for visualizing the semantic network. A depth may be provided so that only those concepts in the semantic network that are related to the active concept within the depth are displayed by the user interface.


Alternatively, to deal with space constraints, the user interface could include the ability to vary resolution, mimicking motion in the z-direction by zooming in and out. The user interface may also provide the ability for panning vertically and horizontally to navigate through different areas of a thought cloud. When panning, a concept that happens to fall near a particular area of the user interface, for example in the centre, can be used as an active concept to further display the semantic network.


The concepts and relationships between concepts can be represented by one or more properties or attributes, including for example visual properties, in accordance with a set of translation rules. These properties may include x, y and z coordinates; size; colour; distance; typeface; underlining; outlining; weight; gradient and any other visually meaningful attribute. Each of the properties is associable with a user context property of the underlying data structure (such as the meaning of a concept or the type of relationship defined by the translation rules, including, for example, order, confidence, time or prevalence). These extensible properties enable the invention to work with any range of complex concepts and complex relationship types (including hierarchical, associative, or equivalence relationships). There may be any number of translation rules provided for associating properties of the underlying data structure with the properties of the visualization and interactions.



FIG. 2 illustrates a thought cloud in one particular example. An importance or depth in the semantic network (from the active node or concept) could be represented in a thought cloud as size of a label, parent-child relationships could be represented by relative position between concepts where the closest larger concept is the parent, the order of sibling concepts could be represented by a distance from a common original or the distance to a parent concept, and the source of the concept (for example, the user providing the concept) could be represented by colour. It should be understood that the specific visual properties used to represent the relationships between concepts in the thought cloud can vary.


For example, consider the semantic network comprising the hierarchy shown in FIG. 8. In this example, concepts with the largest text size are at the root level of the semantic network while concepts with the smallest text size are at the leaf level. For any other concepts, the nearest larger concept is its parent (for example, a concept in font size 8 has a parent that is the nearest concept whose size is greater than 8).


As shown in FIG. 2, for example, the active concept ‘Miles Davis’ is the parent concept for ‘Improvisation’, ‘Trumpet’ and ‘Jazz’, as it is the nearest larger concept for each of these child concept. Similarly, both ‘Fusion’ and ‘Bebop’ have ‘Jazz’ as a parent, ‘Columbia’ has ‘Trumpet’ as a parent and ‘Modal’ has ‘Improvisation’ as a parent. Additionally, it can be seen that ‘Bebop’ is more closely related to ‘Jazz’ than to ‘Modal’.


The depth of relationships to be displayed by the user interface could be constrained by a depth, as previously mentioned.


Thus, the user interface as provided can be used for visualizing a hierarchical or polyhierarchical semantic network. In a strict hierarchy, each concept can have one and only one parent (except for the root concept, which has no parent), while in a polyhierarchy, each concept can have multiple parents. For example, a polyhierarchy can be visualized by placing a child concept equidistant from multiple same sized parent concepts.


Optionally, as it could be difficult to arrange concepts so that they are truly equidistant or of the same size, these or other properties could be divided into configurable units and ranges. For example, within a unit or range the actual property value may vary but the semantic network may treat them as the same. Alternatively, distances and text sizes could be made to ‘snap’ to the nearest value on a configurable grid, where the grid would allow for a simpler placement of concepts.


Adding Concepts


Embodiments disclosed herein also comprise means for using the thought cloud as a means for the user creating concepts in the semantic network, and for automatically generating corresponding relationships in the semantic network. Once the thought cloud has been visualized by a user, the user can add concepts to the thought cloud, which then are reflected as new concepts in the semantic network. The new concepts will have the relationships in the semantic network based on the user's configuration of the properties of the label representing the concept. Again, the user does not have to understand that the properties are reflective of the relationships of the concept to other concepts in the semantic network.



FIG. 3 illustrates the addition of a new concept. A user could select a command for adding a new concept, which results in a new concept field being displayed in the user interface. FIG. 3A shows a label for a new concept added to the user interface. The label may be clearly marked as new for enabling the user to easily determine that it is a new and configurable label.


The user could create a meaningful label for the new concept, as shown in FIG. 3B. Once the user enters a label, the label can become part of the thought cloud, as shown in FIG. 3C. Once added, one or more of the new concept's properties in the underlying semantic network are automatically generated based on the label's properties in the thought cloud, as previously described. A user can further manipulate the generated properties or create the properties for a label.


Manipulating or Deleting Concepts


Embodiments disclosed herein also comprise a user interface to enable a user to manipulate concepts in a thought cloud, resulting in automatic management of the concept and its relationships in a semantic network.


For example, properties or position of a new or existing label could be changed by the user. FIG. 4A shows a thought cloud. The user could select the concept to be manipulated by, for example, hovering a mouse pointer over the label corresponding to the concept or clicking the label. FIG. 4B shows a selected label. Once selected, a means for configuring the property, which may for example be a drop down list or a text box listing all possible configurations of the property, can be provided in proximity of the label to be configured. Options including means for configuring each property could be given to the user. The user could also be presented with options for deleting the label altogether, which would remove the concept from the semantic network.


For example, the user could configure the font size of a label. The user interface could provide up and down arrows to increase and decrease the font size, respectively. The user could, for example select a down arrow to reduce the font size of the label and, once the user is satisfied with the changes to the properties, deselect the label to apply the configuration as shown in FIG. 4C.


Labels could also be repositioned to alter the relationships between concepts of the semantic network. FIG. 5A illustrates a label corresponding to a concept that a user desires to move. The user could select and move the label, for example by clicking the label and dragging it to a desired location as shown in FIG. 5B. The user could deselect the label to apply the configuration of the concept as shown in FIG. 5C.


The user's manipulation of the properties of labels can be processed by automatically altering the relationships between concepts in the underlying semantic network in accordance with the translation rules.


Similarly, concepts could also be removed from the semantic network. For example, the user may select a deletion command to remove the concept. Concepts related to the active concept of the semantic network only through the removed concept could optionally also be removed, however concepts having polyhierarchical links to the active concept of the semantic network may only have links through the removed concept severed while other links not through the removed concept may remain intact.


Similar processes could be provided for configuring any of the properties of the labels including, for example, geometric attributes including colour, weight, gradient, etc.


Once the configuration has been applied, the user's manipulation of the properties can be processed by automatically manipulating the concepts and relationships between the concepts of the semantic network in accordance with the translation rules.


In certain data structures, particularly those that are not polyhierarchical data structures, there may be a need for resolving ties caused by ambiguous manipulations of new or existing labels. However, resolving ties can also be used with polyhierarchical data structures, for example to limit the number of polyhierarchical relationships.


For example, in a semantic network, a concept placed equidistant between two potential parent concepts could cause a tie. The potential for ties increases as the underlying semantic network is developed. A set of tiebreaking rules could, therefore, be provided for resolving ties. For example, when using the earlier set of sample translation rules, the system may need to break ties when determining parent-child relationships. If a concept has multiple larger concepts that are the same shortest distance from it, these criteria could determine which one is the parent. An example set of tie-breaking rules for this purpose could include, in order: (1) the larger concept is the parent; (2) rank on some other established property of the concepts, such as the time of the most recent change, the confidence of the relationship, the distance from a common origin or the alphabetical order (the property chosen for this ranking can be anything that makes sense for the particular thought cloud); (3) the number of generated or user-defined properties for the label corresponding to the concept; (4) if polyhierarchies are not allowed, then make a random selection; (5) prompt the user to select the parent; (6) if polyhierarchies are allowed and the tie is not broken by the first or second of the above criteria, then all the candidate concepts may be deemed to be parents of the child concept.


Continuing with the same example, the system may also need to break ties when determining sibling order. If two or more concepts are the same distance from the common origin (or from their parent concept, if using it instead of an origin), these criteria could determine which sibling has precedence: (1) The larger concept is the earlier sibling; (2) rank on some other established property of the concepts, such as the time of the most recent change, the confidence of the relationship, the distance from a common origin or the alphabetical order (the property chosen for this ranking can be anything that makes sense for the particular thought cloud); (3) make a random selection or prompt the user for the sibling order.


Creation and Manipulation


Additionally, concepts of a semantic network can be manipulated by adding detail to the concepts. For example, while the primary representation of a concept is its label, the concept may also comprise content, normally hidden from view, such as tags, notes, graphics, links to audio or video, or other content.


For example, a user could select a label for a concept of the semantic network which may result in a means for viewing content, which may for example be presented by a drop down menu or a text box displaying the content or a portion thereof, related to the concept, as shown in FIG. 6A. The user could select to view the content, which results in the content being displayed, as shown in FIG. 6B. The content could be displayed in an editable format, enabling the user to manipulate the content. Once the label is deselected, the content may again be hidden and the content could be applied back to the semantic network. The user could also access the means for viewing content for the purpose of creating content corresponding to the label.


Augmentation


The user interface may also provide users with a plurality of suggested labels representing concepts that are potentially related to one or more labels input by the user to the user interface and/or labels displayed to the user on the user interface. The user may select one or more of these inputted and/or displayed labels, whether directly or indirectly related in the underlying data structure, to be presented with suggested labels from the semantic engine.


Suggested labels may change based on the one or more properties of the one or more labels input by the user, for example based on a label's position or size. The suggested labels may changed based on changes to the one or more properties, and may changed due to addition, deletion or modification to other labels.


The user can optionally approve the suggested labels, enabling the automatic generation of new relationships from the existing concepts to the concept represented by the suggested label. Again a semantic network will be used as an example data structure however the following could apply to any other data structure.


Concepts related to those represented in the thought cloud may be suggested by an associated semantic engine, such as that described in PCT/CA2009/000567. These concepts can be from another semantic network (created by the user or someone/something else), from a domain of information, or from a combination thereof. The user can set a minimum and/or a maximum value of some criterion (for example, density or confidence) for the inclusion of concepts in the suggested set. If a minimum and maximum are set the user can create a window for inclusion rather than a simple hurdle.


Related concepts can optionally be distinguished from the user's own concept by using a distinguishing means, for example colour, shade, opacity, gradient, or a geometric attribute. Consider, for example, the thought cloud of FIG. 7A. Labels for related concepts could be presented in a panel of the user interface that is separate from the one containing the user's original thought cloud. Alternatively, as shown in FIG. 7B, labels for the user's concepts can be displayed using different font attributes than labels for the suggested concepts.


A Global/Local Interface Implementation


Turning to FIG. 9, there is shown another embodiment of a user interface having three sections (e.g., windows or frames)—top section 92, bottom section 94 and left-hand section 96 on a display (e.g., computer display screen). Sections 92 and 94 comprise workspaces where a tag cloud (or a portion thereof) is depicted visually. The upper portion 92 represents a global (i.e., high level) view of the tag cloud such that all (or a selected portion of) the data curated by a user using the interface is contained within the view. A user may navigate the global view with conventional zoom, scroll or cursor functionality. The lower portion 94 represents a “local” view—i.e., a view of a selected portion of the global view. In the example of FIG. 9, the tag “fruit” in section 92 has been selected, as indicated by the dashed lines behind that word, suggesting selection of the tag and surrounding area. In the lower portion 94, a view is shown of other tags associated with the tag “fruit,” such as at lower levels in the network hierarchy.


The lower section 94 of the interface may have the functionalities and capabilities of the user interface (UI) as described above relative to FIGS. 2-8. By using this enhanced UI, a user is able to move seamlessly from a global view to a local view by just selecting a part of the global view in upper section 92 and seeing a corresponding local view in lower section 94. The user can then curate the selected cloud area now visible in the lower portion 94 of the interface (i.e., tags—and even portions or whole semantic networks) can be added to, edited from and generally modified). This seamless transition from global to local views allows a user to explore and drill down into the user's concepts, simplifying the building of better knowledge networks.


Such an interface can be further enhanced to allow tag or thought clouds of different users to be merged in part or in their entireties by a user. Preferably, such a merging operation is conditional on the user obtaining permission from the other users whose clouds are being merged, of course. This turns such an embodiment into an environment or platform supporting collaborative activity wherein a user is able leverage the insights of other users when allowed to refine his or her thoughts about a concept.


As it is generally known that concepts often are defined differently by different individuals, having access to differing points of view may bring about different relationship suggestions and, as such, enhanced brainstorming about a topic.


Yet another variation on the interface embodiment(s) is to add the capability for a user to transport a portion (or its entirety) of the user's knowledge network (tag cloud) into someone else's knowledge network and from this operation find the linkages (e.g. concepts in common or concepts that are semantically related) between the transported content and the host content (tag cloud) and attach the two clouds at the linkages. This allows a user, for example, to look at the concept explorations conducted by other users about specific topic and will allow the other users to share portions of their knowledge networks. For example, multiple researchers may be exploring the same concepts and each of them may have built his or her own thought (tag) cloud. They can then export/import the thought clouds of one or more of the others to look at the different concepts that have been explored, or even the different points of view they bring to the study, and then selectively enhance their own semantic networks. This approach can also be extended to exporting/importing complete thought clouds from other users.


To consider another use of the foregoing system and method, the upper portion 92 of the FIG. 9 interface may display represents a global view of curated data from all users in a community—i.e., a thought cloud for a collective sematic network. In such a scenario, anyone in the community can select a portion of the global view and edit it before archiving. Further, using the lower portion 94 of the interface of FIG. 9, any user from the community can add concepts, curate and then archive those changes to the global view. Optionally, community users may be given the ability (rights) to see the changes that (all or select) other users have made. Within a corporation for example, permissions to view and change data might be given on a departmental basis. That is, different departments (from an enterprise perspective) may participate through creating legends or other identifiers for their changes (e.g., blue represents the legal team, red represents the design team, green represents the research group, etc.).


The lower portion 94 may also show the evolution of concepts that deal with specific topics (e.g., everything that relates to the topic of aviation, where “aviation” is a tag from the global view), relationship between these concepts (e.g. parent, associative, etc. as evidenced by the visual properties as described in FIGS. 2-8), concepts which are only related to particular geographic regions (e.g., aviation in Germany) and show the evolution of concepts through time (e.g. when the ideas were conceived, when relationships were explored, when new concepts were added, omitted, etc.)


When it is desired to be able to keep track of changes to the data over time, by time and who made the change, the underlying semantic network data structure may be augmented to include or to couple to a database having time-series capability, such that each change is recorded in the database and labelled by a time stamp and user identification. If users are organized into groups, an aspect of the database or a related database or table would include user-group affiliations.


In some embodiments, a variant of the foregoing example is to facilitate collaboration by making the lower portion 94 of FIG. 9 independent of the upper portion 92, as a standalone UI. With this arrangement, any user from the community can add concepts, curate and then archive them within the independent lower portion 94.


In some embodiments, changes to a thought (tag) cloud may be tracked over time, including the identity of each person making a change. Changes may be tracked by, for example, a time-series database or in a separate database table of a relational database or in any other satisfactory manner that will readily be known to those skilled in computer science. In some embodiments, the user interface shown in FIG. 9 may include a slider bar 98 which can be manipulated by a user to position an indicator 99 along the slider bar to select a time along the time line represented by the slider bar, with a corresponding display in selected upper section 92 or lower section 94 then showing the state of the selected workspace cloud at the selected time.


There are at least two ways to generate the display of the state of a cloud at a selected time. In a first approach, the state of the cloud can be recorded at various times, either automatically as by taking “snapshots” at predetermined intervals or when certain events occur (e.g., changes to the cloud). In a second approach, the display may be generated by searching the records of changes to the cloud from a beginning time and dynamically re-creating the state of the cloud at the desired time. Both approaches are contemplated, the selection being a matter of design choice based on the tradeoffs between factors such as memory and processing overhead.


By time-tracking and contributor-tracking changes to a tag cloud, a user may be able to see evolution of the concept through time, when ideas were conceived and by whom, when relationships were explored, when new concepts were added, omitted, etc. Further, if the tag cloud content of another user was imported, the evolution shows this contribution as well. This capability thus enables a user is able to move through the user's thought cloud an such a way that the user is provided with the ability to perceive on a linear path how he or she thought of (or came up with) their ideas, when they were generated and who influenced an idea. This ability will provide the user to move through the tag cloud to see how it expands, changes and grows.


The left section 96 comprises a list of terms/concepts, showing their relationships in a linear way. When a given concept is selected, and sub-cepts are displayed, the other concepts are deselected. Selection may be indicated by a color or font change for ease of visualization, for example. Concepts listed this way are intended to differ from each other in kind. Use of the list form of displaying concepts is optional with respect to sections 92 and 94, but may be a helpful way to identify and select concepts for the workspaces.


Such embodiments also subjectively allow users to define large concepts and see how these evolve. They also show users the semantic connections between one part of their global thought cloud to the other; and from each relationship, infer other possible relationships. In addition, the UI allows users to scan root level concepts, find relationships and consequently explore these relationships. In finding relationships, the UI provides users with a semantic intersection between two terms. This can be done through query expansion, which can allow synthesis of new relationships. Alternatively, instead of waiting for manual searching for relationships, the UI may automatically show the users deep connections between various concepts in their thought clouds. The system optionally may provide the user with an indication that these relationships exist and the choice of being able to explore these relationships if they so wish to do so.


A UI such as is illustrated in FIG. 9 may be utilized in a variety of ways, including allowing the user to look at the highest level concepts and see the commonalities between disparate areas of knowledge. For example, the user may drag or select “Detroit” and “nanotechnology” from the top, global user-pane containing a variety of disparate concepts, into the lower, local user-pane, which will thus contain concepts in common to, or at the intersection of, those disparate concepts. Although “Detroit” and “nanotechnology” appear at first blush to be disparate areas of knowledge, concepts in common or at the intersection of both “Detroit” and “nanotechnology” would be synthesized and added to the lower pane. Such common concepts may include nanotechnology companies residing in the city of Detroit or the Michigan Nanotechnology Institute for Biological Sciences located in the greater Detroit region (i.e., Ann Arbor).


The UI may also let users identify relationships and organize concepts according to how important they are to the user. The UI provides to the user a feedback mechanism which allows users to infer possible relationships between tags and then to explore them. Thus, the curated data which may be part of the global view may be analyzed by a user for relationships or, alternatively, relationships can be detected automatically and suggested to the user (who may adopt or reject them)


The UI of FIG. 9 further may be provided with the capability of allowing the user to attach links, notes, files, etc. to the concepts in the user's thought cloud and even allowing the user to have an expanded view whereby it can tell the user more about what is written for each concept. This additional information may include an indication of what is “underneath” a concept (i.e., how deep and how many concepts there are for each particular concept). There also may be an indication of which underlying concepts a user already has explored and which have not been explored, using any desired legend or indicia; as an example, using bold type for those already explored or providing a window listing those concepts which have been explored). The idea is, preferably, to make the user aware of this information without necessarily drilling in or navigating away from the main view.


An additional advantage or use at least some embodiments of the above-described UI is that it can be used to curate ideas from other applications to be related to a thought cloud. Such other application may, for example, be an online media service. For example, a data mining application may parse through a page (e.g., article, website, online journal, etc.) and capture links and place them in a relationship box, as a result of which the whole semantic network then is augmented by meaningful relationships from basically single words or phrases.


In another use, such a UI may be employed to harvest the meaning attached to a particular word and from these meanings, identify other concepts. Indeed, one may automate the whole process—i.e., build the knowledge network without doing any work. An automated process may comprise the steps of semantically understanding a page, ranking the concepts within the page and taking the top-ranked concepts and relating them to the user's network.


The benefits of collaborative activity also may be obtained by using user visitations to the thought clouds/web pages of others to “cross pollinate” that user's own thought clouds. By tracking the user's visitations, on return to the user's thought cloud, the interface may show possible relevant additions based on the visits.


With this background, attention is now directed to FIGS. 10-12.


In the example process reflected in FIG. 10, starting with step 1002, a user provides new concept input or input is selected from an existing global workspace 92 of previously curated data. A local workspace 94 then is created, step 1004, providing a place for the user to organically organize related concepts around the selected/newly input concept (called the “root” concept).


The user then curates system-suggested and newly created concepts that relate to the root concept. Step 1006.


At some point, the user finishes curating local concepts and archives changes at both the global and local view (workspace) levels. Step 1008. This results in a new global workspace 92A and a new local workspace 94B.


In FIG. 11's process, two users, A and B, each curates his own local workspace and then store their resulting global workspaces collectively, so that the global workspace becomes a collaborative workspace.


In FIG. 12, by contrast, two users, A and B, create a shared local workspace, allowing them to collaboratively organize concepts (tags) around a concept they selected.


Additional Implementations


Instead of the proximity of labels representing the closeness of relation between concepts and size representing a concept's level in a semantic network, one could augment this approach by representing different levels of confidence in each concept with a gradient of dark to light text for their labels. If a concept suggested by the semantic networking system has a confidence of 0.1, its label could be shown in lighter text. If the user selects the concept, its confidence could change to 1.0, and its label would become dark.


Further extensions of a user interface of embodiments disclosed herein for creating, visualizing and manipulating data structures can be provided beyond thought clouds for semantic networking.


A different example could have a goal-planning application translate x-coordinate as date, y as projected cost and z (size) as importance. The user interface could include explicit axes for these physical dimensions to help the user interpret their meaning. Here, the date dimension could employ a text gradient in parallel with x-coordinate, with older goals having lighter labels.


The representation of time with a text gradient could also be applied to these examples. This way, the gradient would represent time instead of confidence.

Claims
  • 1. A computer network-implemented method for displaying a data structure and enabling one or more users to interact with the data structure via a user interface, the data structure including data entities and relationships between the data entities, the method comprising: receiving, via one or more computer processors executing stored program instructions, a first user interaction in which a user: selects a first text label from a set of text labels populated on the user interface, the first text label representing a first data entity from the data structure, and/oradds a second text label to the set of text labels populated on the user interface, the second text label establishing a second data entity in the data structure;receiving a second user interaction in which the user selects one or more suggested text labels via the user interface, the one or more suggested text labels representing data entities relevant to the first data entity and/or to the second data entity;determining a relationship in the data structure between the relevant data entities by applying translational rules to at least one positional visual property and/or at least one textual visual property of the one or more selected suggested text labels; andstoring the relevant data entities and the relationship in the data structure based on the at least one positional visual property and/or the at least one textual visual property of the one or more selected suggested text labels.
  • 2. The computer network-implemented method of claim 1, wherein the second user interaction includes an approval by the user of the one or more suggested text labels.
  • 3. The computer network-implemented method of claim 1, wherein the data entities are associated with content including tags, notes, graphics, audio links and/or video links.
  • 4. The computer network-implemented method of claim 1, further comprising generating, via a semantic engine, the data entities relevant to the first data entity and/or to the second data entity.
  • 5. The computer network-implemented method of claim 1, further comprising providing a global view and a local view via the user interface.
  • 6. The computer network-implemented method of claim 5, wherein the global view represents data entities curated from a community of users.
  • 7. The computer network-implemented method of claim 1, wherein the relevant data entities are linked to a third-party knowledge network.
  • 8. A system for executing a computer-implemented method for displaying a data structure including data entities and relationships between the data entities, and for enabling one or more users to interact with the data structure, the system comprising one or more computer devices including or being operatively linked to: at least one display, anda user interface utility operable to: present to one or more users a user interface, via the at least one display;receive a first user interaction in which a user: selects a first text label from a set of text labels populated on the user interface, the first text label representing a first data entity from the data structure, and/oradds a second text label to the set of text labels populated on the user interface, the second text label establishing a second data entity in the data structure;receive a second user interaction in which the user selects one or more suggested text labels via the user interface, the one or more suggested text labels representing data entities relevant to the first data entity and/or to the second data entity;determine a relationship in the data structure relating the relevant data entities by applying translational rules to at least one positioning visual property and/or at least one textual visual property of the one or more selected suggested text labels; andstore the relevant data entities and the relationship in the data structure based on the at least one positional visual property and/or the at least one textual visual property of the one or more selected suggested text labels.
  • 9. The system of claim 8, wherein the second user interaction includes an approval by the user of the one or more suggested text labels.
  • 10. The system of claim 8, wherein the data entities are associated with content including tags, notes, graphics, audio links and/or video links.
  • 11. The system of claim 8, further comprising a semantic engine operable to generate the data entities relevant to the first data entity and/or to the second data entity.
  • 12. The system of claim 8, wherein the user interface utility is further operable to provide a global view and a local view via the user interface.
  • 13. The system of claim 12, wherein the global view represents a data entities curated from a community of users.
  • 14. The system of claim 8, wherein the relevant data entities are linked to a third-party knowledge network.
  • 15. A non-transitory computer program product containing executable computer program instructions which, when executed by one or more computers having a display, perform a method for displaying a data structure including data entities and relationships between the data entities, the method comprising: receiving a first user interaction in which a user: selects a first text label from a set of text labels populated on the user interface, the first text label representing a first data entity from the data structure, and/oradds a second text label to the set of text labels populated on the user interface, the second text label establishing a second data entity in the data structure;receiving a second user interaction in which the user selects one or more suggested text labels via the user interface, the one or more suggested text labels representing data entities relevant to the first data entity and/or to the second data entity;determining a relationship in the data structure between the relevant data entities by applying translational rules to at least one positional visual property and/or at least one textual visual property of the one or more selected suggested text labels; andstoring the relevant data entities and the relationship in the data structure based on the at least one positional visual property and/or the at least one textual visual property of the one or more selected suggested text labels.
  • 16. The non-transitory computer program product of claim 15, wherein the second user interaction includes an approval by the user of the one or more suggested text labels.
  • 17. The non-transitory computer program product of claim 15, wherein the method further comprises generating, via a semantic engine, the data entities relevant to the first data entity and/or to the second data entity.
  • 18. The non-transitory computer program product of claim 17, wherein the relevant data entities are linked to a third-party knowledge network.
RELATED APPLICATIONS

This application is a continuation and claims the benefit under 35 U.S.C. § 120 of U.S. patent application Ser. No. 13/105,890, filed 11 May 2011, and titled “System, Method and Computer Program For Creating And Manipulating Data Structures Using An Interactive Graphical Interface,” which is incorporated herein by reference in its entirety. U.S. patent application Ser. No. 13/105,890 is a continuation-in-part and claims the benefit under 35 U.S.C. § 120 of U.S. patent application Ser. No. 12/615,703, filed 10 Nov. 2009, and titled System, Method And Computer Program For Creating And Manipulating Data Structures Using An Interactive Graphical Interface, now U.S. Pat. No. 8,281,238.

US Referenced Citations (215)
Number Name Date Kind
3943462 Thompson Mar 1976 A
4532813 Rinehart Aug 1985 A
4972328 Wu et al. Nov 1990 A
5056021 Ausborn Oct 1991 A
5193185 Lanter Mar 1993 A
5369763 Biles Nov 1994 A
5745910 Piersol et al. Apr 1998 A
5793376 Tanaka et al. Aug 1998 A
5835758 Nochur et al. Nov 1998 A
5911145 Arora et al. Jun 1999 A
5937400 Au Aug 1999 A
5953726 Carter et al. Sep 1999 A
6006222 Culliss Dec 1999 A
6078916 Culliss Jun 2000 A
6098033 Richardson et al. Aug 2000 A
6138085 Richardson et al. Oct 2000 A
6167390 Brady et al. Dec 2000 A
6173276 Kant et al. Jan 2001 B1
6233575 Agrawal et al. May 2001 B1
6292792 Belles et al. Sep 2001 B1
6295066 Tanizaki et al. Sep 2001 B1
6334131 Chakrabarti et al. Dec 2001 B2
6349275 Schumacher et al. Feb 2002 B1
6356899 Chakrabarti et al. Mar 2002 B1
6396864 O'Brien et al. May 2002 B1
6401061 Zieman Jun 2002 B1
6499024 Stier et al. Dec 2002 B1
6539376 Sundaresan et al. Mar 2003 B1
6539395 Gjerdingen et al. Mar 2003 B1
6556983 Altschuler et al. Apr 2003 B1
6571240 Ho et al. May 2003 B1
6694329 Murray Feb 2004 B2
6751611 Krupin et al. Jun 2004 B2
6751621 Calistri-Yeh et al. Jun 2004 B1
6768982 Collins et al. Jul 2004 B1
6772136 Kant et al. Aug 2004 B2
6785683 Zodik et al. Aug 2004 B1
6868525 Szabo Mar 2005 B1
6976020 Anthony et al. Dec 2005 B2
6980984 Huffman et al. Dec 2005 B1
7007074 Radwin Feb 2006 B2
7035864 Ferrari et al. Apr 2006 B1
7051023 Kapur et al. May 2006 B2
7062466 Wagner et al. Jun 2006 B2
7062483 Ferrari et al. Jun 2006 B2
7089237 Turnbull et al. Aug 2006 B2
7120646 Streepy, Jr. Oct 2006 B2
7152065 Behrens et al. Dec 2006 B2
7181465 Maze et al. Feb 2007 B2
7209922 Maze et al. Apr 2007 B2
7225183 Gardner May 2007 B2
7249117 Estes Jul 2007 B2
7280991 Beams et al. Oct 2007 B1
7283992 Liu et al. Oct 2007 B2
7302418 Asahara Nov 2007 B2
7319951 Rising, III et al. Jan 2008 B2
7392250 Dash et al. Jun 2008 B1
7406456 Calistri-Yeh et al. Jul 2008 B2
7418452 Maze Aug 2008 B2
7440940 Chen et al. Oct 2008 B2
7478089 Henkin et al. Jan 2009 B2
7490073 Qureshi et al. Feb 2009 B1
7493319 Dash et al. Feb 2009 B1
7496593 Gardner et al. Feb 2009 B2
7502810 Acevedo-Aviles et al. Mar 2009 B2
7580918 Chang et al. Aug 2009 B2
7596374 Katou Sep 2009 B2
7596574 Sweeney Sep 2009 B2
7606168 Robinson et al. Oct 2009 B2
7606781 Sweeney et al. Oct 2009 B2
7627582 Ershov Dec 2009 B1
7668737 Streepy, Jr. Feb 2010 B2
7711672 Au May 2010 B2
7716207 Odom et al. May 2010 B2
7716216 Harik et al. May 2010 B1
7720857 Beringer et al. May 2010 B2
7752199 Farrell Jul 2010 B2
7752534 Blanchard, III et al. Jul 2010 B2
7827125 Rennison Nov 2010 B1
7844565 Sweeney Nov 2010 B2
7849090 Sweeney Dec 2010 B2
7860817 Sweeney et al. Dec 2010 B2
7945555 Sankaran et al. May 2011 B2
7970764 Ershov Jun 2011 B1
8010570 Sweeney Aug 2011 B2
8281238 Sweeney et al. Oct 2012 B2
9262520 Sweeney et al. Feb 2016 B2
20020069197 Katayama et al. Jun 2002 A1
20020078044 Song et al. Jun 2002 A1
20020133483 Klenk et al. Sep 2002 A1
20020194187 McNeil et al. Dec 2002 A1
20030177112 Gardner Sep 2003 A1
20030196094 Hillis et al. Oct 2003 A1
20030217023 Cui et al. Nov 2003 A1
20030217335 Chung et al. Nov 2003 A1
20040024739 Cooperman et al. Feb 2004 A1
20040049522 Streepy, Jr. Mar 2004 A1
20050010428 Bergeron et al. Jan 2005 A1
20050065955 Babikov et al. Mar 2005 A1
20050086188 Hillis et al. Apr 2005 A1
20050149518 Duan et al. Jul 2005 A1
20050154708 Sun Jul 2005 A1
20050209874 Rossini Sep 2005 A1
20050216335 Fikes et al. Sep 2005 A1
20050223109 Mamou et al. Oct 2005 A1
20050289524 McGinnes Dec 2005 A1
20060010117 Bonabeau et al. Jan 2006 A1
20060026147 Cone et al. Feb 2006 A1
20060053172 Gardner et al. Mar 2006 A1
20060074980 Sarkar Apr 2006 A1
20060085489 Tomic et al. Apr 2006 A1
20060129906 Wall Jun 2006 A1
20060153083 Wallenius Jul 2006 A1
20060195407 Athelogou et al. Aug 2006 A1
20060242564 Egger et al. Oct 2006 A1
20060271520 Ragan Nov 2006 A1
20070033531 Marsh Feb 2007 A1
20070036440 Schaepe et al. Feb 2007 A1
20070038500 Hammitt et al. Feb 2007 A1
20070061195 Liu et al. Mar 2007 A1
20070078889 Hoskinson Apr 2007 A1
20070083492 Hohimer et al. Apr 2007 A1
20070094221 Au Apr 2007 A1
20070106658 Ferrari et al. May 2007 A1
20070118542 Sweeney May 2007 A1
20070136221 Sweeney et al. Jun 2007 A1
20070143300 Gulli et al. Jun 2007 A1
20070174041 Yeske Jul 2007 A1
20070192272 Elfayoumy et al. Aug 2007 A1
20070203865 Hirsch Aug 2007 A1
20070208719 Tran Sep 2007 A1
20070208764 Grisinger Sep 2007 A1
20070288503 Taylor Dec 2007 A1
20070294200 Au Dec 2007 A1
20070300142 King et al. Dec 2007 A1
20080001948 Hirsch Jan 2008 A1
20080004864 Gabrilovich et al. Jan 2008 A1
20080021925 Sweeney Jan 2008 A1
20080072145 Blanchard et al. Mar 2008 A1
20080086465 Fontenot et al. Apr 2008 A1
20080092044 Lewis et al. Apr 2008 A1
20080126303 Park et al. May 2008 A1
20080137668 Rodriguez et al. Jun 2008 A1
20080154906 McDavid et al. Jun 2008 A1
20080162498 Omoigui Jul 2008 A1
20080228568 Williams et al. Sep 2008 A1
20080243480 Bartz et al. Oct 2008 A1
20080270120 Pestian et al. Oct 2008 A1
20080275694 Varone Nov 2008 A1
20080281814 Calistri-Yeh et al. Nov 2008 A1
20080294584 Herz Nov 2008 A1
20090012842 Srinivasan et al. Jan 2009 A1
20090018988 Abrams et al. Jan 2009 A1
20090024385 Hirsch Jan 2009 A1
20090024556 Hirsch Jan 2009 A1
20090028164 Hirsch Jan 2009 A1
20090055342 Gong et al. Feb 2009 A1
20090070219 D'Angelo et al. Mar 2009 A1
20090083140 Phan Mar 2009 A1
20090106234 Siedlecki et al. Apr 2009 A1
20090138454 Rayner et al. May 2009 A1
20090144059 Yu et al. Jun 2009 A1
20090150809 Hirsch Jun 2009 A1
20090157442 Tesler Jun 2009 A1
20090157616 Barber et al. Jun 2009 A1
20090182725 Govani et al. Jul 2009 A1
20090192954 Katukuri et al. Jul 2009 A1
20090192968 Tunstall-Pedoe Jul 2009 A1
20090198561 Otto et al. Aug 2009 A1
20090228425 Goraya Sep 2009 A1
20090300326 Sweeney Dec 2009 A1
20090307581 Jaepel et al. Dec 2009 A1
20090327205 Sweeney Dec 2009 A1
20090327417 Chakra et al. Dec 2009 A1
20100030552 Chen et al. Feb 2010 A1
20100036783 Rodriguez Feb 2010 A1
20100036790 Sweeney et al. Feb 2010 A1
20100036829 Leyba Feb 2010 A1
20100049702 Martinez et al. Feb 2010 A1
20100049766 Sweeney et al. Feb 2010 A1
20100057664 Sweeney et al. Mar 2010 A1
20100070448 Omoigui Mar 2010 A1
20100100546 Kohler Apr 2010 A1
20100107094 Steelberg et al. Apr 2010 A1
20100122151 Mendelson et al. May 2010 A1
20100153219 Mei et al. Jun 2010 A1
20100161317 Au Jun 2010 A1
20100198724 Thomas Aug 2010 A1
20100205061 Karmarkar Aug 2010 A1
20100217745 Song et al. Aug 2010 A1
20100223295 Stanley et al. Sep 2010 A1
20100228693 Dawson et al. Sep 2010 A1
20100235307 Sweeney et al. Sep 2010 A1
20100250526 Prochazka et al. Sep 2010 A1
20100257171 Shekhawat Oct 2010 A1
20100262456 Feng et al. Oct 2010 A1
20100268596 Wissner et al. Oct 2010 A1
20100280860 Iskold et al. Nov 2010 A1
20100285818 Crawford Nov 2010 A1
20100287011 Muchkaev Nov 2010 A1
20110040749 Ceri et al. Feb 2011 A1
20110060644 Sweeney Mar 2011 A1
20110060645 Sweeney Mar 2011 A1
20110060794 Sweeney Mar 2011 A1
20110113386 Sweeney et al. May 2011 A1
20110173176 Christensen et al. Jul 2011 A1
20110282919 Sweeney et al. Nov 2011 A1
20110314006 Sweeney et al. Dec 2011 A1
20110314382 Sweeney Dec 2011 A1
20110320396 Hunt et al. Dec 2011 A1
20120143880 Sweeney et al. Jun 2012 A1
20120150874 Sweeney et al. Jun 2012 A1
20120166371 Sweeney et al. Jun 2012 A1
20120166372 Ilyas et al. Jun 2012 A1
20120166373 Sweeney et al. Jun 2012 A1
Foreign Referenced Citations (28)
Number Date Country
2734756 Mar 2010 CA
1325513 Dec 2001 CN
1395193 May 2003 CN
1536483 Oct 2004 CN
0 962 873 Dec 1999 EP
2005316699 Nov 2005 JP
WO 02054292 Jul 2002 WO
WO 2004075466 Sep 2004 WO
WO 2005020093 Mar 2005 WO
WO 2005020094 Mar 2005 WO
WO 2007047971 Apr 2007 WO
WO 2008025167 Mar 2008 WO
WO 2008076438 Jun 2008 WO
WO 2008076438 Jun 2008 WO
WO 2009014837 Jan 2009 WO
WO 2009132442 Nov 2009 WO
WO 2010022505 Mar 2010 WO
WO 2010149427 Dec 2010 WO
WO 2011029177 Mar 2011 WO
WO 2011029177 Mar 2011 WO
WO 2011057396 May 2011 WO
WO 2011160204 Dec 2011 WO
WO 2011160205 Dec 2011 WO
WO 2011160214 Dec 2011 WO
WO 2012088590 Jul 2012 WO
WO 2012088591 Jul 2012 WO
WO 2012088611 Jul 2012 WO
WO 2012092669 Jul 2012 WO
Non-Patent Literature Citations (82)
Entry
Israel Patent Office, Office Action for IL Application No. 219705 dated Feb. 7, 2016.
Canadian Intellectual Property Office, Office Action for CA Application No. 2,780,330 dated Oct. 4, 2016.
Chinese Office Action for Chinese Application No. 200780032062.9, dated May 17, 2011.
International Preliminary Report on Patentability for International application No. PCT/CA2010/001772, dated May 24, 2012.
International Preliminary Report on Patentability for PCT/CA2007/001546 dated Dec. 19, 2008.
International Preliminary Report on Patentability for PCT/CA2009/000567 dated Nov. 11, 2010.
International Preliminary Report on Patentability for PCT/CA2009/001185 dated Mar. 10, 2011.
International Preliminary Report on Patentability for PCT/CA2010/001382 dated Mar. 22, 2012.
International search report and written opinion for International application No. PCT/CA2010/001772, dated Apr. 28, 2011.
International Search Report and Written Opinion for International Application No. PCT/CA2010/001382 dated Jan. 13, 2011.
International Search Report and Written Opinion for International Application No. PCT/CA2011/000718 dated Oct. 13, 2011.
International Search Report and Written Opinion for International Application No. PCT/CA2011/000745 dated Sep. 22, 2011.
International Search Report and Written Opinion for International Application No. PCT/CA2011/001382 dated Apr. 24, 2012.
International Search Report and Written Opinion for International Application No. PCT/CA2011/001403 dated May 23, 2012.
International Search Report and Written Opinion for PCT/CA2007/001546 dated Dec. 28, 2007.
International Search Report and Written Opinion for PCT/CA2009/000567 dated Aug. 24, 2009.
International Search Report and Written Opinion for PCT/CA2009/001185 dated Dec. 3, 2009.
International Search Report and Written Opinion of the International Searching Authority for International Application No. PCT/CA2011/000719, dated Sep. 28, 2011.
International Search Report and Written Opinion of the International Searching Authority for International Application No. PCT/CA2011/001402, dated Apr. 24, 2012.
International Search Report and Written Opinion of the International Searching Authority for International Application No. PCT/CA2012/000007, dated Apr. 20, 2012.
International Search Report and Written Opinion of the International Searching Authority for International Application No. PCT/CA2012/000009, dated May 1, 2012.
Interview Summary for U.S. Appl. No. 11/469,258 dated Dec. 16, 2008.
Japanese Office Action for Japanese Application No. 2012-538155 dated Jun. 24, 2014.
Japanese Office Action for Japanese Application No. 2012-538155 dated Oct. 29, 2013.
Office Action for U.S. Appl. No. 11/469,258 dated Aug. 21, 2008.
Advisory Action for U.S. Appl. No. 12/441,100 dated May 4, 2012.
Office Action for U.S. Appl. No. 12/555,222 dated Jan. 27, 2012.
Office Action for U.S. Appl. No. 12/555,222 dated Oct. 31, 2012.
Office Action for U.S. Appl. No. 12/555,341 dated Aug. 1, 2012.
Office Action for U.S. Appl. No. 12/555,341 dated Feb. 9, 2012.
Office Action for U.S. Appl. No. 12/615,703 dated Aug. 1, 2012.
Office Action for U.S. Appl. No. 12/615,703 dated Feb. 1, 2012.
Office Action for U.S. Appl. No. 13/105,890 dated Jun. 26, 2012.
Office Action for U.S. Appl. No. 11/550,457 dated Dec. 15, 2008.
Office Action for U.S. Appl. No. 11/625,452 dated Dec. 7, 2009.
Office Action for U.S. Appl. No. 11/625,452 dated Mar. 26, 2010.
Office Action for U.S. Appl. No. 11/625,452 dated Mar. 30, 2009.
Office Action for U.S. Appl. No. 12/441,100 dated Jan. 24, 2012.
Office Action for U.S. Appl. No. 12/441,100 dated Jun. 9, 2011.
Office Action for U.S. Appl. No. 12/477,977 dated Sep. 28, 2010.
Office Action for U.S. Appl. No. 12/477,994 dated Aug. 31, 2010.
Office Action for U.S. Appl. No. 12/549,812 dated Oct. 1, 2012.
Office Action for U.S. Appl. No. 12/556,349 dated Jun. 29, 2010.
Office Action for U.S. Appl. No. 13/105,890 dated Aug. 29, 2013.
Office Action for U.S. Appl. No. 13/105,890 dated Dec. 26, 2012.
Office Action for U.S. Appl. No. 13/105,890 dated Jun. 17, 2015.
Office Action for U.S. Appl. No. 13/105,890 dated Jun. 18, 2014.
Office Action for U.S. Appl. No. 13/105,890 dated Oct. 5, 2015.
Office Communication dated May 12, 2015 for Chinese Application No. 201080055700.0.
Office Communication dated Oct. 8, 2014 for Chinese Application No. 201080055700.0.
[No Author Listed] “Faceted Classification and Adaptive Concept Matching,” Gemstone Business Intelligence Ltd., Feb. 2006. pp. 1-7. 7 pages.
Anick et al., Interactive document retrieval using faceted terminological feedback. HICSS-32. Proceedings of the 32nd Annual Hawaii International Conference on Systems Sciences. 1999;2(2):2036-2048. Digital Object Identifier: 10.1109/HICSS.1999.772692.
Blei et al., Hierarchical Bayesian models for applications in information retrieval. In: Bayesian Statistics 7. Bernardo et al., eds. 2003:25-43.
Bollegala et al., Measuring semantic similarity between words using web searches engines. Proceedings of 16th International Conference on World Wide Web. 2007;757-66.
Brewster et al., User-Centered Ontology Learning for Knowledge Management. 7th International Workshop on Applications of Natural Language to Information Systems, Stockholm, Jun. 27-28, 2002. Lecture Notes in Computer Sciences, Springer Verlag. 2002:12 pages.
Brewster et al., User-Centered Ontology Learning for Knowledge Management. 7th International Workshop on Applications of Natural Language to Information Systems, Stockholm, Jun. 27-28, 2002. Lecture Notes in Computer Sciences, Springer Verlag. 2002:203-207. 5 pages.
Dakka et al., Automatic Extraction of Useful Facet Hierarchies from Text Databases. Data Engineering. IEEE 24th International Conference on Apr. 7-12, 2008. ICDE 2008:466-475. Digital Object Identifier 10.1109/ICDE.2008.4467455.
Fikadu et al., A Framework for Personalized Information Retrieval Model. Conference Proceedings, Second International Conference on Computer and Network Technology (ICCNT), IEEE, Piscataway, NJ, USA Apr. 23, 2010, pp. 500-505.
Gabrilovich et al., Computing semantic relatedness using wikipedia-based explicit semantic analysis. Proceedings of 20th International Joint Conference on Artificial Intelligence. 2007;1606-11.
Hassan-Montero et al., Improving tag-clouds as visual information retrieval interfaces, International Conference on Multidisciplinary Information Sciences and Technologies, InSciT2006. Oct. 25-28, 2006, Merida, Spain. 6 pages.
Hiemstra, A probabilistic justification for using tf-idf term weighting in information retrieval. International Journal on Digital Libraries. 2000;3(2):131-39.
Jiang et al., Semantic similarity based on corpus statistics and lexical taxonomy. Proceedings of International Conference Research on Computational Linguistics. 1997; 15 pages.
Jones, A statistical interpretation of term specificity and its applications in retrieval. Journal of Documentation. 2004;60(5):493-502.
Kaser et al., Tag-Cloud Drawing: Algorithms for Cloud Visualization, arXiv:cs/0703109v2 [cs.DS] May 7, 2007.
Lewis, Naive (Bayes) at forty: The independence assumption in information retrieval. Lecture Notes in Computer Science. 1998;1398:4-15.
Ma et al., Semantic Information Extraction of Video Based on Ontology and Inference. ICSC 2007. International Conference on Semantic Computing. 2007;1:721-726. Digital Object Identifier: 10.1109/ ICSC.2007.12.
Metzler et al., A Markov random field model for term dependencies. Proceedings of SIGIR 2005. 2005:472-79.
Ozcan et al., Concept-based information access. Proceedings of the International Conference on Information Technology: Coding and Computing. ITCC 2005;1:794-799. Digital Object Identifier: 10.1109/ITCC.2005.111.
Payne et al., Calendar Agents on the Semantic Web. IEEE Intelligent Systems. Jun. 2002;17(3):84-86.
Robertson, Understanding inverse document frequency: On theoretical arguments for ids. Journal of Documentation. 2004;60(5):503-20.
Rocha, Adaptive Webs for Heterarchies with Diverse Communities of Users. Paper prepared for the workshop from Intelligent Networks to the Global Brain: Evolutionary Social Organization through Knowledge Technology, Brussels, Jul. 3-5, 2001. LAUR005173. 35 pages.
Seco et al., An intrinsic information content metric for semantic similarity in wordnet. Proceedings of 16th European Conference on Artifical Intelligence. 2004;1089-90.
Slavic et al., Core Requirements for Automation of Analytico-Synthetic Classifications. Advances in Knowledge Organization. 2004;9:187-192.
Song et al., A conceptual graph approach to semantic similarity computation method for e-service discovery. International Journal on Knowledge Engineering and Data Mining. 2010;1(1):50-68.
Storey, Comparing Relationships in Conceptual Modeling: Mapping to Semantic Classifications. IEEE Transactions on Knowledge and Data Engineering. 2005;17(11):1478-1489. Digital Object Identifier: 10.1109/.
Terra et al., Frequency estimates for statistical word similarity measures. Proceedings of 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology. 2003:165-172.
Wang et al., Gene expression correlation and gene ontology-based similarity: An assessment of quantitative relationships. Proceedings of IEEE Symposium on Computational Intelligence in Bioinformatics and Computational Biology. 2004:25-31.
Wu et al., Interpreting tf-idf term weights as making relevance decisions. ACM Transactions on Information Systems. 2008;26(3):Article No. 13.
Zhai, Statistical language models for information retrieval—a critical review. Foundations and Trends in Information Retrieval. 2008;2(3):137-213.
Zhang et al., Bootstrapping Ontology Learning for Information Retrieval Using Formal Concept Analysis and Information Anchors. 14th International Conference on Conceptual Structures. Aalborg, Denmark. Jul. 2006. 14 pages.
SIPO, Office Action and Search Report for CN Application No. 201610097151.2 dated Aug. 28, 2018.
IPI, Office Action for IN Application No. 1084/MUMNP/2012 dated Sep. 11, 2018.
Related Publications (1)
Number Date Country
20160210296 A1 Jul 2016 US
Continuations (1)
Number Date Country
Parent 13105890 May 2011 US
Child 14988168 US
Continuation in Parts (1)
Number Date Country
Parent 12615703 Nov 2009 US
Child 13105890 US