The present disclosure generally relates to knowledge automation. More particularly, techniques are disclosed for transforming data content into knowledge suitable for consumption by users.
With the vast amount of data content available, users often suffer from information overload. For example, in an enterprise environment, a large corporation may store all the data that users need to complete their tasks. However, finding the right data for the right user can be challenging. Users may often spend substantial amount of time looking for a needle in a haystack in trying to find the right data to fill their particular needs from thousands of data files. In a collaborative environment, even after the right data is found, substantial amount of time may be needed to synthesis that data into a suitable output that can be consumed by others. The amount of time that users spend searching and synthesizing the data may also create excessive load on the enterprise computing systems and slow down the processing of other tasks.
Data content can be represented and described using a number of different keys, such as a title of a document, published date, summary of the document, tags etc. However, finding an accurate representation of the data content within a document that provides useful information to a user about the content of the document can oftentimes be challenging.
Embodiments of the present invention address these and other problems individually and collectively.
The present disclosure generally relates to knowledge automation. More particularly, knowledge automation techniques are disclosed for transforming data content within documents into knowledge units and knowledge packs suitable for consumption by users. In an embodiment, the knowledge automation techniques include generating a thumbnail image for a knowledge unit and/or a knowledge pack.
In certain embodiments, techniques are provided (e.g., a method, a system, non-transitory computer-readable medium storing code or instructions executable by one or more processors) for generating thumbnail images for knowledge units and knowledge packs. In an embodiment, a method for generating a thumbnail image for a knowledge unit is disclosed. The method may include receiving, by a data processing system, a request for determining a representative image for a knowledge unit. The method may include determining a set of one or more images associated with the knowledge unit by analyzing text and non-text regions in the knowledge unit. In some embodiments, the method may include providing the set of one or more images to a user on a client device and receiving user input indicative of a selection of a first image (e.g., a representative image) from the set of one or more images. The set of one or more images may include visual representations of the data contents within the knowledge unit. The method may then include generating a thumbnail image for the knowledge unit based on the first image. In some examples, the “thumbnail image” may be a reduced size version of the representative image of the knowledge unit and/or knowledge pack. The method may then include associating the thumbnail image with the knowledge unit and displaying the thumbnail image to the user via the client device. In some examples, the method may include displaying the knowledge unit to the user on the client device when the thumbnail is displayed to the user.
In some embodiments, the method may include receiving a selection of multiple (i.e., more than one) images from the set of one or more images from the user. The method may include combining the selected one or more images into a single representative image for the knowledge pack. For instance, the method may include merging the selected one or more images onto a collage or animated image to generate a representative image for the knowledge unit. The method may then include generating a thumbnail image for the knowledge unit based on the representative image and associating the thumbnail image with the knowledge unit. In some embodiments, the method may then include displaying the thumbnail image to the user via the client device.
In some embodiments, the method may include automatically determining a representative image for a knowledge unit. In this embodiment, the method may include identifying a plurality of features corresponding to the set of one or more images and assigning a plurality of weights to the plurality of features. The method may further include determining a score for each image in the set of one or more images based on the plurality of weights and identifying an image in the set of one or more images with the highest score. The method may then include identifying an image in the set of one or more images with the highest score and determining the identified image as the representative image for the knowledge unit. In some embodiments, the method may then include generating the thumbnail image for the knowledge unit based at least in part on the representative image.
In some embodiments, the method may include generating a thumbnail image for a knowledge unit that does not contain any extractable images. In this embodiment, the method may include determining a set of tags associated with the knowledge unit. In some examples, the set of tags identify one or more terms that describe data content within the knowledge unit. The method may then include generating the thumbnail image for the knowledge unit based at least in part on the set of tags. For instance, in some embodiments, the method may include identifying a stored set of one or more images and comparing the set of tags associated with the knowledge unit to one or more sets of tags associated with the stored set of one or more images. The method may then include determining one or more matching sets of tags based on the comparing and determining a best match set of tags from the one or more matching sets of tags. In some examples, the method may further include identifying an image from the stored set of one or more images that corresponds to the best match set of tags, determining the identified image as the representative image for the knowledge unit and generating the thumbnail image for the knowledge unit based at least in part on the representative image.
In some embodiments, the method may include identifying multiple sets of tags (instead of a single best match set of tags) from the one or more matching sets of tags. In this embodiment, the method may include identifying images from the stored set of one or more images that correspond to each set of tags in the multiple sets of tags and providing the identified images to a user on the client device. The method may further include receiving user input indicative of a user-selected image from the identified images and identifying the user-selected image as the representative image for the knowledge unit. The method may then include generating the thumbnail image for the knowledge unit based at least in part on the representative image.
In certain embodiments, a non-transitory computer-readable storage memory storing a plurality of instructions executable by one or more processors is disclosed. The instructions include instructions that cause the one or more processors to receive a request for determining a representative image for a knowledge unit, determine a set of one or more images associated with the knowledge unit and provide the set of one or more images to a user on a client device. The instructions further include instructions that cause the one or more processors to receive user input indicative of a selection of a first image from the set of one or more images and generate a thumbnail image for the knowledge unit based at least in part on the first image. In some embodiments, the instructions further include instructions that cause the one or more processors to associate the thumbnail image with the knowledge unit and display the thumbnail image to the user via the client device.
In accordance with certain embodiments, a system for generating a thumbnail image for a knowledge pack is provided. The system includes one or more processors and a memory coupled with and readable by the one or more processors. The memory is configured to store a set of instructions which, when executed by the one or more processors, causes the one or more processors to receive a request for determining a representative image for a knowledge pack, determine a set of tags associated with the knowledge pack, determine a set of one or more images for the knowledge pack based at least in part on the tags, determine a representative image for the knowledge pack based on the set of one or more images, generate a thumbnail image for the knowledge pack based on the representative image, associate the thumbnail image with the knowledge pack and display the thumbnail image for the knowledge pack to a user via a client device.
The techniques described above and below may be implemented in a number of ways and in a number of contexts. Several example implementations and contexts are provided with reference to the following figures, as described below in more detail. However, the following implementations and contexts are but a few of many.
The present disclosure relates generally to knowledge automation. Certain techniques are disclosed for discovering data content and transforming information in the data content into knowledge units. Techniques are also disclosed for composing individual knowledge units into knowledge packs, and mapping the knowledge to the appropriate target audience for consumption. Techniques are further disclosed for generating thumbnail images for knowledge units and knowledge packs.
Substantial amounts of data (e.g., data files such as documents, emails, images, code, and other content, etc.) may be available to users in an enterprise. These users may rely on information contained in the data to assist them in performing their tasks. The users may also rely on information contained in the data to generate useful knowledge that is consumed by other users. For example, a team of users may take technical specifications related to a new product release, and generate a set of training materials for the technicians who will install the new product. However, the large quantities of data available to these users may make it difficult to identify the right information to use.
Machine learning techniques can analyze content at scale (e.g., enterprise-wide and beyond) and identify patterns of what is most useful to which users. Machine learning can be used to model both the content accessible by an enterprise system (e.g., local storage, remote storage, and cloud storage services, such as SharePoint, Google Drive, Box, etc.), and the users who request, view, and otherwise interact with the content. Based on a user's profile and how the user interacts with the available content, each user's interests, expertise, and peers can be modeled. The data content can then be matched to the appropriate users who would most likely be interested in that content. In this manner, the right knowledge can be provided to the right users at the right time. This not only improves the efficiency of the users in identifying and consuming knowledge relevant for each user, but also improves the efficiency of computing systems by freeing up computing resources that would otherwise be consumed by efforts to search and locate the right knowledge, and allowing these computing resources to be allocated for other tasks.
To make effective use of the content available to the user, knowledge units and/or knowledge packs can be presented to the user through a graphical user interface. In addition to information about the content (such as title, publication information, authorship information, etc.) in some embodiments, a thumbnail image can be displayed. For image and/or video-based content, thumbnail images can be generated by conventional means. However, for text-based content (research papers, white papers, instruction manuals, and other documents), thumbnail images are typically generated based on a first page of the content. This often results in a thumbnail image that provides very little information to the user about the content of the knowledge unit and/or knowledge pack.
Embodiments of the present invention present techniques for generating representative images for non-multimedia documents such as text documents, PDF documents, and the like by analyzing the data contents of a knowledge unit and/or knowledge pack. Multiple representative images for a knowledge unit and/or knowledge pack can be generated and merged to form a single representative image (e.g., a collage or animated image) for the knowledge unit and/or knowledge pack. A thumbnail image may then be generated for the knowledge unit and/or knowledge pack based on the representative image. In some examples, the thumbnail image may be a reduced sized image version of the representative image of the knowledge unit and/or knowledge pack. In some embodiments, a user can manually select one or more representative images for a knowledge unit and/or knowledge pack. In other embodiments, a representative image for a knowledge unit and/or knowledge pack can be automatically generated by a knowledge automation system and/or identified from external image sources and presented to a user on the user's client device. In some embodiments, dynamic metadata can be integrated into a representative image for a knowledge unit and/or knowledge pack.
Knowledge automation system 100 can be implemented as a data processing system, and may discover and analyze content from one or more content sources 195 stored in one or more data repositories, such as a databases, file systems, management systems, email servers, object stores, and/or other repositories or data stores. In some embodiments, client devices 160-1, 160-2, . . . 160-n can access the services provided by knowledge automation system 100 through a network such as the Internet, a wide area network (WAN), a local area network (LAN), an Ethernet network, a public or private network, a wired network, a wireless network, or a combination thereof. Content sources 195 may include enterprise content 170 maintained by an enterprise, remote content 180 maintained at one or more remote locations (e.g., the Internet), cloud services content 190 maintained by cloud storage service providers, etc. Content sources 195 can be accessible to knowledge automation system 100 through a local interface, or through a network interface connecting knowledge automation system 100 to the content sources via one or more of the networks described above. In some embodiments, one or more of the content sources 195, one or more of the client devices 160-1, 160-2, . . . 160-n, and knowledge automation system 100 can be part of the same network, or can be part of different networks.
Each client device can request and receive knowledge automation services from knowledge automation system 100. Knowledge automation system 100 may include various software applications that provide knowledge-based services to the client devices. In some embodiments, the client devices can access knowledge automation system 100 through a thin client or web browser executing on each client device. Such software as a service (SaaS) models allow multiple different clients (e.g., clients corresponding to different customer entities) to receive services provided by the software applications without installing, hosting, and maintaining the software themselves on the client device.
Knowledge automation system 100 may include a content ingestion module 110, a knowledge modeler 130, and a user modeler 150, which collectively may extract information from data content accessible from content sources 195, derive knowledge from the extracted information, and provide recommendation of particular knowledge to particular clients. Knowledge automation system 100 can provide a number of knowledge services based on the ingested content. For example, a corporate dictionary can automatically be generated, maintained, and shared among users in the enterprise. A user's interest patterns (e.g., the content the user typically views) can be identified and used to provide personalized search results to the user. In some embodiments, user requests can be monitored to detect missing content, and knowledge automation system 100 may perform knowledge brokering to fill these knowledge gaps. In some embodiments, users can define knowledge campaigns to generate and distribute content to users in an enterprise, monitor the usefulness of the content to the users, and make changes to the content to improve its usefulness.
Content ingestion module 110 can identify and analyze enterprise content 170 (e.g., files and documents, other data such as e-mails, web pages, enterprise records, code, etc. maintained by the enterprise), remote content 180 (e.g., files, documents, and other data, etc. stored in remote databases), cloud services content 190 (e.g., files, documents, and other data, etc. accessible form the cloud), and/or content from other sources. For example, content ingestion module 110 may crawl or mine one or more of the content sources to identify the content stored therein, and/or monitor the content sources to identify content as they are being modified or added to the content sources. Content ingestion module 110 may parse and synthesize the content to identify the information contained in the content and the relationships of such information. In some embodiments, ingestion can include normalizing the content into a common format, and storing the content as one or more knowledge units in a knowledge bank 140 (e.g., a knowledge data store). In some embodiments, content can be divided into one or more portions during ingestion. For example, a new product manual may describe a number of new features associated with a new product launch. During ingestion, those portions of the product manual directed to the new features may be extracted from the manual and stored as separate knowledge units. These knowledge units can be tagged or otherwise be associated with metadata that can be used to indicate that these knowledge units are related to the new product features. In some embodiments, content ingestion module 110 may also perform access control mapping to restrict certain users from being able to access certain knowledge units.
Knowledge modeler 130 may analyze the knowledge units generated by content ingestion module 120, and combine or group knowledge units together to form knowledge packs. A knowledge pack may include various related knowledge units (e.g., several knowledge units related to a new product launch can be combined into a new product knowledge pack). In some embodiments, a knowledge pack can be formed by combining other knowledge packs, or a mixture of knowledge unit(s) and knowledge pack(s). The knowledge packs can be stored in knowledge bank 140 together with the knowledge units, or be stored separately. Knowledge modeler 130 may automatically generate knowledge packs by analyzing the topics covered by each knowledge unit, and combining knowledge units covering a similar topic into a knowledge pack. In some embodiments, knowledge modeler 130 may allow a user (e.g., a knowledge publisher) to build custom knowledge packs, and to publish custom knowledge packs for consumption by other users.
User modeler 150 may monitor user activities on the system as they interact with the knowledge bank 140 and the knowledge units and knowledge packs stored therein (e.g., the user's search history, knowledge units and knowledge packs consumed, knowledge packs published, time spent viewing each knowledge pack and/or search results, etc.). User modeler 150 may maintain a profile database 160 that stores user profiles for users of knowledge automation system 100. User modeler 150 may augment the user profiles with behavioral information based on user activities. By analyzing the user profile information, user modeler 150 can match a particular user to knowledge packs that the user may be interested in, and provide the recommendations to that user. For example, if a user has a recent history of viewing knowledge packs directed to a wireless networks, user modeler module 150 may recommend other knowledge packs directed to wireless networks to the user. As the user interacts with the system, user modeler 150 can dynamically modify the recommendations based on the user's behavior. User modeler 150 may also analyze search results performed by users to determine the effectiveness of the search results successful (e.g., did the user select and use the results), and to identify potential knowledge gaps in the system. In some embodiments, user modeler 150 may provide these knowledge gaps to content ingestion module 310 to find useful content to fill the knowledge gaps.
The processing illustrated in flow diagram 200 may begin with content ingestion 201. Content ingestion 201 may include content discovery 202, content synthesis 204, and knowledge units generation 206. Content ingestion 201 can be initiated at block 202 by performing content discovery to identify and discover data content (e.g., data files) at one or more data sources such as one or more data repositories. At block 204, content synthesis is performed on the discovered data content to identify information contained in the content. The content synthesis may analyze text, patterns, and metadata variables of the data content.
At block 206, knowledge units are generated from the data content based on the synthesized content. Each knowledge unit may represent a chunk of information that covers one or more related subjects. The knowledge units can be of varying sizes. For example, each knowledge unit may correspond to a portion of a data file (e.g., a section of a document) or to an entire data file (e.g., an entire document, an image, etc.). In some embodiments, multiple portions of data files or multiple data files can also be merged to generate a knowledge unit. By way of example, if an entire document is focused on a particular subject, a knowledge unit corresponding to the entire document can be generated. If different sections of a document are focused on different subjects, then different knowledge units can be generated from the different sections of the document. A single document may also result in both a knowledge unit generated for the entire document as well as knowledge units generated from portions of the document. As another example, various email threads relating to a common subject can be merged into a knowledge unit. The generated knowledge units are then indexed and stored in a searchable knowledge bank.
At block 208, content analysis is performed on the knowledge units. The content analysis may include performing semantics and linguistics analyses and/or contextual analysis on the knowledge units to infer concepts and topics covered by the knowledge units. Key terms (e.g., keywords and key phrases) can be extracted, and each knowledge unit can be associated with a term vector of key terms representing the content of the knowledge unit. In some embodiments, named entities can be identified from the extracted key terms. Examples of named entities may include place names, people's names, phone numbers, social security numbers, business names, dates and time values, etc. Knowledge units covering similar concepts can be clustered, categorized, and tagged as pertaining to a particular topic or topics. Taxonomy generation can also be performed to derive a corporate dictionary identifying key terms and how the key terms are used within an enterprise.
At block 210, knowledge packs are generated from individual knowledge units. The knowledge packs can be automatically generated by combining knowledge units based on similarity mapping of key terms, topics, concepts, metadata such as authors, etc. In some embodiments, a knowledge publisher can also access the knowledge units generated at block 206 to build custom knowledge packs. A knowledge map representing relationships between the knowledge packs can also be generated to provide a graphical representation of the knowledge corpus in an enterprise.
At block 212, the generated knowledge packs are mapped to knowledge consumers who are likely to be interested in the particular knowledge packs. This mapping can be performed based on information about the user (e.g., user's title, job function, etc.), as well as learned behavior of the user interacting with the system (e.g., knowledge packs that the user has viewed and consumed in the past, etc.). The user mapping can also take into account user feedback (e.g., adjusting relative interest levels, search queries, ratings, etc.) to tailor future results for the user. Knowledge packs mapped to a particular knowledge consumer can be distributed to the knowledge consumer by presenting the knowledge packs on a recommendations page for the knowledge consumer.
Content ingestion module 310 may include a content discovery module 312, a content synthesizer 314, and a knowledge unit generator 316. Content discovery module 312 interfaces with one or more content sources to discover contents stored at the content sources, and to retrieve the content for analysis. In some embodiments, knowledge automation system 300 can be deployed to an enterprise that already has a pre-existing content library. In such scenarios, content discovery module 312 can crawl or mine the content library for existing data files, and retrieve the data files for ingestion. In some embodiments, the content sources can be continuously monitored to detect the addition, removal, and/or updating of content. When new content is added to a content source or a pre-existing content is updated or modified, content discovery module 312 may retrieve the new or updated content for analysis. New content may result in new knowledge units being generated, and updated content may result in modifications being made to affected knowledge units and/or new knowledge units being generated. When content is removed from a content source, content discovery module 312 may identify the knowledge units that were derived from the removed content, and either remove the affected knowledge units from the knowledge bank, or tag the affected knowledge units as being potentially invalid or outdated.
Content synthesizer 314 receives content retrieved by content discovery module 312, and synthesizes the content to extract information contained in the content. The content retrieved by content discovery module 312 may include different types of content having different formats, storage requirements, etc. As such, content synthesizer 314 may convert the content into a common format for analysis. Content synthesizer 314 may identify key terms (e.g., keywords and/or key phrases) in the content, determine a frequency of occurrence of the key terms in the content, and determining locations of the key terms in the content. In addition to analyzing information contained in the content, content synthesizer 314 may also extract metadata associated with the content (e.g., author, creation date, title, revision history, etc.).
Knowledge unit generator 314 may then generate knowledge units from the content based on patterns of key terms used in the content and the metadata associated with the content. For example, if a document has a large frequency of occurrence of a key term in the first three paragraphs of the document, but a much lower frequency of occurrence of that same key term in the remaining portions of the document, the first three paragraphs of the document can be extracted and formed into a knowledge unit. As another example, if there is a large frequency of occurrence of a key term distributed throughout a document, the entire document can be formed into a knowledge unit. The generated knowledge units are stored in a knowledge bank 340, and indexed based on the identified key terms (also referred to herein as “tags”) and metadata to make the knowledge units searchable in knowledge bank 340.
Knowledge modeler 330 may include content analyzer 332, knowledge bank 340, knowledge pack generator 334, and knowledge pack builder 336. Content analyzer 332 may perform various types of analyses on the knowledge units to model the knowledge contained in the knowledge units. For example, content analyzer 332 may perform key term extraction and entity (e.g., names, companies, organizations, etc.) extraction on the knowledge units, and build a taxonomy of key terms and entities representing how the key terms and entities are used in the knowledge units. Content analyzer 332 may also perform contextual, sematic, and linguistic analyses on the knowledge units to infer concepts and topics covered by the knowledge units. For example, natural language processing can be performed on the knowledge units to derive concepts and topics covered by the knowledge units. Based on the various analyses, content analyzer 332 may derive a term vector for each knowledge unit to represent the knowledge contained in each knowledge unit. The term vector for a knowledge unit may include key terms, entities, and dates associated with the knowledge unit, topic and concepts associated with the knowledge unit, and/or other metadata such as authors associated with the knowledge unit. Using the term vectors, content analyzer 332 may perform similarity mapping between the knowledge units to identify knowledge units that cover similar topics or concepts.
Knowledge pack generator 334 may analyze the similarity mapping performed by content analyzer 332, and automatically form knowledge packs by combining similar knowledge units. For example, knowledge units that share at least five common key terms can be combined to form a knowledge pack. As another example, knowledge units covering the same topic can be combined to form a knowledge pack. In some embodiments, a knowledge pack may include other knowledge packs, or a combination of knowledge pack(s) and knowledge unit(s). For example, knowledge packs that are viewed and consumed by the a set of users can be combined into a knowledge pack. The generated knowledge packs can be tagged with their own term vectors to represent the knowledge contain in the knowledge pack, and be stored in knowledge bank 340.
Knowledge pack builder 336 may provide a user interface to allow knowledge publishers to create custom knowledge packs. Knowledge pack builder 336 may present a list of available knowledge units to a knowledge publisher to allow the knowledge publisher to select specific knowledge units to include in a knowledge pack. In this manner, a knowledge publisher can create a knowledge pack targeted to specific knowledge consumers. For example, a technical trainer can create a custom knowledge pack containing knowledge units covering specific new features of a produce to train a technical support staff. The custom knowledge packs can also be tagged and stored in knowledge bank 340.
Knowledge bank 340 is used for storing knowledge units 342 and knowledge packs 344. Knowledge bank 340 can be implemented as one or more data stores. Although knowledge bank 340 is shown as being local to knowledge automation system 300, in some embodiments, knowledge bank 340, or part of knowledge bank 340 can be remote to knowledge automation system 300. In some embodiments, frequently requested, or otherwise highly active or valuable knowledge units and/or knowledge packs, can be maintained in a low latency, multiple redundancy data store. This makes the knowledge units and/or knowledge packs quickly available when requested by a user. Infrequently accessed knowledge units and/or knowledge packs may be stored separately in slower storage.
Each knowledge unit and knowledge pack can be assigned an identifier that is used to identify and access the knowledge unit or knowledge pack. In some embodiments, to reduce memory usage, instead of storing the actual content of each knowledge unit in knowledge bank 340, the knowledge unit identifier referencing the knowledge unit and the location of the content source of the content associated with the knowledge unit can be stored. In this manner, when a knowledge unit is accessed, the content associated with the knowledge unit can be retrieved from the corresponding content source. For a knowledge pack, an knowledge pack identifier referencing the knowledge pack, and the identifiers and locations of the knowledge units and/or knowledge packs that make up the knowledge pack can be stored. Thus, a particular knowledge pack can be thought of as a container or a wrapper object for the knowledge units and/or knowledge packs that make up the particular knowledge pack. In some embodiments, knowledge bank 340 may also store the actual content of the knowledge units, for example, in a common data format. In some embodiments, knowledge bank 340 may selectively store some content while not storing other content (e.g., content of new or frequently accessed knowledge units can be stored, whereas stale or less frequently accessed content are not stored in knowledge bank 340).
Knowledge units 342 can be indexed in knowledge bank 340 according to key terms contained in the knowledge unit (e.g., may include key words, key phrases, entities, dates, etc. and number of occurrences of such in the knowledge unit) and/or associated metadata (e.g., author, location such as URL or identifier of the content, date, language, subject, title, file or document type, etc.). In some embodiments, the metadata associated with a knowledge unit may also include metadata derived by knowledge automation system 300. For example, this may include information such as access control information (e.g., which user or user group can view the knowledge unit), topics and concepts covered by the knowledge unit, knowledge consumers who have viewed and consumed the knowledge unit, knowledge packs that the knowledge unit is part of, time and frequency of access, etc.). Knowledge packs 344 stored in knowledge bank may include knowledge packs automatically generated by the system, and/or custom knowledge packs created by users (e.g., knowledge publishers). Knowledge packs 344 may also be indexed in a similar manner as for knowledge packs described above. In some embodiments, the metadata for a knowledge pack may include additional information that a knowledge unit may not have. For example, these may include a category type (e.g., newsletter, emailer, training material, etc.), editors, target audience, etc.
In some embodiments, a term vector can be associated with each knowledge element (e.g., a knowledge unit and/or a knowledge pack). The term vector may include key terms, metadata, and derived metadata associated with the each knowledge element. In some embodiments, instead of including all key terms present in a knowledge element, the term vector may include a predetermined number of key terms with the highest occurrence count in the knowledge element (e.g., the top five key terms in the knowledge element, etc.), or key terms that have greater than a minimum number of occurrences (e.g., key terms that appear more than ten times in a knowledge element, etc.).
User modeler 350 may include an event tracker 352, an event pattern generator 354, a profiler 356, a knowledge gap analyzer 364, a recommendations generator 366, and a profile database 360 that stores a user profile for each user of knowledge automation system 300. Event tracker 352 monitors user activities and interactions with knowledge automation system 300. For example, the user activities and interactions may include knowledge consumption information such as which knowledge unit or knowledge pack that a user has viewed, the length of time spent on the knowledge unit/pack, and when did the user access the knowledge unit/pack. The user activities and interactions tracked by event tracker 352 may also include search queries performed by the users, and user responses to the search results (e.g., number and frequency of similar searches performed by the same user and by other users, amount of time a user spends on reviewing the search result, how deep into a result list the user traversed, the number of items in the result list the user accessed and length of time spend on each item, etc.). If a user is a knowledge publisher, event tracker 352 may also track the frequency that the knowledge publisher publishes, when the knowledge publisher publishes, and topics or categories that the knowledge publisher publishes in, etc.
Event pattern generator 354 may analyze the user activities and interactions tracked by event tracker 352, and derive usage or event patterns for users or user groups. Profiler 356 may analyze these patterns and augment the user profiles stored in profile database 360. For example, if a user has a recent history of accessing a large number of knowledge packs relating to a particular topic, profiler 356 may augment the user profile of this user with an indication that this user has an interest in the particular topic. For patterns relating to search queries, knowledge gap analyzer 364 may analyze the search query patterns and identify potential knowledge gaps relating to certain topics in which useful information may be lacking in the knowledge corpus. Knowledge gap analyzer 364 may also identify potential content sources to fill the identified knowledge gaps. For example, a potential content source that may fill a knowledge gap can be a knowledge publisher who frequently publishes in a related topic, the Internet, or some other source from which information pertaining to the knowledge gap topic can be obtained.
Recommendations generator 366 may provide a knowledge mapping service that provides knowledge pack recommendations to knowledge consumers of knowledge automation system 300. Recommendations generator 366 may compare the user profile of a user with the available knowledge packs in knowledge bank 340, and based on the interests of the user, recommend knowledge packs to the user that may be relevant for the user. For example, when a new product is released and a product training knowledge pack is published for the new product, recommendations generator 366 may identify knowledge consumers who are part of a sales team, and recommend the product training knowledge pack to those users. In some embodiments, recommendations generator 366 may generate user signatures form the user profiles and knowledge signatures from the knowledge elements (e.g., knowledge units and/or knowledge packs), and make recommendations based on comparisons of the user signatures to the knowledge signatures. The analysis can be performed by recommendations generator 366, for example, when a new knowledge pack is published, when a new user is added, and/or when the user profile of a user changes.
In an embodiment of the present invention, knowledge automation system 402 may be configured to store the tenant-specific data in separate data stores 404, 406 within a knowledge bank 408 of knowledge automation system 402. For instance, the knowledge units and knowledge packs associated with a first tenant (tenant-1) may be stored in a first data store 404, the knowledge units and knowledge packs associated with a second tenant (tenant-2) may be stored in a second data store 406 and so on. In an embodiment, the knowledge units (KU-1, KU-2 and so on) may be stored in a sub-data store 410 within a data store (e.g., 404 or 406) and the knowledge packs (KP-1, KP-2 and so on) may be stored in a sub-data store 412 within a data store (e.g., 404 or 406). In certain embodiments, and as will be discussed in detail below, the tenant-specific knowledge units and knowledge packs may be associated with one or more sets of tags and thumbnail information. As noted above, the sets of tags may represent key terms contained in the knowledge unit or knowledge pack (e.g., may include key words, key phrases, entities, dates, etc. and number of occurrences of such in the knowledge unit) and/or associated metadata (e.g., author, location such as URL or identifier of the content, date, language, subject, title, file or document type, etc.). Thumbnail information, discussed in greater detail below, is also associated with knowledge units and knowledge packs and may include a representative image for the knowledge unit or knowledge pack and its associated thumbnail image.
In an embodiment, knowledge automation system 402 may include a knowledge unit generator 416, a knowledge pack builder 418, a thumbnail image generator 420, a user interface subsystem 422, an image manager 424 and a knowledge bank 408. The components of knowledge automation system 402 shown in
In certain embodiments, thumbnail image generator 420 may be configured to analyze the data contents of a knowledge unit and/or knowledge pack generated by knowledge unit generator 416 and/or stored in knowledge bank 408 and generate a “representative image” for the knowledge unit and/or knowledge pack. As noted above, a representative image may refer to a visual representation of the data contents within the knowledge unit and/or knowledge pack. Based on the generated representative image, in some embodiments, thumbnail image generator 420 may be configured to generate a “thumbnail image” for the knowledge unit and/or knowledge pack. In some instances, the “thumbnail image” may refer to a reduced size version of the representative image of the knowledge unit and/or knowledge pack.
In an embodiment, thumbnail image generator 420 may receive a request to determine a representative image for a knowledge unit (KU). For example, such a KU may be generated by knowledge unit generator 416 and/or stored in knowledge bank 408. Knowledge unit generator 416 may be the same or similar to knowledge unit generator 306 discussed in
Upon receiving the request, thumbnail image generator 420 may analyze the contents of the KU to determine a representative image for the KU. In an embodiment, the analysis of the KU may involve identifying text and non-text portions (e.g., graphical content such as images, figures, graphs, tables, and the like) of the KU and characteristics of these identified portions. For instance, the analysis may involve searching the KU for keywords such as “graph”, “figure” or “Fig” to locate the position of such regions within the KU. For example, for a text portion in the KU, thumbnail image generator 420 may identify key terms or tags associated with the KU. For instance, a key term may be identified based upon the frequency of the occurrence of the term in the KU, or based upon where the terms occur in KU and the like.
In certain embodiments, thumbnail image generator 420 can look for terms typically associated with graphical content to identify non-text content (e.g., graphical content such as images, figures, graphs, tables, and the like) within a knowledge unit. For example, thumbnail image generator 420 can include various object detection modules configured to identify different objects in an image. For example, object detection modules may apply image analysis techniques such as edge detection, curve detection, face detection, etc., to identify non-text objects within a document.
Thumbnail image generator 420 may then extract the identified non-text content to generate a set of one or more candidate representative images for the KU. In an embodiment, thumbnail image generator 420 may present the set of one or more candidate representative images to a user via a graphical user interface on client device 428 for selection by the user. For instance, when the user accesses knowledge automation system 402 using client device 428, user interface subsystem 422 may cause a graphical user interface to be displayed on client device 428 (e.g., via a browser application). Thumbnail image generator 420 may then receive user input indicative of a user selected representative image from the set of candidate images via the graphical user interface and generate a thumbnail image for the KU based on the user selected image. In alternate embodiments, thumbnail image generator 420 may also automatically select a particular representative image from the set of one or more candidate representative images and generate a thumbnail image for the KU based on the selected image. Additional details regarding the manner in which thumbnail image generator 420 may generate representative images and thumbnail images for a KU is discussed in detail in relation to
In some embodiments, thumbnail image generator 420 may provide the representative image and the thumbnail image for the KU to knowledge unit generator 416. Knowledge unit generator may then associate thumbnail information with the KU and store the KU and thumbnail information associated with the KU in knowledge bank 408.
In certain embodiments, and as noted above, thumbnail image generator 420 may also generate thumbnail images for a knowledge pack. For instance, thumbnail image generator 420 may receive a request to determine a representative image for a knowledge pack (KP) built by knowledge pack builder 418 and/or stored in knowledge bank 408. Knowledge pack builder 418 may be the same or similar to knowledge pack builder 336 discussed in
In certain embodiments, knowledge bank 408 may include a global image inventory 412. Global image inventory 414 may maintain and/or store an inventory of representative images for KUs and/or KPs from the different tenant data stores 404, 406. Global image inventory 424 may also store tag information (sets of tags) for each representative image for a KU and/or KP. In some embodiments, global image inventory 414 may also store images obtained from image sources 426. These image sources may include, for instance, images obtained from third party sources such as third-party images, graphics and content. In certain embodiments, and as will be discussed in detail below, thumbnail image generator 420 may utilize the image information stored in global image inventory 414 to generate a representative image and/or thumbnail image for a KU and/or KP.
In certain embodiments, image manager 424 in knowledge automation system 402 may populate global image inventory 412 with image information. For instance, image manager 424 may populate global image inventory 412 with representative images of KUs and/or KPs from tenant data stores 404, 406 and/or from third party images obtained from image sources 426. Additional details of the processes performed by knowledge automation system 402 to generate representative images and thumbnail images for KUs and/or KPs is discussed in
Additionally, some, any, or all of the processes may be performed under the control of one or more computer systems configured with executable instructions and may be implemented as code (e.g., executable instructions, one or more computer programs, or one or more applications) executing collectively on one or more processors, by hardware, or combinations thereof. As noted above, the code may be stored on a computer-readable storage medium, for example, in the form of a computer program including a plurality of instructions executable by one or more processors. The computer-readable storage medium may be non-transitory. In some examples, the knowledge automation system (e.g., utilizing at knowledge unit generator 416, thumbnail image generator 420, knowledge pack builder 418, user interface subsystem 422, image manager 424 and knowledge bank 408) shown in at least
In some embodiments, the user may select multiple images instead of a single image. In this embodiment, the process at 508 may include receiving a selection of the multiple (i.e., more than one) images and combining the selected one or more images into a single representative image for the knowledge pack. For instance, the process may include merging the selected one or more images onto a collage or animated image to generate a representative image for the knowledge unit.
In some embodiments, the analysis of the data contents of the KU by thumbnail mage generator 420 may involve, extracting key terms from the KU from text regions and/or portions of the KU and converting each page of the KU into an image. The analysis may involve analyzing and separating the regions containing text and non-text regions. For each non-image region, the presence of lines, curves, and the like may checked to determine the presence of graphs in the region. Additionally, as noted above, a keyword search for “graph”, “figure” or “Fig” (that provide a good understating of the summary of the contents of the KU) can be performed to locate the position of such regions in the KU. In some embodiments, these locations may be identified as candidates of high interest to be ultimately used as thumbnail images for the KU.
In some embodiments, thumbnail mage generator 420 may analyze the text regions in the KU to identify terms which relate to the terms “graph” or “figure” to enhance the “graph” or “figure” by highlighting the related terms on it. In some examples, thumbnail mage generator 420 may execute object detection algorithms like face detection to localize regions for each non-image region. These localized regions may then be used as candidates for thumbnails. In some examples, thumbnail mage generator 420 may determine a class of object detectors to be used for the KU based on the category of the KU or based on keywords found in the textual region of the document.
In some embodiments, the analysis may involve analyzing the text regions to identify terms or phrases referring to the objects found by the object detector and adding the terms or phrases to the image of the object to enhance the image of the object. In some examples, a sliding window may be utilized for each such region and the window may be moved in the image. Additionally, image features such as color diversity, histogram, grey scale version of the region, text density and the like can be computed for a particular window position. In some examples, region growing scheme for the window may be used so that multiple windows may be combined (grown) into a larger window. The processing described above for text region may also be performed for color-based features in the KU.
In certain embodiments, the above features may be used to cluster the windows into groups based on text density and color diversity. Images within two groups may further be clustered and an image from each group may be selected as representative images for the KU. In some embodiments, thumbnail image generator 420 may automatically choose an image with the highest color, diversity, or highest texture variation or other similar feature as the default representative image.
In some embodiments, the analysis of the KU may involve extracting key terms from the KU and using the key terms as search keys and performing a search for images from external resources. In some examples, a crawler may be used to download some of the matching images and the downloaded images may be used as candidate images for the KU. In some embodiments, the key terms of the KU and the image selected by the user may be tagged and stored in knowledge bank 408.
Based on the analysis performed by thumbnail image generator 420 as discussed above, at 606, thumbnail image generator 420 determines if the KU contains at least one extractable image. If it is determined the KU contains at least one extractable image, then, in some embodiments, thumbnail image generator 420 performs the process 608 described in
If it is determined that the KU contains a single extracted image, then, in some embodiments, at 612, thumbnail image generator 420 selects the extracted image as the representative image for the KU. At 614, thumbnail image generator 420 generates a thumbnail image for the KU based on the representative image. As noted above, a “thumbnail image” may refer to a reduced size version of the representative image of the KU. At 616, thumbnail image generator 420 associates the thumbnail image with the KU. In some embodiments, and as noted above, thumbnail image generator 420 may provide thumbnail information comprising the representative image and the thumbnail image to knowledge unit generator 416. Knowledge unit generator 416 may then store the thumbnail information for the KU in knowledge bank 408.
Image Score (I1)=aX1+bX2+cX3
wherein, a, b, c and so on are weights that are assigned to image features, X1, X2, X3 and so on. The weights assigned to each image feature may be pre-determined by thumbnail image generator 420 in some embodiments, or determined manually by a user, e.g., an administrator of knowledge automation system 402. Thumbnail image generator 420 may then select an image from the multiple extractable images having the highest image score. At 622, thumbnail image generator 420 may perform the processes 614 and 616 described in
In an alternate embodiment, in response to determining that the KU contains multiple extractable images (at 618), thumbnail image generator 420 may perform the processes described in steps 624-632. For instance, at 624, thumbnail image generator 420 may select multiple (e.g., more than one) images from the images extracted at 618. In an embodiment, the selection may be based on a score determined for the images as discussed above. For instance, if it is determined that the KU contains 10 extractable images at 618, thumbnail image generator 420 may select 4 images out of the 10 images based on their respective scores. At 626, thumbnail image generator 420 may output the selected images to a user for example via a graphical user interface on the user's client. At 628, thumbnail image generator 420 may receive user input indicative of a user-selected image from the output images. In some examples, at 630, thumbnail image generator 420 may mark the user-selected image as the selected image. At 632, thumbnail image generator 420 may perform the processes 614 and 616 described in
In some embodiments, at 640, thumbnail image generator 420 may determine a best match set of tags from the matching sets of tags. In an example, thumbnail image generator 420 may determine the best match set of tags for the KU by identifying a set of tags from the matching sets of tags having the maximum number of tags that match the tags associated with the KU. For example, consider that a KU is associated with the following set of tags:
KU={T1, T2, T3, T4}
where T1, T2, T3 and T3 represent different tags (terms) that identify the KU.
Further, consider that the matching sets of tags associated with the sets of images stored in the global image inventory determined by thumbnail image generator 420 are as follows:
I1={T1, T2, T5, T6}
I2={T1, T3, T7, T8}
I3={T1, T2, T3}
I4={T1, T2, T3, T7, T8}
In an embodiment, thumbnail image generator 420 may determine the best match set of tags from the matching sets of tags to be {T1, T2, T3} associated with image I3. This is because, I3 is associated with a maximum number of tags {T1, T2, T3} that match the tags {T1, T2, T3, T4} associated with the KU.
At 642, thumbnail image generator 420 may mark or identify the image in the global image inventory corresponding to the best match set of tags as the representative image for the KU. For instance, per the example discussed above, thumbnail image generator 420 may mark image I3 as the representative image for the KU. At 644, thumbnail image generator 420 may perform the processes 614 and 616 described in
In an alternate embodiment, at 646, instead of identifying a single best match set of tags as discussed above, thumbnail image generator 420 may identify multiple sets of tags from the matching sets of tags determined in 638 as best match sets of tags. For instance, as per the example discussed above, thumbnail image generator 420 may identify that the set of tags {T1, T2, T3} associated with image I3 and the set of tags {T1, T2, T3, T7, T8} associated with I4 to be best match sets of tags for the KU.
At 648, thumbnail image generator 420 may mark the images (e.g., I3, I4) in the global image inventory that correspond to the best match sets of tags. At 650, thumbnail image generator 420 may output the images to the user on the user's client device. At 652, thumbnail image generator 420 may receive user input indicative of a user-selected image from the output images. At 654, thumbnail image generator 420 may mark the user-selected image as the representative image for the KU. In some embodiments, at 656, thumbnail image generator 420 may then perform the processes 614 and 616 of
The above discussion related to the generation of a representative image and a thumbnail image for a knowledge unit (KU) generated by knowledge automation system 402. In an alternate embodiments, thumbnail image generator 420 may also generate representative images and thumbnail images for a knowledge pack (KP) built by knowledge automation system 402. These processes are discussed in
At 712, thumbnail image generator 420 may associate the thumbnail image with the KP. At 714, thumbnail image generator 420 may display the thumbnail image for the KP via the user's client device.
In some embodiments, the processes 704-710 performed by thumbnail image generator 420 to determine a set of one or more candidate images for the KP, identify a representative image for the KP and generate a thumbnail image for the KP respectively may be similar to the processes 636-656 described in
If, for example, thumbnail image generator 420 identifies multiple sets of tags from the matching sets of tags, then thumbnail image generator 420 may determine images in the global image inventory that correspond to the multiple sets of tags and output the images to a user. Then, thumbnail image generator 420 may receive user input indicative of a user-selected image from the output images, mark the user-selected image as a representative image for the KP and generate a thumbnail image for the KP.
The example illustrated in
In an alternate embodiment, based on the obtained sets of images (from 802, 804), thumbnail image generator 420 may perform the processing described in 812-816 to representative image and/or a thumbnail image for the KP. For instance, at 812, thumbnail image generator 420 may select a set of one or more images from the obtained images. At 814, thumbnail image generator 420 may output the set of one or more images to a user for selection. At 816, thumbnail image generator 420 may receive user input indicative of a user-selected image. At 818, thumbnail image generator 420 may perform processes (e.g., 808, 810) discussed above to generate a thumbnail image for the KP based on the selected (representative) image and associate the generated thumbnail image with the KP.
In some embodiments, processing subsystem 1008 can include one or more processing units 1006, 1008. Processing units 1006, 1008 can include one or more of a general purpose or specialized microprocessor, FPGA, DSP, or other processor. In some embodiments, processing unit 1006, 1008 can be a single core or multicore processor.
In some embodiments, storage subsystem can include system memory 1012 which can include various forms of non-transitory computer readable storage media, including volatile (e.g., RAM, DRAM, cache memory, etc.) and non-volatile (flash memory, ROM, EEPROM, etc.) memory. Memory may be physical or virtual. System memory 1012 can include system software 1014 (e.g., BIOS, firmware, various software applications, etc.) and operating system data 1016. In some embodiments, storage subsystem 1010 can include non-transitory computer readable storage media 1018 (e.g., hard disk drives, floppy disks, optical media, magnetic media, and other media). A storage interface 1020 can allow other subsystems within computing system 1000 and other computing systems to store and/or access data from storage subsystem 1010.
In some embodiments, I/O subsystem 1022 can interface with various input/output devices, including displays (such as monitors, televisions, and other devices operable to display data), keyboards, mice, voice recognition devices, biometric devices, printers, plotters, and other input/output devices. I/O subsystem can include a variety of interfaces for communicating with I/O devices, including wireless connections (e.g., Wi-Fi, Bluetooth, Zigbee, and other wireless communication technologies) and physical connections (e.g., USB, SCSI, VGA, SVGA, HDMI, DVI, serial, parallel, and other physical ports).
In some embodiments, communication subsystem 1024 can include various communication interfaces including wireless connections (e.g., Wi-Fi, Bluetooth, Zigbee, and other wireless communication technologies) and physical connections (e.g., USB, SCSI, VGA, SVGA, HDMI, DVI, serial, parallel, and other physical ports). The communication interfaces can enable computing system 1000 to communicate with other computing systems and devices over local area networks wide area networks, ad hoc networks, mesh networks, mobile data networks, the internet, and other communication networks.
In certain embodiments, the various processing performed by a knowledge modeling system as described above may be provided as a service under the Software as a Service (SaaS) model. According this model, the one or more services may be provided by a service provider system in response to service requests received by the service provider system from one or more user or client devices (service requestor devices). A service provider system can provide services to multiple service requestors who may be communicatively coupled with the service provider system via a communication network, such as the Internet.
In a SaaS model, the IT infrastructure needed for providing the services, including the hardware and software involved for providing the services and the associated updates/upgrades, is all provided and managed by the service provider system. As a result, a service requester does not have to worry about procuring or managing IT resources needed for provisioning of the services. This significantly increases the service requestor's access to these services in an expedient manner at a much lower cost point.
In a SaaS model, services are generally provided based upon a subscription model. In a subscription model, a user can subscribe to one or more services provided by the service provider system. The subscriber can then request and receive services provided by the service provider system under the subscription. Payments by the subscriber to providers of the service provider system are generally done based upon the amount or level of services used by the subscriber.
Network 1112 can include one or more networks or any mechanism that enables communications between service provider system 1110 and service requestor devices 1102, 1104. Examples of network 1112 include without restriction a local area network, a wide area network, a mobile data network, the Internet, or other network or combinations thereof. Wired or wireless communication links may be used to facilitate communications between the service requestor devices and service provider system 1110.
In the embodiment depicted in
In some embodiments, service manager component 1116 is configured to manage provision of services to one or more service requesters. Service manager component 1116 may be configured to receive service requests received by service provider system 1110 via access interface 1114, manage resources for providing the services, and deliver the services to the requesting requesters. Service manager component 1116 may also be configured to receive requests to establish new service subscriptions with service requestors, terminate service subscriptions with service requestors, and/or update existing service subscriptions. For example, a service requestor device can request to change a subscription to one or more service applications 1122-1126, change the application or applications to which a user is subscribed, etc.).
Service provider system 1110 may use a subscription model for providing services to service requestors according to which a subscriber pays providers of the service provider system based upon the amount or level of services used by the subscriber. In some embodiments, billing component 1118 is responsible for managing the financial aspects related to the subscriptions. For example, billing component 1110, in association with other components of service provider system 1110, may be configured to determine amounts owed by subscribers, send billing statements to subscribers, process payments from subscribers, and the like.
In some embodiments, service applications 1120 can include various applications that provide various SaaS services. For example, one more applications 1120 can provide the various functionalities described above and provided by a knowledge modeling system.
In some embodiments, tenant-specific data 1132 comprises data for various subscribers or customers (tenants) of service provider system 1110. Data for one tenant is typically isolated from data for another tenant. For example, tenant 1's data 1134 is isolated from tenant 2's data 1136. The data for a tenant may include without restriction subscription data for the tenant, data used as input for various services subscribed to by the tenant, data generated by service provider system 1110 for the tenant, customizations made for or by the tenant, configuration information for the tenant, and the like. Customizations made by one tenant can be isolated from the customizations made by another tenant. The tenant data may be stored service provider system 1110 (e.g., 1134, 1136) or may be in one or more data repositories 1138 accessible to service provider system 1110.
It should be understood that the methods and processes described herein are exemplary in nature, and that the methods and processes in accordance with some embodiments may perform one or more of the steps in a different order than those described herein, include one or more additional steps not specially described, omit one or more steps, combine one or more steps into a single step, split up one or more steps into multiple steps, and/or any combination thereof.
It should also be understood that the components (e.g., functional blocks, modules, units, or other elements, etc.) of the devices, apparatuses, and systems described herein are exemplary in nature, and that the components in accordance with some embodiments may include one or more additional elements not specially described, omit one or more elements, combine one or more elements into a single element, split up one or more elements into multiple elements, and/or any combination thereof.
Although specific embodiments of the invention have been described, various modifications, alterations, alternative constructions, and equivalents are also encompassed within the scope of the invention. Embodiments of the present invention are not restricted to operation within certain specific data processing environments, but are free to operate within a plurality of data processing environments. Additionally, although embodiments of the present invention have been described using a particular series of transactions and steps, it should be apparent to those skilled in the art that the scope of the present invention is not limited to the described series of transactions and steps. Various features and aspects of the above-described embodiments may be used individually or jointly.
Further, while embodiments of the present invention have been described using a particular combination of hardware and software, it should be recognized that other combinations of hardware and software are also within the scope of the present invention. Embodiments of the present invention may be implemented only in hardware, or only in software, or using combinations thereof. The various processes described herein can be implemented on the same processor or different processors in any combination. Accordingly, where components or modules are described as being configured to perform certain operations, such configuration can be accomplished, e.g., by designing electronic circuits to perform the operation, by programming programmable electronic circuits (such as microprocessors) to perform the operation, or any combination thereof. Processes can communicate using a variety of techniques including but not limited to conventional techniques for inter-process communication, and different pairs of processes may use different techniques, or the same pair of processes may use different techniques at different times.
The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense. It will, however, be evident that additions, subtractions, deletions, and other modifications and changes may be made thereunto without departing from the broader spirit and scope as set forth in the claims. Thus, although specific invention embodiments have been described, these are not intended to be limiting. Various modifications and equivalents are within the scope of the following claims. For example, one or more features from any embodiment may be combined with one or more features of any other embodiment without departing from the scope of the invention.
The present application is a non-provisional of and claims the benefit and priority of U.S. Provisional Application No. 62/054,333, filed Sep. 23, 2014, entitled “Automatic Thumbnail Generation,” the entire contents of all of which are incorporated herein by reference for all purposes.
Number | Date | Country | |
---|---|---|---|
62054333 | Sep 2014 | US |