Chatbots are a useful tool to help customer support teams. For example, chatbots can include rules to route users to correct units of customer support/success organizations. As another example, chatbots can directly handle simple user queries based on explicit rules where a specific intent (e.g., a distinct user goal or request in the query) can be identified in the query. However, in order to implement chatbots that can directly handle simple user queries, chatbots require a significant amount of configuration prior to implementation.
Various aspects of the technology described herein are generally directed to systems, methods, and computer storage media for, among other things, programmatically implementing a specialized chatbot platform that uses language models to determine answers from semantically similar documents of knowledge bases. In this regard, embodiments described herein facilitate using a language model to determine aspects of a conversation with a user in order to compute the semantic similarity of the conversation to answers provided in documents of a knowledge base (e.g., manually curated files, historical conversation files, extracted snippets from conversations, Uniform Resource Locators (“URLs”), and/or any other data within a knowledge base). For example, aspects of a conversation, such as a query occurring within the conversation or a summary of the conversation or portions thereof, can be programmatically identified and extracted from conversations between a user (e.g., an existing customer, a potential customer or any individual providing questions to customer support) and a chatbot. The similarity of sentence embeddings between the aspects of the conversation and sentences of documents in the knowledge base can be computed in order to determine whether the answer within a threshold semantic similarity is provided in a document in the database. The language model can then utilize the semantically similar document(s) in the database to generate an answer to the user query through the chatbot.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
Chatbots are a useful tool to help customer support teams. For example, chatbots can include rules to route users to correct units of customer support/success organizations. As another example, chatbots can directly handle simple user queries based on explicit rules where a specific intent (e.g., a distinct user goal or request in the query) can be identified in the query. However, in order to implement chatbots that can directly handle simple user queries, chatbots require a significant amount of configuration prior to implementation.
For example, in order to configure intent so that a specific intent can be identified, examples of previous user interactions representing the specific problem/intent must be manually curated by a developer so that the rules or patterns associated with the intent can be manually input into the system. The developer must then manually hard code the response that the chatbot delivers once the intent is identified. In this regard, after the intents and associated responses of the chatbot are configured, the developer must continually update the intents and responses to match up-to-date knowledge and policies.
Currently, in order to implement a chatbot, humans, such as a developer, must manually configure intents through predefined rules or patterns associated with specific user queries and must manually define the response to the intent. In this regard, a human must manually curate examples of user interactions to match the intent, manually define the rules or patterns for each intent, manually write predefined responses to the intent, and manually update each intent to match any changes to the intent and/or predefined responses. Accordingly, unnecessary computing resources are utilized during the manually intensive process of defining intents and/or responses in conventional implementations. For example, the increase in the number of computer input/output operations related to the manually intensive process of defining intents and/or responses is computationally intensive and unnecessarily consumes computing resources. Similarly, when the information is located over a network, network resources are unnecessarily consumed, thereby decreasing throughput, increasing latency, and increasing packet generation costs.
As such, embodiments of the present disclosure are directed to programmatically implementing a specialized chatbot platform that uses language models to determine answers from semantically similar documents of knowledge bases in an efficient and effective manner. In this regard, a language model can be efficiently and effectively utilized by a specialized chatbot platform in order to retrieve answers to queries from customers from an existing knowledge base without having to manually curate examples of user interactions to match the intent, manually define the rules or patterns for each intent, manually write predefined responses to the intent, and manually update each intent to match any changes to the intent and/or predefined responses.
Generally, and at a high level, embodiments described herein facilitate programmatically implementing a specialized chatbot platform that uses language models to determine answers from semantically similar documents of knowledge bases. In this regard, embodiments described herein facilitate using a language model to determine aspects of a conversation with a user in order to compute the semantic similarity of the conversation to answers provided in documents of a knowledge base (e.g., manually curated files, historical conversation files, extracted snippets from conversations, Uniform Resource Locators (“URLs”), and/or any other data within a knowledge base). The user may comprise a customer interacting with the chatbot platform, such as a user visiting a website or app and engaging in a chat session with a chatbot, or a customer support agent (“CSA”). For example, aspects of a conversation, such as a query occurring within the conversation or a summary of the conversation or portions thereof, can be programmatically identified and extracted from conversations between a user (e.g., an existing customer, a potential customer or any individual providing questions to customer support) and a chatbot. The similarity of sentence embeddings between the aspects of the conversation and sentences of documents in the knowledge base can be computed in order to determine whether the answer within a threshold semantic similarity is provided in a document in the database. The language model can then utilize the semantically similar document(s) in the database to generate an answer to the user query through the chatbot.
In some embodiments, the specialized chatbot platform, utilizing prompts to the language model, disambiguates (e.g., request clarification of) the question and answers user questions only if the question and answers are relevant to the domain of the business and can be answered from sources provided by a business implementing the chatbot. In some embodiments, utilizing prompts to the language model, the specialized chatbot platform is able to determine when the user wants to talk to a human, and escalates to the human, such as a CSA (e.g., or any support personnel). In some embodiments, the specialized chatbot platform collects feedback from users on their interactions to understand whether the delivered answer was useful and to generate reports presenting metrics regarding the implementation of the chatbot.
Advantageously, efficiencies of computing and network resource utilization can be enhanced using implementations described herein. In particular, the use of a language model by a specialized chatbot platform in order to retrieve answers to queries from customers from semantically similar documents of an existing knowledge base provides for a more efficient use of computing and network resources than conventional methods of manually curating examples of user interactions to match the intent, manually defining the rules or patterns for each intent, manually writing predefined responses to the intent, and manually updating each intent to match any changes to the intent and/or predefined responses. The technology described herein decreases the number of computer input/output operations related to otherwise manually intensive process of manually defining intents and/or responses, thereby decreasing computation costs and decreasing network resource utilization (e.g., higher throughput, lower latency, and decreasing packet generation costs due to fewer packets being sent) when the information is located over a computer network.
Turning to
In various implementations, the components of system architecture 100 include computer storage media that stores information including data, data structures, computer instructions (e.g., software program instructions, routines, or services), and/or models (e.g., machine learning models) used in some embodiments of the technologies described herein. For example, in some implementations, customer device 102, customer support device 112, language model 110, server 150, and/or storage 130 may comprise one or more data stores (or computer data memory). Further, although customer device 102, customer support device 112, server 150, language model 110, and storage 130 are each depicted as a single component in
The components of system architecture 100 communicate with each other via a network 104. In some embodiments, network 104 includes one or more local area networks (LANs), wide area networks (WANs), and/or other networks. Such networking environments are commonplace in offices, enterprise-wide computer networks, intranets, and the Internet.
In the example illustrated in
In various embodiments, the functionality described herein is allocated across any number of devices. In some embodiments, application(s) 106 and/or 116 are hosted at least partially server-side, such that chat interface 108, communication tool 120, chatbot tool 170, chatbot component 151, customer support component 165, and/or any of the elements illustrated in
An example workflow of the configuration illustrated in
Customer support device 112 is a desktop, laptop, or mobile device such as a tablet or smart phone, and application 116 provides one or more user interfaces. In some embodiments, an end user, such as a CSA of the business, chats, or accesses a chat (e.g., a conversation with the customer), with a customer through chat interface 124 of communication tool 120. Additionally or alternatively, a chatbot via chatbot component 151 chats, or accesses a chat (e.g., a conversation) with a customer through chat interface 108 of application 106 and an end user, such as a CSA of the business, chats, or accesses a chat between the chatbot and the customer through chat interface 124 of communication tool 120.
In some embodiments, chatbot component 151 facilitates programmatically implements a specialized chatbot platform that uses language models to determine answers from semantically similar documents of knowledge bases. For example, chatbot component 151 facilitates using a language model 110 to determine aspects of a conversation with a user in order to compute the semantic similarity of the conversation to answers provided in documents of a knowledge base 132 (e.g., extracted snippet files 133, manually curated files 134, such as public content 135 and/or private content 136, URLs 137, historical conversation files 138, and/or the like) and provide responses to the customer through chat interface 108.
In some embodiments, chatbot component 151 facilitates providing multimodal responses to the customer through chat interface 108. For example, chatbot component 151 facilitates using language model 110 as a multimodal language model to determine semantically similar images, audio, and/or video provided in documents of a knowledge base 132 (e.g., images, audio, and/or video stored in extracted snippet files 133, manually curated files 134, such as public content 135 and/or private content 136, URLs 137, historical conversation files 138, and/or the like). In another example, chatbot component 151 facilitates providing responses that include images, audio, and/or video that are included in the semantically similar answers of documents of a knowledge base 132. In another example, chatbot component 151 can provide responses that include images, audio, and/or video, generated by language model 110 where the language model is a multimodal generative language model.
In some embodiments, chatbot component 151 facilitates handling multimedia input (e.g., images, videos, gifs, voice notes, etc.), both in the ingested content and/or as the end-user input. An example of chatbot component 151 facilitating the handling of multimedia input in the end-user input is shown in diagram 600F of
An example of chatbot component 151 facilitating the handling of multimedia input in the ingested content is shown in diagram 600G of
Another example of chatbot component facilitating the handling of multimedia input in the end-user input is shown in diagram 600H of
In some embodiments, chatbot component 151 facilitates using a language model 110 to determine aspects of a conversation with a user in order to interact with external systems (e.g., external sources), such as an external application through third-party application configuration files 144, in order to provide a response or take an action with respect to the user. For example, chatbot component 151 may retrieve information from an ecommerce store, such as a price of an item or an answer to the user that allows the user to purchase an item, and provide the relevant response to the user. As another example, chatbot component 151 may retrieve information from an ecommerce system to determine whether historical customer data of the user indicates whether the customer qualifies for a particular offer or discount.
In some embodiments, chatbot component 151 facilitates augmenting the user context by utilizing data from external systems; for example, reading a user's order information from an ecommerce platform, or accessing tracking information for user's order. An example of chatbot component 151 facilitating the augmenting of the user context by reading the data from external sources is shown in diagram 600I of
In the specific example of the get_reply call above, there is one action defined. However, a get_reply call can be generated with any number of actions available. In some embodiments, if action has parameters, the action can be described with the following fields: name: text; description: text; type: enum (data type); required: boolean (true/false); and default_value: any value.
Continuing with diagram 600I, in some embodiments, the action selection process can include chatbot component 6020 facilitating choosing which actions to call in order to augment the user's context based on the user's issue summary, current conversation, and available actions. In this regard, the action selection process can choose to call 0, 1, or more actions of the given conversation state. The output of the action selection process, performed by action selections 6024, can comprise a list of actions to call, and arguments for each action call (e.g., call action “Get ECOMMERCE PLATFORM Order” with “order_id: <some_order_id_from_conversation>”). In some embodiments, the action calling process performed by action calling 6026, can include chatbot component 6020 facilitating calling back into server 6030, asking for the response of called actions (e.g., after action selection is performed). Server 6030 can call external system (e.g., via third-party application configuration files 144) and proxy the response. In some embodiments the response is redacted as users (e.g., teammates) can define which fields are returned back in the response. In some embodiments, the context augmentation process can include chatbot component 6020 facilitating augmenting the context in the “answer finding” stage (e.g., as described herein with respect to answer search state component 154 of
In some embodiments, by connecting with external tools, information from the external tools can be retrieved to provide more personalized replies. For example, actions can include, without limitation: (1) get order info from an ecommerce platform, (2) get subscription status from a payment processing platform, (3) check for ongoing incidents in a communication platform, and/or the like. Customer support teams of a business can also set up an action that connects to the business's own backend system. In some embodiments, actions are API calls that can be manually set up (e.g., read actions) to retrieve data from external systems. In some embodiments, actions can include write actions to update data in external systems.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include a set of functionalities designed to enhance customer control, security, and monitoring capabilities to allow customers to define, monitor, and control the operations that the chatbot (e.g., chatbot component 151) can perform, ensuring a secure and customizable experience. In this regard, a read action can allow the gathering of data from an external system, returning relevant information to the AI Agent (e.g., chatbot component 151) to allow for personalized and contextual responses to end users.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include customization and/or control configurations, such as (1) action specification where users can specify which actions chatbot component 151 is allowed to initiate and under what circumstances; (2) frequency and limits where users have control over how often and/or how much resources or money, chatbot component 151 can handle; (3) custom actions setup where users are required to use Identity Verification (IdV) and other secure defaults to prevent unauthorized updates; and/or the like.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include security measure configurations, such as (1) where IdV is mandatory for setting up custom actions, ensuring that only verified users can configure sensitive operations; (2) attribute protection where attributes that are not protected can be hidden, and updates from the messenger can be disabled to prevent abuse; (3) logging and monitoring where comprehensive logging is in place to track actions and interrogate incidents where chatbot component 151 might take unauthorized actions; and/or the like.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include monitoring configurations, such as where users can monitor the actions that chatbot component 151 is taking through an inbox view (e.g., messages inbox interface 122) that can be filtered to show conversations where actions, such as custom actions, were used.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include security recommendations, such as (1) risk awareness where users setting up chatbot component 151 to take custom actions are made aware of the risks of attribute abuse and are prompted to disable unnecessary updates from the messenger; (2) workspace security where the system can track which workspaces are in insecure states, ensuring that users are informed and can take necessary actions to secure their environments; (3) custom authentication models that can be deleted upon uninstall or workspace deletion to prevent unauthorized access; and/or the like.
With references to
In some embodiments, read actions (e.g., and/or write actions) for external systems can include security features, such as (1) attribute protection to implement attribute protection mechanisms; (2) IdV enforcement to enforce identity verification; (3) drop available actions if IdV is disabled to disable actions if IdV is not enabled; (4) API update for attribute protection to update APIs to enable attribute protection on create/update; (5) protected attributes for an ecommerce platform to ensure all existing attributes of the ecommerce platform are protected; (6) teammate activity log to log teammate activities related to action creation and updates; and/or the like.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include action preview features, such as (1) mocked responses to generate and provide mocked responses for actions; (2) action level preview to preview actions at the action level; (3) user journey to define user journeys to inform whether only draft actions should have their responses mocked in preview; (4) action configuration to allow users to configure actions, including adding audience rules and mocked responses; (5) interaction with redaction/formatting to determine how mocked responses interact with redaction and formatting; and/or the like.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include data controls features, such as (1) trusted identity and data to ensure data reliability by requiring IdV and controlling which data can be used; (2) namespace application data to isolate application-specific data and restrict updates via messenger; (3) data source tracking to track the source of data updates to improve reliability; (4) API default update to implement default updates for APIs to enhance data security; and/or the like.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include IdV and attribute abuse protection features, such as (1) IdV enforcement to enforce IdV for all messenger installations; (2) attribute protection to protect attributes from being updated via messenger and ensure they are managed securely; (3) error messaging to provide error messages for actions using unprotected attributes; (4) inform beta customers to notify beta customers about the enforcement of IdV and attribute protection; and/or the like.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include action access control features, such as (1) control access to actions to manage the ability of chatbot component 151 to access and use actions based on user and company data; (2) audience management to define audience predicates to control which users can trigger specific actions; (3) action availability to set actions to be available to chatbot component 151 by default or restrict them based on audience rules; and/or the like.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include secure by default features, such as (1) secure defaults to implement secure defaults for IdV and data usage; (2) namespace application data to restrict updates to application-specific data via messenger; (3) data validation to validate input fields and add allow lists to control data updates; (4) detailed webhooks to capture detailed information about data updates for better security and auditing; and/or the like.
In some embodiments, read actions (e.g., and/or write actions) for external systems can include action agent flow change features, such as (1) answerbot flow (e.g., answer search state component 154) to integrate actions into the answerbot flow; (2) conversation funnel to track the conversation funnel from start to resolution; (3) security to ensure secure handling of data and actions; (4) LangChain integration to use LangChain for LLM prompts and action execution; (5) write actions; (6) parameter extraction; (7) action chaining; and/or the like.
With reference to
At block 6220, based on user input from the chat session, an available action is determined. The available action may be determined based on the user's input, such as an input reflecting a query or action to be performed. For example, the user may ask about tracking information regarding an order or the status of a refund. In both of these examples, the user is requesting information, thus the user request represents a query. In some embodiments of block 6220, chatbot component 151 or a subcomponent, such as described in connection with
At block 6230, one or more of the available actions that are relevant to the user's input (e.g., a query representing the user's input) are determined. For example, chatbot component 151 may determine relevant actions that are responsive to a query representing the user's input, based on information provided at block 6235 including information about the available actions, question summaries from conversation history, recent passages in the user's chat session, the entire chat record of the chat session, and/or full conversation history with the user.
At block 6240, the one or more actions determined at block 6230 are executed, and the responses to the actions are provided in an answer prompt to a language model. In some embodiments, action, workspace, user, company, and conversation data from block 6245 is also utilized. The language model may comprise an LLM that is used to generate a conversational response from the chatbot to the user in the chat session.
At block 6250, it is determined that the answer cites the context of the action(s). Block 6250 uses, at block 6255, action responses determined from block 6250. At block 6260, content uses in the answer are tracked. Block 6260 uses content references of block 6265. At block 6270, the conversational response is evaluated to determine whether the response is sufficient to resolve the conversation with the user. For example, the response is evaluated to determine whether it is responsive to the user's query. At block 6277, the conversation response is determined not to resolve the conversation. In some instances, block 6277 (or processes occurring subsequent to block 6277) of process flow 600J includes generating a subsequent response to the user in the chat session requesting clarification or additional information in order to more accurately resolve the conversation. In some instances, block 6277 (or processes occurring subsequent to block 6277) includes escalating the chat session to a human CSA to assist the user. On the other hand, if the conversation response is determined at block 6270 to resolve the conversation, then process flow 600J proceeds to block 6275 indicating the response has resolved the conversation. At block 6280 a resolution is counted.
Turning to
As shown in
Agent executor 6410 is generally responsible for determining and facilitating execution of a tool agent 6420 to call, based on information regarding a chat session such as a context of the conversation with a user who is interacting with a chatbot and/or recent passages of the chat inputted by the user. In some embodiments, agent executor 6410 determines or facilitates determining a query from the user chat session, based on recent passage(s) of user input in the chat. For example, a user request for information corresponds to a query for the requested information. Similarly, a user request for help to solve a problem may correspond to a query for information needed to address the problem. Tool agents 6420 include various software agents operating in an MAS, such as, by way of example and without limitation, agents 6422 for routing the user to a customer support agent, greeting the customer, saying goodbye, and other actions. Tool agents 6420 also includes Search for Answers Agent 6424.
Search for answers agent 6424 is generally responsible for determining retrieval of information for answering a particular query corresponding to the user input in the chat session. For example, search for answers agent 6424 can access a knowledge base (e.g., knowledge base 132 of
Actions manager 6430 is generally responsible for planning and managing one or more actions needed to address the user's query. In some scenarios or contexts, multiple actions are used to address a user's query. Accordingly, actions manager 6430 determines or facilitates determining (e.g., using Actions LLM 6435) which actions are needed. Actions manager 6430 receives from search for answers agent 6424 information relating to the user query, such as recent passage(s) in the chat or a query reflecting recent passages, and may also receive information regarding the user and/or context, and information retrieved from knowledge based by search for answers agent 6424. As shown in example process flow 600K, actions manager 6430 also receives available actions, which may be accessed from actions database 6431. In some instances, actions manager 6430 determines which actions are available to the user based on the context or other information regarding the user.
Using the received information, actions manager 6430 generates a prompt for the Actions LLM 6435 using a prompt generator, such as prompt generator 6432. Prompt generator 6432 includes computer logic for formulating a prompt to a particular language model, such as Actions LLM 6435 regarding actions to be performed to address the user's query. For example, the prompt generated by prompt generator 6432 can include information received from the Search for Answers Agent 6424, such as information about the user and the query, as well as available actions to consider using to address the query.
Actions LLM 6435 comprises a language model configured or use for determining one or more actions based on an input regarding a user query. Although the embodiment of actions LLM 6435 shown in process flow 600K comprises an LLM, other embodiments may use a medium or small language model. Actions LLM 6435 receives the prompt from prompt generator 6432 (or more generally from actions manager 6430) and determines one or more specific actions or other external applications or other resources to use in order to generate an answer to address the user's query. For example, actions LLM 6435 may provide an output, in response to the prompt, that indicates a particular action to call and API (or third-party template) parameters or formatting information used for calling the particular action.
The output of actions LLM 6435 is received by Actions manager 6430. In some instances, the output of actions LLM 6435 is parsed or formatted using output formatter 6434. In particular, output formatter 6434 (or more generally actions manager 6430) uses the output of actions LLM 6435 and the data received from search for answers agent 6424 to format API parameters (or third-party template parameters) for an API call (or to access a third-party resource using a template) to an action identified in the output of actions LLM 6435. For example, if the output of actions LLM 6435 includes an explanation and JSON formatted data, output formatter 6434 may parse the output to extract the necessary information and then format it in a way suitable for the API call corresponding to an action indicated in the output.
Next, with the formatted parameters for an API call, action(s) executor 6436 (or more generally actions manager 6430) executes the action(s) by calling the API(s) (or third-party resource) corresponding to the action(s). Action(s) executor 6436 (or more generally actions manager 6430) receives a response from the API call (or third-party resource), such as information regarding the user's query, and provides the response (or a portion thereof) to search for answers agent 6424.
Continuing the process flow 600K, search for answers agent 6424 provides the information received from actions manager 6430 as a prompt to conversation LLM 6455. Search for answers agent 6424 also may provide other data it received from knowledge bases or internal sources and/or information about from the chat session, such as a recent passage inputted by the user.
Conversation LLM 6455 comprises a language model that generates a chat response to the user that is responsive to the user's input. Conversation LLM 6455 receives a prompt from a tool agent 6420, such as search for answers agent 6424. In some embodiments, conversation LLM 6455 also receives other information as input such as user chat history, chatbot information, app information or other contextual information. Although the embodiment of conversation LLM 6455 shown in process flow 600K comprises an LLM, other embodiments may use a medium or small language model.
In some embodiments, the output of conversation LLM 6455 is parsed before being presented to the user in the chat session. Accordingly, process flow 6K shows the output of conversation LLM 6455 being received by output parser 6460. Output parser 6460 is generally responsible for parsing the output of the conversation LLM 6455, as needed, to provide an answer that is responsive to the user input. For example, where the output of conversation LLM 6455 includes content in addition to a response to the user, then output parser 6460 parses out the portion of the output that is responsive.
At block 6470 of process flow 600K, the output of conversation LLM 6455 (which may have been parsed, in some scenarios) is provided to the user. In particular, the output is used to generate an answer to the user's input in the chat session. In some embodiments, the output is provided from chatbot component 151 via a chat interface 108, as described in connection to
In some embodiments of process flow 600K, actions manager 6430 performs parameter collection for the API call or template to access a third-party resource. For example, action manager 6430 determines that a call to an action needs a particular parameter, and that parameter is not available. So actions manager 6430 communicates the needed parameter to search for answers agent 6424. Search for answers agent 6424 then may perform a search for data corresponding to the parameter, such as a RAG on internal databases, or may determine to request from the user the data corresponding to the parameter. If the user is asked for the data, then search for answers agent 6424 may generate a request to be provided directly in the chat session or may prompt conversation LLM 6455 to generate a request for the user to provide the data. For example, if a customer ID is needed as an API parameter for calling a particular action, and the customer ID data is not available, then action manager 6430 may request the data from search for answers agent 6424, which can cause a chatbot request to be provided to the user in the chatbot session requesting the customer ID. The parameter collection operations performed by actions manager 6430 are useful for planning chaining actions, as described below.
In some embodiments, action manager 6430 accesses a risk setting indicating to a risk level associated with a particular transaction corresponding to the user input or user query. In particular, certain queries or operations requested by the user may have varying risks, such as requesting a refund for a very expensive item, or making a payment for a very large amount of money. Accordingly in some embodiments, an administrator, such as a teammate, can configure an indication of a risk-level associated with certain transaction or certain types of transactions. Based on the risk level, action manager 6430 may determine not to invoke an action, but instead to facilitate implementation of a deterministic workflow. Specifically, in some instances, an action may result in a probabilistic outcome that potentially could be detrimental in certain high-risk transactions. In contrast, a deterministic workflow will always have an outcome that is known or certain. By way of example, in one embodiment, where a user requests a payment exceeding a certain dollar amount (e.g., $5000.00) action manager 6430, accessing a risk setting and determining that there an indication of a risk level is sufficiently high, determines to implement a workflow with a certain outcome such as passing the user onto to a human CSA to verify the payment, rather than perform an action where the outcome could be the chatbot approving a payment that should not occur given a context associated with the user. An example of a process flow of an action manager 6430 deciding to follow a deterministic workflow or a probabilistic path using an action is illustratively depicted in diagram 600N of
Turning back to
In this example, search for answers agent 6424 first performs a Retrieval-Augmented Generation (RAG) to one or more knowledge bases of the merchant company in order to retrieve information regarding the customer, such as a return process that has started, information about the customer such as their address, prior communication, and order details. If this information is available, the search for answers agent 6424 retrieves it (e.g., using RAG) provides it, along with the query or passage(s) inputted by the user in the chat session, to actions manager 6430.
Actions manager 6430 receives the information from search for answers agent 6424 and determines one or more actions and a plan for addressing the issue. For example, actions manager 6430 leverages actions LLM 6435 to determine an action that will address the user's query. In some embodiments or situations, actions manager 6430 facilitates action chaining, where multiple actions are sequentially called. For instance, action chaining may be used where output from calling a first action is needed as input for a second action. Further and as described herein, where the output of LLM 6435 indicates a chain of actions to be performed, actions manager 6430 performs planning to determine that data, as API parameters, is available for the actions.
Actions manager 6430 generates a prompt for the actions LLM 6435 using prompt generator 6432. The prompt includes information received from the search for answers agent 6424, such as information about the user and the query, as well as available actions to consider using to address the query, which may be received from an actions database 6431.
The Actions LLM 6435 receives the prompt from actions manager 6430 and in response, outputs one or more specific actions to address the query. The output from the Actions LLM 6435 is formatted by the Output formatter 6434 to be usable to call an API corresponding to the action. The formatted API parameter information is then passed to the action(s) executor 6436, which calls the API(s) to execute the action(s) and receives from the API(s) a response comprising the information. For instance, here a first API may be called to provide tracking details for the customer's return shipment, and a second API may be called to provide information regarding the payment for the order so that an appropriate refund can be issued to the user.
The information returned from the API call(s) is then used to generate a response for the user in the chat session. In particular, continuing this example, actions manager 6430 returns the API output to the search for answers Agent 6424, which takes that information and generates a prompt for the conversation LLM 6455. The conversation LLM 6455 generates a conversational response for the chat that is responsive to the user's query. The response is then provided to the user through the Generate Answer in Chat component 6470. In some instances, the output of conversation LLM 6455 is parsed by output parser 6460 before it the parsed portion is provided as a response to the user in the chat session.
Turning briefly to
Turning to
Returning back to
In some embodiments, chatbot component 151 and/or chatbot analysis tool 181 facilitates using a language model 110 to facilitate analysis of customer satisfaction across all conversations. In some embodiments, chatbot component 151 and/or chatbot analysis tool 181 facilitates using a language model 110 to facilitate identification of areas of poor quality across customer support conversations (e.g., AI agent and/or human interactions).
Continuing with
At a high level, chatbot (e.g., implemented by chatbot component 151) can include a number of agents (e.g., a program that, for each user turn, based on the conversation transcript and/or other factors, computes the next appropriate action to take in the conversation, and performs the action), such as a start agent (e.g., start state chatbot agent component 153) and a feedback agent (e.g., feedback state chatbot agent component 155). In this regard, computing the next action (e.g., a sub procedure of an agent that causes the effect on the conversation and can be implemented in a general programming language, such as Python, and can perform any operation a given language can perform, such as another call to a large language model (LLM), such as GPT3, GPT4, Anthropic's Claude, and/or the like) can be implemented as an LLM call with lightweight post-processing. In some examples, based on the chosen action and a history of previous actions, the computed action is overridden and another action is used. An example of an agent flow can include: (1) user (e.g., customer) sends a message; (2) API endpoint invokes the active agent (e.g., start state chatbot agent component 153 or feedback state chatbot agent component 155) of the conversation; (3) agent computes the best next action (e.g., from predefined set of actions); (4) agent executes the action (e.g., which can include as part of action execution, an agent yielding the control to another agent).
Turning now to
With reference to
In some embodiments, start state chatbot agent component 153 is programmatically designed to elicit a clear question from the user (e.g., a customer via chat interface 108) and find an answer to the given question. An example diagram 300 of a model implementing a start state agent of a chatbot that uses a language model to determine answers from semantically similar documents of a knowledge base is shown in
Continuing with reference to
In some embodiments, start state chatbot agent component 153 includes a goodbye action (e.g., goodbye action 316 of
In some embodiments, start state chatbot agent component 153 includes a route to teammate/escalate action (e.g., escalate action 318 of
In some embodiments, start state chatbot agent component 153 includes a search for answer action (e.g., search for answers action 308 of
In some embodiments, start state chatbot agent component 153 includes a clarify action (e.g., clarify action 310 of
In some embodiments, start state chatbot agent component 153 includes a default action (e.g., default action 314 of
A specific example of a prompt provided by start agent (e.g., start state chatbot agent component 153) to a language model 110 (e.g., with a conversation with the customer) is as follows:
A specific example of response to the specific example of the prompt provided by start agent (e.g., start state chatbot agent component 153) by a language model 110 is as follows:
In some embodiments, after output from language model 110 is received, the output of the language model can be parsed with regular expressions by chatbot component 151. For example, in the specific example of response to the specific example of the prompt, the returned action (e.g., the search for answer action) is invoked with returned action input (e.g., “When am I eligible to get a bonus?”). Following the returned action and action input, in some embodiments, a programming function (e.g., a python function) representing the action (e.g., the search for answer action) can be called with a textual argument corresponding to the action input (e.g., “When am I eligible to get a bonus?”). In some embodiments, if the output of the language model 110 does not conform to the specifications and/or the returned action name of the language model 110 is not a valid action name, a recovery procedure is started.
In some embodiments, answer search state component 154 facilitates programmatically implementing the search for answer action of start state chatbot agent component 153. Generally, answer search state component 154 facilitates the delivery of truthful and grounded resolution to the user's issue in textual form. As a high level overview, in some embodiments, answer search state component 154 initially determines whether a custom answer (e.g., a manually-defined answer) exists based on aspects of the conversation that triggered the search for answer action (e.g., via custom answers component 160). If no custom answer is determined, an issue summary is generated based on aspects of the conversation with the user. A language model determines whether the answer can be found in portions of documents of the knowledge base (e.g., knowledge base 132 of
Returning to the high level overview of answer search state component 154, in some embodiments, if the answer can be found, the answer is provided to the user via chat interface 108. If no answer can be found, a set of summaries corresponding to a set of documents that are most similar to the issue summary are provided to the language model to determine whether any of the documents may contain the answer. If the language model determines that documents within the set of documents may contain the answer, a threshold number of document most likely to contain the answer, along with summaries of the documents, are provided to the user via chat interface 108. If the language model determines that documents within the set of documents will not contain the answer, the language model indicates that the answer could not be found. The start state chatbot agent component 153 can then determine the next action, such as a clarify action or escalation action.
In some embodiments, different language models can be utilized for different actions or portions of actions. For example, one language model (e.g., GPT3) can be utilized to summarize the issue by answer search state component 154 whereas a different language model (e.g., GPT4) can be utilized to evaluate whether an article in the knowledge contains the answer by answer search state component 154. In some embodiments, one language model (e.g., GPT3) is utilized to screen articles as an initial check whether a top article contains an answer before asking a different language model (e.g., GPT4) to determine whether the top article contains an answer in order to reduce calls to the different language model (e.g., GPT4). In some embodiments, frequently asked questions (FAQ) are pre-computed using a language model to reduce live calls to the language model. In some embodiments, answer search state component 154 can utilize information from beyond the knowledge base. In some embodiments, answer search state component 154 and other components are multilingual.
In some embodiments, answer search state component 154 can request an issue summary based on the conversation with user (e.g., the customer) from language model 110. In some embodiments, the summary of the user's issue as action input to the search for answer action provided by language model 110 (e.g., the issue summary utilized to determine the appropriate action) is ignored and a second call is provided to the language model to summarize the issue again in order to increase the quality and/or reliability of the summary of the user's issues in certain scenarios. In some embodiments, aspects of the conversation with the user, such as the issue summary generated by language model 110 and/or historical customer data of the user, can be utilized to find the relevant context for searching the knowledge base 132. In some embodiments, the issue summary as generated by language model 110 can be utilized to engineer around an LLM's limitation on input length (e.g., the context window), which can cause the computation of the output to not be based on all parts of the input over a certain length and/or cause certain portions of input over a certain length to get a different level of attention while computing the output. In some embodiments, the issue summary can be used to stuff the prompt when the answer is being extracted by answer search state component 154 by utilizing the retrieved context and the issue summary to answer the user question at hand.
In some embodiments, answer search state component 154 facilitates programmatically finding an answer from a knowledge base. An example diagram 400 of a model implementing a specialized chatbot platform that uses language models to determine answers from knowledge bases is shown in
At block 406, it is determined whether aspects of the input correspond to a custom answer (e.g., via custom answers component 160 of
In some embodiments, at block 410, a call is made to the language model 110 to summarize the issue based on aspects of input 404. A sentence embedding of the issue summary generated from block 410, or a portion thereof, is computed (e.g., utilizing SBERT). In some embodiments, historical customer data for the user chatting with the chatbot (e.g., through chat interface 108) can be utilized to provide contextual information for the issue summary. For example, the contextual information related to the historical customer data can be encoded into the embedding of the issue summary.
A specific example of a prompt (e.g., as provided by answer search state component 154 via start state chatbot agent component 153) to a language model (e.g., language model 110) to request an issue summary (e.g., a summary of an issue within the conversation) is as follows:
A specific example of response from the language model to the specific example of the prompt requesting an issue summary is as follows:
At block 412, a threshold number of portions from documents of a knowledge base (e.g., knowledge base 132 as accessed by knowledge base accessing component 157 of
Returning to
At block 416, an LLM (e.g., language model 110 of
A specific example of a prompt (e.g. as provided by answer search state component 154 via start state chatbot agent component 153 of
A specific example of a response from the language model to the specific example of the prompt to find an answer given an issue summary and a threshold number of portions of documents of a knowledge base similar to the issue summary is as follows:
In some embodiments, if there is an answer tag (e.g., <answer>) present in the response provided by the language model, and is different from “UNKNOWN”, the answer is determined to be found and the answer is delivered to the user. In some embodiments, the answer tag comprises a textual answer corresponding to whatever the language model returned inside answer tag and/or a link to the documents used to form an answer. An example chat interface between a chatbot and a user showing a response from the chatbot with a link to the corresponding source of the response is shown in
In some embodiments, if the language model did not find the answer to the issue summary at block 416, at block 422, a threshold number of most similar documents, or portions thereof, are identified. In this regard, if the previous stage didn't find a direct, inline answer, a prompt is provided to the language model to find content that is not directly answering the question, but might be relevant. In some embodiments, semantic search is used to determine the portions of documents of the knowledge base that are most similar to the issue summary based on the semantic similarity of the issue summary, or a portion thereof, to portions of documents of the knowledge base. For example, the semantic similarity of an embedding corresponding to the issue summary and an embedding corresponding to each portion of documents of the knowledge base customer can be computed (e.g., based on a vector search query, a dot product of the embeddings, using MS MARCO, and/or the like) to determine the portions of documents of the knowledge base most similar to the issue summary. In some embodiments, utilizing a Q&A asymmetrical search model (e.g., MS MARCO), the embedding of the issue summary can be utilized as a question and the answer can be found based on embeddings of corresponding portions of the knowledge base. A threshold number of passages (e.g., the 20 most similar passages) can be identified based on the dot product of the embeddings. In some embodiments, the location of the documents, or portions thereof, such as the identification of the article, can be determined.
In some embodiments, at block 424, a prompt is provided to a language model to determine whether an article might contain an answer relevant to the issue based on the issue summary and summaries 426 of each of the threshold number of most similar documents, or portions thereof, as identified at block 422. In some embodiments, the summaries of each of the documents of the knowledge base are precomputed and stored in the knowledge base. At block 428, if the language model determines that documents within the threshold number of documents may contain the answer, a threshold number of document most likely to contain the answer are provided to the user (e.g., via chat interface 108 of
A specific example of a prompt (e.g., as provided by answer search state component 154 via start state chatbot agent component 153 of
A specific example of a response from the language model to the specific example of the prompt to determine documents likely to contain an answer to an issue summary is as follows:
In some embodiments, article indices are then tied back to article identification from the knowledge base (e.g., knowledge base 132 of
Returning to
An example diagram 500 of a model implementing a feedback state agent of a chatbot is shown in
Returning to
In some embodiments, feedback state chatbot agent component 155 includes a greet action (e.g., greet action 510 of
In some embodiments, feedback state chatbot agent component 155 includes a goodbye action (e.g., goodbye action 512 of
In some embodiments, feedback state chatbot agent component 155 includes a route to teammate/escalate action (e.g., escalate action 318 of
In some embodiments, feedback state chatbot agent component 155 includes a route to start agent action (e.g., route to main action 516 of
Data regarding the prompts utilized by chatbot component 151 and/or data communicated to/from the language model 110, customer device 102, and/or customer support device 112 can be stored in any suitable storage location, such as storage 130, customer support device 112, server 150, some combination thereof, and/or other locations as communication records files 131 and/or chatbot configuration files 140.
In an example implementation, chat interface 108 provides interface functionality that allows a user (e.g., a customer) to communicate with a chatbot and/or a CSA. Generally, chat interface 108 presents one or more interaction elements that provide various interaction modalities for chatting with a chatbot and/or a CSA. In various embodiments, these tools are implemented using code that causes a presentation of a corresponding interaction element(s), and detects and interprets inputs interacting with the interaction element(s).
In an example implementation, chatbot preview tool 128 provides interface functionality that allows a user (e.g., CSA) to communicate with a chatbot, for example, to preview the functionality of the chatbot. Generally, chatbot preview tool 128 presents one or more interaction elements that provide various interaction modalities for chatting with a chatbot. In various embodiments, these tools are implemented using code that causes a presentation of a corresponding interaction element(s), and detects and interprets inputs interacting with the interaction element(s).
Examples of a chat interface displayed to a customer (e.g., via chat interface 108) and/or CSA (e.g., via chatbot preview tool 128) are shown in
Returning to
Customer data accessing component 156 facilitates programmatically accessing customer data (e.g., customer data files 141) in order to implement a chatbot configuration of a chatbot implemented by chatbot component 151 based on the customer data of the customer. For example, a chatbot may access certain portions of the knowledge based on the status of the customer as a VIP. Customer data can be stored in any suitable storage location, such as storage 130, customer support device 112, server 150, some combination thereof, and/or other locations as customer data files 141. Customer data within a dataset may include, by way of example and not limitation, data that is sensed or determined from one or more sensors, such as location information of mobile device(s), smartphone data (such as phone state, charging data, date/time, or other information derived from a smartphone), activity information (for example: app usage; online activity; searches; browsing certain types of webpages; listening to music; taking pictures; voice data such as automatic speech recognition; activity logs; communications data including calls, texts, instant messages, and emails; website posts; other user data associated with communication events) including activity that occurs over more than one device, user history, session logs, application data, contacts data, calendar and schedule data, notification data, social network data, news (including popular or trending items on search engines or social networks), online gaming data, ecommerce activity, sports data, health data, and nearly any other source of data that may be used to identify the customer.
Knowledge base accessing component 157 facilitates programmatically implementing access to the knowledge base. In some embodiments, knowledge base accessing component 157 accesses portions of the knowledge base based on the configuration of the chatbot implemented by chatbot component 151. For example, the configuration of chatbot component 151 may indicate that only a specific set of documents of the knowledge base should be searched by the chatbot. In some embodiments, knowledge base accessing component 157 automatically syncs content (e.g., such as content provided by content management tool 173 and/or snippets tool 129) provided by the customer support application 116 and/or stored in knowledge base 132. Data regarding the configuration of the chatbot facilitating access to different portions of the knowledge base can be stored in any suitable storage location, such as storage 130, customer support device 112, server 150, some combination thereof, and/or other locations as chatbot configuration files 140.
Route to agent component 158 facilitates programmatically implementing routing of a conversation from a chatbot to a CSA via the customer support component 165 to the communication tool 120 of the customer support application 116. For example, if a customer request to speak to a person or the chatbot decides to route to teammate/escalation action, route to agent component 158 automatically implements the routing of the conversation from the chatbot to a CSA. Data regarding the configuration of the chatbot to route to a CSA can be stored in any suitable storage location, such as storage 130, customer support device 112, server 150, some combination thereof, and/or other locations as chatbot configuration files 140.
Chat workflow implementation component 159 facilitates programmatically implementing chat workflows designed to implement bots, triggers, conditions, and/or rules for chats. For example, a CSA can designed a chat workflow through cat workflow design tool so that the chatbot automatically implements certain actions, such as asking for specific information, before routing to a CSA. Data regarding chat workflows of the chatbot can be stored in any suitable storage location, such as storage 130, customer support device 112, server 150, some combination thereof, and/or other locations as workflow configuration files 142.
Custom answers component 160 facilitates programmatically implementing manually written answers and/or chat workflows to specific questions from customers by prioritizing the custom answers over AI-generated answers from the chatbot. In some embodiments, custom answers are searched using SBERT directly. For example, an embedding generated based on the customer's query can be searched against embeddings of one or more example questions corresponding to a custom answer (e.g., a user of customer support application 116 can input multiple examples of a single question. In some embodiments, the custom answer can include examples of questions, keywords, and/or phrases to automatically call a chat workflow designed in chat workflow design tool 180 and implemented by chat workflow implementation component 159. For example, if a customer requests a refund through chat interface 108, the custom answer for a refund can be determined through SBERT, which implements a chat workflow that automatically calls an API to trigger a refund. Data regarding custom answers can be stored in any suitable storage location, such as storage 130, customer support device 112, server 150, some combination thereof, and/or other locations as custom answer files 139.
Chatbot analysis component 161 facilitates programmatically generating reports by analyzing chatbot performance chatbot conversation monitoring, chatbot customer satisfaction score (CSAT), usage metrics, usage limits notifications, performance/return on investment (ROI) metrics, custom reports, and/or content-level metrics regarding access to content. Data regarding analysis of chatbots can be stored in any suitable storage location, such as storage 130, customer support device 112, server 150, some combination thereof, and/or other locations as metrics files 145.
In an example implementation, chatbot tool 170 provides interface functionality that allows a user (e.g., a CSA) of customer support application 116 to implement settings and/or configuration of the chatbot in order to customize the chatbot of chatbot component. Generally, chatbot tool 170 presents one or more interaction elements that provide various interaction modalities for customizing the chatbot. In various embodiments, these tools are implemented using code that causes a presentation of a corresponding interaction element(s), and detects and interprets inputs interacting with the interaction element(s). An example of chatbot tool 170 is shown in
In the example implementation in
Chatbot tool 170 includes chatbot identity customization tool 172 that allows the user to rename the chatbot and/or change the image/icon for the chatbot, which can be stored in chatbot configuration files 140. An example of chatbot identity customization tool 172 is shown in
Chatbot tool 170 includes content management tool 173 that allows the selection of content of the knowledge base 132 that the chatbot uses to search for answers. For example, the CSA may not want to utilize certain content as the content may be outdated. As another example, different content can be utilized for different audiences to target the chatbot to different audiences, such as based on customer data or customer segment (e.g., VIP customer). In some embodiments, knowledge base accessing component 157 automatically syncs any content (e.g., such as content provided by content management tool 173) provided by the customer support application 116 and/or stored in knowledge base 132. In some embodiments, content management tool 173 allows the selection of multimodal content stored in knowledge base 132 in order for the chatbot to provide multimodal answers, such as images, audio, and/or video in responses. In some embodiments, content management tool 173 allows the selection of external systems, such as an external application (e.g., an ecommerce store and/or system) through third-party application configuration files 144, in order for the chatbot to utilize external systems to provide answers. An example of content management tool 173 is shown in
Chatbot tool 170 includes audience selection tool 174 that allows the selection of audience to target the chatbot, such as based on customer data or customer segment (e.g., VIP customer) based on customer data stored in customer data files 141. Examples of audience selection tool 174 are shown in
Chatbot tool 170 includes communication channel selection tool 175 that allows the selection of various communication channels, such as Web, mobile, application, telephonic, SMS, messaging applications, social media, email, and/or any communication channel (e.g., as stored in third-party application configuration files 144). In some embodiments, the chatbots can be deployed on any number of communication channels. In some embodiments, different settings/configuration of the chatbot can be applied to different communication channels. For example, only some content of knowledge base 132 may be utilized by chatbot for customers communicating through a specific communication channel. In some embodiments, communication channel selection tool 175 allows the selection of multimodal receiving input user queries and/or outputting responses by the chatbot via audio. An example of communication channel selection tool 175 is shown in
Chatbot tool 170 includes chatbot behavior settings tool 176 that allows a user to provide settings for the chatbot regarding how the chatbot should behave when answering multiple questions and/or when to handover the chatbot to a teammate, which can be stored in chatbot configuration files 140. Examples of chatbot behavior settings tool 176 are shown in
Chatbot tool 170 includes inactive conversations settings tool 177 that allows a user of customer support application 116 to set automatic replies to inactive conversations, which can be stored in chatbot configuration files 140 and/or workflow configuration files 142. For example, inactive conversations settings tool 177 can send auto-replies to snoozed conversations. As another example, inactive conversations settings tool 177 can automate snoozing and set conversation pacing with new actions. As yet another example, inactive conversations settings tool 177 can automatically reroute conversations that receive no response to a human, such as a CSA. In some embodiments, inactive conversations settings tool 177 is implemented through chat workflow design tool 180 as a part of a chat workflow. An example of inactive conversations settings tool 177 is shown in
Chatbot tool 170 includes language settings tool 178 allows the chatbot to be multilingual and/or target the chatbot differently to customers in different languages, which can be stored in language configuration files 143. For example, only some content of knowledge base 132 may be utilized by chatbot for customers communicating in a specific language. Example of language settings tool 178 are shown in
Chatbot tool 170 includes scheduling settings tool 179 that allows a user to provide settings as to when to enable a chatbot, which can be stored in chatbot configuration files 140. For example, the chatbot can be implemented only for certain timeframe. An example of scheduling settings tool 179 is shown in
Chatbot tool 170 includes chat workflow design tool 180 allows the user of the customer support application 116 to design chat workflows, which can be stored as workflow configuration files 142. For example, the user can design chat workflows to handoff to support teams or to other bots. For example, a chatbot can include a chat workflow can be designed to automatically handover to a CSA when the chatbot cannot answer the question. As another example, a chat workflow can be designed to automatically implement a chat workflow to request additional information from the customer before handing the chat over to a CSA when the chatbot cannot answer a question. In some embodiments, the chatbot that uses a language model to determine answers from a knowledge base (e.g., knowledge base 132 of
Chatbot tool 170 includes chatbot analysis tool 181 that provides an interface for generating reports analyzing chatbot performance, such as through chatbot conversation monitoring, chatbot CSAT, usage metrics, usage limits notifications, performance/ROI metrics, custom reports, and/or content-level metrics, which can be stored as metrics files 145. Examples of chatbot analysis tool 181 and output of chatbot analysis tool 181 are shown in
Chatbot tool 170 includes custom answers tool 182 that provides an interface for a user to implement manually written answers and/or chat workflows to specific questions from customers. For example, a user of the customer support application 116 may provide a specific written answer and examples of potential customer questions that would require the specific written answer so that specific written answer is provided without requiring the chatbot to call the language model. In some embodiments, the custom answer can include examples of questions, keywords, and/or phrases to automatically call a chat workflow designed in chat workflow design tool 180 and implemented by chat workflow implementation component 159. For example, if a customer requests a refund through chat interface 108, the custom answer for a refund can be determined through SBERT based on aspects of the conversation, which implements a chat workflow that automatically calls an API to trigger a refund. An example of custom answers tool 182 is shown in
Continuing with
In some embodiments, messages routing component 166 routes messages from conversations from customers, such as conversations between a customer and a chatbot (e.g., through route to agent component 158 of chatbot component 151). In some embodiments, messages routing component 166 routes messages to and/or from customers from various communication channels, such as Web, mobile, application, telephonic, SMS, messaging applications, social media, email, and/or any communication channel. Messages routing component 166 can route messages to messages inbox interface 122 and/or chat interface 124 of a customer support application 116 of a CSA.
In some embodiments, AI-assisted chat component 167 assists customer support in responding to customers in chats using AI. For example, a language model may be called to revise an initial response drafted by the CSA to expand the response, rephrase the response, make the response more formal, make the response friendlier, and/or the like. As another example, using a similar process as the answer search state component 154 of chatbot component 151, the AI-assisted chat component 167 can determine answers to customer queries and provide the answers as suggestions to the CSA to respond to the customer query.
In some embodiments, snippets component 168 can programmatically extract conversational snippets from conversations between a customer and a CSA (e.g., from a communication record of communication records files 131, such as a chat log, chat transcript, meeting transcript, email, or other communication between a customer and a CSA). For example, a customer initiates a chat with a CSA through a chat interface 108 of application 106 executing on customer device 102. During the chat, the customer asks the CSA a series of questions and the CSA provides answers to each question through chat interface 124 of application 116 executing on customer support device 112 before the chat is ended. After the chat ends, the chat is stored (e.g., communication records files 131) so that the chat can be accessed in order to extract each question and corresponding answer (“Q&A pairs”) from the chat in a subsequent block. In some embodiments, the conversation can be accessed from the communication record by snippets component 168 during the conversation (e.g., chat) between the customer and the CSA in order to extract Q&A pairs from the conversation during the conversation. Snippets component 168 generates a prompt to language model 110 in order to extract Q&A pairs from the portion of the conversation record and to generate a single, summarized Q&A pair based on the Q&A pairs extracted from the conversation record. In some embodiments, snippets component generates a prompt to language model 110 to tag each Q&A pair of the conversation with contextual metadata corresponding to each of the Q&A pairs and filter irrelevant Q&A pairs of the conversation based on the metadata of each Q&A pair before generating a single, summarized Q&A pair based on the remaining Q&A pairs extracted from the conversation record. Examples of contextual metadata include: (1) contextual metadata corresponding to whether the question was answered by a human (e.g., a CSA) or a bot (e.g., the chatbot); (2) contextual metadata corresponding to the topic of the Q&A pair; (3) contextual metadata corresponding to a score indicating the relevance of the Q&A pair to other customers; and/or (4) contextual metadata corresponding to a dialog classification corresponding to a category of the type of dialogue. Examples of the type of dialogue can include: (a) informational content where an answer in the Q&A pair is directed to general information, knowledge and/or instructions; (b) a clarification where an answer in the Q&A pair is a request for clarification from the user; (c) a CSA action where an answer in the Q&A pair required the CSA to take some action on the background of the conversation; (d) a feature request where an answer in the Q&A pair is directed to future product improvements or feature requests; and/or (e) other categories or a category for an answer in the Q&A pair does not fit into the previous specified types of dialogue.
In an example implementation, communication tool 120 provides interface functionality that allows a user (e.g., a CSA) to chat with a customer, chat with a customer with AI-assisted chat capabilities, trigger extractions of snippets, interact with messages from various customers and/or communication channels through their inbox, and/or preview a chatbot through interactions with an interface controlled by communication tool 120. Generally, communication tool 120 presents one or more interaction elements that provide various interaction modalities for its functionality. In various embodiments, these tools are implemented using code that causes a presentation of a corresponding interaction element(s), and detects and interprets inputs interacting with the interaction element(s). An example of communication tool 120 is shown in
In the example implementation in
Communication tool 120 includes chatbot preview tool 128 allows the user of the customer support application 116 to preview the chatbot, such as before setting the chatbot live for customers. An example of chatbot preview tool 128 is shown in
Communication tool 120 includes snippets tool 129 that provides an interface so that users of customer support application 116 can add snippets determined from conversational data (e.g., communication records files 131) by snippets component 168 to add to the knowledge base 132 (e.g., snippets files 133). In this regard, the answers in snippets of snippets files 133 can be edited, targeted (e.g., by content management tool 173), and managed in the knowledge base utilized by the chatbot to provide answers. An example of snippets tool 129 is shown in
With reference now to
Having briefly described an overview of aspects of the technology described herein, an exemplary operating environment in which aspects of the technology described herein may be implemented is described below in order to provide a general context for various aspects of the technology described herein. Referring to the drawings in general, and initially to
The technology described herein may be described in the general context of computer code or machine-usable instructions, including computer-executable instructions such as program components, being executed by a computer or other machine, such as a personal data assistant or other handheld device. Generally, program components, including routines, programs, objects, components, data structures, and the like, refer to code that performs particular tasks or implements particular abstract data types. Aspects of the technology described herein may be practiced in a variety of system configurations, including handheld devices, consumer electronics, general-purpose computers, and specialty computing devices. Aspects of the technology described herein may also be practiced in distributed computing environments where tasks are performed by remote-processing devices that are linked through a communications network.
With continued reference to
Computing device 1400 typically includes a variety of computer-readable media. Computer-readable media can be any available media that can be accessed by computing device 1400 and includes both volatile and nonvolatile, removable and non-removable media. By way of example, and not limitation, computer-readable media may comprise computer storage media and communication media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program sub-modules, or other data.
Computer storage media includes RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage, or other magnetic storage devices. Computer storage media does not comprise a propagated data signal.
Communication media typically embodies computer-readable instructions, data structures, program sub-modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared, and other wireless media. Combinations of any of the above should also be included within the scope of computer-readable media.
Memory 1412 includes computer storage media in the form of volatile and/or nonvolatile memory. The memory 1412 may be removable, non-removable, or a combination thereof. Exemplary memory includes solid-state memory, hard drives, and optical-disc drives. Computing device 1400 includes one or more processors 1414 that read data from various entities such as bus 1410, memory 1412, or I/O components 1420. Presentation component(s) 1416 present data indications to a user or other device. Exemplary presentation components 1416 include a display device, speaker, printing component, and vibrating component. I/O port(s) 1418 allow computing device 1400 to be logically coupled to other devices including I/O components 1420, some of which may be built in.
Illustrative I/O components include a microphone, joystick, game pad, satellite dish, scanner, printer, display device, wireless device, a controller (such as a keyboard, and a mouse), a natural user interface (NUI) (such as touch interaction, pen (or stylus) gesture, and gaze detection), and the like. In aspects, a pen digitizer (not shown) and accompanying input instrument (also not shown but which may include, by way of example only, a pen or a stylus) are provided in order to digitally capture freehand user input. The connection between the pen digitizer and processor(s) 1414 may be direct or via a coupling utilizing a serial port, parallel port, and/or other interface and/or system bus known in the art. Furthermore, the digitizer input component may be a component separated from an output component such as a display device, or in some aspects, the usable input area of a digitizer may be coextensive with the display area of a display device, integrated with the display device, or may exist as a separate device overlaying or otherwise appended to a display device. Any and all such variations, and any combination thereof, are contemplated to be within the scope of aspects of the technology described herein.
A NUI processes air gestures, voice, or other physiological inputs generated by a user. Appropriate NUI inputs may be interpreted as ink strokes for presentation in association with the computing device 1400. These requests may be transmitted to the appropriate network element for further processing. A NUI implements any combination of speech recognition, touch and stylus recognition, facial recognition, biometric recognition, gesture recognition both on screen and adjacent to the screen, air gestures, head and eye tracking, and touch recognition associated with displays on the computing device 1400. The computing device 1400 may be equipped with depth cameras, such as stereoscopic camera systems, infrared camera systems, RGB camera systems, and combinations of these, for gesture detection and recognition. Additionally, the computing device 1400 may be equipped with accelerometers or gyroscopes that enable detection of motion. The output of the accelerometers or gyroscopes may be provided to the display of the computing device 1400 to render immersive augmented reality or virtual reality.
A computing device may include radio(s) 1424. The radio 1424 transmits and receives radio communications. The computing device may be a wireless terminal adapted to receive communications and media over various wireless networks. Computing device 1400 may communicate via wireless protocols, such as code division multiple access (“CDMA”), global system for mobiles (“GSM”), or time division multiple access (“TDMA”), as well as others, to communicate with other devices. The radio communications may be a short-range connection, a long-range connection, or a combination of both a short-range and a long-range wireless telecommunications connection. When we refer to “short” and “long” types of connections, we do not mean to refer to the spatial relation between two devices. Instead, we are generally referring to short range and long range as different categories, or types, of connections (i.e., a primary connection and a secondary connection). A short-range connection may include a Wi-Fi® connection to a device (e.g., mobile hotspot) that provides access to a wireless communications network, such as a WLAN connection using the 802.11 protocol. A Bluetooth connection to another computing device is a second example of a short-range connection. A long-range connection may include a connection using one or more of CDMA, GPRS, GSM, TDMA, and 802.16 protocols.
The technology described herein has been described in relation to particular aspects, which are intended in all respects to be illustrative rather than restrictive. The technology described herein is described with specificity to meet statutory requirements. However, the description itself is not intended to limit the scope of this patent. Rather, the inventors have contemplated that the claimed subject matter might also be embodied in other ways, to include different steps or combinations of steps similar to the ones described in this document, in conjunction with other present or future technologies. Moreover, although the terms “step” and “block” may be used herein to connote different elements of methods employed, the terms should not be interpreted as implying any particular order among or between various steps herein disclosed unless and except when the order of individual steps is explicitly described.
This application claims the benefit of U.S. Provisional Application No. 63/564,879, filed on Mar. 13, 2024. The entire contents of which are incorporated herein.
Number | Date | Country | |
---|---|---|---|
63564879 | Mar 2024 | US | |
63613969 | Dec 2023 | US | |
63705435 | Oct 2024 | US |