Aspects of various embodiments are directed to capturing and analyzing digital voice data in a data-communication server system such as a broadband server and/or a Voice over Internet Protocol (VoIP) system, which is used to identify an outcome of a conversation based on a tone/sentiment and/or keywords. In the following discussion, various implementations and applications are disclosed to provide an understanding of the instant disclosure by way of non-limiting example embodiments exemplified by way of VoIP-type data communications which typically involve a data-communication server communicating with an endpoint device, such as a VoIP-enabled endpoint device (“VoIP device”) via a broadband network (e.g., Internet, WiFi, cellular) to connect with the data-communication server that is managed by an VoIP provider such as 8×8, Inc. and/or an Internet Service Provider (ISP) such as Yahoo or Google. Through such a server, call routing and other data communications services are managed for the endpoint device.
Computer servers are increasingly being used to provide various services over a network including, but not limited to, VoIP communications services such as VoIP calls, video conferencing, call exchange servers, packet switching, and traffic management as well as non-VoIP services including, but not limited to: website hosting, remote data storage, remote computing services, and virtual computing environments.
The use of VoIP telecommunications services has been widespread and significant in terms of both numbers of users and types of services being made available. This growth can be attributable to any of a wide variety of socio-economic changes such as the mobility of users of these services, the types and reduced costs of portable telecommunication tools, and the ever-evolving technology adapting to the personal and business needs of the telecommunications users.
Various example embodiments are directed to issues such as those addressed above and/or others which may become apparent from the following disclosure concerning analyzing digital voice data in a data-communications system.
A number of aspects are directed to a data-communications system that includes a data-communication server that interfaces with a plurality of remotely-situation client entities to provide data communication services and processing circuitry coupled to a memory circuit and to the data-communication server. The data-communications server is operated by a telecommunications provider, where the data-communications server on the data communication provider side (e.g., a VoIP provider server as commercially available services from 8×8, Inc.), is used to provide data-communication services to a multitude of client entities. The processing circuit, which can be part of or separate from the data-communication server, is in communication with the data-communications server and can analyze digital voice data that represents audio (e.g., spoken) conversations between agents of the client entities and other personnel, such as customers. For example, the processing circuitry can receive digital voice data indicative of transcribed audio conversations between a plurality of agents and customers of a remotely-situated client entity and identify keywords and speech characteristic parameters from the digital voice data and which are from the data communication services provided by the data-communications server. The speech characteristic parameters are indicative of a tone or sentiment of the transcribed audio conversations. Example speech characteristic parameters include a frequency of audio of the transcribed audio conversation, wavelength or velocity of the audio, amplitude of the audio, and a combination thereof.
The processing circuitry can provide associations of at least one of the keywords and speech characteristic parameters with outcomes of the transcribed audio conversations, and provide the associations to a client data-communications server as feedback for the outcomes of the transcribed audio conversations with the customer(s). The associations can indicate which keywords and speech characteristics have a greater probability of a negative or positive outcome, which can be defined by the entity. Each entity and/or different divisions of an entity can have different negative and positive outcomes. Example outcomes can include a sale to a customer, a renewal of a service, a discontinuation of service, resolving of an issue, not resolving an issue, and customer hang up, among other outcomes. Other example outcomes can include a change in tone or sentiment of the customer during the conversation, such as a customer that starts the conversation angry or upset and ends the conversation happy. The associations can be stored in a database as an archive that is accessible and that ties the keywords with dates of the conversations and the speech characteristic parameters of the particular transcribed audio conversations. In various specific aspects, the processing circuitry can adjust the associations over time based on further analysis of digital voice data and provide the adjustments as feedback.
The outcomes can be determined based on keywords in the digital voice data. For example, the processing circuitry, which can form part of the data-communications server or be separate therefrom (such as forming part of a client server), can identify the outcome based on the keywords. Alternatively and/or in addition, metadata can be located in the digital voice data that can be used to determine the outcome. In some instances, an agent can insert the metadata during the conversation and which can be tied to the digital voice data of the conversation and provided to the processing circuitry by the data-communications server. In various specific aspects, the association of the keywords and speech characteristic parameters can be provided to the agent associated with the audio conversation via a graphical user interface that is accessible to the agent. The associations can be adjusted in response to the agent providing an input indicating the association is incorrect. Similarly, the association can be stored in response to the agent providing an input indicating the association is correct or not providing an input within a threshold amount of time. In some specific aspects, the associations can be locked and a manager reviews the recommended adjustments (by the agent) and approves or denies the adjustment.
In a number of specific aspects, the analysis can be based on or further correlated with different classifications. The different classification can be based on characteristics of the audio conversation, the agents, and/or the customer, such as geographic region of the customer, sex of the customer, age of the customer, type of customer, and a combination thereof. As previously described, the associations are stored in a database as an archive. The database can be used to identify keywords and phrases to avoid for particular types of customers and keywords and phrases to use for the particular types of customers.
In a number of related and more specific aspects, the processing circuitry can analyze the digital voice data while the audio conversation is occurring. The analysis can be used to provide real-time feedback for the agent to use during the conversation and/or to trigger particular responses. For example, the processing circuitry can identify a speech characteristic parameter outside of a threshold value (e.g., indicates problem) or a keyword that is associated with a (manager) trigger, and in response, automatically bridge a manager of the respective agent into the audio conversation. The processing circuitry and/or data-communications server can access a database to identify the manager and information for bridging the manager into the audio conversation. Additionally, the agent can be provided with feedback on customer interactions, which is sometimes herein referred to as “customer interaction feedback.” The processing circuitry can provide customer interaction feedback using the associations by generating a report indicating negative and positive outcomes and associated keywords and/or speech characteristic parameters.
A number of related aspects are directed to a method of using such an example apparatus. An example method includes receiving, at processing circuitry in communication with a data-communications server, digital voice data indicative of transcribed audio conversations between a plurality of agents and customers of a remotely-situated client entity and analyzing, via the processing circuitry, the digital voice data. For example, the digital voice data can be analyzed by identifying keywords and speech characteristic parameters of the digital voice data, identifying outcomes of the respective transcribed audio conversations, and generating associations with at least one of the keywords and speech characteristic parameters with the identified outcomes. The method further includes providing the associations to a client data-communications server as feedback on audio conversations with the customers. As described above, the speech characteristic parameters are indicative of a tone or sentiment of the audio conversations, and the method can further include transcribing the audio conversation to digital voice data by endpoint devices of the remotely-situated client entity, the transcription including capturing the speech characteristic parameters in the digital voice data, the speech characteristic parameters selected from the group consisting of: frequency of the audio conversation, wavelength or velocity of the audio conversation, amplitude of the audio conversation, and a combination thereof. Identifying the outcome of the transcribed audio conversations can include identifying metadata in the digital voice data and/or determining the outcome from the keywords in the digital voice data.
In various specific aspects, the method can further include providing different classifications for the analysis based on characteristics of the particular audio conversations, including characteristics selected from the group consisting of: geographic region of the customer, sex of the customer, age of the customer, type of customer, and a combination thereof. The method can additionally include storing the associations in a database as an archive that is accessible and that ties the keywords with dates of the conversations and the classifications, and which identifies keywords and phrases to avoid for particular types of customers and keywords and phrases to use for the particular types of customers. In other specific embodiments and/or in addition, the method includes providing a respective association with at least one of the keywords and speech characteristics with the identified outcome to an agent associated with the audio conversation, and adjusting the association, in response to the agent providing an input indicating the associated is incorrect or storing the indication in response to the agent providing an indication indicating the associated in correct or not providing any indication.
The above discussion/summary is not intended to describe each embodiment or every implementation of the present disclosure. The figures and detailed description that follow also exemplify various embodiments.
Various example embodiments may be more completely understood in consideration of the following detailed description and in connection with the drawings, in which:
While various embodiments discussed herein are amenable to modifications and alternative forms, aspects thereof have been shown by way of example in the drawings and will be described in detail. It should be understood, however, that the intention is not to limit the disclosure to the particular embodiments described. On the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the scope of the disclosure including aspects defined in the claims. In addition, the term “example” as used throughout this application is only by way of illustration, and not limitation.
Aspects of the present disclosure are believed to be applicable to a variety of different types of apparatuses, systems and methods involving analysis of digital voice data in a data-communication system. In certain implementations, aspects of the present disclosure have been shown to be beneficial when used in a Voice over Internet Protocol (VoIP) system including a VoIP communications server that interfaces with a plurality of remote-situation client entities. While the present disclosure is not necessarily limited to such VoIP communications systems as described herein, for purposes of facilitating understanding and appreciation of certain embodiments, the following discussion uses such VoIP-based services and systems in the context and on behalf of telecommunications platforms of client entities which subscribe to such services from a VoIP telecommunications service provider (with a server).
Users of a data-communications system can use a plurality of communication services to communicate with clients and other professionals and to carry out various tasks. For example, agents of a client entity can use a private network application to generate work products, a cloud-based application to manage service issues, another cloud-based application to manage direct communications such as email and chat messages, and yet another cloud-based application to manage financial matters such as billing and invoicing. In many instances, it can be beneficial for an entity subscribed to or that otherwise uses the data-communications system to analyze audio conversations across the entity (or across many entities) to improve subsequent interactions with the client or professional. For example, as a particular agent of an entity is conversing vocally with a client, the spoken conversation is captured and transcribed from audio to digital voice data (e.g., audio to data). In specific aspects, a data-communications server operated by the data-communications provider can capture and analyze conversations of many (e.g., 10,000 or more) agents, which can be stored in a directory. A client on the agent side can communicate with a web client, such as a Virtual Contact Center (VCC) web client, and a client on the provider side. In some embodiments, the client on the agent side can transcribe the conversation from audio to the digital voice data. In other embodiments, the client on the agent side can communicate the audio to a data-communications server operated by the entity, which transcribes the audio to digital voice data using processing circuitry in communication therewith. The processing circuitry can form part of the data-communications server or be separate therefrom. As used herein, the digital voice data is indicative of transcribed audio conversations between agents and others (e.g., customers or other external personal) of remotely-situated client entities.
In specific embodiments, the processing circuitry analyzes the received digital voice data, the digital voice data being from the data communication services provided by the data-communications server. The analysis can include identifying keywords and speech characteristic parameters from the digital voice data, which can be, alone or together, indicative of a tone or a sentiment of a conversation (e.g., is the conversation going well or not). Example speech characteristic parameters can include a frequency, wavelength or velocity, and/or amplitude of the audio. In some embodiments, the speech characteristic parameters can be compared to a baseline of the agent to determine a tone. In other embodiments and/or in addition, speech characteristic parameters of the customer can be compared to thresholds to determine a tone. In a number of embodiments, the threshold can be different for different geographic regions, different types of customers (e.g., age, sex), different industries, among other types of classifications. In addition, particular keywords can be associated with and/or indicate a greater probability of a particular tone, sentiment, and/or outcome. As a particular example, the phrases “I want to speak to a manager” or “What is your name” can be associated with a conversation that is not going well for the agent. In other instances, “What is your name” at a particular frequency can be indicative of a positive outcome (e.g., an agent resolves an issue and the customer would like to commend the agent). Although as may be appreciated by one of ordinary skill in the art, embodiments are not so limited and can include a variety of different associations. Further, in various embodiments, the same keywords or phrases can indicate different tones or outcomes for different types of customers.
The identified keywords and speech characteristics can be associated with outcomes of the transcribed audio conversation. The associations can be provided in an archive stored on a memory circuit and which can be updated over time. In various embodiments, the outcome can be determined based on the keywords and/or metadata in the digital voice data. For example, the processing circuitry can identify the keywords and compare them to previously-identified associations of keywords and outcomes. As a specific example, an entity can initially set-up an archive to include associated words and phrases that indicate a sale has occurred, a subscription is continued, a customer hangs up prior to a sale, etc. Alternatively and/or additionally, the archive can include thresholds for speech characteristics parameters that are indicative of different tones or sentiments. The thresholds can indicate or otherwise be associated with a positive conversation, a negative conversation, and changes in threshold indicating a change from a positive to negative or negative to positive conversation. Such thresholds can be adjusted over time based on feedback into the system and corrections by agents. For example, prior to storing an outcome as being associated with a keyword and/or a speech characteristics parameter, the outcome and the association can be provided to the agent and/or to a manager for review and approval. In response to an input verifying the outcome, the association is stored. In response to an input correcting the outcome, the association is not stored or a revised outcome is associated with the speech characteristics parameter and stored.
The analysis of the digital voice data can be used for training purposes. For example, a particular agent is provided feedback after the call on what keywords to avoid and/or strategies for the next call. More specifically, the feedback can include providing the association(s) to the client data-communications server and/or to the endpoint device of the agent. In a number of specific embodiments, the feedback or training can be provided in real time or near-real time during the audio conversation. For example, the agent, via an endpoint device (e.g., a computer or otherwise) can be provided a display or audio that cannot be heard by the customer, that indicates keywords to avoid, keywords to use, and/or directs the user to change their speech pattern (e.g., slow your speech down, take a deep breath). In some specific embodiments, specific keywords, alone or in combination with specific speech characteristic parameters, can cause an auto-trigger for connecting a manager to the audio conversation. For example, another endpoint device that is used by the manager can be bridged to the data communication, such as a VoIP call. The data-communications server can access a database to identify the manager and information for connecting the manager into the audio conversation. In this way, a manager can be bridged into a call without additional action by the agent. In some instances, different customers can have different keywords or speech characteristic parameters that trigger the connection with the manager. These keywords or speech characteristic parameters can be based on previous audio conversations with the customer, identification of a category of customer (e.g., important customer that is ranked 10 on a scale of 1-10), among other analytics.
In various embodiments, the processing circuitry and/or the data-communications server operated by the data-communications provider or a client side server can analyze the associations over a period of time, and generate a report indicating different outcomes and associated keywords and/or speech characteristics parameters. Such reports can be used for training purposes and can also be used to identify different patterns. For example, customers from different geographic regions can interact similarly or differently from one another. As previously described, specific keywords or tones can indicate different outcomes based on geographic regions, markets, type of customer, etc. The processing circuitry and/or data-communications server can provide assessment of a quality of calls based on the analysis, and provide feedback to the entity. The feedback can be used for training, which can be provided in real time during the call or after.
In some specific embodiments, the analytics can be provided as a service by the data-communication (e.g., VoIP) provider. For example, the analytics can identify keywords and/or tones/sentiments that result in positive outcomes, and also keywords that provide faster outcomes. Additionally, the metrics used to assess the outcome and/or the quality of the call can adjust over time. The adjustment can be responsive to additional digital voice data, and/or verification or adjustment by the agents or a manager to ensure the determined outcomes are correct. In a more specific example, the feedback can be provided by identifying customer-interaction metrics in the digital voice data. In some embodiments, the associations can be locked and a manager reviews the recommended adjustments (by the agent) and approves or denies the adjustment. The customer-interaction metrics include values or ratings of an interaction of an agent with a customer. For example, the customer-interaction metrics can include different ratings which can be impacted by the use or non-use of specific keywords, phrases and speech characteristic parameters.
The processing circuitry can analyze the data over time and across a plurality of agents. For example, particular agents can be identified for an entity that have customer-interaction metrics outside a threshold value. Such agents, for example, can be provided feedback, as described above, for training purposes. The feedback can include identification of customer-interaction metrics to adjust for potentially better outcomes or specific outcomes (e.g., a sale or customer retention). In some instances, the feedback is provided in real time and/or during the conversation, such as recommended phrases to the say to the customer and/or recommended adjustments in the agent's tone. The customer-interaction metrics can be tracked over time and/or adjusted using digital voice data of additional audio conversation.
In various embodiments, the assessment of call quality can be provided as a service to an entity. For example, a data-communications provider can provide the analytics as a service and/or package. The analysis can be based on geographic regions, markets, type of customers, among other types of categorizations as specified by the entity. The assessment can be used to generate reports on different agents, regions, and to use for training purposes. For example, feedback can be provided to agents to indicate specific phrases used that can have resulted in a particular outcome, suggested phrases to use, and which can be provided in real time and/or after the call.
Turning now to the figures,
The endpoint devices are circuit-based instruments that are used by personnel (or users) and include data communications-enabled circuitry, such as VoIP-enabled endpoint devices (e.g., IP phones, smart phones, tablets, and/or desktop computers with appropriate VoIP software applications) and/or non-data communication/VoIP enabled endpoint devices (e.g., plain old telephone service (POTS) telephones and cellular-capable devices). Each endpoint device is respectively associated with an account of a respective client. Endpoint devices can be associated with a particular client account by registering the endpoint device with a particular client account serviced by the data-communications server 110. Registered devices for each client account can be listed in a respective account settings file (not shown) stored by the data-communications server 110. In this example, endpoint devices 118, 120, and 122 are associated with an account 113 for a first client A and endpoint devices 124, 126, 128 are associated with an account 115 for a second client B. In such a manner, a plurality of endpoint devices can each be serviced by the data-communications server 110 in accordance with the present disclosure. One or more of the clients can have client servers and/or databases 121 used to implement a variety of different services.
The system can include one or more processing circuits configured to implement client-specific control engines 112, which are configured to adjust the data-communications provided for each client account according to a respective set of control directives. For instance, the client-specific control engines 112 can adjust routing of a data (e.g., VoIP) communication for a client account by generating client-specific sets of control data to the data-communications server 110.
Different embodiments can implement the client-specific control engines 112 in various locations. For example, client-specific control engines 112 for one or more client accounts can be implemented in a central server connected to, or incorporated with, the data-communications server(s) 110. Additionally or alternatively, one or more client-specific control engines 112 can be implemented by one or more processing circuits maintained by the client (e.g., server/database 101). Similarly, the control directives can be stored locally within the client-specific control engines, or stored remotely (e.g., in a centralized database, in a database maintained by the client or a combination thereof).
In various embodiments, the data-communications system can track a variety of parameters and/or metrics related to calls made or received by the agents of a client entity via processing circuitry 106 that is communicatively coupled to the data-communications server 110. Although the processing circuitry 106 is illustrated as a component of the data-communications server 110, embodiments are not so limited and the processing circuitry can form part of or be separate from the data-communications server 110. The parameters can include information such as average call duration, compliance of call opening and account information, identification of issues and troubleshooting, resolution, professionalism, and other metrics. The parameters can be scored (e.g., percentage or other value scored) to rate the particular agent on each particular call and to form a metric used to assess an agent. In some embodiments, the parameters and/or metrics can be assessed automatically by the system using keywords. In other embodiments and/or in addition, the calls are recorded and can be replayed for a person to score (or update an automatic score) on the metrics. The parameters and/or metrics can be used to determine an outcome of the call. For example, the resolution can be indicative of the outcome. The scores can be reviewed upon recording the calls which can be reviewed by a reviewer. For example, the reviewer can listen to the call recording while a user interface is displayed on a computing device associated with the reviewer. The user interface displayed can illustrate a timeline of the recorded call, which can show timing of the current voice data being played. The timeline can allow the reviewer to navigate within the call by selecting portions of the timeline to change what the reviewer is listening to. Additionally, the agent and the customer can have different waveforms for efficient reviewing. In other embodiments, other types of analysis are used, as further described herein.
In a number of embodiments, a particular client can customize the metrics that are tracked. Such customization can include the type of metrics, values given, and/or particular phrases or statements for complying with the metric (e.g., a specific call opening). For example, the customization can include definitions of performance expectations and scores, as well as performance thresholds for various metrics.
The parameters and/or metrics can be tracked across a plurality of agents and over time. Such metrics can be used to generate reports on individual agents and/or comparing one or more agents. For example, the reports can be generated based on filters, such as agent's name, queue, length of call, customer name, transferred calls, transaction codes, custom fields, among other filters. The reports can be used to set performance baselines and measure changes over time. Additionally, in some embodiments, the parameters can be scored to form metrics or the metrics can be revised by reviewers manually and the scores can be normalized. For example, multiple reviewers can score the same call. Differences between the scores can be used to normalize the review scoring between the different reviewers. The reports can be periodically generated and stored. For example, the reports can be generated and scored weekly, monthly, quarterly, and/or yearly, and which can be indicative of customer feedback (provided without the user actively answering a questionnaire).
As previously described, client-specific control engines can be used to facilitate adjustment of a variety of remote services including, for example, data-communication services such as VoIP calls, audio and/or video conferencing, general private branch exchange services, packet switching, chat, and traffic management as well as non-VoIP services including, but not limited to, website hosting, remote data storage, remote computing services, and virtual computing environments. One or more of such services can be provided, for example, by a cloud computing network having one or more servers configurable for a plurality of clients.
The data-communications server 110 interfaces with the plurality of remotely-situated client entities to provide data communications services. As illustrated in the lower portion of
As may be appreciated by one of ordinary skill in the art, audio can be transcribed to text using a variety of techniques. As an example, an audio file can be generated and provided to speech recognition circuitry, which can be part of the endpoint device, the data-communications server 110, or other external circuitry. The audio file, which includes an acoustic signal received by a microphone of the endpoint, is converted or transcribed to text (e.g., a set of text words) by the speech recognition circuitry. In various embodiments, the speech recognition circuitry can use a voice model and/or database of words for converting or transcribing the audio to text. For example, the speech recognition circuitry can index the words in the audio file to identify words or phrases, such as using an extensible markup langue (XML), structured query language (SGL), mySQL, idx, and other database language. For more general and specific teachings on transcribing audio to test, reference is made to U.S. Publication No. 2009/0276215, filed on Apr. 17, 2007, entitled “Methods and Systems for Correcting Transcribed Audio Files;” U.S. Pat. No. 7,236,932, filed Sep. 12, 2000, entitled “Method of and Apparatus for Improving Productivity of Human Reviewers of Automatically Transcribed Documents by Media Conversion Systems;” and U.S. Pat. No. 6,424,935, filed Jul. 31, 2000, entitled “Two-way Speech Recognition and Dialect System,” each of which are fully incorporated by reference for their teachings.
As illustrated, the data-communications server 110 interfaces with a plurality of remotely-situated client entities and includes or is otherwise in communication with processing circuitry 106. The processing circuitry 106 can receive digital voice data indicative of transcribed audio conversations between a plurality of agents and customers of a remotely-situated client entity and identify keywords and speech characteristic parameters from the digital voice data. In specific embodiments, the data-communications server 110 captures the digital voice data via a client on the agent side communicating with a VCC web client and a client on the server side (e.g., the provider-side client 102) and provides the digital voice data to the processing circuitry 106. In other embodiments, the agent-side client communicates the digital voice data to the provider-side client 102. The provider-side client 102 then provides the digital voice data to the analytics circuitry 103 for identification of keywords and/or speech characteristic parameters. The identification can include analyzing the digital voice data for matches to keywords and speech characteristic parameters stored in an archive and/or database 101. The database 101 can include a relational database management system (RDMS) which stores data securely and returns the data in response to requests from other applications, as implemented by a database management server, and which can temporarily store data in cache 105.
The stored keywords and/or speech characteristic parameters can be associated with outcomes, in some embodiments. Example outcomes can include sale, no sale, positive or negative tone/sentiment. A tone or sentiment of the audio conversation can indicate how the call is perceived by the customer. A tone or sentiment can be identified based on the speech characteristic parameters. Example speech characteristic parameters include frequency, velocity, and amplitude of the conversation.
In a number of specific embodiments, the speech characteristic parameters can be compared to one or more thresholds. For example, the processing circuitry 106 via the analytics circuitry 103 and the provider database 101 can identify a speech characteristic parameter is outside of a threshold value (e.g., indicates problem), and which can be indicative of a tone of sentiment of the conversation. As a specific example, audio above a particular amplitude and velocity can be indicative of a customer or agent who is upset or otherwise agitated. The thresholds can include generic thresholds (anything above a particular value), thresholds that are specific to geographic regions or types of customers, and/or baseline values of the specific agent. For example, if the agent's speech is faster and louder than their baseline, it can be indicative of an issue in the conversation.
The processing circuitry 106 can provide association of the keywords and/or speech characteristic parameters with outcomes based on the analysis. The associations can be stored in the database 101 and/or provided to a client data-communications server 121 as feedback via the feedback circuitry 104. The feedback can be provided to a manager of the client entity and used for training purposes. Alternatively and/or in addition, the feedback can be provided to the particular agent in real time or near real time, such as while the audio conversation is ongoing. The feedback can include specific phrases to use and/or suggested changes in speech characteristic parameters.
In other embodiments, the feedback is provided to a manager, such as via a report that summarizes a subset of agents for the client entity. The report can include customer interaction feedback using the associations indicating negative and positive outcomes and associated keywords and/or speech characteristic parameters.
In some embodiments, the processing circuitry 106 can identify a speech characteristic parameter outside of a threshold value (i.e., indicates problem) or a keyword that is associated with a (manager) trigger, and in response, automatically bridge a manager of the respective agent into the audio conversation. The processing circuitry 106 and/or the data-communications server 110 can access a database to identify the manager and information for bridging the manager into the audio conversation.
In some examples, the processing circuitry 106 can store the associations in a database as an archive that is accessible and that ties the keywords with dates of the conversations and the speech characteristic parameters of the particular transcribed audio conversations. The archive can be adjusted over time based on additional audio conversations. For example, the processing circuitry 106 can adjust the associations over time based on further analysis of digital voice data and provide the adjustments as feedback. Alternatively and/or in addition, the adjustments can be based on user input, such as an input from an agent or a manager, as further described herein.
The call routing and other services for the VoIP telephone calls can be provided by one or more data-communications servers 232 within a cloud services system 230 (e.g., configured to provide a PaaS to customers of the VoIP provider). In particular example embodiments, the data-communications servers 232 can be located within platform as a service (PaaS) computing servers, which are part of the cloud services system 230. The cloud services system 230 also includes one or more provider hosted client-specific control engines 236. A client-specific control engine can also be implemented locally by a client (e.g., 246). In some embodiments, data centers can be part of a cloud-based system where the hardware providing the cloud services is located in a number of different data centers with different physical locations. Consistent with embodiments, the cloud services can include session initiation protocol (SIP) servers, media servers, and servers providing other services to both VoIP endpoint devices and the users of the VoIP endpoint devices. In some instances, the various servers, including both the data-communications servers and data analytic servers discussed herein, can have their functions spread across different physical and logical components. For instance, a cloud-based solution can implement virtual servers that can share common hardware and can be migrated between different underlying hardware. Moreover, separate servers or modules can be configured to work together so that they collectively function as a single unified server.
A particular example of a data-communications server which uses Session Initiation Protocol (SIP) to handle various call functions (e.g., call setup and tear down); however, the various embodiments discussed herein are not necessarily limited thereto. Consistent with the above and other embodiments disclosed herein, the data-communication servers are VoIP communications servers that are configured to establish a leg of the call from the VoIP endpoint devices (or dial peers) to another VoIP endpoint device, or to a gateway.
According to various embodiments, one or more data-communications servers 232 can monitor and analyze call data relating to digital call data of calls occurring using the VoIP endpoint devices 239, 241, 243, 245 via processing circuitry 248. For example, a data-communications server can be designed to receive digital voice data, such as directly from an agent-side client associated with particular endpoint devices. The agent-side client can communicate the audio or the digital voice data to the provider-side client. The provider-side client then provides the audio or digital voice data to processing circuitry (e.g., processing circuitry 248, although embodiments are not so limited and can include processing circuitry external to the data-communications server) for further analysis, such as transcribing to digital voice data, identifying keywords and/or speech characteristic parameters, identifying an outcome, and providing an association between the keywords and/or speech characteristic parameter using the identified outcome. The association can be stored in an archive in a provider-side database 234. In some specific embodiments, new keywords and/or parameter values are identified as having an association with an outcome. In other embodiments and/or in addition, a stronger correlation (e.g., probability) between the keywords and/or parameters is provided over time responsive to multiple verifications of an association. For example, a script running the data-communications server 232 can parse call digital call data and stored association to generate database queries that direct the data-communications server to provide a new association and/or update an existing association. The script can use the information to generate a report that can be used for training, promotions, and/or other analysis of agents. According to various embodiments, the database queries can be sent to a client database 237. The feedback can be provided in real time or near real time to the endpoint device of the agent and/or can be accessed by a manager.
Consistent with certain embodiments, the data-communications server 232 can be configured to interface with a third party server 233 such as a customer server, or other third party servers. For instance, a CPXML document stored by the data network 231 can identify, based upon a received call, a Uniform Resource Identifier (URI) that points to customer servers, or to other third party servers.
The above described data-communications system, as illustrated by
At 340, the method includes receiving, at processing circuitry in communication with a data-communications server, digital voice data indicative of transcribed audio conversations between a plurality of agents and customers of a remotely-situated client entity. The digital voice data can be transcribed from audio by the endpoint device of the agents and communicated to the processing circuitry and/or the data-communications server via clients on the agent and provide side. In other embodiments, a client server can transcribe the digital voice data and communicate the same to the data-communications server and/or the data-communications server can transcribe communicated audio waves.
At 342, the digital voice data can be analyzed, via the processing circuitry, using a database, such as a database of archived data including associations of keywords and/or speech characteristic parameters and outcomes. The digital voice data can be analyzed to identify keywords and speech characteristic parameters of the digital voice data, identify outcomes of the respective transcribed audio conversations, generate associations with at least one of the keywords and speech characteristic parameters with the identified outcomes, and/or providing the associations to a client data-communications server as feedback on audio conversations with the customers.
For example, at 344, the digital voice data is analyzed and used to identify keywords and speech characteristic parameters of the digital voice data and compare the identified keywords and parameters to a database. The database has stored correlations or associations of keywords and parameters to tones or sentiments of a conversation and outcomes of the conversations. The associations or correlations can include a probability of the match indicating the outcome. In response to a match, at 346, the correlated archived outcomes is identified. In response to no match and/or in addition (can have matching and nonmatching keywords), at 348, new keywords can be identified. At 350, the outcome of the respective transcribed audio conversation can be identified. The outcome can be identified based on a match and a respective correlation and/or based on analysis of metadata in the digital voice data. In some embodiments, an agent can enter metadata that is correlated with the digital voice data and/or that otherwise ties to the transcript. The metadata can identify the outcome.
The use of the archived database, in various embodiments, can include performing artificial intelligence to determine the outcome. The processing circuitry and/or data-communications server can analyze patterns to identify and associate different keywords, phrases, and/or speech characteristic parameters with particular outcomes, and which can be different for different types of customers (e.g., different geographic regions, fields, sexes, ages, and other categorizations). The processing circuitry and/or data-communications server can analyze resulting data to determine the outcome, such as identifying a sale, retention of a customer, hang-up, loss of a customer, resolution of a problem, which can include a specific and identified problem or a general problem, among other outcomes. Overtime, the processing circuitry and/or data-communications server can update the database based on additional digital voice data. For example, at 352, the method includes generating associations with at least one of the keywords and speech characteristic parameters with the identified outcomes. The associations can include new keywords that are associated with particular outcomes, new outcomes, and increasing or decreasing a correlation of keywords or parameters with an outcome. For more general information on artificial intelligence and specific information on achieving artificial intelligence, reference to made to U.S. Publication No. 2004/0143559, filed on Jan. 13, 2004, entitled “System and Method for Developing Artificial Intelligence,” and U.S. Pat. No. 5,473,732 filed on Feb. 7, 1995, entitled “Relational Artificial Intelligence System,” each of which are fully incorporated herein for their teachings.
In specific embodiments, the associations can be classified differently for different types of customers. For example, the processing circuitry can provide different associations based on characteristics of the particular audio conversations, including characteristics selected from the group consisting of: geographic region of the customer, sex of the customer, age of the customer, type of customer, and a combination thereof. The associations are stored in the provider database as an archive that is accessible and that ties the keywords with dates of the conversations and the classifications, and which identifies keywords and phrases to avoid for particular types of customers and keywords and phrases to use for the particular types of customers.
As a specific example, and not intended to be limiting, a first region can include a Northeastern region of the United States and a second region can include a Southern region of the United states. It may be identified that customers from the first region speak at a greater frequency and amplitude that customers from the second region. As another example, it may be identified that women speak at a lower amplitude than men. A third region can include Southern Pacific Coast of the United States (e.g., parts of California) which may include customers that speak at a similar frequency and amplitude as customers from the first region. However, embodiments are not so limited and the above is provided only as an illustrative example. Additionally each of the about may have specific words and phrases to avoid, as they have a greater probability then other words at causing a positive or negative outcome. As a specific example, the second region may have a positive outcome associated with calling the customer “Miss” or “Mr.” whereas customers from the third region may have negative outcome associated with the same or may have a greater probability of a positive outcome when addressing customers by their first name.
In various embodiments, feedback can be provided. For example, at 354, the method further includes providing the associations to a client data-communications server as feedback on audio conversations with the customers. The feedback can be provided to the agent in real time and used by the agent to assist in achieving a positive outcome or to otherwise adjust an outcome of the conversation. In other embodiments, the feedback can be provided after the call is finished and used to train the user. For example, the feedback can be used to train the user in how to handle different situations. In some instances, a report is generated that can be reviewed by a manager, as previous described. Further, in specific embodiments, a manager can be automatically bridged into a VoIP call responsive to particular keywords (e.g., swearing), based on the specific customer, and/or speech characteristics parameters outside a threshold. In such a manner, the feedback can include or otherwise be indicative of customer feedback which is provided to the client entity without actively querying the customer, such as with the customer actively providing answers to a questionnaire.
In a number of related embodiments, the associations between keywords and/or speech characteristic parameters can be adjusted over time. The adjustment can be responsive to different outcomes or input from the agent or other users. For example, the processing circuitry can provide a respective association with at least one of the keywords and speech characteristics with the identified outcome to an agent associated with the audio conversation or a manager associated with the agent. In a number of embodiments, the associations can be locked and a manager reviews the recommended adjustments (by the agent) and approves or denies the adjustment. In response to the agent or manager providing an input indicating the associated is incorrect, the association is adjusted. In other embodiments and/or in addition, the association can be stored in response to the agent or manager providing an indication indicating the associated is correct or not providing any indication within a threshold period of time.
Additionally, as previously described, one or more data-communications servers 432 can monitor and analyze call data relating to digital call data of calls occurring using the VoIP endpoint devices 439, 441, 443, 445 and processing circuitry. For example, a data-communications server 432 can be designed to receive digital voice data, such as directly from an agent-side client associated with particular endpoint device(s). The agent-side client 439, 441, 443, 445 can communicate the audio or the digital voice data to the provider-side client 446. The provider-side client 446 then provides the audio or digital voice data to processing circuitry 448 for further analysis, such as that previously described above. The association can be stored in an archive in a provider-side database 434. In some embodiments new keywords and/or parameter values are identified as having an association with an outcome. In other embodiments and/or in addition, a stronger correlation (e.g., probability) between the keywords and/or parameters is provider over time responsive to multiple verifications of an association. A script running the data-communications server 432 can parse call digital call data and stored association to generate database queries that direct the data-communications server 432 to provide a new association and/or update an existing association. The script can then use the information to generate a report that can be used for training, promotions, and other analysis of agents. According to various embodiments, the database queries could be sent to a third party server 433, as previously described. The feedback can be provided in real time or near real time to the endpoint device of the agent and/or can be accessed by a manager.
The feedback, in various embodiment, can include or be based on an assessment of a quality of agent interactions with customers, which is interchangeable referred to as “customer interaction feedback” herein. The digital voice data, as well as the keywords and speech characteristic parameters, can be used to analyze and assess agents. The feedback can be provided in real time, such as during a VoIP call between the agent and a customer. For example, an agent can be provided guidance during a live call in order to change an outcome of the call. In other instances and/or in addition, the feedback can be used to provide an assessment of the quality of the call, which can be used for training purposes by a manager.
The feedback and/or assessment can be different for different geographic regions and/or types of customers. For example, as previous described in the example above, customers from the Southern region of the United States may respond better to different speech characteristics than customers from the Northeastern region of the United States. Such associations can be determined over time using the database that captures the digital voice data, and can be across different entities. The analytics can indicate that certain tones reflect an upset or happy customer, particular phrases or words that result in positive or negative outcomes for different types of customers, among other analytics. The feedback can be provided as a report and/or is provided in real time to the agent, such as suggested phrases or changes in tone that can result in a particular outcome, such as a positive (or faster) outcome or a greater probability of positive outcome. The feedback can be displayed on a user device that the agent has access to (e.g., GUI of a computer) and/or as voice that the agent can hear but the customer cannot hear. In such example embodiments, the agent can determine whether or not to follow the feedback, and which can be used to further adjust the associations. For example, if a user does not follow the feedback and has a positive outcome, additional keywords or speech characteristics parameters can be identified from the conversation and associated with the positive outcome. Additionally and/or alternatively, data in the feedback can have an adjusted (e.g., lowered probability) association with the outcome, although embodiments are not so limited.
The one or more data-communications servers 432 monitor and analyze call data relating to digital call data of calls occurring using the VoIP endpoint devices 439, 441, 443, 445 can further analyze such assessment for different customers over many entities and can provide the same as data analytic packages 436. The packages can be provided as a service to entities and can be based on geographic regions, markets, and type of customers. In specific embodiments, the packages 436 can be provided as a service and can be used for training agents for providing better customer interactions, such as suggested phrasing and words that result in particular outcomes and/or outcomes that are efficient. The packages 436 can be based on various customer-interaction metrics used to analyze the quality of calls.
In some embodiments, the metrics or criteria can be adjusted over time. Example customer-indication metrics include number of outcomes identified as positive, use of particular keywords, audio outside a threshold frequency/velocity (can be based on agents baseline). As described above, the adjustment can be based on artificial intelligence used to analyze digital voice data across a plurality of entities over time. Such analysis can identify patterns between types of customers and outcomes, such as particular tones, frequencies, phrase, etc., that result in particular outcomes and/or at a greater frequency. A particular agent, such as a manager, can verify the adjustments to ensure the outcome determined via the analysis by the data-communications server 432 is correct.
As an example, the data analytics packages 436 can be offered as a cloud-based service to an entity to use for training purposes and/or otherwise improve or monitor customer interactions by agents. In some instances, the cloud based services can include automatic antitrust blockers. For example, the data-communication (e.g., VoIP) service provider can have data blockers 451 which can automatically block particular data provides as part of the data analytics packages 436 to particular entities based on antitrust concerns or privacy for other entities. Example feedback can include outputting a phrase to say during a live audio conversation to an endpoint device currently in use by the agent for the live audio conversation between the agent and the customer, outputting the phrase provided as data for display on a user display of the endpoint device and/or audio provided to a speaker of the endpoint device, a customer feedback report using the associations indicative of negative and positive outcomes and associated keywords and/or speech characteristic parameters associations.
The metrics and associations can be analyzed over time and used to adjust the archive by adjusting keywords and speech characteristic parameters associated with outcomes for a geographic location of the customer or the agent of the current audio conversation (adjust tones/phrases that result in positive/negative for the specific geographic location), and/or outcomes for different categorization of the customer or market associated with the customer of the current conversation (adjust tones/phrases that result in positive/negative for the type of customer). Additionally, the feedback can be used to promote agents, such as indicating agents having customer-interaction metrics above a threshold to the client data-communications server (identify agents with good outcomes for promotion).
In accordance with various embodiments, the system can be used to implement a method, including accessing, at processing circuitry in communication with a data-communications server, digital voice data indicative of transcribed audio conversations between a plurality of agents and customers of a remotely-situated client entity. The digital voice data can include keywords and speech characteristic parameters associated with outcomes of the transcribed audio conversations. The method can further include analyzing, via the processing circuitry, the digital voice data to identify customer-interaction metrics including values of specific keywords and speech characteristic parameters of the digital voice data associated with particular outcomes, and provide identification of particular agents for particular remotely-situated client entities that have customer-interaction metrics outside a threshold value to client data-communications servers associated the particular agents (e.g., good or bad agents). Additionally, the customer-interaction metrics can be adjusted. For example, the adjustment can be based on digital voice data of additional audio conversations and/or corrections by the agent and/or a manager of the agent. In some specific embodiments, the adjustment is verified in response to receiving an input indicating approval, such as allowing for a manager to approve or further adjust based on real-life experience. Additionally, the adjustments over time can be based on different categories of customers, such as geography, sex, time of day, market, and age of the customer.
Feedback, as previously described, can be provided to the agent. For example, feedback to an agent of the plurality of agents can be provided during a live audio conversation based on the adjusted customer-interaction metrics, such as outputting a phrase to say during the live audio conversation to an endpoint device currently in use by the agent for the current audio conversation between the agent and the customer. In other embodiments, the analytics can be provided as the service or package as service to a third party.
Various embodiments are implemented in accordance with the underlying Provisional Application (Ser. No. 62/650,652), entitled “Analysis of Digital Voice Data in a Data-Communication Server System,” filed Mar. 30, 2018, to which benefit is claimed and which are fully incorporated herein by reference. For instance, embodiments herein and/or in the provisional application may be combined in varying degrees (including wholly). Reference may also be made to the teachings and underlying references provided in the underlying provisional application. Embodiments discussed therein are not intended, in any way, to be limiting to the overall technical disclosure, or to any part of the claimed invention unless specifically noted.
Various blocks, modules or other circuits can be implemented to carry out one or more of the operations and activities described herein and/or shown in the figures. As examples, the Specification describes and/or illustrates aspects useful for implementing the claimed invention by way of various circuits or circuitry using terms such as blocks, modules, device, system, unit, controller, and the like. In these contexts, a “block” (also sometimes “logic circuitry” or “module”) is a circuit that carries out one or more of these or related operations/activities (e.g., a call control circuit). For example, in certain ones of the above-discussed embodiments, one or more modules are discrete logic circuits, computer processing circuits, or programmable logic circuits configured and arranged for implementing these operations/activities, as in the blocks shown in the figures.
Similarly, it will be apparent that a server includes a computer processing circuit that is configured to provide services to other circuit-based devices. Moreover, various other circuit-related terminology is used in a similar context as apparent to the skilled artisan, as is the case with each such apparatus which refers to or includes otherwise known circuit-based structures. As a first example, a (VoIP) endpoint device (or endpoint) includes a communication circuit and (computer) processing circuits which are configured to establish (VoIP) communication sessions with other endpoint devices. The communications circuit can be enabled to communicate over a broadband network such as the Internet or a cellular communications network (e.g., computer) processing circuits as configured to establish data communications sessions with other endpoint devices and such endpoints include, e.g., personal computers, IP-enabled mobile phones, and tablet computers. Also, a client entity refers to or includes an endpoint device (as above) which is linked/associated with a client of a provider/operator of the company overseeing the data-communications server or data-center communications server. Further, a data-center communications server or data-communications server refers to or includes a computer processing circuit that is configured to provide data-communications services to other circuit-based devices. In certain embodiments, a processing circuit or circuitry, as provided above, is one or more computer processing circuits programmed to execute a set (or sets) of instructions (and/or configuration data). The instructions (and/or configuration data) can be in the form of software stored in and accessible from a memory circuit, and where such circuits are directly associated with one or more algorithms (or processes), the activities pertaining to such algorithms are not necessarily limited to the specific flows such as shown in the flow charts illustrated in the figures (e.g., where a circuit is programmed to perform the related steps, functions, operations, activities, etc., the flow charts are merely specific detailed examples). The skilled artisan would also appreciate that different (e.g., first and second) modules can include a combination of a central processing unit (CPU) hardware-based circuitry and a set of computer-executable instructions, in which the first module includes a first CPU hardware circuit with one set of instructions and the second module includes a second CPU hardware circuit with another set of instructions.
Certain embodiments are directed to a computer program product (e.g., nonvolatile memory device), which includes a machine or computer-readable medium having stored thereon, instructions which can be executed by a computer (or other electronic device) that includes a computer processor circuit to perform these operations/activities. For example, these instructions reflect activities or data flows as may be exemplified in figures, flow charts, and the detailed description.
Based upon the above discussion and illustrations, those skilled in the art will readily recognize that various modifications and changes may be made to the various embodiments without strictly following the exemplary embodiments and applications illustrated and described herein. For example, although aspects and features may in some cases be described in individual figures, it will be appreciated that features from one figure can be combined with features of another figure even though the combination is not explicitly shown or explicitly described as a combination. Such modifications do not depart from the true spirit and scope of various aspects of the disclosure, including aspects set forth in the claims.
Number | Name | Date | Kind |
---|---|---|---|
5473732 | Chang | Dec 1995 | A |
6424935 | Taylor | Jul 2002 | B1 |
7236932 | Grajski | Jun 2007 | B1 |
7457404 | Hession et al. | Nov 2008 | B1 |
7602900 | Croak et al. | Oct 2009 | B1 |
8126136 | Tong | Feb 2012 | B2 |
8537983 | Haggerty | Sep 2013 | B1 |
8577018 | Ben-Yair | Nov 2013 | B1 |
8600034 | Teitelman et al. | Dec 2013 | B2 |
8605878 | Claudatos et al. | Dec 2013 | B2 |
8804577 | Kondapalli | Aug 2014 | B1 |
9031222 | Wolfeld | May 2015 | B2 |
9053211 | Goldfarb | Jun 2015 | B2 |
9112974 | Wilsie et al. | Aug 2015 | B1 |
9160853 | Daddi | Oct 2015 | B1 |
9160854 | Daddi et al. | Oct 2015 | B1 |
9178999 | Hegde et al. | Nov 2015 | B1 |
9203652 | Petit-Huguenin et al. | Dec 2015 | B2 |
9602666 | Ponting et al. | Mar 2017 | B2 |
9848082 | Lillard et al. | Dec 2017 | B1 |
9894578 | Jawaharlal | Feb 2018 | B1 |
9936066 | Mammen | Apr 2018 | B1 |
10194027 | Daddi et al. | Jan 2019 | B1 |
10198427 | Yishay | Feb 2019 | B2 |
10419609 | Harris | Sep 2019 | B1 |
10546008 | Yishay | Jan 2020 | B2 |
10574822 | Sheshaiahgari et al. | Feb 2020 | B1 |
10608929 | Zaslow et al. | Mar 2020 | B2 |
10614107 | Yishay | Apr 2020 | B2 |
10642889 | Reshef et al. | May 2020 | B2 |
10735592 | Liu | Aug 2020 | B1 |
11140110 | Kasireddy et al. | Oct 2021 | B2 |
20020019864 | Mayer | Feb 2002 | A1 |
20040098274 | Dezonno et al. | May 2004 | A1 |
20040143559 | Ayala | Jul 2004 | A1 |
20050246211 | Kaiser | Nov 2005 | A1 |
20060026067 | Nicholas | Feb 2006 | A1 |
20060233347 | Tong et al. | Oct 2006 | A1 |
20060235994 | Wu | Oct 2006 | A1 |
20060285665 | Wasserblat | Dec 2006 | A1 |
20060285670 | Chin | Dec 2006 | A1 |
20070025537 | Claudatos | Feb 2007 | A1 |
20070025539 | Claudatos | Feb 2007 | A1 |
20080082341 | Blair | Apr 2008 | A1 |
20080118052 | Houmaidi | May 2008 | A1 |
20080130842 | Johri | Jun 2008 | A1 |
20080219429 | Mandalia | Sep 2008 | A1 |
20080228296 | Eilam | Sep 2008 | A1 |
20090210228 | George | Aug 2009 | A1 |
20090254404 | Eggenberger | Oct 2009 | A1 |
20090276215 | Hager | Nov 2009 | A1 |
20090292583 | Eilam et al. | Nov 2009 | A1 |
20090306984 | Mark et al. | Dec 2009 | A1 |
20100197329 | Claudatos et al. | Aug 2010 | A1 |
20100322399 | Couse | Dec 2010 | A1 |
20110078113 | Franz | Mar 2011 | A1 |
20110305331 | Hughes et al. | Dec 2011 | A1 |
20130016823 | Odinak | Jan 2013 | A1 |
20130018967 | Gannu | Jan 2013 | A1 |
20130176413 | Lowry | Jul 2013 | A1 |
20140254775 | O'Connor | Sep 2014 | A1 |
20140304365 | Khanna | Oct 2014 | A1 |
20140324420 | Sorensen et al. | Oct 2014 | A1 |
20150046514 | Madan et al. | Feb 2015 | A1 |
20150170674 | Ishibashi | Jun 2015 | A1 |
20150256677 | Konig | Sep 2015 | A1 |
20150365725 | Belyaev | Dec 2015 | A1 |
20150373196 | Scott et al. | Dec 2015 | A1 |
20160006871 | O'Connor | Jan 2016 | A1 |
20160036972 | Ristock et al. | Feb 2016 | A1 |
20160071517 | Beaver | Mar 2016 | A1 |
20160127553 | McCormack et al. | May 2016 | A1 |
20160182718 | Yoakum et al. | Jun 2016 | A1 |
20170011740 | Gauci | Jan 2017 | A1 |
20170019529 | Bostick et al. | Jan 2017 | A1 |
20170104876 | Hibbard | Apr 2017 | A1 |
20170214779 | Moran et al. | Jul 2017 | A1 |
20180082112 | Keller | Mar 2018 | A1 |
20180165723 | Wright | Jun 2018 | A1 |
20190158671 | Feast | May 2019 | A1 |
20210350793 | Yuasa | Nov 2021 | A1 |
20220292518 | Kilicoglu | Sep 2022 | A1 |
Number | Date | Country |
---|---|---|
2779538 | Sep 2014 | EP |
2546433 | Jul 2017 | GB |
2010069567 | Jun 2010 | WO |
Number | Date | Country | |
---|---|---|---|
62650652 | Mar 2018 | US |