Integration of contact center surveys

Information

  • Patent Grant
  • 7792278
  • Patent Number
    7,792,278
  • Date Filed
    Friday, March 31, 2006
    18 years ago
  • Date Issued
    Tuesday, September 7, 2010
    13 years ago
Abstract
Systems and methods for integrating contact center surveys are provided. In this regard, a representative method comprises: obtaining performance data pertaining to a contact center agent, the agent being designated to assist a customer of the contact center; providing a survey to the customer; receiving survey data responsive to the survey and corresponding to interaction with the customer by the contact center agent; and correlating the performance data with the survey data.
Description
BACKGROUND

Contact center agents communicate with customers in a variety of scenarios. By way of example, a customer may contact a contact center in order to receive consumer support services related to a purchased product. In response, a contact center agent typically interacts with the customer to acquire enough information so that the customer's requests can be accommodated. Such interaction may be via telephone, Internet, chat and messenger, and/or email. As should be understood, the training, skills and experience of a contact center agent can dramatically affect the manner in which this interaction takes place.


In this regard, significant management resources can be devoted to ensuring that contact center agents are properly evaluated and trained. For instance, contact center agents typically are evaluated by quality monitoring personnel, such as supervisors. These supervisors oftentimes review audio recordings or desktop events of the interactions between the contact center agents and customers, resulting in some form of quality scoring by the supervisor. The scoring is indicative of the performance characteristics of the contact center agents and can be used to schedule a contact center agent for training. For example, if the scoring indicates that a contact center agent is a substandard performer with respect to a particular skill set, the contact center agent can be assigned training for improving that skill set. Thus, there remains a need for an actionable contact center survey system and method.


SUMMARY

Systems and methods for integrating actionable contact center surveys are provided. In this regard, an exemplary embodiment of such a method comprises: obtaining performance data pertaining to a contact center agent, the agent being designated to assist a customer of the contact center; providing a survey to the customer; receiving survey data responsive to the survey and corresponding to interaction with the customer by the contact center agent; and correlating the performance data with the survey data. Therefore, for example, if customers communicate dissatisfaction with a specific interaction, event, policy, or process, this may be correlated with performance data and/or shared with the relevant operating area to ensure a timely response, without upsetting more customers.


In an enhancement, if, for example, a customer is dissatisfied, the system of the representative method above may transfer the unhappy customer back to a contact center supervisor so that the problem/dissatisfaction can be immediately addressed before the conclusion of the customer's interaction with the contact center. This real-time actionable intervention based on survey responses allows for a contact center to turn a bad experience into a positive experience.


An exemplary embodiment of a system for integrating contact center surveys comprises a quality monitoring system, a data repository and a survey system. The quality monitoring system is operative to obtain performance data pertaining to a contact center agent, the agent being designated to assist a customer of the contact center. The data repository is configured to store the performance data obtained by the quality monitoring system. The survey system is operative to provide a survey to the customer, to receive survey data responsive to the survey and corresponding to the contact center agent, and to provide the survey data to the data repository. The quality monitoring system is further operative to access the data repository and correlate the performance data with the survey data. This, therefore, allows for immediate insights into threats and opportunities.


In an enhancement, the quality monitoring system may further analyze the survey data to identify when customer satisfaction drops below a predefined threshold. This may be done in real-time or near real-time, such as within minutes or hours. Yet in another enhancement, the system provides at least one actionable report, item or alert for the contact center based on at least the survey data or the correlation of the performance data and the survey data.


Computer-readable media also are provided. In this regard, an exemplary embodiment of a computer-readable medium includes a computer program that comprises computer-executable instructions for performing the computer-implemented steps of: obtaining performance data pertaining to a contact center agent; providing a survey to a customer of the contact center agent; receiving survey data responsive to the survey and corresponding to the contact center agent; and correlating the performance data with the survey data.


Other systems, methods, features and/or advantages of this disclosure will be or may become apparent to one with skill in the art upon examination of the following drawings and detailed description. It is intended that all such additional systems, methods, features and/or advantages be included within this description and be within the scope of the present disclosure.





BRIEF DESCRIPTION

Many aspects of the disclosure can be better understood with reference to the following drawings. The components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present disclosure. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views. While several embodiments are described in connection with these drawings, there is no intent to limit the disclosure to the embodiment or embodiments disclosed herein. On the contrary, the intent is to cover all alternatives, modifications, and equivalents.



FIG. 1 is a schematic diagram illustrating an exemplary embodiment of a system for integrating contact center surveys.



FIG. 2 is a flowchart illustrating functionality (or method steps) that can be performed by an exemplary embodiment of a system for integrating contact center surveys.



FIG. 3 is a schematic diagram illustrating an embodiment of a computer-implemented system that can be used for performing functionality, such as the functionality depicted in FIG. 2.



FIG. 4 is flowchart illustrating functionality (or method steps) that can be performed by an exemplary embodiment of a quality monitoring system.



FIG. 5 is flowchart illustrating functionality (or method steps) that can be performed by an exemplary embodiment of a survey system.



FIG. 6 is a schematic diagram illustrating another exemplary embodiment of a system for integrating contact center surveys.



FIGS. 7-10 are flowcharts illustrating functionality (or method steps) associated with exemplary embodiments of systems for integrating contact center surveys.





DETAILED DESCRIPTION

As will be described here with reference to several exemplary embodiments, systems and methods for integrating contact center surveys are provided. In this regard, some embodiments can potentially provide one or more perceived advantages over conventional contact center operations. By way of example, at least some embodiments provide the ability to correlate customer responses with monitoring personnel (supervisor) scoring. That is, customer responses pertaining to the quality of contact center interactions with one or more contact center agents can be obtained. The responses then can be correlated with supervisor scoring of the contact center agents that interact with the customers. In so doing, various anomalies can potentially be detected. For instance, if customer responses tend to vary dramatically from the supervisor scoring, such an anomaly could indicate supervisor scoring bias or latent deficiencies in contact handling procedures. In some embodiments, information corresponding to customer responses is obtained by use of a survey that can be provided in various formats, such as an interactive voice response (IVR) survey. In other embodiments, various other formats of surveys, such as email, web-based, mail or courier-based and live agent callback, can be used.


Notably, the desire to provide a customer with a survey may be based, at least in part, on historical data associated with an agent. For instance, if a particular agent has been experiencing difficulty in a particular skill set, such a product knowledge, historical data pertaining to this lack of knowledge should be available. This historical data could then be used to provide a customer with a survey that is intended to determine whether the customer's interaction with the agent also indicates a lack of product knowledge. Multiple surveys provided to multiple customers could provide the basis for trend analysis to determine whether the agent was improving in this or other skill sets, for example.


Referring now to the drawings, FIG. 1 is a schematic diagram illustrating an exemplary embodiment of a system for integrating contact center surveys. In particular, system 100 involves interaction between contact center agent 102 and a customer 104 via a communication network 106. Note that network can include one or more of various communication systems including the Internet and PSTN, for example. System 100 also incorporates a quality monitoring system 108 that enables a supervisor 109, for example, to monitor various aspects of the interaction between the agent and the customer. In addition, the quality monitoring system can facilitate scheduling and training of the agents.


A recording system 110 is able to record at least some of the interactions between the customer and one or more of the agents involved, in this case, agent 102. For instance, the recording system is configured to record audio communications occurring between the customer and agent. In other embodiments, various other facets of such an interaction could be recorded. By way of example, one or more display screens of data can be captured from the computer that the agent is using during the interaction.


In this embodiment, recording of communications is facilitated when the customer is to participate in a survey. That is, when a survey is to be provided to the customer for obtaining responses pertaining to the customer's interactions with the contact center agent, the recording system ensures that communications associated with those interactions are recorded. In other embodiments, recording of communications could take place whether or not a survey is to be provided to a customer. Such recording can be screen, e.g., desktop events, and/or voice/audio. However, provisions for ensuring that recordings can be correlated with any acquired survey data should be used, such as by tagging the recordings and the survey data with a form of identification.


In the embodiment of FIG. 1, designation of a particular customer for being provided with a survey, as well as providing of the survey itself, is facilitated by a survey system 112, representative functionality of which will be described in detail later. Notably, however, other embodiments could use various other systems and/or combinations of systems to facilitate designation of a customer and/or administration of such a survey. By way of example, a rules engine associated with a quality monitoring system could be used.


In this regard, FIG. 2 is a flowchart illustrating functionality (or method steps) that can be performed by the embodiment of the system for integrating contact center surveys of FIG. 1. In particular, the functionality may be construed as beginning at block 150, where historical performance data pertaining to a contact center agent is obtained. By way of example, such performance data can include data representative of various performance characteristics of an agent, such as product knowledge. Such performance data can be obtained by inputs provided to a quality monitoring system, such as by a supervisor of the agent.


In block 152, a survey is provided to a customer of the contact center. Notably, interaction between the agent and the customer could take place before a survey is given. However, pre-interaction surveys, intra-interaction surveys and/or post interaction surveys could be provided. In some embodiments, the format of the survey could depend, at least in part, on the time that a survey is given. For example, a pre-interaction survey given in response to a contact could be provided via an IVR survey. In contrast, a post-interaction survey could be provided via email, for example.


In block 154, survey data responsive to the survey is received. Preferably, the survey data includes information pertaining to the customer's interaction with the agent, such as data indicating whether the agent appeared to have sufficient product knowledge to answer the customer's questions, for example. Additionally or alternatively, various other types of information can be obtained.


In block 156, the performance data can be correlated with the survey data. As mentioned before, the survey data can be correlated with the performance data for various purposes. By way of example, the correlated data can be analyzed to determine whether internal (supervisor) quality scoring procedures are consistent with customer responses. Notably, most conventional contact center monitoring systems permit quality scoring of contact center agents by supervisors. However, correlating third party data, such as customer responses to surveys, has not been utilized to determine whether the internal scoring is an accurate measure of contact center agent performance. Additionally or alternatively, processing algorithms can be used to identify scores (or elements of scores) that seem to be out of sync with the customer perception captured in the survey.



FIG. 3 is a schematic diagram illustrating an embodiment of a computer-implemented system that can be used for performing various functions associated with a system for integrating contact center surveys. Notably, although this embodiment involves a single platform performing multiple functions, other embodiments can involve distributed architectures that include various components distributed across multiple systems.


Generally, in terms of hardware architecture, system 200 includes a processor 202, a memory 204, and one or more user and/or communication (I/O) device interface(s) 206 that are communicatively coupled via a local interface 208. The local interface can include, for example but not limited to, one or more buses or other wired or wireless connections. The local interface may have additional elements, which are omitted for simplicity, such as controllers, buffers (caches), drivers, repeaters, and receivers to enable communications. Further, the local interface may include address, control, and/or data connections to enable appropriate communications among the aforementioned components. The processor may be a hardware device for executing software, particularly software stored in memory.


The processor can be any custom made or commercially available processor, a central processing unit (CPU), an auxiliary processor among several processors associated with the recorder, a semiconductor based microprocessor (in the form of a microchip or chip set), a macroprocessor, or generally any device for executing software instructions. Examples of suitable commercially available microprocessors are as follows: a PA-RISC series microprocessor from Hewlett-Packard® Company, an 80×86 or Pentium® series microprocessor from Intel® Corporation, a PowerPC® microprocessor from IBM®, a Sparc® microprocessor from Sun Microsystems®, Inc, or a 68xxx series microprocessor from Motorola® Corporation.


The memory can include any one or combination of volatile memory elements (e.g., random access memory (RAM, such as DRAM, SRAM, SDRAM, etc.)) and nonvolatile memory elements (e.g., ROM, hard drive, tape, CDROM, etc.). Moreover, the memory may incorporate electronic, magnetic, optical, and/or other types of storage media. Note that the memory (as well as various other components) can have a distributed architecture, where various components are situated remote from one another, but can be accessed by the processor 202. Additionally, memory 204 can also include an operating system 210, as well as instructions associated with various subsystems, such as a quality monitoring system 212, a recording system 214 and a survey system 216.


The software in memory may include one or more separate programs, each of which includes an ordered listing of executable instructions for implementing logical functions. In this regard, a nonexhaustive list of examples of suitable commercially available operating systems is as follows: (a) a Windows® operating system available from Microsoft® Corporation; (b) a Netware® operating system available from Novell®, Inc.; (c) a Macintosh® operating system available from Apple® Computer, Inc.; (d) a UNIX operating system, which is available for purchase from many vendors, such as the Hewlett-Packard® Company, Sun Microsystems®, Inc., and AT&T® Corporation; (e) a LINUX operating system, which is freeware that is readily available on the Internet 100; (f) a run time Vxworks® operating system from WindRiver® Systems, Inc.; or (g) an appliance-based operating system, such as that implemented in handheld computers or personal data assistants (PDAs) (e.g., PalmOS® available from Palm® Computing, Inc., and Windows CE® available from Microsoft® Corporation). The operating system 486 can be configured to control the execution of other computer programs and provides scheduling, input-communication control, file and data management, memory management, and communication control and/or related services.


A system component embodied as software may also be construed as a source program, executable program (object code), script, or any other entity comprising a set of instructions to be performed. When constructed as a source program, the program is translated via a compiler, assembler, interpreter, or the like, which may or may not be included within the memory, so as to operate properly in connection with the operating system.


When the system for integrating contact center surveys 200 is in operation, the processor is configured to execute software stored within the memory, to communicate data to and from the memory, and to generally control operations of the system pursuant to the software. Software in memory, in whole or in part, is read by the processor, perhaps buffered, and then executed.



FIG. 4 is a flowchart illustrating functionality associated with the embodiment of the system of FIG. 3. Specifically, the functionality depicted in FIG. 4 is associated with the survey system 216. In this regard, the functionality may be construed as beginning at block 250, where a determination is made as to whether a survey is to be provided to a customer. In block 252, if it is determined that a survey is to be provided, such a survey is provided to the customer. In block 254, survey data responsive to the survey is obtained and, thereafter, such as depicted in block 256, the survey data is provided to a data repository. By way of example, the survey data can be provided to a memory device so that the survey data can be stored and accessed for later use.



FIG. 5 is a flowchart illustrating additional functionality associated with the embodiment of FIG. 3. Specifically, the functionality depicted in FIG. 4 is associated with the quality monitoring system 212. In this regard, the functionality may be construed as beginning at block 260, wherein performance data corresponding to a contact center agent is received. For instance, the performance data can include a quality score associated with a performance characteristic of an agent.


In block 262, the performance data is provided to a data repository. In block 264, the data repository is accessed to obtain survey data associated with the contact center agent. Thereafter, such as depicted in block 266, the performance data is correlated with the survey data. By way of example, an element of a quality score for an agent corresponding to product knowledge can be correlated with survey data relevant to the perceived level of product knowledge of that agent during an interaction with the customer.



FIG. 6 is a schematic diagram illustrating another exemplary embodiment of a system for integrating contact center surveys. In particular, system 300 utilizes a communication network 302 that facilitates interaction of various callers and agents, such as customer 304 and agent 306. System 300 also incorporates a quality monitoring system 308 that enables monitoring of the interactions of the callers and agents, such as by a supervisor 309. The system also incorporates a recording system 310, survey system 312, data repository 314, computer telephony integration (CTI) 316 and business rules engine 318 that facilitate enhanced quality monitoring as will be described below.


In operation, customer 304 connects to a contact center associated with system 300, thereby notifying the CTI 316 of the incoming contact. The recording system 310 and CTI interact to determine whether the interaction, e.g., the incoming contact, is to be recorded. As will be described in more detail later, this determination can be based, at least in part, on whether a survey is to be provided to the customer. However, regardless of whether the recording system is to record the interaction, the contact is connected to agent 306.


In this embodiment, at the end of the agent interaction with the customer, an instruction is provided to the agent to transfer the contact to the survey system 312. This is accomplished by functionality associated with the CTI and business rules engine. By way of example, an agent involved in the interaction may have received low quality scoring relating to product knowledge. If it is determined that the current interaction involves product knowledge (as may be determined by the particular display screens of information that the agent is accessing during the interaction), a determination may be made to provide the customer with a survey. Note that such a survey need not be tailored to specifically address historically noted deficiencies in the agent (in this case, lack of product knowledge), although such could be addressed in some embodiments.


Once the contact is transferred, the survey system queries the customer about his experience with the contact center, such as the interactions with the contact center agent. Responses to the survey are converted to survey data, which can be in the form of scaled scores or “yes/no” responses, for example. The survey data corresponding to the customer's responses are then provided to the data repository 314. In some embodiments, providing of the survey data to the data repository can be facilitated, at least in part, by the recording system 310. Specifically, the recording system can be responsible for identifying the survey data so that a correlation can be made between any recording associated with the customer and the survey data obtained from the customer.


Once the survey data has been obtained, the quality monitoring system can be utilized for analysis and/or review. By way of example, processing algorithms can be used to correlate the survey data with other information, such as performance data pertaining to the contact center agent associated with the customer that responded to the survey. With respect to reviewing of the data, the quality monitoring system can enable a supervisor to access the stored survey data and compare that data to results of agent scoring, for example, which could have been previously provided to the data repository.



FIG. 7 is a flowchart illustrating functionality associated with the embodiment of FIG. 6. Specifically, the functionality depicted in FIG. 7 relates to an implementation in which a contact is manually transferred to a survey.


In this regard, the functionality may be construed as beginning at block 350, where a determination is made as to whether agent interaction with a customer is nearing completion or is complete. By way of example, such a determination can be based on application driven events that are known to occur near the completion of an interaction. In block 352, a determination is made as to whether the customer is to be provided with a survey. In some embodiments, an agent can explicitly ask a customer regarding willingness to participate in a survey. In other embodiments, an agent may check the status of a customer's response to a survey opt-in query or programmatically be notified that the customer is to participate in a survey. That is, if a customer has been previously queried regarding willingness to participate in a survey, that information can be provided to the agent. By way of example, a pop-up or other notification can be provided to the agent notifying that agent that the customer is to participate in a survey. In this embodiment, once a determination is made that the customer is to participate in a survey, the agent manually transfers the contact to the survey, such as depicted in block 354.



FIG. 8 is another flowchart illustrating functionality associated with the embodiment of FIG. 6. Specifically, the functionality depicted in FIG. 8 is associated with enabling survey data to be correlated with other information. In this regard, the functionality may be construed as beginning at block 360, where an indication that a customer is being transferred to a survey is received. In block 362, information provided by the customer is enabled to be correlated with other data. In some embodiments, this can be accomplished by a transfer process attaching an identifier to the transfer so that survey data provided by the customer can by synched up with results (such as future survey data, for example) that are later provided to the quality monitoring system.


After the appropriate steps have been taken to ensure that the obtained data can be correlated, the customer is transferred to the survey (block 364). In block 366, the survey is executed with the customer and, thereafter (block 368), survey data is obtained.



FIG. 9 is another flowchart illustrating functionality associated with the embodiment of FIG. 6. Specifically, the functionality depicted in FIG. 9 is associated with use of a customer opt-in procedure during which a customer is queried as to their desire to participate in a survey. In this regard, the functionality may be construed as beginning at block 370, where a contact is designated for an agent queue. A determination is then made, such as depicted in block 372, as to whether a survey is to be offered to the customer associated with the contact. In some embodiments, this determination can be facilitated by referring to business rules, such as those provided by a business rules engine, and/or historical contact data pertaining to the customer and/or agent. Such data could include historical agent performance, customer satisfaction information and level/status of customer. This information may be gathered from quality monitoring and/or customer relationship management systems.


If it is determined in block 372 that a survey is not to be offered, the contact can be placed in the agent queue for routing to an available agent, such as depicted in block 374. However, if it is determined that a survey is to be offered, the process may proceed to block 376, in which a determination can be made as to whether the customer accepted the survey offer. If the customer declined the survey offer, the process can proceed as described before with respect to block 374. However, if the customer accepts the survey offer, the process can proceed to block 378.


In block 378, the contact can be designated for a survey. In some embodiments, this can include updating information contained with or associated to the contact via CTI or other contact routing entity so that the contact is routed to the survey following agent interaction. Notably, transfer of the contact can take place either manually, i.e., through agent interaction, or automatically, whereby the contact would be transferred to the survey once agent interaction has completed. In other embodiments, the contact could be placed in a specialized queue to facilitate survey participation. Such would likely be involved in embodiments that provide surveys via a different communication medium than the one used to facilitate the interaction. For instance, if an interaction takes place via voice communication and the survey is to be provided via a subsequent email, information pertaining to the contact could be placed in an email survey queue.



FIG. 10 is another flowchart illustrating functionality associated with the embodiment of FIG. 6. Specifically, the functionality depicted in FIG. 10 is associated with use of an automatic transfer of a contact to a survey. In this regard, the functionality may be construed as beginning at block 390, where it is determined that interaction of an agent with a customer has completed. By way of example, a determination can be made as to whether the agent has hung up the extension associated with the contact while the customer is still held on the line. In block 392, a determination is made as to whether the customer accepted to participate in a survey. As mentioned before, this determination can be facilitated by accessing information associated with a previously provided opt-in query. Irrespective of the particular manner in which such a determination is made, if the customer has accepted participation in a survey, the process can proceed to block 394, where the customer is transferred to the survey. In block 396, the survey is executed and survey data is then obtained as depicted in block 398. Typically, the survey data is stored in a data repository. As depicted in block 400, once the survey is completed, or alternatively, if the customer declined to participate in the survey, the contact is terminated.


In addition to providing surveys, integrated workforce optimization platforms can also integrate other capabilities in support of a greater customer service strategy: (1) Quality Monitoring/Contact Recording—voice of the customer; the complete customer experience across multimedia touch points; (2) Workforce Management—strategic forecasting and scheduling that drives efficiency and adherence, aids in planning, and helps facilitate optimum staffing and service levels; (3) Performance Management—key performance indicators (KPIs) and scorecards that analyze and help identify synergies, opportunities and improvement areas; (4) e-Learning—training, new information and protocol disseminated to staff, leveraging best practice customer interactions and delivering learning to support development; and/or (5) Analytics—deliver insights from customer interactions to drive business performance. These five segments can become part of an interwoven and interoperable solution, enabling contact centers to transition from reactive cost centers to proactive, information-rich departments that deliver strategic value to the organization. Workforce optimization is discussed in greater detail in the U.S. Patent Application entitled “Systems and Methods for Workforce Optimization,” filed Feb. 22, 2006, and assigned Ser. No. 11/359,356, which is entirely incorporated herein by reference.


It should be noted that the flowcharts included herein show the architecture, functionality and/or operation of implementations that may be configured using software. In this regard, each block can be interpreted to represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of the order. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.


It should be noted that any of the executable instructions, such as those depicted functionally in the accompanying flowcharts, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. In the context of this document, a “computer-readable medium” can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. The computer readable medium can be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device. More specific examples (a nonexhaustive list) of the computer-readable medium could include an electrical connection (electronic) having one or more wires, a portable computer diskette (magnetic), a random access memory (RAM) (electronic), a read-only memory (ROM) (electronic), an erasable programmable read-only memory (EPROM or Flash memory) (electronic), an optical fiber (optical), and a portable compact disc read-only memory (CDROM) (optical). In addition, the scope of the certain embodiments of this disclosure can include embodying the functionality described in logic embodied in hardware or software-configured mediums.


It should be emphasized that the above-described embodiments are merely possible examples of implementations, merely set forth for a clear understanding of the principles of this disclosure. Many variations and modifications may be made to the above-described embodiments without departing substantially from the spirit and principles of the disclosure. By way of example, some embodiments may not include a pre-contact opt-in mechanism. In such an embodiment, an agent may determine or be instructed to offer a survey to a customer. The instruction to offer a survey may be an operational procedure where every agent is to offer a survey on every contact, for example, or an automated one in which the system decides that a survey is to be provided. For example, such a system may dynamically decide to offer a survey based on the fact that the contact dealt with payment issues (potentially determined by application activity) and that agent has shown poor handling of payment issues based on historical quality data. All such modifications and variations are intended to be included herein within the scope of this disclosure.

Claims
  • 1. A method for integrating contact center surveys, comprising: obtaining performance data pertaining to a contact center agent;providing a survey to a customer of the contact center agent;receiving survey data responsive to the survey and corresponding to interaction with the customer by the contact center agent;correlating the performance data with the survey data; anddesignating a detected anomaly if it is determined that the performance data does not correspond to the survey data.
  • 2. The method of claim 1, further comprising: providing at least one actionable report for the contact center based on at least the survey data or the correlation of the performance data and the survey data.
  • 3. The method of claim 1, further comprising: querying the customer regarding participation in a survey; andresponsive to a positive indication from the customer regarding participation, enabling the survey to be provided.
  • 4. The method of claim 1, further comprising: determining whether the survey is to be provided to the customer such that, if the survey is to be provided, communications between the customer and the contact center agent are automatically designated to be recorded.
  • 5. The method of claim 4, wherein the recording of said interaction between the customer and the contact center agent is comprised of at least one of desktop events of the contact center agent and voice dialog between the customer and the contact center agent.
  • 6. The method of claim 1, further comprising: determining whether the performance data corresponds to the survey data.
  • 7. The method of claim 6, further comprising: scheduling the contact center agent for training, the training being selected based, at least in part, on the survey data.
  • 8. The method of claim 1, wherein, in designating the detected anomaly, information corresponding to the anomaly is displayed to monitoring personnel.
  • 9. The method of claim 1, wherein the performance data comprises scoring, provided by monitoring personnel associated with the contact center agent, the scoring corresponding to performance characteristics of the contact center agent.
  • 10. The method of claim 1, wherein the survey data comprises responses of the customer pertaining to interaction of the customer and the contact center agent.
  • 11. The method of claim 1, wherein providing the survey comprises providing an interactive voice response survey.
  • 12. The method of claim 1, wherein providing the survey comprises providing a web-based survey.
  • 13. The method of claim 1, further comprising: determining whether to provide the survey to the customer prior to permitting communication between the contact center agent and the customer.
  • 14. The method of claim 1, wherein the contact center agent is not informed that the survey is being provided to the customer.
  • 15. A computer-readable medium having a computer program stored thereon, the computer program comprising computer-executable instructions for performing the computer-implemented steps of: obtaining performance data pertaining to a contact center agent;providing a survey to a customer of the contact center agent;receiving survey data responsive to the survey and corresponding to interaction with the customer by the contact center agent;correlating the performance data with the survey data; anddesignating a detected anomaly if it is determined that the performance data does not correspond to the survey data.
  • 16. A system for integrating contact center surveys comprising: a quality monitoring system operative to obtain performance data pertaining to a contact center agent;a data repository configured to store the performance data obtained by the quality monitoring system; anda survey system operative to provide a survey to a customer of the contact center agent, to receive survey data responsive to the survey and corresponding to interaction with the customer by the contact center agent, and to provide the survey data to the data repository;wherein the quality monitoring system is further operative to access the data repository, correlate the performance data with the survey data, and designate a detected anomaly if it is determined that the performance data does not correspond to the survey data.
  • 17. The system of claim 16, wherein the quality monitoring system is operative to analyze the survey data to identify when customer satisfaction drops below a predefined threshold.
  • 18. The system of claim 16, further comprising: means for recording communications between the customer and the contact center agent.
  • 19. The system of claim 16, further comprising: a recording system operative to record communications between the customer and the contact center agent.
  • 20. The system of claim 16, wherein the survey system is operative to determine whether to provide the survey to the customer prior to the customer communicating with the contact center agent.
  • 21. The system of claim 16, wherein the survey system is operative to indicate to the contact center agent that the customer is to be provided with the survey.
  • 22. The system of claim 16, wherein the survey system is operative to provide the survey as an interactive voice response survey.
  • 23. The system of claim 16, further comprising: a computer telephony integration system operative to facilitate communication between the customer with the contact center agent.
US Referenced Citations (162)
Number Name Date Kind
3594919 De Bell et al. Jul 1971 A
3705271 De Bell et al. Dec 1972 A
4510351 Costello et al. Apr 1985 A
4684349 Ferguson et al. Aug 1987 A
4694483 Cheung Sep 1987 A
4763353 Canale et al. Aug 1988 A
4815120 Kosich Mar 1989 A
4924488 Kosich May 1990 A
4953159 Hayden et al. Aug 1990 A
5016272 Stubbs et al. May 1991 A
5101402 Chiu et al. Mar 1992 A
5117225 Wang May 1992 A
5210789 Jeffus et al. May 1993 A
5239460 LaRoche Aug 1993 A
5241625 Epard et al. Aug 1993 A
5267865 Lee et al. Dec 1993 A
5299260 Shaio Mar 1994 A
5311422 Loftin et al. May 1994 A
5315711 Barone et al. May 1994 A
5317628 Misholi et al. May 1994 A
5347306 Nitta Sep 1994 A
5388252 Dreste et al. Feb 1995 A
5396371 Henits et al. Mar 1995 A
5432715 Shigematsu et al. Jul 1995 A
5465286 Clare et al. Nov 1995 A
5475625 Glaschick Dec 1995 A
5485569 Goldman et al. Jan 1996 A
5491780 Fyles et al. Feb 1996 A
5499291 Kepley Mar 1996 A
5535256 Maloney et al. Jul 1996 A
5572652 Robusto et al. Nov 1996 A
5577112 Cambray et al. Nov 1996 A
5590171 Howe et al. Dec 1996 A
5597312 Bloom et al. Jan 1997 A
5619183 Ziegra et al. Apr 1997 A
5696906 Peters et al. Dec 1997 A
5717879 Moran et al. Feb 1998 A
5721842 Beasley et al. Feb 1998 A
5742670 Bennett Apr 1998 A
5748499 Trueblood May 1998 A
5778182 Cathey et al. Jul 1998 A
5784452 Carney Jul 1998 A
5790798 Beckett, II et al. Aug 1998 A
5796952 Davis et al. Aug 1998 A
5809247 Richardson et al. Sep 1998 A
5809250 Kisor Sep 1998 A
5825869 Brooks et al. Oct 1998 A
5835572 Richardson, Jr. et al. Nov 1998 A
5862330 Anupam et al. Jan 1999 A
5864772 Alvarado et al. Jan 1999 A
5884032 Bateman et al. Mar 1999 A
5907680 Nielsen May 1999 A
5918214 Perkowski Jun 1999 A
5923746 Baker et al. Jul 1999 A
5933811 Angles et al. Aug 1999 A
5944791 Scherpbier Aug 1999 A
5948061 Merriman et al. Sep 1999 A
5958016 Chang et al. Sep 1999 A
5964836 Rowe et al. Oct 1999 A
5978648 George et al. Nov 1999 A
5982857 Brady Nov 1999 A
5987466 Greer et al. Nov 1999 A
5990852 Szamrej Nov 1999 A
5991373 Pattison et al. Nov 1999 A
5991796 Anupam et al. Nov 1999 A
6005932 Bloom Dec 1999 A
6009429 Greer et al. Dec 1999 A
6014134 Bell et al. Jan 2000 A
6014647 Nizzari et al. Jan 2000 A
6018619 Allard et al. Jan 2000 A
6035332 Ingrassia et al. Mar 2000 A
6038544 Machin et al. Mar 2000 A
6039575 L'Allier et al. Mar 2000 A
6057841 Thurlow et al. May 2000 A
6058163 Pattison et al. May 2000 A
6061798 Coley et al. May 2000 A
6072860 Kek et al. Jun 2000 A
6076099 Chen et al. Jun 2000 A
6078894 Clawson et al. Jun 2000 A
6091712 Pope et al. Jul 2000 A
6108711 Beck et al. Aug 2000 A
6122665 Bar et al. Sep 2000 A
6122668 Teng et al. Sep 2000 A
6130668 Stein Oct 2000 A
6138139 Beck et al. Oct 2000 A
6144991 England Nov 2000 A
6146148 Stuppy Nov 2000 A
6151622 Fraenkel et al. Nov 2000 A
6154771 Rangan et al. Nov 2000 A
6157808 Hollingsworth Dec 2000 A
6171109 Ohsuga Jan 2001 B1
6182094 Humpleman et al. Jan 2001 B1
6195679 Bauersfeld et al. Feb 2001 B1
6201948 Cook et al. Mar 2001 B1
6211451 Tohgi et al. Apr 2001 B1
6225993 Lindblad et al. May 2001 B1
6230197 Beck et al. May 2001 B1
6236977 Verba et al. May 2001 B1
6244758 Solymar et al. Jun 2001 B1
6282548 Burner et al. Aug 2001 B1
6286030 Wenig et al. Sep 2001 B1
6286046 Bryant Sep 2001 B1
6288753 DeNicola et al. Sep 2001 B1
6289340 Purnam et al. Sep 2001 B1
6301462 Freeman et al. Oct 2001 B1
6301573 McIlwaine et al. Oct 2001 B1
6324282 McIlwaine et al. Nov 2001 B1
6347374 Drake et al. Feb 2002 B1
6351467 Dillon Feb 2002 B1
6353851 Anupam et al. Mar 2002 B1
6360250 Anupam et al. Mar 2002 B1
6370574 House et al. Apr 2002 B1
6404857 Blair et al. Jun 2002 B1
6411989 Anupam et al. Jun 2002 B1
6418471 Shelton et al. Jul 2002 B1
6459787 McIlwaine et al. Oct 2002 B2
6487195 Choung et al. Nov 2002 B1
6493758 McLain Dec 2002 B1
6502131 Vaid et al. Dec 2002 B1
6510220 Beckett, II et al. Jan 2003 B1
6535909 Rust Mar 2003 B1
6542602 Elazar Apr 2003 B1
6546405 Gupta et al. Apr 2003 B2
6560328 Bondarenko et al. May 2003 B1
6583806 Ludwig et al. Jun 2003 B2
6606657 Zilberstein et al. Aug 2003 B1
6665644 Kanevsky et al. Dec 2003 B1
6674447 Chiang et al. Jan 2004 B1
6683633 Holtzblatt et al. Jan 2004 B2
6697858 Ezerzer et al. Feb 2004 B1
6724887 Eilbacher et al. Apr 2004 B1
6738456 Wrona et al. May 2004 B2
6757361 Blair et al. Jun 2004 B2
6772396 Cronin et al. Aug 2004 B1
6775377 McIlwaine et al. Aug 2004 B2
6792575 Samaniego et al. Sep 2004 B1
6810414 Brittain Oct 2004 B1
6820083 Nagy et al. Nov 2004 B1
6823384 Wilson et al. Nov 2004 B1
6870916 Henrikson et al. Mar 2005 B2
6901438 Davis et al. May 2005 B1
6959078 Eilbacher et al. Oct 2005 B1
6965886 Govrin et al. Nov 2005 B2
20010000962 Rajan May 2001 A1
20010032335 Jones Oct 2001 A1
20010043697 Cox et al. Nov 2001 A1
20020038363 MacLean Mar 2002 A1
20020052948 Baudu et al. May 2002 A1
20020065911 Von Klopp et al. May 2002 A1
20020065912 Catchpole et al. May 2002 A1
20020128925 Angeles Sep 2002 A1
20020143925 Pricer et al. Oct 2002 A1
20020165954 Eshghi et al. Nov 2002 A1
20030055883 Wiles et al. Mar 2003 A1
20030079020 Gourraud et al. Apr 2003 A1
20030144900 Whitmer Jul 2003 A1
20030154240 Nygren et al. Aug 2003 A1
20040100507 Hayner et al. May 2004 A1
20040165717 McIlwaine et al. Aug 2004 A1
20050131732 Potenza Jun 2005 A1
20050138560 Lee et al. Jun 2005 A1
20060265090 Conway et al. Nov 2006 A1
Foreign Referenced Citations (6)
Number Date Country
0453128 Oct 1991 EP
0773687 May 1997 EP
0989720 Mar 2000 EP
2369263 May 2002 GB
WO 9843380 Nov 1998 WO
WO 0016207 Mar 2000 WO
Related Publications (1)
Number Date Country
20070233549 A1 Oct 2007 US