Method and apparatus for automatically taking action based on the content of call center communications

Information

  • Patent Grant
  • 11736616
  • Patent Number
    11,736,616
  • Date Filed
    Friday, May 27, 2022
    2 years ago
  • Date Issued
    Tuesday, August 22, 2023
    10 months ago
Abstract
A method and system for automatically executing an action within a call center environment. Data is aggregated from multiple data sources into a combined data stream. One of the data sources is a source of data corresponding to at least one communication processed by the call center one of the data sources can be a source of external data representing activity that is external to the call center. The combined data stream is processed into successive batches of data corresponding to one or more communications between a call center agent and a communicating party received by the call center. A sensor data structure specifying at least one rule is applied to the batches of data. The at least one rule can include a machine learning model and a configuration data structure based on historical data from the multiple data sources. When it is determined that at least one of the batches satisfies the at least one rule, a notification message relating to the one or more communications is generated. The call center executes an action specified by the sensor data structure based on the notification message. The action can address a situation corresponding the at least one communication. The actions can be various actions such as notifying a specified party, generating an API call, or the like.
Description
BACKGROUND

Contact centers, also referred to as “call centers”, in which agents are assigned to queues based on skills and customer requirements are well known. FIG. 1 is an example system architecture 100, of a cloud-based contact center system 100. Customers 110 interact with a contact center 150 using, for example, voice, email, text, and web interfaces to communicate with the agents 120 through a network 130 and at least one or more of text, voice or multimedia channels. The platform that controls the operation of the contact center 150 including the routing and handling of communications between customers 110 and agents 120 for the contact center 150 is referred herein as the contact routing system 153. The contact routing system 153 could be any of a contact center as a service (CCaS) system, an automated call distributor (ACD) system, or a case system, for example.


The agents 120 may be remote from the contact center 150 and handle communications (also referred to as “interactions” or “calls” herein) with customers 110 on behalf of an enterprise. The agents 120 may utilize devices, such as but not limited to, workstations, desktop computers, laptops, telephones, a mobile smartphone and/or a tablet. Similarly, customers 110 may communicate using a plurality of devices, including but not limited to, a telephone, a mobile smartphone, a tablet, a laptop, a desktop computer, or other. For example, telephone communication may traverse networks such as a public switched telephone networks (PSTN), Voice over Internet Protocol (VoIP) telephony (via the Internet), a Wide Area Network (WAN) or a Large Area Network (LAN). The network types are provided by way of example and are not intended to limit types of networks used for communications.


The agents 120 may be assigned to one or more queues representing call categories and/or agent skill levels. The agents 120 assigned to a queue may handle communications that are placed in the queue by the contact routing system 153. For example, there may be queues associated with a language (e.g., English or Chinese), topic (e.g., technical support or billing), or a particular country of origin. When a communication is received by the contact routing system 153, the communication may be placed in a relevant queue, and one of the agents 120 associated with the relevant queue may handle the communication.


Agents may be assigned to one or more entities using the cloud-based contact center. Therefore, it is possible that agents, on any given day or shift, are providing support/service for customers of various entities. For example, an agent may handle a communication from a customer of a computer supplier and then immediately thereafter handle a communication from a customer of an automobile company. Accordingly, agents might not be trained in all aspects of customer service for each entity. The term “customer”, as used herein, refers to the party contacting the call center for support or other information and includes actual customers, potential customers, or any other party contacting the call center. Further, agents may be employees of the call center provider, employees of the entity using the call center service, contractors, or freelancers. Therefore, in order to provide a high level of service, it is important to be able to take various actions in real time, such as provide the agent with information or notify a supervisor, based on the content and sentiment of a communication from a customer. Known call center systems are not able to reliably detect the need for taking actions in an automated manner in real-time.


SUMMARY OF THE INVENTION

The disclosed implementations can leverage combined data streams relating to communications, historical data, Natural Language Processing (NLP), and future looking Machine Learning (ML) models to predict possible events and/or automatically take actions in a call center. A first aspect of the invention is a method for executing actions related to communications received in a call center, the method comprising: aggregating data from multiple data sources into a combined data stream, at least a first data source of the multiple data sources being a source of data corresponding to at least one communication processed by the call center and at least a second data source of the multiple data sources being a source of external data, wherein the external data is data representing activity that is external to the call center; processing the combined data stream into successive batches of data corresponding to one or more communications between a call center agent and a communicating party received by the call center; applying a sensor data structure defining at least one rule to the batches of data, wherein the at least one rule includes a machine learning model and a configuration data structure based on historical data from the multiple data sources; determining that at least one of the batches satisfies the at least one rule and generating a notification message relating to the one or more communications in response to the determining; and the call center executing an action based on the notification message, wherein the action addresses a situation corresponding the at least one communication.





BRIEF DESCRIPTION OF THE DRAWINGS

The foregoing summary, as well as the following detailed description of the invention, will be better understood when read in conjunction with the appended drawings. For the purpose of illustrating the invention, there are shown in the appended drawings various illustrative embodiments. It should be understood, however, that the invention is not limited to the precise arrangements and instrumentalities shown. In the drawings:



FIG. 1 is an architectural diagram of a conventional cloud based contact center computing environment.



FIG. 2 is a diagram of pipeline data processing in accordance with disclosed implementations.



FIG. 3 is a diagram of structured streaming processing modes in accordance with disclosed implementations.



FIG. 4 is a diagram illustrating a stream join use case in accordance with disclosed implementations.



FIG. 5 is a block diagram of a process for creating an action in accordance with disclosed implementations.



FIG. 6 illustrates a more detailed computing architecture for providing automated actions within a call center in accordance with disclosed implementations.



FIGS. 7A-7E illustrate a user interface in accordance with disclosed implementations.



FIG. 8 illustrates method 800 of automatically executing an action within a call center environment in accordance with the disclosed implementations.





DETAILED DESCRIPTION


FIG. 2 illustrates a data pipeline 200 of a disclosed implementation. The pipeline 200 of this example processes two types of datasets: utterance data 202 and “sensor” configuration data 204. Utterance data 202 is derived from communications between agents and customers, as described in more detail below. Sensor configuration data represents conditions under which events (also referred to as “hits” herein) should be generated, also as described in greater detail below. Pipeline 200 can use a Machine Learning model, from the Spark NLP library for example, to compute a sentiment for communications. The sentiment can be included in the utterance data. Additionally, the pipeline can generate timestamps for the generation of hit and notification events. The term “sensor” as used herein includes data structures and processing models that can be applied to data streams and which set forth rules for generating an event.


Sensor configuration data 202 includes conditions under which an event is to be generated, such as key words, sentiment, frequency of key words, and the like. Sensor configuration data can also include an account identification or other identification data of the customer, entity, and/or call center. Sensor configuration data can also include a sensor name, applicable time ranges, and a time stamp. An example schema for sensor data is set forth below.


SensorConfigurations(

    • account_id: String,
    • active: Boolean,
    • configured_hits: Integer,
    • filter: (
    • keywords: Array[String],
    • sentiment: String


),

    • frequency: String,
    • sensor_id: String,
    • sensor_name: String,
    • time_range: String
    • timestamp: Timestamp,


)


Utterance data 204 is collected and derived from the content of communications between customers and agents. For example, a transcription of the communication (a phone call for example) can be created by recording the communication and using known speech to text processing. The text can then be parsed to identify keywords and/or keyphrases (referred to collectively herein as “keywords”). The utterance data can also include sentiment data of a communication and other identifying metadata, as described further below. An example schema for sensor data is set forth below.


Utterance(

    • account_id: String,
    • agent_id: String,
    • channel_id: String,
    • id: String,
    • interaction_id: String,
    • text: String,
    • timestamp: Timestamp,
    • ‘type’: String,
    • timestamp: Timestamp


)


Sensor configuration data 204 is delivered into pipeline 200 as a stream and can be stored in a Delta Table, such as a Databricks delta table created in Delta Lake. A delta table maintains an entry for each “sensor” describing the state of the sensor, including whether it is active or inactive. Pipeline 200 continuously receives utterance information, such as transcriptions of communications, as data streams, as they are generated, by a speech-to-text service for example. Transcriptions can be processed in Spark Structured Streaming micro batches. In each batch, the sensor configuration table retrieves the configurations for active sensors. Active configurations are then joined with the transcription stream at 206 in FIG. 2.


Pipeline 200 matches a transcription and a configuration, at 208, when the transcription data contains at least one word specified as a key word in the sensor configuration. Additional filters can also be defined in the sensor. For example the filters can include:

    • a sentiment filter which selects only transcriptions with the sentiment—computed as described in the added sentiment metadata—specified in the sensor configuration; and/or
    • a speaker filter which selects only transcriptions of utterances from the speaker specified in the sensor configuration


This matching process yields a stream of transcription-configuration matches or hits at 210. These hits can be published to a Kafka topic, enabling a hit count of the sensor. The pipeline also keeps, for each active configuration, a count of the hits that fall within a user-defined time window, at 212, to trigger a notification event when hits fall within a user-defined maximum frequency. This count can be kept using a Spark state store functionality on top of a RocksDB, using the flatMapGroups. Pipeline 200 can publish a notification event to Kafka when the hit count for a sensor is greater or equal to the value set in the sensor's configuration within the intended time window (time_range), when the maximum frequency (frequency) is not violated, enabling a notification to the user or any other event/signal to be generated in order to take a desired action.


The pipeline outputs, hit events and notification events can be published to Kafka in json format with the following data structures:


SensorHits

    • account_id: String,
    • interaction_id: String,
    • sensor_id: String,
    • sensor_name: String,
    • speaker: String,
    • utterance_id: String,
    • utterance_started: Timestamp,
    • timestamp: Timestamp


)


SensorNotification

    • account_id: String,
    • sensor_id: String,
    • sensor_name: String,
    • time_range: String,
    • count: Integer,
    • timestamp: Timestamp


)


The disclosed implementations can leverage Spark Structured Streaming, the Apache Spark API that allows expression of computation on streaming data in the same way and in batch computation on static data. The data is treated as never ending tables on which queries or other processing can be performed. Queries and other processes can be performed on the tables continuously on new data as it arrives. Stateful transformations are also possible. This allows an SQL Engine to operate on data streams with a high throughput, high fault-tolerance, and high scalability. Each new record in a data stream can be stored as a new row in the corresponding table.



FIG. 3 illustrates micro-batch and continuous processing modes of a structured streaming architecture. Architecture 300 include input streams 302 and input tables 304.



FIG. 4 illustrates an example 400 of joining data streams. In this example, data streams 402 and 404 are joined. As an example, data stream 402 could represent keywords and data stream 404 could represent sentiment. Each data stream is buffered to handle late/delayed data because corresponding events in data streams 402 and 404 could arrive out of order with arbitrary delays between them. Buffer size can be managed by dropping delayed data beyond a certain threshold. A joint time range condition is used to limit the time range of other events that each event can join against.



FIG. 5 illustrates a more specific example of a process 500 for generating an action within a call center in accordance with disclosed implementations. At 502, disparate data streams from within and outside of the contact center are combined. It this example, the data streams include: voice, text (e.g., chat/sms/chatbot), arbitrary events, web events, IOT events, and stored historical data. At 504, a sensor is applied to the combined data streams. The sensor can be configured to define conditions to make a decision based on any of the metrics calculated, intents detected, keywords matched, or conditions recognized in the data streams. Intent data can be produced using known intent engines applied to the utterance data.


Predictive models can be used to predict a trigger condition based on past data. For example, if a customer was browsing kidney disease on a website for re predetermined time or number of visits over a period of time, has an overdue kidney prescription, and just called with intent matched “insurance bill”, it can be predicted that there will be a sudden increase in medical payments for the customer (the “supervised event”). Once sensor conditions are triggered, then an action can be taken at 506. For example, the action can include a notification to an appropriate person or a call to a specified API. As an example, the API call could cause the customer to be adder to a “call immediately” list.



FIG. 6 illustrates and example of the overall architecture 600 of a system for automatically generating actions in a call center environment.



FIG. 7A illustrates a user interface for configuring and managing sensors. The various sensors, defined as the data structures described above, are displayed by name at column 702. Column 704 shows the total number of hits corresponding to the sensor. Column 706 shows the channels which the sensor is active. Column 708 shows a category of the sensor which can be used to organize and manage sensors. Column 710 indicates whether notifications for the sensor are currently activated (on) or not activated (off). Column 714 provides a selection tool for editing the corresponding sensor. FIG. 7B shows the popup user interface 711 of the UI when Notifications are selected. As shown, notifications can be turned on or off and persons, groups of persons, or APIs can be designated for receiving notifications.



FIG. 7C illustrates the Create Sensor popup user interface 720. As illustrated in FIG. 7C, popup user interface 720 allows a user to enter a sensor name, a sensor category, a sentiment label, channels for which the sensor applies, speakers for which the sensor applies, ring groups for which the sensor applies, and agents for which the sensor applies. All of this data can be stored in the sensor data structure described above. FIG. 7D illustrates a Checkout popup user interface 722 which allows additional data to be specified and stored in the sensor data structure. User interface 722 allows entry of the number of hits in the sensor that will trigger a notification, the time range for the number of hits specified, time frequency of notifications for the sensor, a notification manner/channel (e.g., through a notification center of the call center, through email, etc.), available integrations, and data triggers specifying data outputs, such as email reports, API, calls and other data that results in desired actions. After entering all data/parameters, a selection of the Create button will cause the sensor to be created. FIG. 7E shows sensor notification report popup 724 which allows a user to view and manage sensor notifications.



FIG. 8 illustrates method 800 of automatically executing an action within a call center environment in accordance with the disclosed implementations. Method 800 can be accomplished by the systems described above. At 802, data is aggregated from multiple data sources into a combined data stream. One of the data sources is a source of data corresponding to at least one communication processed by the call center one of the data sources can be a source of external data representing activity that is external to the call center. At 804, the combined data stream is processed into successive batches of data corresponding to one or more communications between a call center agent and a communicating party received by the call center. At 806, a sensor data structure specifying at least one rule is applied to the batches of data. The at least one rule can include a machine learning model and a configuration data structure based on historical data from the multiple data sources. At 808 it is determined that at least one of the batches satisfies the at least one rule and a notification message relating to the one or more communications is generated. At 810, the call center executes an action specified by the sensor data structure based on the notification message. The action can address a situation corresponding the at least one communication. As noted above, the actions can be various actions such as notifying a specified party, generating an API call, or the like.


It will be appreciated by those skilled in the art that changes could be made to the embodiments described above without departing from the broad inventive concept thereof. It is understood, therefore, that this invention is not limited to the particular implementations disclosed, but it is intended to cover modifications within the spirit and scope of the present invention as defined by the appended claims.

Claims
  • 1. A method for executing actions related to communications received in a call center, the method comprising: aggregating data from multiple data sources into a combined data stream, at least a first data source of the multiple data sources being a source of data corresponding to at least one communication processed by the call center and at least a second data source of the multiple data sources being a source of external data, wherein the external data is data representing activity that is external to the call center;processing the combined data stream into successive batches of data corresponding to one or more communications between a call center agent and a communicating party received by the call center;applying a sensor data structure defining at least one rule to the batches of data, wherein the at least one rule includes a machine learning model and a configuration data structure based on historical data from the multiple data sources;determining that at least one of the batches satisfies the at least one rule and generating a notification message relating to the one or more communications in response to the determining; andthe call center executing an action based on the notification message, wherein the action addresses a situation corresponding the at least one communication.
  • 2. The method of claim 1, wherein the configuration data structure is updated multiple times during the communication and the configuration data structure is a delta table.
  • 3. The method of claim 2, wherein the configuration data structure includes a keyword, a frequency designation, a time range, and a flag threshold value indicator for each of the successive batches of data.
  • 4. The method of claim 2, wherein applying at least one rule includes: comparing the batch of utterance data with the keyword and frequency designation; andgenerating a flag if the batch of utterance data satisfies criteria, the criteria including the batch of data containing the keyword at a frequency specified by the frequency designation.
  • 5. The method of claim 4, wherein the configuration data structure further includes a speaker indicator and the criteria further includes that the keyword is uttered by a speaker specified by the speaker indicator.
  • 6. The method of claim 1, wherein the action based on the notification message is at least one of: sending a notification to an agent;generating an API call;sending a notification to an agent supervisor;adding a party to an email campaign; and/orsending data to an external system.
  • 7. The method of claim 1, wherein the first data source includes a source of voice and/or text data relating to communications processed by the call center.
  • 8. The method of claim 1, wherein the second data source includes at least one of: a source of events occurring external to the call center;a web server; and/orat least one loT device.
US Referenced Citations (622)
Number Name Date Kind
5862203 Wulkan et al. Jan 1999 A
5897616 Kanevsky et al. Apr 1999 A
5966691 Kibre et al. Oct 1999 A
5970124 Csaszar et al. Oct 1999 A
6100891 Thorne Aug 2000 A
6128415 Hultgren et al. Oct 2000 A
6163607 Bogart et al. Dec 2000 A
6230197 Beck et al. May 2001 B1
6263057 Silverman Jul 2001 B1
6345093 Lee et al. Feb 2002 B1
6385584 McAlister et al. May 2002 B1
6411687 Bohacek et al. Jun 2002 B1
6493695 Pickering et al. Dec 2002 B1
6560222 Pounds et al. May 2003 B1
6587831 O'Brien Jul 2003 B1
6639982 Stuart et al. Oct 2003 B1
6721416 Farrell Apr 2004 B1
6754333 Flockhart et al. Jun 2004 B1
6859776 Cohen et al. Feb 2005 B1
6970829 Leamon Nov 2005 B1
7023979 Wu et al. Apr 2006 B1
7076047 Brennan et al. Jul 2006 B1
7110525 Heller et al. Sep 2006 B1
7209475 Shaffer et al. Apr 2007 B1
7274787 Schoeneberger Sep 2007 B1
7292689 Odinak et al. Nov 2007 B2
7343406 Buonanno et al. Mar 2008 B1
7372952 Wu et al. May 2008 B1
7382773 Schoeneberger et al. Jun 2008 B2
7409336 Pak et al. Aug 2008 B2
7426268 Walker et al. Sep 2008 B2
7466334 Baba Dec 2008 B1
7537154 Ramachandran May 2009 B2
7634422 Andre et al. Dec 2009 B1
7657263 Chahrouri Feb 2010 B1
7672746 Hamilton et al. Mar 2010 B1
7672845 Beranek et al. Mar 2010 B2
7676034 Wu et al. Mar 2010 B1
7698163 Reed et al. Apr 2010 B2
7752159 Nelken et al. Jul 2010 B2
7774790 Jirman et al. Aug 2010 B1
7788286 Nourbakhsh et al. Aug 2010 B2
7853006 Fama et al. Dec 2010 B1
7864946 Fama et al. Jan 2011 B1
7869998 Di Fabbrizio et al. Jan 2011 B1
7949123 Flockhart et al. May 2011 B1
7953219 Freedman et al. May 2011 B2
7966369 Briere et al. Jun 2011 B1
8060394 Woodings et al. Nov 2011 B2
8073129 Kalavar Dec 2011 B1
8116446 Kalavar Feb 2012 B1
8135125 Sidhu et al. Mar 2012 B2
8160233 Keren et al. Apr 2012 B2
8184782 Vatland et al. May 2012 B1
8223951 Edelhaus et al. Jul 2012 B1
8229761 Backhaus et al. Jul 2012 B2
8243896 Rae Aug 2012 B1
8300798 Wu et al. Oct 2012 B1
8369338 Peng et al. Feb 2013 B1
8370155 Byrd et al. Feb 2013 B2
8391466 Noble, Jr. Mar 2013 B1
8447279 Peng et al. May 2013 B1
8488769 Noble et al. Jul 2013 B1
8526576 Deich et al. Sep 2013 B1
8535059 Noble, Jr. et al. Sep 2013 B1
8583466 Margulies et al. Nov 2013 B2
8594306 Laredo et al. Nov 2013 B2
8626137 Devitt et al. Jan 2014 B1
8635226 Chang et al. Jan 2014 B2
8644489 Noble et al. Feb 2014 B1
8671020 Morrison et al. Mar 2014 B1
8688557 Rose et al. Apr 2014 B2
8738739 Makar et al. May 2014 B2
8767948 Riahi et al. Jul 2014 B1
8811597 Hackbarth et al. Aug 2014 B1
8861691 De et al. Oct 2014 B1
8869245 Ranganathan et al. Oct 2014 B2
8898219 Ricci Nov 2014 B2
8898290 Siemsgluess Nov 2014 B2
8909693 Frissora et al. Dec 2014 B2
8935172 Noble, Jr. et al. Jan 2015 B1
8996509 Sundaram Mar 2015 B1
9020142 Kosiba et al. Apr 2015 B2
9026431 Moreno Mengibar et al. May 2015 B1
9060057 Danis Jun 2015 B1
9065915 Lillard et al. Jun 2015 B1
9082094 Etter et al. Jul 2015 B1
9100483 Snedden Aug 2015 B1
9117450 Cook et al. Aug 2015 B2
9123009 Etter et al. Sep 2015 B1
9137366 Medina et al. Sep 2015 B2
9152737 Micali et al. Oct 2015 B1
9160853 Daddi et al. Oct 2015 B1
9178999 Hegde et al. Nov 2015 B1
9185222 Govindarajan et al. Nov 2015 B1
9237232 Williams et al. Jan 2016 B1
9280754 Schwartz et al. Mar 2016 B1
9286413 Coates et al. Mar 2016 B1
9300801 Warford et al. Mar 2016 B1
9319524 Webster Apr 2016 B1
9386152 Riahi et al. Jul 2016 B2
9397985 Seger et al. Jul 2016 B1
9426291 Ouimette et al. Aug 2016 B1
9473637 Venkatapathy et al. Oct 2016 B1
9514463 Grigg et al. Dec 2016 B2
9595049 Showers et al. Mar 2017 B2
9609131 Placiakis et al. Mar 2017 B2
9674361 Ristock et al. Jun 2017 B2
9679265 Schwartz et al. Jun 2017 B1
9774731 Haltom et al. Sep 2017 B1
9787840 Neuer, III et al. Oct 2017 B1
9813495 Van et al. Nov 2017 B1
9823949 Ristock et al. Nov 2017 B2
9883037 Lewis et al. Jan 2018 B1
9894478 Deluca et al. Feb 2018 B1
9930181 Moran et al. Mar 2018 B1
9955021 Liu et al. Apr 2018 B1
RE46852 Petrovykh May 2018 E
9998596 Dunmire et al. Jun 2018 B1
10009465 Fang et al. Jun 2018 B1
10038788 Khalatian Jul 2018 B1
10044862 Cai et al. Aug 2018 B1
10079939 Bostick et al. Sep 2018 B1
10085073 Ray et al. Sep 2018 B2
10101974 Ristock et al. Oct 2018 B2
10115065 Fama et al. Oct 2018 B1
10135973 Algard et al. Nov 2018 B2
10154138 Te Booij et al. Dec 2018 B2
10194027 Daddi et al. Jan 2019 B1
10235999 Naughton et al. Mar 2019 B1
10241752 Lemay et al. Mar 2019 B2
10242019 Shan et al. Mar 2019 B1
10276170 Gruber et al. Apr 2019 B2
10277745 Araujo et al. Apr 2019 B1
10290017 Traasdahl et al. May 2019 B2
10331402 Spector et al. Jun 2019 B1
10380246 Clark et al. Aug 2019 B2
10440180 Jayapalan et al. Oct 2019 B1
10445742 Prendki et al. Oct 2019 B2
10460728 Anbazhagan et al. Oct 2019 B2
10497361 Rule et al. Dec 2019 B1
10554590 Cabrera-Cordon et al. Feb 2020 B2
10554817 Sullivan et al. Feb 2020 B1
10572879 Hunter et al. Feb 2020 B1
10574822 Sheshaiahgari et al. Feb 2020 B1
10601992 Dwyer et al. Mar 2020 B2
10623572 Copeland Apr 2020 B1
10635973 Dirac et al. Apr 2020 B1
10636425 Naughton et al. Apr 2020 B2
10699303 Ismail et al. Jun 2020 B2
10715648 Vashisht et al. Jul 2020 B1
10718031 Wu et al. Jul 2020 B1
10728384 Channakeshava et al. Jul 2020 B1
10735586 Johnston Aug 2020 B1
10742806 Kotak Aug 2020 B2
10750019 Petrovykh et al. Aug 2020 B1
10783568 Chandra et al. Sep 2020 B1
10789956 Dube Sep 2020 B1
10803865 Naughton et al. Oct 2020 B2
10812654 Wozniak Oct 2020 B2
10812655 Adibi et al. Oct 2020 B1
10827069 Paiva Nov 2020 B1
10827071 Adibi et al. Nov 2020 B1
10839432 Konig et al. Nov 2020 B1
10841425 Langley et al. Nov 2020 B1
10855844 Smith et al. Dec 2020 B1
10861031 Sullivan et al. Dec 2020 B2
10878479 Wu et al. Dec 2020 B2
10943589 Naughton et al. Mar 2021 B2
10970682 Aykin Apr 2021 B1
11017176 Ayers et al. May 2021 B2
11089158 Holland et al. Aug 2021 B1
20010008999 Bull Jul 2001 A1
20010024497 Campbell Sep 2001 A1
20010054072 Discolo et al. Dec 2001 A1
20020019737 Stuart et al. Feb 2002 A1
20020029272 Weller Mar 2002 A1
20020034304 Yang Mar 2002 A1
20020038420 Collins et al. Mar 2002 A1
20020067823 Walker et al. Jun 2002 A1
20020143599 Nourbakhsh et al. Oct 2002 A1
20020169664 Walker et al. Nov 2002 A1
20020174182 Wilkinson et al. Nov 2002 A1
20020181689 Rupe et al. Dec 2002 A1
20030007621 Graves et al. Jan 2003 A1
20030009520 Nourbakhsh et al. Jan 2003 A1
20030032409 Hutcheson et al. Feb 2003 A1
20030061068 Curtis Mar 2003 A1
20030112927 Brown et al. Jun 2003 A1
20030126136 Omoigui Jul 2003 A1
20030167167 Gong Sep 2003 A1
20040044585 Franco Mar 2004 A1
20040044664 Cash et al. Mar 2004 A1
20040062364 Dezonno et al. Apr 2004 A1
20040078257 Schweitzer et al. Apr 2004 A1
20040098274 Dezonno et al. May 2004 A1
20040103051 Reed et al. May 2004 A1
20040141508 Schoeneberger et al. Jul 2004 A1
20040162724 Hill et al. Aug 2004 A1
20040162753 Vogel et al. Aug 2004 A1
20040174980 Knott et al. Sep 2004 A1
20040215451 Macleod Oct 2004 A1
20050033957 Enokida Feb 2005 A1
20050043986 Mcconnell et al. Feb 2005 A1
20050063365 Mathew et al. Mar 2005 A1
20050071178 Beckstrom et al. Mar 2005 A1
20050105712 Williams et al. May 2005 A1
20050177368 Odinak et al. Aug 2005 A1
20050226220 Kilkki et al. Oct 2005 A1
20050228774 Ronnewinkel Oct 2005 A1
20050246511 Willman et al. Nov 2005 A1
20050271198 Chin et al. Dec 2005 A1
20060095575 Sureka et al. May 2006 A1
20060126818 Berger et al. Jun 2006 A1
20060153357 Acharya et al. Jul 2006 A1
20060166669 Claussen Jul 2006 A1
20060188086 Busey et al. Aug 2006 A1
20060209797 Anisimov et al. Sep 2006 A1
20060215831 Knott et al. Sep 2006 A1
20060229931 Fligler et al. Oct 2006 A1
20060256953 Pulaski et al. Nov 2006 A1
20060271361 Vora et al. Nov 2006 A1
20060274856 Dun et al. Dec 2006 A1
20060277108 Altberg et al. Dec 2006 A1
20070016565 Evans et al. Jan 2007 A1
20070036334 Culbertson et al. Feb 2007 A1
20070038499 Margulies et al. Feb 2007 A1
20070041519 Erhart et al. Feb 2007 A1
20070061183 Seetharaman et al. Mar 2007 A1
20070078725 Koszewski et al. Apr 2007 A1
20070121902 Stoica et al. May 2007 A1
20070121903 Moore, Jr. et al. May 2007 A1
20070136284 Cobb et al. Jun 2007 A1
20070155411 Morrison Jul 2007 A1
20070157021 Whitfield Jul 2007 A1
20070160188 Sharpe et al. Jul 2007 A1
20070162296 Altberg et al. Jul 2007 A1
20070198329 Lyerly et al. Aug 2007 A1
20070201636 Gilbert et al. Aug 2007 A1
20070211881 Parker-Stephen Sep 2007 A1
20070263810 Sterns Nov 2007 A1
20070265990 Sidhu et al. Nov 2007 A1
20070269031 Honig et al. Nov 2007 A1
20070280460 Harris et al. Dec 2007 A1
20070287430 Hosain et al. Dec 2007 A1
20080002823 Fama et al. Jan 2008 A1
20080043976 Maximo et al. Feb 2008 A1
20080065902 Spohrer et al. Mar 2008 A1
20080095355 Mahalaha et al. Apr 2008 A1
20080126957 Tysowski et al. May 2008 A1
20080205620 Odinak et al. Aug 2008 A1
20080225872 Collins et al. Sep 2008 A1
20080254774 Lee Oct 2008 A1
20080255944 Shah et al. Oct 2008 A1
20080260138 Chen et al. Oct 2008 A1
20080288770 Kline et al. Nov 2008 A1
20080300955 Hamilton et al. Dec 2008 A1
20090018996 Hunt et al. Jan 2009 A1
20090080411 Lyman Mar 2009 A1
20090086945 Buchanan et al. Apr 2009 A1
20090086949 Caspi et al. Apr 2009 A1
20090086953 Vendrow Apr 2009 A1
20090110182 Knight, Jr. et al. Apr 2009 A1
20090171164 Jung et al. Jul 2009 A1
20090222551 Neely et al. Sep 2009 A1
20090228264 Williams et al. Sep 2009 A1
20090234710 Belgaied et al. Sep 2009 A1
20090234732 Zorman et al. Sep 2009 A1
20090245479 Surendran Oct 2009 A1
20090285384 Pollock et al. Nov 2009 A1
20090306981 Cromack et al. Dec 2009 A1
20090307052 Mankani et al. Dec 2009 A1
20100106568 Grimes Apr 2010 A1
20100114646 Mcilwain et al. May 2010 A1
20100189250 Williams et al. Jul 2010 A1
20100211515 Woodings et al. Aug 2010 A1
20100235341 Bennett Sep 2010 A1
20100250196 Lawler et al. Sep 2010 A1
20100262549 Kannan et al. Oct 2010 A1
20100266115 Fedorov et al. Oct 2010 A1
20100266116 Stolyar et al. Oct 2010 A1
20100274618 Byrd et al. Oct 2010 A1
20100287131 Church Nov 2010 A1
20100293033 Hall et al. Nov 2010 A1
20100299268 Guha et al. Nov 2010 A1
20100332287 Gates et al. Dec 2010 A1
20110014932 Estevez Jan 2011 A1
20110022461 Simeonov Jan 2011 A1
20110071870 Gong Mar 2011 A1
20110077994 Segev et al. Mar 2011 A1
20110082688 Kim et al. Apr 2011 A1
20110116618 Zyarko et al. May 2011 A1
20110125697 Erhart et al. May 2011 A1
20110143323 Cohen Jun 2011 A1
20110182283 Van et al. Jul 2011 A1
20110185293 Barnett et al. Jul 2011 A1
20110216897 Laredo et al. Sep 2011 A1
20110264581 Clyne Oct 2011 A1
20110267985 Wilkinson et al. Nov 2011 A1
20110286592 Nimmagadda Nov 2011 A1
20110288897 Erhart et al. Nov 2011 A1
20120046996 Shah et al. Feb 2012 A1
20120051537 Chishti et al. Mar 2012 A1
20120084217 Kohler et al. Apr 2012 A1
20120087486 Guerrero et al. Apr 2012 A1
20120095835 Makar et al. Apr 2012 A1
20120109830 Vogel May 2012 A1
20120257116 Hendrickson et al. Oct 2012 A1
20120265587 Kinkead Oct 2012 A1
20120290373 Ferzacca et al. Nov 2012 A1
20120321073 Flockhart et al. Dec 2012 A1
20130023235 Fan et al. Jan 2013 A1
20130073361 Silver Mar 2013 A1
20130085785 Rogers et al. Apr 2013 A1
20130090963 Sharma et al. Apr 2013 A1
20130124361 Bryson May 2013 A1
20130136252 Kosiba et al. May 2013 A1
20130223608 Flockhart et al. Aug 2013 A1
20130223610 Kohler et al. Aug 2013 A1
20130236002 Jennings et al. Sep 2013 A1
20130257877 Davis Oct 2013 A1
20130304581 Soroca et al. Nov 2013 A1
20130325972 Boston et al. Dec 2013 A1
20140012603 Scanlon et al. Jan 2014 A1
20140016762 Mitchell et al. Jan 2014 A1
20140039944 Humbert et al. Feb 2014 A1
20140039962 Nudd et al. Feb 2014 A1
20140067375 Wooters Mar 2014 A1
20140079195 Srivastava et al. Mar 2014 A1
20140079207 Zhakov et al. Mar 2014 A1
20140099916 Mallikarjunan et al. Apr 2014 A1
20140101261 Wu et al. Apr 2014 A1
20140136346 Teso May 2014 A1
20140140494 Zhakov May 2014 A1
20140143018 Nies et al. May 2014 A1
20140143249 Cazzanti et al. May 2014 A1
20140161241 Baranovsky et al. Jun 2014 A1
20140164502 Khodorenko et al. Jun 2014 A1
20140177819 Vymenets et al. Jun 2014 A1
20140188477 Zhang Jul 2014 A1
20140200988 Kassko et al. Jul 2014 A1
20140219132 Delveaux et al. Aug 2014 A1
20140219438 Brown et al. Aug 2014 A1
20140233719 Vyemenets et al. Aug 2014 A1
20140244712 Walters et al. Aug 2014 A1
20140254790 Shaffer et al. Sep 2014 A1
20140257908 Steiner et al. Sep 2014 A1
20140270108 Riahi et al. Sep 2014 A1
20140270138 Uba et al. Sep 2014 A1
20140270142 Bischoff et al. Sep 2014 A1
20140270145 Erhart et al. Sep 2014 A1
20140278605 Borucki et al. Sep 2014 A1
20140278649 Guerinik et al. Sep 2014 A1
20140279045 Shottan et al. Sep 2014 A1
20140279050 Makar et al. Sep 2014 A1
20140314225 Riahi et al. Oct 2014 A1
20140335480 Asenjo et al. Nov 2014 A1
20140372171 Martin et al. Dec 2014 A1
20140379424 Shroff Dec 2014 A1
20150006400 Eng et al. Jan 2015 A1
20150010134 Erel et al. Jan 2015 A1
20150012278 Metcalf Jan 2015 A1
20150016600 Desai et al. Jan 2015 A1
20150023484 Ni et al. Jan 2015 A1
20150030151 Bellini et al. Jan 2015 A1
20150030152 Waxman et al. Jan 2015 A1
20150051957 Griebeler et al. Feb 2015 A1
20150066632 Gonzalez et al. Mar 2015 A1
20150071418 Shaffer et al. Mar 2015 A1
20150078538 Jain Mar 2015 A1
20150100473 Manoharan et al. Apr 2015 A1
20150127400 Chan et al. May 2015 A1
20150127441 Feldman May 2015 A1
20150127677 Wang et al. May 2015 A1
20150142704 London May 2015 A1
20150172463 Quast et al. Jun 2015 A1
20150178371 Seth et al. Jun 2015 A1
20150195406 Dwyer et al. Jul 2015 A1
20150213454 Vedula Jul 2015 A1
20150215464 Shaffer et al. Jul 2015 A1
20150222751 Odinak et al. Aug 2015 A1
20150256677 Konig et al. Sep 2015 A1
20150262188 Franco Sep 2015 A1
20150262208 Bjontegard et al. Sep 2015 A1
20150269377 Gaddipati Sep 2015 A1
20150271334 Wawrzynowicz Sep 2015 A1
20150281445 Kumar et al. Oct 2015 A1
20150281449 Milstein et al. Oct 2015 A1
20150281450 Shapiro et al. Oct 2015 A1
20150281454 Milstein et al. Oct 2015 A1
20150287410 Mengibar et al. Oct 2015 A1
20150295788 Witzman et al. Oct 2015 A1
20150296081 Jeong Oct 2015 A1
20150334230 Volzke Nov 2015 A1
20150339446 Sperling et al. Nov 2015 A1
20150339620 Esposito et al. Nov 2015 A1
20150339769 Deoliveira et al. Nov 2015 A1
20150347900 Bell et al. Dec 2015 A1
20150350429 Kumar et al. Dec 2015 A1
20150350440 Steiner et al. Dec 2015 A1
20150350443 Kumar et al. Dec 2015 A1
20150379562 Spievak et al. Dec 2015 A1
20160026629 Clifford et al. Jan 2016 A1
20160034260 Ristock et al. Feb 2016 A1
20160034995 Williams et al. Feb 2016 A1
20160036981 Hollenberg et al. Feb 2016 A1
20160036983 Korolev et al. Feb 2016 A1
20160042419 Singh Feb 2016 A1
20160042749 Hirose Feb 2016 A1
20160055499 Hawkins et al. Feb 2016 A1
20160057284 Nagpal et al. Feb 2016 A1
20160065739 Brimshan et al. Mar 2016 A1
20160080567 Hooshiari et al. Mar 2016 A1
20160085891 Ter et al. Mar 2016 A1
20160112867 Martinez Apr 2016 A1
20160124937 Elhaddad May 2016 A1
20160125456 Wu et al. May 2016 A1
20160134624 Jacobson et al. May 2016 A1
20160140627 Moreau et al. May 2016 A1
20160150086 Pickford May 2016 A1
20160155080 Gnanasambandam et al. Jun 2016 A1
20160173692 Wicaksono et al. Jun 2016 A1
20160180381 Kaiser et al. Jun 2016 A1
20160191699 Agrawal et al. Jun 2016 A1
20160191709 Pullamplavil et al. Jun 2016 A1
20160191712 Bouzid et al. Jun 2016 A1
20160234386 Wawrzynowicz Aug 2016 A1
20160247165 Ryabchun et al. Aug 2016 A1
20160261747 Thirugnanasundaram et al. Aug 2016 A1
20160295018 Loftus et al. Oct 2016 A1
20160300573 Carbune et al. Oct 2016 A1
20160335576 Peng Nov 2016 A1
20160349960 Kumar et al. Dec 2016 A1
20160358611 Abel Dec 2016 A1
20160360033 Kocan Dec 2016 A1
20160378569 Ristock et al. Dec 2016 A1
20160381222 Ristock et al. Dec 2016 A1
20170004178 Ponting et al. Jan 2017 A1
20170006135 Siebel et al. Jan 2017 A1
20170006161 Riahi et al. Jan 2017 A9
20170011311 Backer et al. Jan 2017 A1
20170024762 Swaminathan Jan 2017 A1
20170032436 Disalvo et al. Feb 2017 A1
20170034226 Bostick et al. Feb 2017 A1
20170068436 Auer et al. Mar 2017 A1
20170068854 Markiewicz et al. Mar 2017 A1
20170098197 Yu et al. Apr 2017 A1
20170104875 Im et al. Apr 2017 A1
20170111505 Mcgann et al. Apr 2017 A1
20170111509 McGann et al. Apr 2017 A1
20170116173 Lev-Tov et al. Apr 2017 A1
20170118336 Tapuhi et al. Apr 2017 A1
20170132536 Goldstein et al. May 2017 A1
20170148073 Nomula et al. May 2017 A1
20170155766 Kumar et al. Jun 2017 A1
20170161439 Raduchel et al. Jun 2017 A1
20170162197 Cohen Jun 2017 A1
20170169325 McCord et al. Jun 2017 A1
20170207916 Luce et al. Jul 2017 A1
20170214795 Charlson Jul 2017 A1
20170220966 Wang Aug 2017 A1
20170223070 Lin Aug 2017 A1
20170236512 Williams et al. Aug 2017 A1
20170286774 Gaidon Oct 2017 A1
20170288866 Vanek et al. Oct 2017 A1
20170308794 Fischerstrom Oct 2017 A1
20170316386 Joshi et al. Nov 2017 A1
20170323344 Nigul Nov 2017 A1
20170337578 Chittilappilly et al. Nov 2017 A1
20170344754 Kumar et al. Nov 2017 A1
20170344988 Cusden et al. Nov 2017 A1
20170359421 Stoops et al. Dec 2017 A1
20170372436 Dalal et al. Dec 2017 A1
20180018705 Tognetti Jan 2018 A1
20180032997 Gordon et al. Feb 2018 A1
20180052664 Zhang et al. Feb 2018 A1
20180053401 Martin et al. Feb 2018 A1
20180054464 Zhang et al. Feb 2018 A1
20180060830 Abramovici et al. Mar 2018 A1
20180061256 Elchik et al. Mar 2018 A1
20180077088 Cabrera-Cordon et al. Mar 2018 A1
20180077250 Prasad et al. Mar 2018 A1
20180097910 D'Agostino et al. Apr 2018 A1
20180114234 Fighel Apr 2018 A1
20180121766 Mccord et al. May 2018 A1
20180137472 Gorzela et al. May 2018 A1
20180137555 Clausse et al. May 2018 A1
20180146093 Kumar et al. May 2018 A1
20180150749 Wu et al. May 2018 A1
20180152558 Chan et al. May 2018 A1
20180165062 Yoo et al. Jun 2018 A1
20180165691 Heater et al. Jun 2018 A1
20180165692 McCoy Jun 2018 A1
20180165723 Wright et al. Jun 2018 A1
20180174198 Wilkinson et al. Jun 2018 A1
20180189273 Campos et al. Jul 2018 A1
20180190144 Corelli et al. Jul 2018 A1
20180198917 Ristock et al. Jul 2018 A1
20180205825 Vymenets et al. Jul 2018 A1
20180248818 Zucker et al. Aug 2018 A1
20180260857 Kar et al. Sep 2018 A1
20180285423 Ciano et al. Oct 2018 A1
20180286000 Berry et al. Oct 2018 A1
20180293327 Miller et al. Oct 2018 A1
20180293532 Singh et al. Oct 2018 A1
20180300295 Maksak et al. Oct 2018 A1
20180300641 Donn et al. Oct 2018 A1
20180308072 Smith et al. Oct 2018 A1
20180309801 Rathod Oct 2018 A1
20180349858 Walker et al. Dec 2018 A1
20180361253 Grosso Dec 2018 A1
20180365651 Sreedhara et al. Dec 2018 A1
20180367672 Ristock et al. Dec 2018 A1
20180372486 Farniok et al. Dec 2018 A1
20180376002 Abraham Dec 2018 A1
20190013017 Kang et al. Jan 2019 A1
20190028587 Unitt et al. Jan 2019 A1
20190028588 Shinseki et al. Jan 2019 A1
20190037077 Konig et al. Jan 2019 A1
20190042988 Brown et al. Feb 2019 A1
20190043106 Talmor et al. Feb 2019 A1
20190058793 Konig et al. Feb 2019 A1
20190104092 Koohmarey et al. Apr 2019 A1
20190108834 Nelson et al. Apr 2019 A1
20190130329 Fama et al. May 2019 A1
20190132443 Munns et al. May 2019 A1
20190146647 Ramachandran et al. May 2019 A1
20190147045 Kim May 2019 A1
20190172291 Naseath Jun 2019 A1
20190180095 Ferguson et al. Jun 2019 A1
20190180747 Back et al. Jun 2019 A1
20190182383 Shaev et al. Jun 2019 A1
20190196676 Hillis et al. Jun 2019 A1
20190197568 Li et al. Jun 2019 A1
20190205389 Tripathi et al. Jul 2019 A1
20190236205 Jia et al. Aug 2019 A1
20190238680 Narayanan et al. Aug 2019 A1
20190253553 Chishti Aug 2019 A1
20190258825 Krishnamurthy Aug 2019 A1
20190287517 Green et al. Sep 2019 A1
20190295027 Dunne et al. Sep 2019 A1
20190306315 Portman et al. Oct 2019 A1
20190335038 Alonso Y Caloca et al. Oct 2019 A1
20190341030 Hammons et al. Nov 2019 A1
20190342450 Kulkarni et al. Nov 2019 A1
20190349477 Kotak Nov 2019 A1
20190377789 Jegannathan et al. Dec 2019 A1
20190378076 O'Gorman et al. Dec 2019 A1
20190385597 Katsamanis et al. Dec 2019 A1
20190386917 Malin Dec 2019 A1
20190392357 Surti et al. Dec 2019 A1
20190394333 Jiron et al. Dec 2019 A1
20200005375 Sharan et al. Jan 2020 A1
20200007680 Wozniak Jan 2020 A1
20200012697 Fan et al. Jan 2020 A1
20200012992 Chan et al. Jan 2020 A1
20200019893 Lu Jan 2020 A1
20200028968 Mendiratta et al. Jan 2020 A1
20200050788 Feuz et al. Feb 2020 A1
20200050996 Generes, Jr. et al. Feb 2020 A1
20200058299 Lee et al. Feb 2020 A1
20200076947 Deole Mar 2020 A1
20200097544 Alexander et al. Mar 2020 A1
20200104801 Kwon et al. Apr 2020 A1
20200118215 Rao et al. Apr 2020 A1
20200119936 Balasaygun et al. Apr 2020 A1
20200125919 Liu et al. Apr 2020 A1
20200126126 Briancon et al. Apr 2020 A1
20200134492 Copeland Apr 2020 A1
20200134648 Qi et al. Apr 2020 A1
20200137097 Zimmermann et al. Apr 2020 A1
20200154170 Wu et al. May 2020 A1
20200160870 Baughman et al. May 2020 A1
20200175478 Lee et al. Jun 2020 A1
20200193335 Sekhar et al. Jun 2020 A1
20200193983 Choi Jun 2020 A1
20200211120 Wang et al. Jul 2020 A1
20200218766 Yaseen et al. Jul 2020 A1
20200219500 Bender et al. Jul 2020 A1
20200242540 Rosati et al. Jul 2020 A1
20200250272 Kantor et al. Aug 2020 A1
20200250557 Kishimoto et al. Aug 2020 A1
20200257996 London Aug 2020 A1
20200280578 Hearty et al. Sep 2020 A1
20200280635 Barinov et al. Sep 2020 A1
20200285936 Sen Sep 2020 A1
20200329154 Baumann et al. Oct 2020 A1
20200336567 Dumaine Oct 2020 A1
20200342868 Lou et al. Oct 2020 A1
20200351375 Lepore et al. Nov 2020 A1
20200351405 Pace Nov 2020 A1
20200357026 Liu et al. Nov 2020 A1
20200364507 Berry Nov 2020 A1
20200365148 Ji et al. Nov 2020 A1
20200395008 Cohen et al. Dec 2020 A1
20200410506 Jones et al. Dec 2020 A1
20210004536 Adibi et al. Jan 2021 A1
20210005206 Adibi et al. Jan 2021 A1
20210042839 Adamec Feb 2021 A1
20210056481 Wicaksono et al. Feb 2021 A1
20210067627 Delker et al. Mar 2021 A1
20210081869 Zeelig et al. Mar 2021 A1
20210081955 Zeelig et al. Mar 2021 A1
20210082417 Zeelig et al. Mar 2021 A1
20210082418 Zeelig et al. Mar 2021 A1
20210084149 Zeelig et al. Mar 2021 A1
20210089762 Rahimi et al. Mar 2021 A1
20210091996 Mcconnell et al. Mar 2021 A1
20210105361 Bergher et al. Apr 2021 A1
20210124843 Vass et al. Apr 2021 A1
20210125275 Adibi Apr 2021 A1
20210133763 Adibi et al. May 2021 A1
20210133765 Adibi et al. May 2021 A1
20210134282 Adibi et al. May 2021 A1
20210134283 Adibi et al. May 2021 A1
20210134284 Adibi et al. May 2021 A1
20210136204 Adibi et al. May 2021 A1
20210136205 Adibi et al. May 2021 A1
20210136206 Adibi et al. May 2021 A1
20210201244 Sella et al. Jul 2021 A1
20210201359 Sekar Jul 2021 A1
20220129905 Sethumadhavan et al. Apr 2022 A1
Foreign Referenced Citations (6)
Number Date Country
1 418 519 May 2004 EP
5986065 Sep 2016 JP
2006037836 Apr 2006 WO
2012024316 Feb 2012 WO
2015099587 Jul 2015 WO
2019142743 Jul 2019 WO
Non-Patent Literature Citations (39)
Entry
Gaietto, Molly., “What is Customer DNA?”,—NGDATA Product News, Oct. 27, 2015, 10 pages.
Fan et al., “Demystifying Big Data Analytics for Business Intelligence Through the Lens of Marketing Mix”, Big Data Research, vol. 2, Issue 1, Mar. 1, 2015, 16 pages.
An et al,, Towards Automatic Persona Generation Using Social Media Aug. 1, 2016, 2016 IEEE 4th International Conference on Future Internet of Things and Cloud Workshops (FiCloudW), 2 pages.
Bean-Mellinger, Barbara., “What Is the Difference Between Marketing and Advertising?”, available on Feb. 12, 2019, retrieved from https://smallbusiness.chron .com/difference-between-marketing-advertising-2504 7 .html, Feb. 12, 2019, 6 pages.
Twin, Alexandra., “Marketing”, URL: https://www.investopedia.com/lerms/m/marketing.asp, Mar. 29, 2019, 5 pages.
dictionary.com, “Marketing”, URL: https://www.dictionary.com/browse/marketing, Apr. 6, 2019, 7 pages.
Ponn et al., “Correlational Analysis between Weather and 311 Service Request Volume”, eil.mie.utoronto.ca., Jan. 1, 2017, 16 pages.
Zhang et al., “A Bayesian approach for modeling and analysis of call center arrivals”, Jan. 1, 2013 Winter Simulations Conference (WSC), ieeexplore.ieee.org, pp. 713-723.
Mehrotra et al., “Call Center Simulation Modeling: Methods, Challenges, and Opportunities”, Proceedings of the 2003 Winter Simulation Conference, vol. 1, Jan. 1, 2003, pp. 135-143.
Mandelbaum et al., “Staffing Many-Server Queues with Impatient Customers: Constraint Satisfaction in Call Center”, Operations Research, Sep.- Oct. 2009, vol. 57, No. 5 (Sep. 1-Oct. 2009), pp. 1189-1205.
Fukunaga et al., “Staff Scheduling for Inbound Call Centers and Customer Contact Centers”, AI Magazine, Winter, vol. 23, No. 4, Jan. 1, 2002, pp. 30-40.
Feldman et al., “Staffing of Time-Varying Queues to Achieve Time-Stable Performance”, Management Science, Feb. 1, 2008, vol. 54, No. 2, Call Center Management, pp. 324-338.
Business Wire, “Rockwell SSD announces Call Center Simulator”, Feb. 4, 1997, 4 pages.
Nathan, Stearns., “Using skills-based routing to the advantage of your contact center”, Customer Inter@ction Solutions, Technology Marketing Corporation, May 1, 2001, vol. 19 No. 11, pp. 54-56.
Aksin et al., “The Modern Call Center: A Multi-Disciplinary Perspective on Operations Management Research”, Production and Operations Management, 2007, vol. 16, No. 6, pp. 665-688.
Aldor-Noiman, et al., “Workload forecasting for a call center: Methodology and a case study.” The Annals of Applied Statistics 3.4 (2009); 1403-1447.
Buesing et al., “Getting the Best Customer Service from your IVR: Fresh eyes on an old problem,” [online] McKinsey and Co., published on Feb. 1, 2019, available at: < https://www.nnckinsey.conn/business-functions/operations/our-insights/ getting-the-best-customer-service-from-your-ivr-fresh-eyes . . . (Year: 2019).
Chiu et al., “A multi-agent infrastructure for mobile workforce management in a service oriented enterprise”, Proceedings of the 38th annual Hawaii international conference on system sciences, IEEE, 2005, pp. 10.
Data Warehousing in the Age of Big Data, Krishnan, 2013, Morgan Kaufmann, Chapter 5.
Diimitrios et al., “An overview of workflow management: From process modeling to workflow automation infrastructure,” Distributed and parallel Databases, 1995, vol. 3, No. 2 pp. 119-153.
Ernst et al. “An Annotated Bibliography of Personnel Scheduling and Rostering”, CSIRO Mathematical and Information Sciences, 2003, 155 pages.
Ernst et al.,“Staff scheduling and rostering: A review of applications, methods and models,” European Journal of Operational Research, 2004, vol. 153, pp. 3-27.
Federal Register, vol. 72, No. 195, Oct. 10, 2007, pp. 57526-57535.
Federal Register, vol. 75, No. 169, Sep. 1, 2010, pp. 53643-53660.
Federal register, vol. 79, No. 241 issued on Dec. 16, 2014, p. 74629, col. 2, Gottschalk v. Benson.
Federal Register, vol. 84, No. 4, Jan. 7, 2019, pp. 50-57.
Federal Register, vol. 84, No. 4, Jan. 7, 2019, p. 53-55.
Grefen et al., “A reference architecture for workflow management systems”, Data & Knowledge Engineering, 1998, vol. 27, No. 1, pp. 31-57.
Huang et al., “Agent-based workflow management in collaborative product development on the Internet”, Computer-Aided Design, 2000, vol. 32, No. 2, pp. 133-144.
Janarthanam, “Hands on Chatbots and conversational UI development: Build chatbots and voice user interfaces with Chatfuel, Dialogflow, Microsoft Bot Framework, Twilio, and Alexa Skills” Dec. 2017.
Koole, et al., “An overview of routing and staffing algorithms in multi-skill customer contact centers.” 2006.
Myers et al., “At the Boundary of Workflow and AI”, Proc. AAAI 1999 Workshop on Agent-Based Systems in The Business Context, 1999, 09 pages.
Niven, “Can music with prosocial lyrics heal the working world? A field intervention in a call center.” Journal of Applied Social Psychology, 2015; 45(3), 132-138. doi:10.1111/jasp.12282).
On Hold Marketing, “Growing Your Business with Customized on-Hold Messaging” (Published on Apr. 5, 2018 at https://adhq.com/about/ad-news/growing-your-business-with-customized-on-hold-messaging) (Year: 2018).
U.S. Appl. No. 16/668,214, NFOA mailed Nov. 10, 2021.
U.S. Appl. No. 16/668,215, NFOA mailed Dec. 7, 2021.
Van Den Bergh et al. “Personnel scheduling: A literature review”, European journal of operational research, 2013, vol. 226, No. 3 pp. 367-385.
United States Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 16/550,961 dated Mar. 2, 2020.
United States Patent and Trademark Office, Final Office Action for U.S. Appl. No. 16/550,961 dated Jun. 17, 2020.