Diarization using acoustic labeling to create an acoustic voiceprint

Information

  • Patent Grant
  • 10692501
  • Patent Number
    10,692,501
  • Date Filed
    Monday, October 7, 2019
    5 years ago
  • Date Issued
    Tuesday, June 23, 2020
    4 years ago
Abstract
Systems and method of diarization of audio files use an acoustic voiceprint model. A plurality of audio files are analyzed to arrive at an acoustic voiceprint model associated to an identified speaker. Metadata associate with an audio file is used to select an acoustic voiceprint model. The selected acoustic voiceprint model is applied in a diarization to identify audio data of the identified speaker.
Description
BACKGROUND

The present disclosure is related to the field of automated transcription. More specifically, the present disclosure is related to diarization using acoustic labeling.


Speech transcription and speech analytics of audio data may be enhanced by a process of diarization wherein audio data that contains multiple speakers is separated into segments of audio data typically to a single speaker. While speaker separation in diarization facilitates later transcription and/or speech analytics, further identification or discrimination between the identified speakers can further facilitate these processes by enabling the association of further context and information in later transcription and speech analytics processes specific to an identified speaker.


Systems and methods as disclosed herein present solutions to improve diarization using acoustic models to identify and label at least one speaker separated from the audio data. Previous attempts to create individualized acoustic voiceprint models are time intensive in that an identified speaker must recorded training speech into the system or the underlying data must be manually separated to ensure that only speech from the identified speak is used. Recorded training speech further has limitation as the speakers are likely to speak differently than when the speaker is in the middle of a live interaction with another person.


BRIEF DISCLOSURE

An embodiment of a method of diarization of audio files includes receiving speaker metadata associated with each of a plurality of audio files. A set of audio files of the plurality belonging to a specific speaker are identified based upon the received speaker metadata. A sub set of the audio files of the identified set of audio files is selected. An acoustic voiceprint for the specific speaker is computed from the selected subset of audio fifes. The acoustic voiceprint is applied to a new audio file to identify a specific speaker in the diarization of the new audio file.


An exemplary embodiment of a method of diarization of audio files of a customer service interaction between at least one agent and at least one customer includes receiving agent metadata associated with each of a plurality of audio files. A set of audio files of the plurality of audio files associated to a specific agent is identified based upon the received agent metadata. A subset of the audio files of the identified set of audio files are selected that maximize an acoustical difference between audio data of an agent and audio data of at least one other speaker in each of the audio files. An acoustic voiceprint is computed from the audio data of the agent in the selected subset. The acoustic voiceprint is applied to a new audio file to identify the agent in diarization of the new audio file.


An exemplary embodiment of a system for diarization of audio data includes a database of audio files, each audio file of the database being associated with metadata identifying at least one speaker in the audio file. A processor is communicatively connected to the database. The processor selects a set of audio files with the same speaker based upon the metadata. The processor fitters the selected set to a subset of the audio files that maximize an acoustical difference between audio data of at least two speakers in an audio file. The processor creates an acoustic voiceprint for the speaker identified by the metadata. A database includes a plurality of acoustic voiceprints, each acoustic voiceprint of the plurality is associated with a speaker. An audio source provides new audio data to the processor with metadata that identified at least one speaker in the audio data. The processor selects an acoustic voiceprint from the plurality of acoustic voiceprints based upon the metadata and applies the selected acoustic voiceprint to the new audio data to identify audio data of the speaker in the new audio data for diarization of the new audio data.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a flow chart that depicts an embodiment of a method of diarization.



FIG. 2 is a flow chart that depicts an embodiment of creating and using an acoustic voiceprint model.



FIG. 3 is a system diagram of an exemplary embodiment of a system for diarization of audio files.





DETAILED DISCLOSURE

Embodiments of a diarization process disclosed herein includes a first optional step of a speech-to-text transcription of an audio file to be diarized. Next, a “blind” diarization of the studio file is performed. The audio file is exemplarily a .WAV file. The blind diarization receives the audio file and optionally the automatically generated transcript. This diarization is characterized as “blind” as the diarization is performed prior to an identification of the speakers. In an exemplary embodiment of a customer service call, the “blind diarization” may only cluster the audio data into speakers while it may still be undetermined which speaker is the agent and which speaker is the customer.


The blind diarization is followed by a speaker diarization wherein a voiceprint model that represents the speech and/or information content of an identified speaker in the audio data is compared to the identified speech segments associated with the separated speakers. Through this comparison, one speaker can be selected as the known speaker, while the other speaker is identified as the other speaker. In an exemplary embodiment of customer service interactions, the customer agent will have a voiceprint model as disclosed herein which is used to identify one of the separated speaker as the agent while the other speaker is the customer.


The identification of segments in an audio file, such as an audio stream or recording (e.g. a telephone call that contains speech) can facilitate increased accuracy in transcription, diarization, speaker adaption, and/or speech analytics of the audio file. An initial transcription, exemplarily from a fast speech-to-text engine, can be used to more accurately identify speech segments in an audio file, such as an audio stream or recording, resulting in more accurate diarization and/or speech adaptation.



FIGS. 1 and 2 are flow charts that respectively depict exemplary embodiments of method 100 of diarization and a method 200 of creating and using an acoustic voiceprint model. FIG. 3 is a system diagram of an exemplary embodiment of a system 300 for creating and using an acoustic voiceprint model. The system 300 is generally a computing system that includes a processing system 306, storage system 304, software 302, communication interface 308 and a user interface 310. The processing system 300 loads and executes software 302 from the storage system 304, including a software module 330. When executed by the computing system 300, software module 330 directs the processing system 306 to operate as described in herein in further detail in accordance with the method 100 and alternatively the method 200.


Although the computing system 100 as depicted in FIG. 3 includes one software module in the present example, it should be understood that one or more modules could provide the same operation. Similarly, while description as provided herein refers to a computing system 300 and a processing system 306, it is to be recognized that implementations of such systems can be performed using one or more processors, which may be communicatively connected, and such implementations are considered to be within the scope of the description.


The processing system 306 can comprise a microprocessor and other circuitry that retrieves and executes software 302 from storage system 304. Processing system 306 can be implemented within a single processing device but can also be distributed across multiple processing devices or sub-systems that cooperate in existing program instructions. Examples of processing system 306 include general purpose central processing units, application specific processors, and logic devices, as well as any other type of processing device, combinations of processing devices, or variations thereof.


The storage system 304 can comprise any storage media readable by processing system 306, and capable of storing software 302. The storage system 304 can include volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data. Storage system 304 can be implemented as a single storage device but may also be implemented across multiple storage devices or sub-systems. Storage system 304 can further include additional elements, such as a controller capable of communicating with the processing system 306.


Examples of storage media include random access memory, read only memory, magnetic discs, optical discs, flash memory, virtual memory, and non-virtual memory, magnetic sets, magnetic tape, magnetic disc storage or other magnetic storage devices, or any other medium which can be used to storage the desired information and that may be accessed by an instruction execution system, as well as any combination or variation thereof, or any other type of storage medium. In some implementations, the storage media can be a non-transitory storage media. In some implementations, at least a portion of the storage media may be transitory. It should be understood that in no case is the storage media a prorogated signal.


User interface 310 can include a mouse, a keyboard, a voice input device, a touch input device for receiving a gesture from a user, a motion input device for detecting non-touch gestures and other motions by a user, and other comparable input devices and associated processing elements capable of receiving user input from a user. Output devices such as a video display or graphical display can display an interface further associated with embodiments of the system and method as disclosed herein. Speakers, printers, haptic devices and other types of output devices may also be included in the user interface 310.


As described in further detail herein, the computing system 200 receives an audio file 320. The audio file 320 may be an audio recording or a conversation, which may exemplarily be between two speakers, although the audio recording may be any of a variety of other audio records, including multiple speakers, a single speaker, or an automated or recorded auditory message. In still further embodiments, the audio file may be streaming audio data received in real lime or near-real time by the computing system 300.



FIG. 1 is a flow chart that depicts an embodiment of a method of diarization 100. Audio data 102 is exemplarily an audio recording of a conversation exemplarily between two or more speakers. The audio file may exemplarily be a .WAV file, but may also be other types of audio or video files, for example, pulse code modulated (PCM) formatted audio, and more specifically, linear pulse code modulated (LPCM) audio files. Furthermore, the audio data is exemplarily a mono audio file; however, it is recognized that embodiments of the method disclosed herein may also be used with stereo audio files. One feature of the method disclosed herein is that speaker separation and diarization can be achieved in mono audio files where stereo speaker separation techniques are not available.


In embodiments, the audio data 102 further comprises or is associated to metadata 108. The metadata 108 can exemplarily include an identification number for one or more of the speakers to the audio data 102. In alternative embodiments, the metadata 108 may provide information regarding context or content of the audio data 102, including a topic, time, date, location etc. In the context of a customer service call center, the metadata 108 provides a customer service agent identification.


In an embodiment, the audio data 102 and the metadata 108 are provided to a speech-to-text (STT) server 104, which may employ any of a variety of method of techniques for automatic speech recognition (ASR) to create an automated speech-to-text transcription 106 from the audio file. The transcription performed by the STT server at 104 can exemplarily be a large-vocabulary continuous speech recognition (LVCSR) and the audio data 102 provided to the STT server 104 can alternatively be a previously recorded audio file or can be streaming audio data obtained from an ongoing communication between two speakers. In an exemplary embodiment, the STT server 104 may use the received metadata 108 to select one or more models or techniques for producing the automated transcription cased upon the metadata 108. In a non-limiting example, an identification of one of the speakers in the audio data can be used to select a topical linguistic model based upon a context area associate with the speaker. In addition to the transcription 106 from the STT server 104, STT server 104 may also output time stamps associate with particular transcription segments, words, or phrases, and may also include a confidence score in the automated transcription. The transcription 106 may also identify homogeneous speaker speech segments. Homogenous speech segments are those segments of the transcription that have a high likelihood of originating from a single speaker. The speech segments may exemplarily be phonemes, words, or sentences.


After the transcription 106 is created, both the audio data 102 and the transcription 106 are used for a blind diarization at 110. However, it is to be recognized that in alternative embodiments, the blind diarization may be performed without the transcription 106 and may be applied directly to the audio data 102. In such embodiments, the features at 104 and 106 as described above may not be used. The diarization is characterized as blind as the identities of the speakers (e.g. agent, customer) are not known at this stage and therefore the diarization 110 merely discriminates between a first speaker (speaker 1) and a second speaker (speaker 2), or more. Additionally, in some embodiments, those segments for which a speaker cannot be reliably determined may be labeled as being of an unknown speaker.


An embodiment of the blind diarization at 110 receives the mono audio data 102 and the transcription 106 and begins with the assumption that there are two main speakers in the audio file. The homogeneous speaker segments from 106 are identified in the audio file. Then, long homogeneous speaker segments can be split into sub-segments if long silent intervals are found within a single segment. The sub-segments are selected to avoid splitting the long speaker segments within a word. The transcription information in the information file 106 can provide context to where individual words start and end. After the audio file has been segmented based upon both the audio file 102 and the information file 106, the identified segments are clustered into speakers (e.g. speaker 1 and speaker 2).


In an embodiment, the blind diarization uses voice activity detection (VAD) to segment the audio data 102 into utterances or short segments of audio data with a likelihood of emanating from a single speaker. In an embodiment, the VAD segments the audio data into utterances by identifying segments of speech separated by segments of non-speech on a frame-by-frame basis. Context provided by the transcription 106 can improve the distinction between speech and not speech segments. In the VAD at 304 an audio frame may be identified as speech or non-speech based upon a plurality of characteristics or probabilities exemplarity based upon mean energy, band energy, peakiness, or residual energy; however, it will be recognized that alternative characteristics or probabilities may be used in alternative embodiments.


Embodiments of the blind diarization 110 may further leverage the received metadata 108 to select an acoustic voiceprint model 116, from a plurality of stored acoustic voiceprint models as well be described in further detail herein. Embodiments that use the acoustic voiceprint model in the blind diarization 110 can improve the clustering of the segmented audio data into speakers, for example by helping to cluster segments that are otherwise indeterminate, or “unknown.”


The blind diarization at 110 results in audio data of separated speaker at 112. In an example, the homogeneous speaker segments in the audio data are tagged as being associated with a first speaker or a second speaker. As mentioned above, in some embodiments, in determinate segments may be tagged as “unknown” and audio data may have more man two speakers tagged.


At 114 a second diarization, “speaker” diarization, is undertaken to identify which of the first speaker and second speaker is the speaker identified by the metadata 108 and which speaker is the at least one other speaker. In the exemplary embodiment of a customer service interaction, the metadata 108 identifies a customer service agent participating in the recorded conversation and the other speaker is identified as the customer. An acoustic voiceprint model 116, which can be derived in a variety of manners or techniques as described in more detail herein, is compared to the homogeneous speaker audio data segments assigned to the first speaker and then compared to the homogeneous speaker audio data segments assigned to the second speaker to determine which separated speaker audio data segments have a greater likelihood of matching the acoustic voiceprint model 116. At 118, the homogeneous speaker segments tagged in the audio file as being the speaker that is most likely the agent based upon the comparison of the acoustic voiceprint model 116 are tagged as the speaker identified in the metadata and the other homogeneous speaker segments are tagged as being the other speaker.


At 120, the diarized and labeled audio data from 118 again undergoes an automated transcription, exemplarily performed by a STT server or other form of ASR, which exemplarily may be LVCSR. With the additional context of both enhanced identification of speaker segments and clustering and labeling of the speaker in the audio data, an automated transcription 122 can be output from the transcription at 120 through the application of improved algorithms and selection of further linguistic or acoustic models tailored to either the identified agent or the customer, or another aspect of the customer service interaction as identified through the identification of one or more of the speakers in the audio data. This improved labeling of the speaker in the audio data and the resulting transcription 122 can also facilitate analytics of the spoken content of the audio data by providing additional context regarding the speaker, as well as improved transcription of the audio data.


It is to be noted that in some embodiments, the acoustic voice prints as described herein may be used in conjunction with one or more linguistic models, exemplarily the linguistic models as disclosed and applied in U.S. Provisional Patent Application No. 61/729,067, which is incorporated herein by reference. In such combined embodiments, the speaker diarization may be performed in parallel with both a linguistic model and an acoustic voice print model and the two resulting speaker diarization are combined or analyzed in combination in order to provide an improved separation of the audio data into known speakers. In an exemplary embodiment, if both models agree on a speaker label, then that label is used, while if the analysis disagrees, then an evaluation may be made to determine which model is the more reliable or more likely model based upon the context of the audio data. Such an exemplary embodiment may offer the advantages of both acoustic and linguistic modeling and speaker separation techniques.


In a still further embodiment, the combination of both an acoustic voiceprint model and a linguistic model can help to identify errors in the blind diarization or the speaker separation phases, exemplarily by highlighting the portions of the audio data above within which the two models disagree and providing for more detailed analysis on those areas in which the models are in disagreement in order to arrive at the correct diarization and speaker labeling. Similarly, the use of an additional linguistic model may provide a backup for an instance wherein an acoustic voiceprint is not available or identified based upon the received metadata. For example, this situation may arrive when there is insufficient audio data regarding a speaker to create an acoustic voiceprint as described in further detail herein.


Alternatively, in embodiments, even if the metadata does not identify a speaker, if an acoustic voiceprint exists for a speaker in the audio data, all of the available acoustic voiceprints may be compared to the audio data in order to identify at leant one of the speakers in the audio data. In a still further embodiment, a combined implantation using a linguistic model and an acoustic model may help to identify an incongruity between the received metadata, which may identify one speaker, while the comparison to that speaker's acoustic voiceprint model reveals that the identified speaker is not in the audio data. In one non-limiting example, in the context of a customer service interaction, this may help to detect an instance wherein a customer service agent enters the wrong agent ID number so that corrective action may be taken. Finally, in still further embodiments the use of a combination of acoustic and linguistic models may help in the identification and separation of speaker in audio data that contain more than two speakers, exemplarily, one customer service agent and two customers; two agents and one customer; or an agent, a customer, and an automated recording such as a voicemail message.



FIG. 2 is a flow chart that depicts an embodiment of the creation and use of an acoustic voiceprint model exemplarily used as the acoustic voiceprint model 116 in FIG. 1. Referring back to FIG. 2, the method 200 is divided into two portions, exemplarily, the creation of the acoustic voiceprint model at 202 and the application or use of the acoustic voiceprint model at 204 to label speakers in an audio file. In an exemplary embodiment of a customer service interaction, the acoustic voiceprint model is of a customer service agent and associated with an agent identification number specific to the customer service agent.


Referring specifically to the features at 202, at 206 a number (N) of files are selected from a repository of files 208. The files selected at 206 all share a common speaker, exemplarily, the customer service agent for which the model is being created. In an embodiment, in order to make this selection, each of the audio files in the repository 208 are stored with or associated to an agent identification number. In exemplary embodiments, N may be 5 files, 100 files, or 1,000; however, these are merely exemplary numbers. In an embodiment, the N files elected at 20 may be further filtered in order to only select audio files in which the speaker, and thus the identified speaker are easy to differentiate, for example due to the frequency of the voices of the different speakers. By selecting only those files in which the acoustic differences between the speakers are maximized, the acoustic voiceprint model as disclosed herein may be started with files that are likely to be accurate in the speaker separation. In one embodiment, the top 50% of the selected files are used to create the acoustic voiceprint, while in other embodiments, the top 20% or top 10% are used; however, these percentages are in no way intended to be limiting on the thresholds that may be used in embodiments in accordance with the present disclosure.


In a still further embodiment, a diarization or transcription of the audio file is received and scored and only the highest scoring audio files are used to create the acoustic voiceprint model. In an embodiment, the score may exemplarily be an automatedly calculated confidence score for the diarization or transcription. Such automated confidence may exemplarily, but not limited to, use an auto correction function.


Each of the files selected at 206 are processed through a diarization at 210. The diarization process may be such as is exemplarily disclosed above with respect to FIG. 1. In an embodiment, the diarization at 210 takes each of the selected audio files and separates the file into a plurality of segments of speech separated by non-speech. In an embodiment, the plurality of speech segments are further divided such that each segment has a high likelihood of containing speech sections from a single speaker. Similar to the blind diarization described above, the diarization at 210 can divide the audio file into segments labeled as a first speaker and a second speaker (or in some embodiments more speakers) at 212.


At 214 the previously identified speaker segments from the plurality of selected audio files are clustered into segments that are similar to one another. The clustering process can be done directly by matching segments based upon similarity to one another or by clustering the speaker segments based upon similarities to a group of segments. The clustered speaker segments are classified at 216. Embodiments of the system and method use one or more metrics to determine which clusters of speaker segments belong to the customer service agent and which speaker segment clusters belong to the customers with whom the customer service agent was speaking. In one non-limiting embodiment, the metric of cluster size may be used to identify the segment clusters associated with the customer service agent as larger clusters may belong to the customer service agent because the customer service agent is a party in each of the audio files selected for use in creating a model at 206. While it will be recognized that other features related to the agent's script, delivery, other factors related to the customer service calls themselves may be used as the classifying metric.


At 218 an acoustic voiceprint model for the identified speaker, exemplarily a customer service agent is built using the segments that have been classified as being from the identified speaker. At 220 a background voiceprint model that is representative of the audio produced from speakers who are not the identified speaker is built from those speech segments identified to not be the identified speaker, and thus may include the other speakers as well as background noise.


Therefore, in some embodiments, the acoustic voiceprint model, such as exemplarily used with respect to FIG. 1 described above, includes both an identified speaker voiceprint 222 that is representative of the speech of the identified speaker and a background voiceprint 224 that is representative of the other speaker with whom the identified speaker speaks, and any background noises to the audio data of the identified speaker.


It will be recognized that in embodiments, the creation of the acoustic voiceprint model 202 may be performed in embodiments to create an acoustic voiceprint model for each of a plurality of identified speakers that will be recorded and analyzed in the diarization method of FIG. 1. Exemplarily in these embodiments, the identified speakers may be a plurality of customer service agents. In some embodiments, each of the created acoustic voiceprint models are stored in a database of acoustic voiceprint models from which specific models are accessed as described above with respect to FIG. 1, exemplarily based upon an identification number in metadata associated with audio data.


In further embodiments, the processes at 202 may be performed at regular intervals using a predefined number of recently obtained audio data, or a stored set of exemplary audio files. Such exemplary audio files may be identified from situations in which the identified speaker is particularly easy to pick out in the audio, perhaps due to differences in the pitch or tone between the identified speaker's voice and the other speaker's voice, or due to a distinctive speech pattern or characteristic or prevalent accent by the other speaker. In still other embodiments, the acoustic voiceprint model is built on an ad hoc basis at the time of diarization of the audio. In such an example, the acoustic model creation process may simply select a predetermined number of the most recent audio recordings that include the identified speaker or may include all audio recordings within a predefined date that include the identified speaker. It will be also noted that once the audio file currently being processed has been diarized, that audio recording may be added to the repository of audio files 208 for training of future models of the speech of the identified speaker.



204 represents an embodiment of the use of the acoustic voiceprint model as created at 202 in performing a speaker diarization, such as represented at 114 in FIG. 1. Referring back to FIG. 2, at 226 new audio data is received. The new audio data received at 226 may be a stream of real-time audio data or may be recorded audio data being processed. Similar to that described above with respect to 110 and 112 in FIG. 1, the new audio data 226 undergoes diarization at 228 to separate the new audio data 226 into segments that can be confidently tagged as being the speech of a single speaker, exemplarily a first speaker and a second speaker. At 230 the selected acoustic voiceprint 222 which may include background voiceprint 224, is compared to the segments identified in the diarization at 228. In one embodiment, each of the identified segments is separately compared to both the acoustic voiceprint 222 and to the background voiceprint 224 and an aggregation of the similarities of the first speaker segments and the second speaker segments to each of the models is compared in order to determine which of the speakers in the diarized audio file is the identified speaker.


In some embodiments, the acoustic voiceprint model is created from a collection of audio files that are selected to provide a sufficient amount of audio data mat can be confidently tagged to belong only to the agent, and these selected audio files are used to create the agent acoustic model. Some considerations that may go into such a selection may be identified files with good speaker separation and sufficient length to provide data to the model and confirm speaker separation. In some embodiments, the audio files are preprocessed to eliminate non-speech data from the audio file that may affect the background model. Such elimination of non-speech data can performed by filtering or concatenation.


In an embodiment, the speakers in an audio file can be represented by a feature vector and the feature vectors can be aggregated into clusters. Such aggregation of the feature vectors may help to identify the customer service agent from the background speech as the feature vector associated with the agent will aggregate into clusters more quickly than those feature vectors representing a number of different customers. In a still further embodiment, an iterative process may be employed whereby a first acoustic voiceprint model is created using some of the techniques disclosed above, the acoustic voiceprint model is tested or verified, and if the model is not deemed to be broad enough or be based upon enough speaker segments, additional audio files and speaker segments can be selected from the repository and the model is recreated.


In one non-limiting example, the speaker in an audio file is represented by a feature vector. An initial super-segment labeling is performed using agglomerative clustering of feature vectors. The feature vectors from the agent will aggregate into clusters more quickly than the feature vectors from the second speaker as the second speaker in each of the audio files is likely to be a different person. A first acoustic voiceprint model is built from the feature vectors found in the largest clusters and the background model is built from all of the other feature vectors. In one embodiment, a diagonal Gaussian can be trained for each large cluster from the super-segments in that cluster. However, other embodiments may use Gaussian Mixture Model (GMM) while still further embodiments may include i-vectors. The Gaussians are then merged where a weighting value of each Gaussian is proportionate to the number of super-segments in the cluster represented by the Gaussian. The background model can be comprised of a single diagonal Gaussian trained on the values of the super segments that are remaining.


Next, the acoustic voiceprint model can be refined by calculating a log-likelihood of each audio file's super-segments with both the acoustic voiceprint and background models, reassigning the super-segments based upon this comparison. The acoustic voiceprint and background models can be rebuilt from the reassigned super-segments in the manner as described above and the models can be iteratively created in the manner described above until the acoustic voiceprint model can be verified.


The acoustic voiceprint model can be verified when a high enough quality match is found between enough of the sample agent super-segments and the agent model. Once the acoustic voiceprint model has been verified, then the final acoustic voiceprint model can be built with a single full Gaussian over the last super-segment assignments from the application of the acoustic voiceprint model to the selected audio files. As noted above, alternative embodiments may use Gaussian Mixture Model (GMM) while still further embodiments may use i-vectors. The background model can be created from the super-segments not assigned to the identified speaker. It will be recognized that in alternative embodiments, an institution, such as a call center, may use a single background model for all agents with the background model being updated in the manner described above at periodic intervals.


Embodiments of the method described above can be performed or implemented in a variety of ways. The SST server, in addition to performing the LVCSR, can also perform the diarization process. Another alternative is to use a centralized server to perform the diarization process. In one embodiment, a stand-alone SST server performs the diarization process locally without any connection to another server for central storage or processing. In an alternative embodiment, the STT server performs the diarization, but relies upon centrally stored or processed models, to perform the initial transcription. In a still further embodiment, a central dedicated diarization server may be used where the output of many STT servers are sent to the centralized diarization server for processing. The centralized diarization server may have locally stored models that build from processing of all of the diarization at a single server.


This written description uses examples to disclose the invention, including the best mode, and also to enable any person skilled in the art to make and use the invention. The patentable scope of the invention is defined by the claims, and may include other examples that occur to those skilled in the art. Such other examples are intended to be within the scope of the claims if they have structural elements that do not differ from the literal language of the claims, or if they include equivalent structural elements with insubstantial differences from the literal languages of the claims.

Claims
  • 1. A method of diarization of audio files, the method comprising: selecting a plurality of audio files from a database server, wherein each audio file is a recording of a customer service interaction including a known speaker and at least one other speaker, wherein each audio file selected maximizes an acoustical difference in voice frequencies between the known speaker and the at least one other speaker in the same audio file;performing a blind diarization on the selected audio files to segment the audio files into a plurality of segments of speech separated by non-speech, such that each segment has a high likelihood of containing speech sections from a single speaker;automatedly applying at least one metric to the segments of speech with a processor to label segments of speech likely to be associated with the known speaker and clustering the selected segments into an audio speaker segment;analyzing the selected audio speaker segments to create an acoustic voiceprint, wherein the acoustic voiceprint is built from all the selected speaker segments; andapplying the acoustic voiceprint to the audio files with the processor to label a portion of the audio file as having been spoken by the known speaker.
  • 2. The method of claim 1, wherein the at least one metric is that of cluster size wherein the larger the cluster the more likely the segment belongs to the known speaker.
  • 3. The method of claim 1, further comprising saving the acoustic voiceprint to a voiceprint database server and associating it with the known speaker.
  • 4. The method of claim 3, further comprising applying the saved acoustic voiceprint from the voiceprint database server to a new audio file from an audio source to perform diarization of the new audio file by blind diarizing the new audio file, comparing each of the new speech segments to the acoustic voiceprint, and labeling each speech segment as belonging to the known speaker associated with the acoustic voiceprint or belonging to an other speaker.
  • 5. The method of claim 4, wherein the new audio file is real time audio data.
  • 6. The method of claim 1, further comprising: labeling the segments of speech not likely to be associated with the known speaker and clustering the segments into an audio background segment;analyzing the audio background segment to create a background acoustic voiceprint, wherein the background acoustic voiceprint is built from the labeled audio background segment; andapplying the background acoustic voiceprint to the audio files with the processor to label a portion of the audio file as having been spoken by an other speaker.
  • 7. The method of claim 6, further comprising removing the non-speech speech segments from the labeled audio background segment prior to creating the background acoustic voiceprint.
  • 8. The method of claim 6, further comprising saving the background acoustic voiceprint to the voiceprint database server and associating it with the other speaker.
  • 9. The method of claim 8, further comprising applying the saved background acoustic voiceprint from the voiceprint database server to a new audio file from an audio source to perform diarization of the new audio file by blind diarizing the new audio file, comparing each of the new speech segments to the acoustic voiceprint, and labeling each speech segment as belonging to the other speaker or as belonging to a known speaker.
  • 10. The method of claim 1, wherein the selected audio files are the top 50% or less of audio files in the database server based upon an acoustical difference between the known speaker and the at least one other speaker.
  • 11. The method of claim 1, wherein the selected audio files are the top 20% or less of audio files in the database server based upon an acoustical difference between the known speaker and the at least one other speaker.
  • 12. The method of claim 1, wherein the selected audio files are further selected by sending audio files in the database server to a transcription server wherein the transcription server transcribes the audio files and automatedly scores each transcribed audio file for a confidence of transcription, only those audio files that maximize the confidence of transcription and maximize an acoustical difference in voice frequencies between the known speaker and the at least one other speaker in the same audio file are selected.
  • 13. A non-transitory computer-readable medium having instructions stored thereon for facilitating diarization of audio files from a customer service interaction, wherein the instructions, when executed by a processing system, direct the processing system to: select a plurality of audio files from a database server, wherein each audio file is a recording of a customer service interaction including a known speaker and at least one other speaker, wherein each audio file selected maximizes an acoustical difference in voice frequencies between the known speaker and the at least one other speaker in the same audio file;perform a blind diarization on the selected audio files to segment the audio files into a plurality of segments of speech separated by non-speech, such that each segment has a high likelihood of containing speech sections from a single speaker;automatedly apply at least one metric to the segments of speech with a processor to label segments of speech likely to be associated with the known speaker and clustering the selected segments into an audio speaker segment;analyze the selected audio speaker segments to create an acoustic voiceprint, wherein the acoustic voiceprint is built from all the selected speaker segments; andapply the acoustic voiceprint to the audio files with the processor to label a portion of the audio file as having been spoken by the known speaker.
  • 14. The non-transitory computer-readable medium of claim 13, wherein the at least one metric is that of cluster size wherein the larger the cluster the more likely the segment belongs to the known speaker.
  • 15. The non-transitory computer-readable medium of claim 13, further directing the processing system to save the acoustic voiceprint to a voiceprint database server and associating it with the known speaker.
  • 16. The non-transitory computer-readable medium of claim 15, further directing the processing system to apply the saved acoustic voiceprint from the voiceprint database server to a new audio file from an audio source to perform diarization of the new audio file by blind diarizing the new audio file, comparing each of the new speech segments to the acoustic voiceprint, and labeling each speech segment as belonging to the known speaker associated with the acoustic voiceprint or belonging to an other speaker.
  • 17. The non-transitory computer-readable medium of claim 13, further directing the processing system to: label the segments of speech not likely to be associated with the known speaker and clustering the segments into an audio background segment;analyze the audio background segment to create a background acoustic voiceprint, wherein the background acoustic voiceprint is built from the labeled audio background segment; andapply the background acoustic voiceprint to the audio files with the processor to label a portion of the audio file as having been spoken by an other speaker.
  • 18. The non-transitory computer-readable medium of claim 17, further directing the processing system to remove the non-speech speech segments from the labeled audio background segment prior to creating the background acoustic voiceprint.
  • 19. The non-transitory computer-readable medium of claim 18, further directing the processing system to save the background acoustic voiceprint to a voiceprint database server and associating it with the other speaker.
  • 20. The non-transitory computer-readable medium of claim 19, further directing the processing system to apply the saved background acoustic voiceprint from the voiceprint database server to a new audio file from an audio source to perform diarization of the new audio file by blind diarizing the new audio file, comparing each of the new speech segments to the acoustic voiceprint, and labeling each speech segment as belonging to the other speaker or as belonging to a known speaker.
CROSS-REFERENCE TO RELATED APPLICATIONS

The present application claims priority of U.S. patent application Ser. No. 16/170,306, filed on Oct. 25, 2018, which application claims priority of U.S. patent application Ser. No. 14/084,974, filed on Nov. 20, 2013, which application claims priority of U.S. Provisional Patent Application Nos. 61/729,064, filed on Nov. 21, 2012, and 61/729,067 filed Nov. 21, 2012, the contents of which are incorporated herein by reference in their entireties.

US Referenced Citations (134)
Number Name Date Kind
4653097 Watanabe et al. Mar 1987 A
4864566 Chauveau Sep 1989 A
5027407 Tsunoda Jun 1991 A
5222147 Koyama Jun 1993 A
5638430 Hogan et al. Jun 1997 A
5805674 Anderson Sep 1998 A
5907602 Peel et al. May 1999 A
5946654 Newman et al. Aug 1999 A
5963908 Chadha Oct 1999 A
5999525 Krishnaswamy et al. Dec 1999 A
6044382 Martino Mar 2000 A
6145083 Shaffer et al. Nov 2000 A
6266640 Fromm Jul 2001 B1
6275806 Pertrushin Aug 2001 B1
6427137 Petrushin Jul 2002 B2
6480825 Sharma et al. Nov 2002 B1
6510415 Talmor et al. Jan 2003 B1
6587552 Zimmerman Jul 2003 B1
6597775 Lawyer et al. Jul 2003 B2
6915259 Rigazio Jul 2005 B2
7006605 Morganstein et al. Feb 2006 B1
7039951 Chaudhari et al. May 2006 B1
7054811 Barzilay May 2006 B2
7106843 Gainsboro et al. Sep 2006 B1
7158622 Lawyer et al. Jan 2007 B2
7212613 Kim et al. May 2007 B2
7299177 Broman et al. Nov 2007 B2
7386105 Wasserblat et al. Jun 2008 B2
7403922 Lewis et al. Jul 2008 B1
7539290 Ortel May 2009 B2
7657431 Hayakawa Feb 2010 B2
7660715 Thambiratnam Feb 2010 B1
7668769 Baker et al. Feb 2010 B2
7693965 Rhoads Apr 2010 B2
7778832 Broman et al. Aug 2010 B2
7822605 Zigel et al. Oct 2010 B2
7908645 Varghese et al. Mar 2011 B2
7940897 Khor et al. May 2011 B2
8036892 Broman et al. Oct 2011 B2
8073691 Rajakumar Dec 2011 B2
8112278 Burke Feb 2012 B2
8311826 Rajakumar Nov 2012 B2
8510215 Gutierrez Aug 2013 B2
8537978 Jaiswal et al. Sep 2013 B2
9001976 Arrowood Apr 2015 B2
10134400 Ziv Nov 2018 B2
20010026632 Tamai Oct 2001 A1
20020022474 Blom et al. Feb 2002 A1
20020099649 Lee et al. Jul 2002 A1
20030009333 Sharma et al. Jan 2003 A1
20030050780 Rigazio Mar 2003 A1
20030050816 Givens et al. Mar 2003 A1
20030097593 Sawa et al. May 2003 A1
20030147516 Lawyer et al. Aug 2003 A1
20030208684 Camacho et al. Nov 2003 A1
20040029087 White Feb 2004 A1
20040111305 Gavan et al. Jun 2004 A1
20040131160 Mardirossian Jul 2004 A1
20040143635 Galea Jul 2004 A1
20040167964 Rounthwaite et al. Aug 2004 A1
20040203575 Chin et al. Oct 2004 A1
20040240631 Broman et al. Dec 2004 A1
20050010411 Rigazio Jan 2005 A1
20050043014 Hodge Feb 2005 A1
20050076084 Loughmiller et al. Apr 2005 A1
20050125226 Magee Jun 2005 A1
20050125339 Tidwell et al. Jun 2005 A1
20050135595 Bushey Jun 2005 A1
20050185779 Toms Aug 2005 A1
20060013372 Russell Jan 2006 A1
20060098803 Bushey et al. May 2006 A1
20060106605 Saunders May 2006 A1
20060149558 Kahn Jul 2006 A1
20060161435 Atef et al. Jul 2006 A1
20060212407 Lyon Sep 2006 A1
20060212925 Shull et al. Sep 2006 A1
20060248019 Rajakumar Nov 2006 A1
20060251226 Hogan et al. Nov 2006 A1
20060282660 Varghese et al. Dec 2006 A1
20060285665 Wasserblat et al. Dec 2006 A1
20060289622 Khor et al. Dec 2006 A1
20060293891 Pathuel Dec 2006 A1
20070041517 Clarke et al. Feb 2007 A1
20070071206 Gainsboro et al. Mar 2007 A1
20070074021 Smithies et al. Mar 2007 A1
20070100608 Gable et al. May 2007 A1
20070124246 Lawyer et al. May 2007 A1
20070244702 Kahn et al. Oct 2007 A1
20070250318 Waserblat et al. Oct 2007 A1
20070280436 Rajakumar Dec 2007 A1
20070282605 Rajakumar Dec 2007 A1
20070288242 Spengler Dec 2007 A1
20080010066 Broman et al. Jan 2008 A1
20080181417 Pereg Jul 2008 A1
20080195387 Zigel et al. Aug 2008 A1
20080222734 Redlich et al. Sep 2008 A1
20090046841 Hodge Feb 2009 A1
20090119106 Rajakumar May 2009 A1
20090147939 Morganstein et al. Jun 2009 A1
20090247131 Champion et al. Oct 2009 A1
20090254971 Herz et al. Oct 2009 A1
20090319269 Aronowitz Dec 2009 A1
20100138282 Kannan Jun 2010 A1
20100228656 Wasserblat et al. Sep 2010 A1
20100303211 Hartig Dec 2010 A1
20100305946 Gutierrez Dec 2010 A1
20100305960 Gutierrez Dec 2010 A1
20100332287 Gates Dec 2010 A1
20110004472 Zlokamik Jan 2011 A1
20110026689 Metz et al. Feb 2011 A1
20110119060 Aronowitz May 2011 A1
20110191106 Khor et al. Aug 2011 A1
20110255676 Marchand et al. Oct 2011 A1
20110282661 Dobry et al. Nov 2011 A1
20110282778 Wright et al. Nov 2011 A1
20110320484 Smithies et al. Dec 2011 A1
20120053939 Gutierrez et al. Mar 2012 A9
20120054202 Rajakumar Mar 2012 A1
20120072453 Guerra et al. Mar 2012 A1
20120130771 Kannan May 2012 A1
20120253805 Rajakumar et al. Oct 2012 A1
20120254243 Zeppenfeld et al. Oct 2012 A1
20120263285 Rajakumar et al. Oct 2012 A1
20120284026 Cardillo et al. Nov 2012 A1
20130163737 Dement et al. Jun 2013 A1
20130197912 Hayakawa et al. Aug 2013 A1
20130253919 Gutierrez et al. Sep 2013 A1
20130300939 Chou et al. Nov 2013 A1
20140067394 Abuzeina Mar 2014 A1
20140142940 Ziv et al. May 2014 A1
20150055763 Guerra et al. Feb 2015 A1
20160364606 Conway Dec 2016 A1
20160379032 Mo Dec 2016 A1
20160379082 Rodriguez et al. Dec 2016 A1
Foreign Referenced Citations (7)
Number Date Country
0598469 May 1994 EP
2004193942 Jul 2004 JP
2006038955 Sep 2006 JP
2000077772 Dec 2000 WO
2004079501 Sep 2004 WO
2006013555 Feb 2006 WO
2007001452 Jan 2007 WO
Non-Patent Literature Citations (11)
Entry
Baum, L.E., et al., “A Maximization Technique Occurring in the Statistical Analysis of Probabilistic Functions of Markov Chains,” The Annals of Mathematical Statistics, vol. 41, No. 1, 1970, pp. 164-171.
Cheng, Y., “Mean Shift, Mode Seeking, and Clustering,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 17, No. 8, 1995, pp. 790-799.
Cohen, I., “Noise Spectrum Estimation in Adverse Environment: Improved Minima Controlled Recursive Averaging,” IEEE Transactions on Speech and Audio Processing, vol. 11, No. 5, 2003, pp. 466-475.
Cohen, I., et al., “Spectral Enhancement by Tracking Speech Presence Probability in Subbands,” Proc. International Workshop in Hand-Free Speech Communication (HSC'01), 2001, pp. 95-98.
Coifman, R.R., et al., “Diffusion maps,” Applied and Computational Harmonic Analysis, vol. 21, 2006, pp. 5-30.
Hayes, M.H., “Statistical Digital Signal Processing and Modeling,” J. Wiley & Sons, Inc., New York, 1996, 200 pages.
Hermansky, H., “Perceptual linear predictive (PLP) analysis of speech,” Journal of the Acoustical Society of America, vol. 87, No. 4, 1990, pp. 1738-1752.
Lailler, C., et al., “Semi-Supervised and Unsupervised Data Extraction Targeting Speakers: From Speaker Roles to Fame?,” Proceedings of the First Workshop on Speech, Language and Audio in Multimedia (SLAM), Marseille, France 2013, 6 pages.
Mermelstein, P., “Distance Measures for Speech Recognition—Psychological and Instrumental,” Pattern Recognition and Artificial Intelligence, 1976, pp. 374-388.
Schmalenstroeer, J., et al., “Online Diarization of Streaming Audio-Visual Data for Smart Environments,” IEEE Journal of Selected Topics in Signal Processing, vol. 4, No. 5, 2010, 12 pages.
Viterbi, A.J., “Error Bounds for Convolutional Codes and an Asymptotically Optimum Decoding Algorithm,” IEEE Transactions on Information Theory, vol. 13, No. 2, 1967, pp. 260-269.
Related Publications (1)
Number Date Country
20200043501 A1 Feb 2020 US
Provisional Applications (2)
Number Date Country
61729067 Nov 2012 US
61729064 Nov 2012 US
Continuations (2)
Number Date Country
Parent 16170306 Oct 2018 US
Child 16594764 US
Parent 14084974 Nov 2013 US
Child 16170306 US