The invention relates to speech recognition and, more particularly, to improving the accuracy and efficiency of speech recognition systems.
Example Applications for Speech Recognition Systems
Speech recognition systems have simplified many tasks particularly for a user in the workplace by permitting the user to perform hands-free communication with a computer as a convenient alternative to communication via conventional peripheral input/output devices. For example, a user could wear a wireless wearable terminal having a speech recognition system that permits communication between the user and a central computer system so that the user can receive work assignments and instructions from the central computer system. The user could also communicate to the central computer system information such as data entries, questions, work progress reports, and working condition reports. In a warehouse or inventory environment, a user can be directed (through an instruction from the central computer system or visually by means of a display) to a particular work area that is labeled with a multiple-digit number (check-digit) such as “1-2-3” and asked to speak the check-digit. The user would then respond with the expected response “1-2-3”. (Note that a “check-digit” can be any word or sequence of words, and is not limited to digits.) Other such examples of communication between a user and speech recognition system are described in U.S. Patent Application Serial No. 2003/0154075 and include environments where a wearable or portable terminal is not required such as in an automobile or a telephone system; environments that are not in a warehouse such as in a pharmacy, retail store, and office; voice-controlled information processing systems that process for example credit card numbers, bank account numbers, social security numbers and personal identification numbers; other applications such as command and control, dictation, data entry and information retrieval applications; and speech recognition system features such as user verification, password verification, quantity verification, and repeat/acknowledge messages. The inventions presented here can be used in those applications. In using a speech recognition system, manual data entry is eliminated or, at the least, reduced, and users can perform their tasks faster, more accurately, and more productively.
Errors can be made by a speech recognition system however, due to for example background noise or a user's unfamiliarity or misuse of the system. The errors made by a system can be classified into various types. A metric, an error rate (which can be defined as the percentage or ratio of observations with speech recognition errors over the number of observations of the system and which can be determined over a window of time and/or data and per user) is often used to evaluate the number and types of errors made by a speech recognition system and is thus useful in evaluating the performance of the system. An observation can be defined as any speech unit by which speech recognition may be measured. An observation may be a syllable, a phoneme, a single word or multiple words (such as in a phrase, utterance or sentence). When counting the number of observations of the system, the observations input to the system may be counted or the observations output by the system may be counted. One skilled in the art will also know and understand that an accuracy rate (which can be defined as the percentage or ratio of correct observations of the system over the number of observations of the system and which can be determined over a window of time and/or date and per user) can be used to evaluate the performance of the system. Therefore, a recognition rate (which can be an error rate, an accuracy rate, or other type of recognition rate) is useful in evaluating the performance of the system. In general, a recognition rate can be determined for a word or for various words among a set of words, or for a user or multiple users. Identification of a system's errors can be done by comparing a reference transcription of a user's input speech to the hypothesis generated by the system (the system's interpretation of the user's input speech). Furthermore, as known to those skilled in the art, the comparison can be time-aligned mode or text-aligned.
One type of speech recognition error is a substitution, in which the speech recognition system's hypothesis replaces a word that is in the reference transcription with an incorrect word. For example, if system recognizes “1-5-3” in response to the user's input speech “1-2-3”, the system made one substitution: substituting the ‘5’ for the ‘2’.
Another type of speech recognition error is a deletion, in which the speech recognition system's hypothesis lacks a word that is in the reference transcription. For example, if system recognizes “1-3” in response to the user's input speech “1-2-3”, the system deleted one word, the ‘2’. One variation of the deletion error is a deletion due to recognizing garbage, in which the system erroneously recognizes a garbage model instead of recognizing an actual word. Another variation of the deletion error is a deletion due to a speech misdetection, where the system fails to detect that the audio input to the system contains speech and as a result does not submit features of the audio input to the system's search algorithm. Another type of deletion occurs when the system rejects a correct observation due to a low confidence score. Yet another variation of the deletion error is a deletion due to a rejected substitution, where a search algorithm of the speech recognition generates a substitution which is later rejected by an acceptance algorithm of the system. Still another type of deletion, occurring in time-aligned comparisons, is a merge: the speech recognition system recognizes two spoken words as one. For example, the user says “four-two” and the system outputs “forty”.
In this application, a garbage model refers to the general class of models for sounds that do not convey information. Examples may include for example models of breath noises, “um”, “uh”, sniffles, wind noise, the sound of a pallet dropping, the sound of a car door slamming, or other general model such as a wildcard that is intended to match the input audio for any audio that doesn't match a model in the library of models.
Yet another type of speech recognition error is an insertion, in which the speech recognition system's hypothesis includes a word (or symbol) that does not correspond to any word in the reference transcription. Insertion errors often occur when the system generates two symbols that correspond to one symbol. One of these symbols may correspond to the reference transcription and be tagged as a correct observation. If it does not correspond to the reference transcription, it can be tagged as a substitution error. In either case, the other symbol can be tagged as an insertion error. Insertion errors are common when noise is mistakenly recognized as speech.
In contrast to determining that an actual error occurred by comparing a system's hypothesis to words actually spoken in a reference transcript, an error can be estimated or deemed to have occurred based on system behavior and user behavior. This application describes methods for determining a recognition rate, wherein the recognition rate is an estimate based on estimated errors or correct observations deemed to have occurred after evaluating system and user behavior. Accordingly, one can estimate or evaluate the performance level of the speech recognition system by detecting in this manner the various errors committed by or correct observations of the system. One way to detect a speech recognition error is based on feedback a user provides to the speech recognition system. Feedback can be requested by the speech recognition system. For example, the system could ask the user to confirm the system's hypothesis by asking the user for example “Did you say 1-5-3?”, and if the user responds “no”, it indicates that the system made an error recognizing “1-5-3”. Another type of feedback is based on a user's emotion detected by speech recognition. For example, if the system recognizes in the user's input speech that the user is sighing or saying words indicating aggravation, it may indicate that an error occurred. Yet another type of feedback is based on a user's correction command to the system, such as the user speaking “back-up” or “erase”, or the user identifying what word was spoken (which could be from a list of possible words displayed by the system). When a correction is commanded to the system, it may be that an error occurred.
Model Adaptation for a Speech Recognition System
A speech recognition system can improve its performance over time, as more speech samples are received and processed by a speech recognition system, by improving its acoustic models through training or other learning or adaptation algorithms. At the same time, it is desirable to prevent the system from adapting in an undesirable way, thereby resulting in a system that performs worse than it did prior to adaptation or a system that degrades over time. Avoiding additional processing by a speech recognition system due to adaptation of acoustic models is particularly useful in applications employing a battery powered mobile computer, wireless network, and server to store models. Adapting models can use significant computational resources to create the adapted models and radio transmission energy to transmit the new models to the server. Example embodiments of the invention disclosed herein can control adaptation of the speech recognition system to avoid inefficient use of computational, storage and/or power resources and to avoid adapting away from well-performing models. Example embodiments of the invention control adaptation by using triggers, which are based on a recognition rate determination or estimation and the accuracy of the recognition rate determination or estimation, to cause the adaptation of prior models or create new models.
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate speech recognition system components and embodiments of the invention and, together with the detailed description of the embodiments given below, serve to explain the principles of the invention.
A speech recognition system can be improved by adaptive learning or training of the acoustic models used by the speech recognition system. The example embodiments described herein involve situations wherein improved methods for determining or estimating an accurate recognition rate are utilized to improve adaptation by a speech recognition system. Note that in this description, references to “one embodiment” or “an embodiment” mean that the feature being referred to is included in at least one embodiment of the invention. Further, separate references to “one embodiment” in this description do not necessarily refer to the same embodiment; however, neither are such embodiments mutually exclusive, unless so stated and except as will be readily apparent to those skilled in the art. Thus, the invention can include any variety of combinations and/or integrations of the embodiments described herein.
One approach for efficient use of resources of a speech recognition system includes determining a recognition rate, corresponding to either recognition of instances of a word or recognition of instances of various words among a set of words, determining an accuracy range of the recognition rate, corresponding to a desired confidence level. In addition, the approach may include using a recognition rate threshold and the desired confidence level upon which to base model adaptation, wherein a model for the word is adapted or various models for the various words are adapted, based upon a comparison of at least one value in the accuracy range with the recognition rate threshold. The approach may be implemented as an apparatus, which may include all or a subset of the following: a processor adapted to determine a recognition rate, corresponding to either recognition of instances of a word or recognition of instances of various words among a set of words, and determine an accuracy range of the recognition rate corresponding to a desired confidence level; and a controller adapted to use a recognition rate threshold and the desired confidence level upon which to base model adaptation, by adapting a model for the word or various models for the various words based upon a comparison of at least one value in the accuracy range with the recognition rate threshold.
Another approach for efficient use of resources of a speech recognition system includes using a recognition rate threshold and a desired confidence level of a recognition rate upon which to base adaptation of a model. The approach may further include determining a recognition rate and determining an accuracy range of the recognition rate corresponding to the desired confidence level, the accuracy range related to a number of observations. Furthermore, the approach may include using a relationship between the recognition rate threshold, the recognition rate and the number of observations, which balances the accuracy of the recognition rate with expediency in making a model adaptation decision. And the approach may include basing a decision for model adaptation on using the relationship. The approach may be implemented as an apparatus, which may include all or a subset of the following: a selector adapted to use a recognition rate threshold and a desired confidence level of a recognition rate upon which to base adaptation of a model; a processor adapted to determine a recognition rate and an accuracy range of the recognition rate corresponding to the desired confidence level, the accuracy range related to a number of observations; and a controller adapted to base a decision for model adaptation on using a relationship between the recognition rate threshold, the recognition rate and the number of observations, which balances the accuracy of the recognition rate with expediency in making a model adaptation decision.
Yet in another approach, a method for optimizing determination of a recognition rate of a speech recognition system includes determining a number of observations that would yield a desired accuracy range of the recognition rate. The approach may be implemented as an apparatus, which may include a processor adapted to determine a number of observations that would yield a desired accuracy range of the determination.
Still another approach for efficient use of resources of a speech recognition system includes determining a recognition rate, corresponding to either recognition of instances of a word or recognition of instances of various words among a set of words, and determining an accuracy range of the recognition rate. The approach may further include adjusting adaptation of a model for the word or various models for the various words, based on a comparison of at least one value in the accuracy range with a recognition rate threshold. The approach may be implemented as an apparatus which may include all or a subset of the following: a processor adapted to determine a recognition rate corresponding to either recognition of instances of a word or recognition of instances of various words among a set of words, and an accuracy range of the recognition rate; and a controller adapted to adjust adaptation of a model for the word or various models for the various words, based on a comparison of at least one value in the accuracy range with a recognition rate threshold.
Another approach for efficient use of resources of a speech recognition system includes selecting a recognition rate threshold upon which to base, at least in part, adaptation or change of a speech recognition model, determining a recognition rate corresponding to either recognition of instances of a word or recognition of instances of various words among a set of words, determining an accuracy range of the recognition rate, and adapting or changing a model for the word or various models for the various words, based on a comparison of the recognition rate, within the limits of the determined accuracy range, with the recognition rate threshold. The approach may be implemented as an apparatus which may include all or a subset of the following: a selector adapted to select a recognition rate threshold upon which to base, at least in part, adaptation or change of a speech recognition model, a processor adapted to determine a recognition rate corresponding to either recognition of instances of a word or recognition of instances of various words among a set of words, and an accuracy range of the recognition rate, and a controller adapted to control adapting or changing a model for the word or various models for the various words, based on a comparison of the recognition rate, within the limits of the determined accuracy range, with the recognition rate threshold.
An approach for expedited model adaptation for a speech recognition system includes selecting a recognition rate threshold upon which to base at least in part adaptation or change of a speech recognition model, determining a recognition rate, corresponding to either recognition of instances of a word or recognition of instances of various words among a set of words, using a process in which increasing numbers of observations result in greater accuracy in determining the recognition rate, and deciding when to adapt or change a model for the word or various models for the various words, as a function of the distance of a determined recognition rate from the recognition rate threshold. The approach may be implemented as an apparatus which may include all or a subset of the following: a selector adapted to select a recognition rate threshold upon which to base, at least in part, adaptation or change of a speech recognition model, a processor adapted to determine a recognition rate corresponding to either recognition of instances of a word or recognition of instances of various words among a set of words, using a process in which increasing numbers of observations result in greater accuracy in determining the recognition rate, and a controller adapted to decide when to adapt or change a model for the word or various models for the various words, as a function of the distance of a determined recognition rate from the recognition rate threshold. In either of the two approaches, less accuracy, and thus fewer observations may be required to make a model adaptation or change based on a determined recognition rate distant from the recognition rate threshold than a determined recognition rate close to the recognition rate threshold.
Referring to
In particular, the signal processor 104 divides the digital stream of data that is created into a sequence of time-slices, or frames 105, each of which is then processed by a feature generator 106, thereby producing features (vector, matrix, or otherwise organized set of numbers representing the acoustic features of the frames) 107. Further explanation of an example speech recognition system is provided in U.S. Pat. No. 4,882,757, entitled “Speech Recognition System”, the disclosure of which is incorporated herein by reference in its entirety. This referenced patent discloses Linear Predictive Coding (LPC) coefficients to represent speech; however, other functionally equivalent methods are contemplated within the scope of the invention as well.
A speech recognition search algorithm function 108, realized by an appropriate circuit and/or software in the system 100 analyzes the features 107 in an attempt to determine what hypothesis to assign to the speech input captured by input device 102. As is known in the art in one recognition algorithm, the recognition search 108 relies on probabilistic models provided through 122 from a library of suitable models 110 to recognize the speech input 102. Each of the models in the library 110 may either be customized to a user or be generic to a set of users.
When in operation, the search algorithm 108 (which can be implemented using Hidden Markov Models with a Viterbi algorithm or other modeling techniques such as template matching dynamic time warping (DTW) or neural networks), in essence, compares the features 107 generated in the generator 106 with reference representations of speech, or speech models, in library 110 in order to determine the word or words that best match the speech input from device 102. Part of this recognition process is to assign a confidence factor for the speech to indicate how closely the sequence of features 107 used in the search algorithm 108 matches the closest or best-matching models in library 110. As such, a hypothesis consisting of one or more vocabulary items and associated confidence factors 111 is directed to an acceptance algorithm 112, which also can take as input a threshold adjustment 116 and one or more expected responses 114. If the confidence factor is above a predetermined acceptance threshold, then the acceptance algorithm 112 makes a decision 118 to accept the hypothesis as recognized speech. If, however, the confidence factor is not above the acceptance threshold, as utilized by the acceptance algorithm, then the acceptance algorithm 112 makes a decision 118 to ignore or reject the recognized speech. The system may then ignore the input or prompt the user to repeat the speech. In this instance, the user may repeat the speech to input device 102. The hypothesis and confidence factors 111, the expected response 114, acceptance algorithm decision 118 and features 107 can also be input to a model adaptation control module 117. Model adaptation control module 117 (which may be implemented in a hardware or software controller or control mechanism) controls the adaptation of library of models 110.
Recognition rate module 210 also determines the accuracy range of the recognition rate that meets the desired confidence level (which may be selected beforehand and be implemented either in hardware or software). Inputs 205 are those needed for a recognition rate determination used for a particular application. In this example embodiment, inputs 205 are a hypothesis and confidence factor (such as 111 of
Adaptation control module 225 controls or adjusts the adaptation of models by model adaptation module 235. Inputs 220 are those needed for the particular control of model adaptation desired for a particular application. In this example embodiment, inputs 220 are a hypothesis and features (such as 107 of
In an example embodiment, an adaptation control module (such as 225 of
The techniques illustrated in the flowcharts of
An example embodiment of 315 in
where,
# errors is the number of observations with speech recognition errors,
n is the number observations of the system, and
RATEerror is the value of the error rate.
In another example embodiment, the recognition rate is an accuracy rate, determined using the equation:
where,
# CorrectObservations is the number of correct observations of the system,
n is the number of observations of the system, and
RATEaccuracy is the value of the accuracy rate.
In other example embodiments, the recognition rate can be determined or estimated or in the following ways: over a window of time; over a window of data observed by the system; per user; over a number of users; per word; for a set of words. Furthermore, the recognition rate need not be a word recognition rate, but can another type of recognition rate such as a syllable recognition rate, a phoneme recognition rate, a phrase recognition rate, an utterance recognition rate, and a sentence recognition rate. Therefore, n can be counted in view of any one or more of these parameters. However, for this discussion of various example embodiments of the invention, a word recognition rate in which n is the number of words input to the system will be used.
An example embodiment of 320 in
Specifically, a two-sided accuracy range of a recognition rate can be determined (assuming a large number of observations: e.g., n is greater than 100) as follows:
where,
z is a value that is based on a desired confidence level,
n is the number of observations of the system,
RATE is the value of the recognition rate, and
RATEaccuracy is the two-sided accuracy range of the recognition rate.
Example z values for certain confidence levels are:
for a 95% confidence level, z=1.96
for a 90% confidence level, z=1.645
for a 80% confidence level, z=1.282
The two values generated by equation (2) are the low-end and the high-end values of the two-sided accuracy range of the recognition rate, for a certain confidence level. (Other equations may be used when n is not greater than 100). One skilled in the art will know and understand that there are other equations which can provide a range based on a confidence level, that equation (2) is an example embodiment method to provide an accuracy range, and the invention is not restricted in just using it.
To illustrate use of equation (2), if RATE=2%, n=600 words, z=1.96 (for a 95% confidence level), the two-sided accuracy range of the recognition rate is 0.9% to 3.1%. In other words, in this example, there is a 95% level of confidence that the recognition rate is in the range of 0.9% to 3.1%.
A one-sided accuracy range of a recognition rate can be determined (assuming n is greater than 100) as follows:
where,
z is a value that is based on a desired confidence level,
n is the number of observations of the system,
RATE is the recognition rate, and
RATEaccuracy is the high-end value of a one-sided accuracy range of the recognition rate.
Example z values for certain confidence levels are:
for a 95% confidence level, z=1.645
for a 90% confidence level, z=1.282
for a 80% confidence level, z=0.482
The value generated by equation (3) is the high-end value of a one-sided accuracy range of the recognition rate. Alternatively, one can generate a low-end value of a one-sided accuracy range of the recognition rate by modifying equation (3) to have a negative sign “−” instead of a positive sign “+”, i.e.,
To illustrate use of equation (3), if RATE=2%, n=600 words, z=1.645 (for a 95% confidence level), the high-end value of a one-sided accuracy range of the recognition rate is 2.9%. In other words, in this example, there is a 95% level of confidence that the recognition rate is a value below 2.9%. To illustrate use of equation (4), if RATE=2%, n=600 words, z=1.645 (for a 95% confidence level), the low-end value of a one-sided accuracy range of the recognition rate is 1.1%. In other words, in this example, there is a 95% level of confidence that the recognition rate is a value above 1.1%.
An example embodiment of 325 in
In Table 1, example error rates are used for the recognition rate. In these examples, the low-end value in the accuracy range of an error rate is used to determine whether to perform model adaptation. If this value is below the recognition rate threshold, the speech recognition system may be performing well and accordingly, the adaptation decision is that adaptation should not be performed. If the value is above or equal to the recognition rate threshold, it is presumed that the system is not performing well and the adaptation decision is that adaptation should be performed.
In Table 2, example accuracy rates are used for the recognition rate. In these examples, the high-end value in the accuracy range of an accuracy rate to determine whether to perform model adaptation. If this value is above the recognition rate threshold, the speech recognition system may be performing well and accordingly, the adaptation decision is that adaptation should not be performed. If the value is below or equal to the recognition rate threshold, it is presumed that the system is not performing well and the adaptation decision is that adaptation should be performed.
It can be understood that values other than the low-end value of an accuracy range of an error rate can be used in a comparison with a recognition rate threshold in order to determine whether to adapt. For example, values in the lower half of the accuracy range of the error rate can be used in comparison with the recognition rate threshold. In another example, the high-end value of an accuracy range or values in the upper half of the accuracy range can be used in comparison with the recognition rate threshold.
Likewise, values other than the high-end value of an accuracy range of an accuracy rate can be used in a comparison with a recognition rate threshold in order to determine whether to adapt.
Relationships Between the Number of Observations of the Speech Recognition System, the Accuracy of the Recognition Rate and the Recognition Rate Threshold
The inventors have recognized that there is a relationship between the number of observations of the speech recognition system and the accuracy of the recognition rate. As the number of observations increases, the accuracy range of the recognition rate decreases thereby producing more reliable determinations of the recognition rate. (In other words, the accuracy range is inversely related to the number of observations.) Equations (1)-(4) above illustrate this relationship. Also, the higher the desired confidence level for the accuracy of recognition rate, the number of observations required to achieve the desired confidence level is greater for the same accuracy range. For example, if the desired confidence level of the recognition rate is selected at a 95%, more observations are needed to achieve the 95% desired confidence level than if the desired confidence level was selected at 80% for the same accuracy range. Using this relationship, model adaptation can be optimized by balancing the selected desired confidence level against the number of observations required to achieve the confidence level for a given accuracy range.
Another relationship that the inventors have recognized is the relationship between the number of observations of the speech recognition system, the recognition rate and the recognition rate threshold. Specifically, the further away the recognition rate is from the recognition rate threshold, fewer observations are needed to be confident in the decision of whether to adapt. Table 3 illustrates this relationship using an error rate for the recognition rate. In all examples shown, there is a decision to adapt because the low-end value of the accuracy range of the error rate is equal to or exceeds the error rate threshold. However, the number of observations n required to arrive at this decision to adapt (at a 95% confidence level), dramatically decreases as the difference between the error rate and the error rate threshold increases. Therefore, as the difference between the error rate and the error rate threshold increases, a decision to adapt is made with increasing expediency, since fewer observations are needed to arrive at the decision to adapt due to the increased difference between the determined error rate and the error rate threshold.
Throughout this present application, there are various example embodiments for determining or estimating the occurrences of possible (or potential or suspected) errors made by a speech recognition system and an error rate (which can be performed by the recognition rate module 210 of
Low Confidence Rate
In an example embodiment of the invention, a count of occurrences of possible errors made by a speech recognition system called a low confidence rate, can be used to determine an estimate of a low confidence rate or an estimate of an error rate.
In
An example embodiment, which uses a low confidence rate, also considers when a word is from a hypothesis generated by the system that matches an expected response in counting errors for an error rate estimation. (U.S. patent application Ser. No. 11/051,825, and the BACKGROUND section of this present application describes scenarios in which an expected response from a user is processed by a speech recognition system. An expected response can be defined as a response that the system expects to receive from the user, as a result of the application in which the system is used). In an example embodiment in the referenced patent application, a matching algorithm of the system normally requires that the system's hypothesis is accepted only if a confidence factor for the hypothesis exceeds an acceptance threshold. However, when the system's most likely hypothesis matches an expected response, the hypothesis is more favorably treated so that the hypothesis may be accepted by the system. The reasoning behind the favorable treatment despite the relatively low confidence factor is that a hypothesis matching an expected response usually indicates a high probability of correct recognition.
Turning back to the example embodiment of the invention, in which the error rate is a low confidence rate, responses that match the expected response and have a relatively low confidence factor for the application in which the system is used are counted as errors for an error rate estimation. Although a recognition error may not have actually occurred (because the system's hypothesis was correctly accepted due to the hypothesis matching the expected response as described in referenced U.S. patent application Ser. No. 11/051,825), in this example embodiment, a word with a relatively low confidence is counted as an error for an error rate estimation due to the relatively low confidence factor. The range of confidence factors for which a word is counted as a low confidence could be, for example, between the adjusted acceptance threshold and the original, unadjusted acceptance threshold. More generally, the confidence factor thresholds or range for the counting low confidence errors do not need to match the acceptance threshold and adjusted acceptance threshold in the referenced patent application. The range could be between two other thresholds, including a high confidence threshold, which is higher than the acceptance threshold and indicates the boundary between low and high confidence. In this example embodiment, the range of confidence factors used for the low confidence rate is determined based on the application in which the speech recognition system is used.
Substitution Rate
In example embodiment of the invention, a count of occurrences of possible substitution errors made by a speech recognition system can be used to determine an estimate of a substitution error rate or an estimate of an error rate. The substitution rate is the rate at which substitution errors (such as the substitution errors defined in the BACKGROUND section of this present application) are made by a system. In an example embodiment, a hypothesis generated by the speech recognition system is compared to an expected response and a substitution error occurs if the system replaces a word in the expected response with an incorrect word in the hypothesis. For example, if the system recognizes “1-5-3” and the expected response is “1-2-3”, a substitution error is counted because it is deemed that the system made one substitution: substituting the ‘5’ for the ‘2’. In other words, if the hypothesis and the expected response do not match word-for-word, but do mostly match, (i.e. the hypothesis and the expected response match except for a predetermined number of words), it is a reasonable assumption that a word substitution error has occurred. (The predetermined number of words depends upon the application. For example, an application that uses three-word hypotheses or utterances may define “mostly match” as matching word-for-word except for one word. An application that uses five-word hypotheses or utterances may define “mostly match” as matching word-for-word except for two words.)
Repeated Utterances
Yet in other example embodiments, the error rate is based on a recognition error made by the speech recognition system that is realized after comparing the speech recognition system's decision on its hypothesis of at least two consecutive or proximate utterances. The decision can occur after the speech recognition system has processed the incoming utterances (such as at 118 of
Reject and Repeat
In
Substitute and Repeat
In
The same approach as in the previous paragraph can be used to detect deletion due to garbage errors where a content word is recognized as garbage in a first utterance, then correctly recognized in the next utterance. By comparing the recognition results of the two utterances and using verifications such as those described above, one can detect the error. For example, if the system's hypothesis of the first utterance is “1-GARBAGE-3” and the system's hypothesis of the second utterance is “1-5-3”, there is a mismatch of one word, and it becomes a reasonable assumption that the speech recognition system made an error in its hypothesis of the first utterance. Again, similar verifications as described above may be used to guard against the system considering a correct observation to be in error.
The same approach as described above in the discussion of
Correction Rate
In an example embodiment error rate for a speech recognition system, a correction rate at which a user provides feedback to the system can be used as an estimate of an error rate or an estimate for part of an error rate. The reasoning behind using a correction rate to estimate an error rate or estimate part of an error rate is that when a correction is commanded to the system, it may indicate that an error occurred. Examples of user feedback are described in the BACKGROUND section of this present application. The correction rate can include the rate at which the user indicates that the system made a mistake. Furthermore, the user may provide feedback in response to the system requesting feedback, such as asking the user to confirm a hypothesis generated by the system or asking the user to identify what word was spoken by the user. The feed back may include a word indicating aggravation by the user or the feed back may be a correction command to the system, such as “back-up” or “erase”.
Recognition Rate Considerations
In determining or estimating the recognition rate (error rate, accuracy rate or other type of recognition rate), considerations can be made for the amount of time and data needed to determine or estimate a recognition rate that is useful for the application in which the speech recognition system is used. One example consideration is that the recognition rate is determined or estimated for speech input to the speech recognition system over a predetermined period of time. Other example considerations are that the recognition rate is determined or estimated for speech input to the speech recognition system over a predetermined number of utterances, words, or hypotheses.
Other example considerations are that the recognition rate is determined or estimated from hypotheses of utterances collected over a moving or sliding window or a collection period that is dynamic in period of time and/or size of data. As a result, the recognition rate is determined or estimated over a period when useful data has been collected. For example, a moving or sliding window can cover a collection of data taken from equal periods in noisy environment and a quiet environment to offset any favoring by the speech recognition system in one of those environments. Other examples of moving, sliding windows are those that collect data only during recent use (e.g. the last half-hour) of the speech recognition system, collecting data for time spent by a particular user (e.g. since the user started using the system), or collecting a certain amount of data regardless of the time spent collecting the data (e.g. the last one-hundred hypotheses).
It can be understood by those skilled in the art that in other example embodiments of the invention, other recognition rates can be used in place of the recognition rate, such as a phrase recognition rate, utterance recognition rate, and sentence recognition rate. For example, a recognition rate can be defined as the percentage or ratio of either utterances with errors or correctly recognized utterances made by a system divided by the number of utterances input to the system and this utterance recognition rate can be used in controlling or adjusting an adaptation of a model.
In addition, it will be evident to one skilled in the art that the various methods to identify possible errors or correct observations can process the same hypotheses, but safeguards must be taken to avoid double counting, as some possible errors or correct observations may be counted by more than one method. Furthermore, in example embodiments of the invention, the desired confidence level and/or accuracy range of the recognition rate and/or recognition rate threshold can be a predetermined value(s), a value(s) settable by a user, a dynamic value(s), or it can be adjusted upwardly or downwardly. Moreover, the desired confidence level and/or accuracy range can be based on factors that affect the achievable recognition rate of the speech recognition system and those that determine an acceptable recognition rate for the application in which the system is used.
Furthermore, the desired confidence level and/or accuracy range and/or recognition rate threshold can be based on a number of words in an utterance input to the speech recognition system, based on a background noise level, based on a signal-to-noise ratio, based on a number of words in the vocabulary of a speech recognition system or based on a recognition rate corresponding to a particular user or users of the system.
At 710, the features observed by a speech recognition system corresponding to an input utterance are aligned with the states in the models for the words of the utterance. In an example embodiment, the Baum-Welch re-estimation algorithm can be used to perform the alignment. At 715, the statistics (for example, means and variances) of the states are updated using the values of the features. At 720, these values are mixed into the models with an appropriate weighting to maintain a balance between previous training data and new features. Thus, in an example embodiment of the invention, new models are created through adaptation by using the observed features of an input utterance to adapt existing or original models. In that scenario, both the observed features of the input utterance and the existing features of the original models, and the statistics associated with each, are used to create the new models. Additionally, in such a scenario, the new statistics might be weighted in various fashions to tailor their effect on the original statistics in the model. In an alternative example embodiment of the invention, only the new observed features, and information therefrom, are utilized to create the new model. That is, a new model is created with the new features only. Furthermore, the adaptation could be performed using data from a single user or multiple users. In one particular embodiment, only speech data from an individual user might be used to perform the adaptation. This generates a model that is adapted for that user and performs well for that user.
The invention, in its various forms, may be implemented directly in the software of a speech recognition system. That is, the improvements are actually part of the speech recognition system. Alternatively, the invention does not have to be built into the speech recognition system. Rather, the invention or parts of the invention may be implemented in a separate program or application may be utilized by a speech recognition system to provide the benefits of the invention. For example, a separate application or software module may be utilized to handle the adaptation in accordance with the principles of the invention. Specifically, an application may interface with a speech recognition system to determine a recognition rate and/or control when and how models are adapted.
In the foregoing description, the invention is described with reference to specific example embodiments thereof. The specification and drawings are accordingly to be regarded in an illustrative rather than in a restrictive sense and it is not the intention of the applicants to restrict or in any way limit the scope of the appended claims to such detail. It will, however, be evident to those skilled in the art that additional advantages and modifications can be made, in a computer program product or software, hardware or any combination thereof, without departing from the broader spirit and scope of the invention will readily appear. Software embodiments may include an article of manufacture on a machine accessible or machine readable medium having instructions. Furthermore, software embodiments may be distributed or downloaded via a network or email. The instructions on the machine accessible or machine readable medium may be used to program a computer system such as for example, a PC, cell phone, industrial mobile computer, PDA, electronic headset or other electronic device to perform example embodiment methods or approaches described herein. The machine-readable medium may include, but is not limited to, floppy diskettes, optical disks, CD-ROMs, and magneto-optical disks or other type of media/machine-readable medium suitable for storing or transmitting electronic instructions. Furthermore, departures may be made from the application in which the invention is described without departing from the spirit and scope of the invention. For example, the example speech recognition system described herein has focused on wearable terminals. However, the principles of the invention are applicable to other speech recognition environments as well.
This application is a continuation-in-part of U.S. patent application Ser. No. 11/539,456, entitled “Methods and Systems for Adapting a Model for a Speech Recognition System” filed Oct. 6, 2006, which is a continuation-in-part of U.S. patent application Ser. No. 11/331,649, entitled “Methods and Systems for Considering Information About an Expected Response When Performing Speech Recognition” filed Jan. 13, 2006, which is a continuation-in-part application of U.S. patent application Ser. No. 11/051,825, entitled “Method and System for Considering Information About an Expected Response When Performing Speech Recognition” filed Feb. 4, 2005, which all applications are incorporated herein by reference in entirety. This application also claims the benefit of U.S. Provisional Application No. 60/788,621, entitled “Methods and Systems for Optimizing Model Adaptation for a Speech Recognition System”, filed Apr. 3, 2006, U.S. Provisional Application No. 60/788,606, entitled “Methods and Systems for Adapting a Model for a Speech Recognition System”, filed Apr. 3, 2006, and U.S. Provisional Application No. 60/788,622, entitled “Method and Systems for Assessing and Improving the Performance of a Speech Recognition System”, filed on Apr. 3, 2006, which all applications are incorporated herein by reference in entirety.
Number | Name | Date | Kind |
---|---|---|---|
4882757 | Fisher et al. | Nov 1989 | A |
4928302 | Kaneuchi et al. | May 1990 | A |
4959864 | Van Nes | Sep 1990 | A |
5127043 | Hunt et al. | Jun 1992 | A |
5127055 | Larkey | Jun 1992 | A |
5230023 | Nakano | Jul 1993 | A |
5297194 | Hunt et al. | Mar 1994 | A |
5428707 | Gould | Jun 1995 | A |
5457768 | Tsuboi et al. | Oct 1995 | A |
5465317 | Epstein | Nov 1995 | A |
5488652 | Bielby | Jan 1996 | A |
5566272 | Brems et al. | Oct 1996 | A |
5602960 | Hon et al. | Feb 1997 | A |
5625748 | McDonough et al. | Apr 1997 | A |
5640485 | Ranta | Jun 1997 | A |
5644680 | Bielby | Jul 1997 | A |
5651094 | Takagi et al. | Jul 1997 | A |
5684925 | Morin et al. | Nov 1997 | A |
5710864 | Juang et al. | Jan 1998 | A |
5717826 | Setiu et al. | Feb 1998 | A |
5737489 | Chou et al. | Apr 1998 | A |
5737724 | Atal et al. | Apr 1998 | A |
5774841 | Salazar | Jun 1998 | A |
5774858 | Taubkin et al. | Jun 1998 | A |
5797123 | Chou et al. | Aug 1998 | A |
5799273 | Mitchell | Aug 1998 | A |
5832430 | Lleida et al. | Nov 1998 | A |
5839103 | Mammone et al. | Nov 1998 | A |
5842163 | Weintraub | Nov 1998 | A |
5893057 | Fujimoto et al. | Apr 1999 | A |
5893902 | Transue et al. | Apr 1999 | A |
5895447 | Ittycheriah et al. | Apr 1999 | A |
5899972 | Miyazawa et al. | May 1999 | A |
5946658 | Miyazawa et al. | Aug 1999 | A |
5960447 | Holt | Sep 1999 | A |
6003002 | Netsch | Dec 1999 | A |
6006183 | Lai et al. | Dec 1999 | A |
6073096 | Gao et al. | Jun 2000 | A |
6088669 | Maes | Jul 2000 | A |
6094632 | Haitori | Jul 2000 | A |
6101467 | Bartosik | Aug 2000 | A |
6122612 | Goldberg | Sep 2000 | A |
6151574 | Lee et al. | Nov 2000 | A |
6182038 | Batakrishnan et al. | Jan 2001 | B1 |
6192343 | Morgan et al. | Feb 2001 | B1 |
6205426 | Nguyen et al. | Mar 2001 | B1 |
6230129 | Morin et al. | May 2001 | B1 |
6233555 | Parthasarathy et al. | May 2001 | B1 |
6233559 | Batakrishnan | May 2001 | B1 |
6243713 | Nelson et al. | Jun 2001 | B1 |
6292782 | Weideman | Sep 2001 | B1 |
6330536 | Parthasarathy et al. | Dec 2001 | B1 |
6374212 | Phillips et al. | Apr 2002 | B2 |
6374221 | Haimi-Cohen | Apr 2002 | B1 |
6377662 | Hunt et al. | Apr 2002 | B1 |
6377949 | Gilmour | Apr 2002 | B1 |
6397180 | Jaramillo et al. | May 2002 | B1 |
6421640 | Dolfing et al. | Jul 2002 | B1 |
6438519 | Campbell et al. | Aug 2002 | B1 |
6438520 | Curt et al. | Aug 2002 | B1 |
6487532 | Schoofs et al. | Nov 2002 | B1 |
6496800 | Kong et al. | Dec 2002 | B1 |
6505155 | Vanbuskirk | Jan 2003 | B1 |
6507816 | Ortega | Jan 2003 | B2 |
6526380 | Thelen et al. | Feb 2003 | B1 |
6539078 | Hunt et al. | Mar 2003 | B1 |
6542866 | Jiang et al. | Apr 2003 | B1 |
6567775 | Maali et al. | May 2003 | B1 |
6571210 | Hon et al. | May 2003 | B2 |
6581036 | Varney, Jr. | Jun 2003 | B1 |
6587824 | Everhart | Jul 2003 | B1 |
6594629 | Basu et al. | Jul 2003 | B1 |
6598017 | Yamamoto et al. | Jul 2003 | B1 |
6606598 | Holthouse | Aug 2003 | B1 |
6629072 | Thelen et al. | Sep 2003 | B1 |
6675142 | Ortega | Jan 2004 | B2 |
6732074 | Kuroda | May 2004 | B1 |
6735562 | Zhang et al. | May 2004 | B1 |
6754627 | Woodward | Jun 2004 | B2 |
6766295 | Murvett | Jul 2004 | B1 |
6799162 | Goronzy | Sep 2004 | B1 |
6832224 | Gilmour | Dec 2004 | B2 |
6834265 | Balasuriya | Dec 2004 | B2 |
6868381 | Peters | Mar 2005 | B1 |
6879956 | Honda et al. | Apr 2005 | B1 |
6882972 | Kompe et al. | Apr 2005 | B2 |
6961700 | Mitchell | Nov 2005 | B2 |
6961702 | Dobler | Nov 2005 | B2 |
7031918 | Hwang | Apr 2006 | B2 |
7039166 | Peterson et al. | May 2006 | B1 |
7050550 | Steinbiss | May 2006 | B2 |
7062441 | Townshend | Jun 2006 | B1 |
7065488 | Yajima et al. | Jun 2006 | B2 |
7072750 | Pi | Jul 2006 | B2 |
7203651 | Baruch | Apr 2007 | B2 |
7216148 | Matsunami | May 2007 | B2 |
7266492 | Goodman | Sep 2007 | B2 |
7392186 | Duan et al. | Jun 2008 | B2 |
7406413 | Geppert et al. | Jul 2008 | B2 |
7457745 | Kadambe et al. | Nov 2008 | B2 |
20010016816 | Lucke | Aug 2001 | A1 |
20020026312 | Tapper | Feb 2002 | A1 |
20020032566 | Tzirkel-Hancock | Mar 2002 | A1 |
20020049593 | Shao | Apr 2002 | A1 |
20020052742 | Thrasher | May 2002 | A1 |
20020103656 | Bahler et al. | Aug 2002 | A1 |
20020135609 | Damiba | Sep 2002 | A1 |
20020138274 | Sharma | Sep 2002 | A1 |
20020143540 | Malayath | Oct 2002 | A1 |
20020152071 | Chaiken et al. | Oct 2002 | A1 |
20020161581 | Morin | Oct 2002 | A1 |
20020173955 | Reich | Nov 2002 | A1 |
20020173956 | Hartley et al. | Nov 2002 | A1 |
20020177999 | Ortega | Nov 2002 | A1 |
20020178004 | Chang | Nov 2002 | A1 |
20020193991 | Bennett et al. | Dec 2002 | A1 |
20020198710 | Hernandez-Abrego et al. | Dec 2002 | A1 |
20020198712 | Hinde et al. | Dec 2002 | A1 |
20030004721 | Zhou | Jan 2003 | A1 |
20030023438 | Schram et al. | Jan 2003 | A1 |
20030036903 | Konopka | Feb 2003 | A1 |
20030120486 | Brittan et al. | Jun 2003 | A1 |
20030125945 | Doyle | Jul 2003 | A1 |
20030139925 | Anderson et al. | Jul 2003 | A1 |
20030149561 | Zhou | Aug 2003 | A1 |
20030154075 | Schalk et al. | Aug 2003 | A1 |
20030154076 | Kemp | Aug 2003 | A1 |
20030191639 | Mazza | Oct 2003 | A1 |
20030220791 | Toyama | Nov 2003 | A1 |
20040024601 | Gopinath | Feb 2004 | A1 |
20040215457 | Meyer | Oct 2004 | A1 |
20050055205 | Jersak et al. | Mar 2005 | A1 |
20050071161 | Shen | Mar 2005 | A1 |
20050080627 | Hennebert et al. | Apr 2005 | A1 |
20050086055 | Sakai et al. | Apr 2005 | A1 |
20050137868 | Epstein et al. | Jun 2005 | A1 |
20070073540 | Hirakawa et al. | Mar 2007 | A1 |
Number | Date | Country |
---|---|---|
0867857 | Sep 1998 | EP |
0905677 | Mar 1999 | EP |
1011094 | Jun 2000 | EP |
1377000 | Jan 2004 | EP |
WO0211121 | Feb 2002 | WO |
WO2005119193 | Dec 2005 | WO |
WO2006031752 | Mar 2006 | WO |
Number | Date | Country | |
---|---|---|---|
20070192101 A1 | Aug 2007 | US |
Number | Date | Country | |
---|---|---|---|
60788621 | Apr 2006 | US | |
60788606 | Apr 2006 | US | |
60788622 | Apr 2006 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11539456 | Oct 2006 | US |
Child | 11688920 | US | |
Parent | 11331649 | Jan 2006 | US |
Child | 11539456 | US | |
Parent | 11051825 | Feb 2005 | US |
Child | 11331649 | US |