The present disclosure relates to multilingual speech recognition
Speech recognition technology is technology that converts and processes a speech signal generated by utterance into text data and is also referred to as speech-to-text (STT). Speech becomes available as an input method of an apparatus due to speech recognition technology so that speech recognition technology is being applied to various technology fields such as an information search and device control through speech. Recently, research on a speech recognition algorithm using machine learning to improve the performance of speech recognition and research to supplement the application of speech recognition technologies, such as technology to separate speech for each speaker from a speech signal including speech of speakers and technology to identify a speaker from a speech signal, are being actively conducted.
Provided is a technique for improving the accuracy of speech recognition of a multilingual speech recognition model through the introduction of a condition for maintaining the consistency of language of tokens. However, the disclosure is not limited to the aforementioned aspects, and other technical aspects may be present.
According to an aspect of the disclosure, a multilingual speech recognition method includes: determining a language homogeneity score (LHS) of a target token based on a language identification result of the target token, wherein the target token is obtained as a speech recognition result of input speech data; and identifying text data corresponding to the input speech data based on the LHS of the target token and a probability that the target token corresponds to the input speech data.
The identifying the text data corresponding to the input speech data may include: determining an automatic speech recognition (ASR) score of the target token based on the probability that the target token corresponds to the input speech data; correcting the ASR score of the target token based on the LHS of the target token; and identifying the text data corresponding to the input speech data based on the ASR score of the target token.
The determining the LHS of the target token may include determining the LHS of the target token based on a degree of similarity between the language identification result of the target token and a language identification result of a token sequence prior to the target token.
The determining the LHS of the target token may include determining the LHS of the target token based on a degree of similarity between the language identification result of the target token and a language identification result of the input speech data.
The determining the LHS of the target token may include: determining a parameter regarding a proportion of the LHS of the target token based on a language change probability corresponding to the target token; and correcting the LHS of the target token based on the parameter.
The parameter may decrease as the language change probability of the target token increases.
According to an aspect of the disclosure, a multilingual speech recognition method includes: obtaining pieces of candidate text data corresponding to input speech data based on a speech recognition result of the input speech data; determining a language homogeneity score (LHS) of each of the pieces of candidate text data based on a degree of similarity between language identification results of a plurality of tokens included in the pieces of candidate text data; and identifying, based on the LHSs of the pieces of candidate text data, one or more pieces of candidate text data corresponding to the input speech data from among the pieces of candidate text data.
The determining the LHS of each of the pieces of candidate text data may include: determining, for each of the plurality of tokens included in the pieces of candidate text data, an LHS of the token; and determining the LHS of each of the pieces of candidate text data based on a sum of the LHSs of the plurality of tokens.
The determining the LHS of each of the plurality of tokens may include determining, for each of the plurality of tokens, an LHS of the token based on a degree of similarity between a language identification result of the token and a language identification result of a token sequence prior to the token.
The determining the LHSs of each of the plurality of tokens may include determining, for each of the plurality of tokens, an LHS of the token based on a degree of similarity between a language identification result of the token and a language identification result of the input speech data.
The identifying the one or more pieces of candidate text data corresponding to the input speech data may include identifying, among the pieces of candidate text data, the one or more pieces of candidate text data corresponding to the input speech data based on respective probabilities that each of the pieces of candidate text data corresponds to the input speech data and the respective LHS of each of the pieces of candidate text data.
The determining the LHS of each of the pieces of candidate text data may include: determining, for each of the pieces of candidate text data, a parameter regarding a proportion of the LHS of the piece of candidate text data based on a language change probability corresponding to each token of the plurality of tokens included in the piece of candidate text data; and correcting the LHS of each of the pieces of candidate text data based on the determined parameters.
The parameter corresponding to a given piece of candidate text data among the pieces of candidate text data may decrease as the language change probability of each token of the plurality of tokens included in the given piece of candidate text data increases.
According to an aspect of the disclosure, a non-transitory computer-readable storage medium has instructions stored therein, which when executed by at least one processor, cause the at least one processor to execute a multilingual speech recognition method including: determining a language homogeneity score (LHS) of a target token based on a language identification result of the target token, wherein the target token is obtained as a speech recognition result of input speech data; and identifying text data corresponding to the input speech data based on the LHS of the target token and a probability that the target token corresponds to the input speech data.
According to an aspect of the disclosure, a multilingual speech recognition apparatus includes: at least one memory storing one or more instructions; and at least one processor configured to execute the one or more instructions, wherein the one or more instructions, when executed by the at least one processor, cause the multilingual speech recognition apparatus to: determine a language homogeneity score (LHS) of a target token based on a language identification result of a target token obtained as a speech recognition result of input speech data, and identify text data corresponding to the input speech data based on the LHS of the target token and a probability that the target token corresponds to the input speech data.
The one or more instructions, when executed by the at least one processor, may cause the multilingual speech recognition apparatus to, in the identification of the text data corresponding to the input speech data: determine an automatic speech recognition (ASR) score of the target token based on the probability that the target token corresponds to the input speech data, correct the ASR score of the target token based on the LHS of the target token, and identify the text data corresponding to the input speech data based on the ASR score of the target token.
The one or more instructions, when executed by the at least one processor, may cause the multilingual speech recognition apparatus to, in the determination of the LHS of the target token, determine the LHS of the target token based on a degree of similarity between the language identification result of the target token and a language identification result of a token sequence prior to the target token.
The one or more instructions, when executed by the at least one processor, may cause the multilingual speech recognition apparatus to, in the determination of the LHS of the target token, determine the LHS of the target token based on a degree of similarity between the language identification result of the target token and a language identification result of the input speech data.
The one or more instructions, when executed by the at least one processor, may cause the multilingual speech recognition apparatus to, in the determination of the LHS of the target token: determine a parameter regarding a proportion of the LHS of the target token, based on a language change probability corresponding to the target token, and correct the LHS of the target token based on the parameter.
According to an aspect of the disclosure, a non-transitory computer-readable storage medium has instructions stored therein, which when executed by at least one processor, cause the at least one processor to execute a multilingual speech recognition method including: obtaining pieces of candidate text data corresponding to input speech data based on a speech recognition result of the input speech data; determining a language homogeneity score (LHS) of each of the pieces of candidate text data based on a degree of similarity between language identification results of a plurality of tokens included in the pieces of candidate text data; and identifying, based on the LHSs of the pieces of candidate text data, one or more pieces of candidate text data corresponding to the input speech data from among the pieces of candidate text data.
According to an aspect of the disclosure, a multilingual speech recognition apparatus includes: at least one memory storing one or more instructions; and at least one processor configured to execute the one or more instructions, wherein the one or more instructions, when executed by the at least one processor, cause the multilingual speech recognition apparatus to: obtain pieces of candidate text data corresponding to input speech data based on a speech recognition result of the input speech data, determine a language homogeneity scores (LHSs) of each of the pieces of candidate text data based on a degree of similarity between language identification results of a plurality of tokens included in the pieces of candidate text data, and identify, based on the LHSs of the pieces of candidate text data, one or more pieces of candidate text data corresponding to the input speech data from among the pieces of candidate text data.
The above and other aspects and features of certain embodiments of the present disclosure will be more apparent the following description taken in conjunction with the accompanying drawings, in which:
The following detailed structural or functional description is provided as an example only and various alterations and modifications may be made to the embodiments described herein. Accordingly, the embodiments described herein should not be construed as limited to the disclosure, but should be understood to include all changes, equivalents, and replacements within the idea and the technical scope of the disclosure.
In connection with the description of the drawings, like reference numerals may be used for similar or related components. It is to be understood that a singular form of a noun corresponding to an item may include one or more of the things, unless the relevant context clearly indicates otherwise.
As used herein, “A or B,” “at least one of A and B,” “at least one of A or B,” “A, B or C,” “at least one of A, B and C,” and “at least one of A, B, or C,” each of which may include any one of the items listed together in the corresponding one of the phrases, or all possible combinations thereof.
Terms, such as “first” or “second”, are simply used to distinguish a component from another component and do not limit the components in other aspects (e.g., importance or sequence).
It is to be understood that if an element (e.g., a first element) is referred to, with or without the term “operatively” or “communicatively,” as “coupled with,” “coupled to,” “connected with,” or “connected to” another element (e.g., a second element), it means that the element may be coupled with the other element directly (e.g., by wire), wirelessly, or via a third element.
The singular forms “a”, “an”, and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises/comprising” and/or “includes/including” when used herein, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components and/or groups thereof.
Unless otherwise defined, all terms, including technical and scientific terms, used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure pertains. Terms, such as those defined in commonly used dictionaries, should be construed to have meanings matching with contextual meanings in the relevant art, and are not to be construed to have an ideal or excessively formal meaning unless otherwise defined herein.
Hereinafter, embodiments will be described in detail with reference to the accompanying drawings. When describing the embodiments with reference to the accompanying drawings, like reference numerals refer to like elements and a repeated description related thereto will be omitted.
According to one or more embodiments, the multilingual speech recognition method may include a method of converting speech data into text data. The multilingual speech recognition method may include a method of recognizing speech data that is uttered in a plurality of languages and converting the speech data into text data.
According to one or more embodiments, the multilingual speech recognition method may be implemented as a multilingual speech recognition model capable of recognizing a plurality of languages. For example, a multilingual speech recognition model that recognizes English and Korean may output data converted from speech data into Korean text when input speech data is speech data that is uttered in Korean and may output data converted from speech data into English text when input speech data is speech data that is uttered in English.
The multilingual speech recognition method may be performed by an apparatus that executes the multilingual speech recognition model. The multilingual speech recognition apparatus that executes the multilingual speech recognition method is described in detail below.
According to one or more embodiments, the multilingual speech recognition method may include, determining a language homogeneity score (LHS) of a target token based on a language identification result of the target token obtained as a speech recognition result of input speech data.
A token is a unit of data output as a speech recognition result, and the multilingual speech recognition model may output the speech recognition result in a token unit. The token may correspond to text data corresponding to a portion of speech data. Tokens output in response to input speech data may correspond to time series data. The target token may correspond to any one of a plurality of tokens output as a speech recognition result of speech data.
The LHS of a token is an indicator of whether a language of the token is similar to a language of another token and may be determined to be a higher value as a degree of similarity between the language of the token and the language of the other token increases.
According to one or more embodiments, operation 110 of determining an LHS of a target token may include determining the LHS of the target token based on a similarity between a language identification result of the target token and a language identification result of a token sequence prior to the target token.
The token sequence prior to the target token may correspond to a sequence including one or more consecutive tokens before the target token when tokens are listed in chronological order. For example, when the target token is an i-th token, the token sequence prior to the target token may correspond to a sequence including an i−1-th token, starting from a j-th token (j is any natural number that is less than i).
The language identification result of the target token may include the probability that the target token corresponds to each of a plurality of languages. The plurality of languages may include languages that may be recognized by the multilingual speech recognition model. For example, the language identification result of the target token may correspond to a vector including the probability that the target token corresponds to English and the probability that the target token corresponds to Korean.
The LHS of the target token may be determined to be a higher value as a degree of similarity between the language identification result of the target token and the language identification result of the token sequence prior to the target token increases.
For example, an LHS LT[i] of the i-th token may be determined as shown in Equation 1 below.
In Equation 1, M[x] denotes a language identification result of an x-th token and may correspond to a vector including the probability that the x-th token corresponds to each of a plurality of languages. The term sim(x, y) denotes a similarity between x and y and may correspond to, for example, a calculated value of a cosine similarity. The term merge (M[j], . . . , M[i−1]) denotes an operation to sum of M[j] to M[i−1] and may correspond to, for example, a weighted average of M[j] to M[i−1]. The term merge (M[j], . . . , M[i−1]) denotes a result of an operation to sum language identification results from the j-th token (j is any natural number that is less than i) to the i−1-th token, which may also be represented as M[j:i]. That is, the LHS LT[i] of the i-th token may be determined based on a similarity between a weighted average of language identification results of tokens prior to the i-th token and a language identification result of the i-th token.
For example,
The LHS of an i-th token may be determined based on a similarity between a language identification result M[i] of the i-th token and a language identification result M[j:i−1] (j is any natural number that is less than i) of a token sequence prior to the i-th token. For example, when there are many tokens with the highest probability of corresponding to English as the language identification results of first to seventh tokens, which are tokens prior to the 8th token, the degree of similarity between the language identification result of the 8th token and the language identification result of the token sequence prior to the 8th token may be determined to be high. In this case, the LHS of the 8th token may be determined to be a high value.
For example, referring to
When there are many tokens with the highest probability of corresponding to English as the language identification results of the first to seventh tokens, which are the tokens prior to the 8th token, a degree of similarity between the language identification result of the 8th token and the language identification result of the token sequence prior to the 8th token may be determined to be low. In this case, the LHS of the 8th token may be determined to be a low value.
According to one or more embodiments, operation 110 of determining the LHS of the target token may include determining the LHS of the target token based on a degree of similarity between the language identification result of the target token and a language identification result of speech data corresponding to the target token. The language identification result of the speech data may include a determination of which language corresponds to the input speech data. For example, the language identification result of the speech data may be obtained by a separate language identification model. For example, the language identification result of the speech data may correspond to data (e.g., a one-hot vector) indicating a language corresponding to the speech data. For example, the language identification result of the speech data may correspond to a vector including the probability that the speech data corresponds to each of a plurality of languages.
For example, when the language identification result of the speech data includes a value of 0, which is the probability of corresponding to German, a value of 1, which is the probability of corresponding to English, a value of 0, which is the probability of corresponding to French, and a value of 0, which is the probability of corresponding to Italian, the LHS LT of each token may be determined as the graph illustrated in
According to one or more embodiments, operation 110 of determining the LHS of the target token may include determining a parameter regarding a proportion of the LHS of the target token, based on a language change probability corresponding to the target token, and correcting the LHS of the target token based on the parameter.
The language change probability corresponding to the target token may correspond to the probability that the language corresponding to the target token is different from a language corresponding to another token. For example, when the speech data is “ Bruno Mars
,” the language change probability of a token corresponding to “Bruno Mars” in English may be determined to be a high value, unlike a token corresponding to other speech data in Korean.
The parameter regarding the proportion of the LHS of the target token may be determined to be a lesser value as the language change probability of the target token increases. For example, the parameter regarding the proportion of the LHS of the target token may be determined to be an inverse number of the language change probability of the target token.
According to one or more embodiments, the language change probability of the target token may be obtained by a model that estimates the language change probability of a token. For example, the model that estimates the language change probability of the token may be an n-gram model. For example, the model that estimates the language change probability of the token may be a neural network trained to estimate the language change probability.
For example, the language change probability of the target token may be obtained using a language model (e.g., a count-based language model such as an n-gram-based language model, a training-based language model, etc.) used for speech recognition. For example, the language change probability of the target token may be obtained using a prediction network of a speech recognition model with a transducer structure or a decoder of a speech recognition model with a transformer structure.
For example, the probability that an i-th token is a certain language may be obtained as P(li|w0:wi−1)=sumw
For example, the probability of the language model may be determined to be an appropriate probability value even for an unseen context.
When the probability of the language model corresponding to a token decreases rapidly, a token with increasing entropy may be determined to be a token with a high language change probability. The increase in entropy may indicate that the probability of the language model corresponding to a certain token has a lower value than the probability of a language model corresponding to a preceding token with a low probability.
For example, an LHS LT[i] of the i-th token may be determined as in Equation 2 below.
In Equation 2, α[i] may correspond to the parameter regarding the proportion of the LHS of the i-th token.
For example, when the speech data is data in which a sentence including English words in the middle of a sentence corresponding to Korean, such as “ Bruno Mars
,” is uttered, the language change probability of a token corresponding to “Bruno Mars” in English may be determined to be a high value by an n-gram model, etc., unlike a token corresponding to the other speech data in Korean. The parameter α[i] regarding the proportion of the LHS of a token i corresponding to “Bruno Mars” with a high language change probability may be determined to be a less value than that of the other token. The proportion in which the LHS is reflected in the final score for determining whether to determine the token i to be a speech recognition result by the parameter α[i] regarding the language proportion may be less than the proportion of an automatic speech recognition (ASR) score. By lowering the proportion of the LHS of a token with a high language change probability through a parameter regarding a language proportion, the token corresponding to the language that is different from the language of the other token may be determined to be the speech recognition result.
According to one or more embodiments, a multilingual speech recognition method may include operation 120 of determining text data corresponding to the speech data, based on the LHS of the target token and a probability that the target token corresponds to the speech data. The probability that the target token corresponds to the speech data may be a probability that the target token determined based on pronunciation in an ASR module of a multilingual speech recognition model corresponds to a speech recognition result of the speech data.
According to one or more embodiments, operation 120 of identifying the text data corresponding to the speech data may include determining an ASR score of the target token based on the probability that the target token corresponds to the speech data, correcting the ASR score of the target token based on the LHS of the target token, and identifying the text data corresponding to the speech data, based on the ASR score of the target token.
The ASR score of the target token may be determined by the probability of which text the target token corresponds to, based on the pronunciation of the speech data. The ASR score may be determined by the speech recognition result output from the ASR module of the multilingual speech recognition model.
For example, the ASR score of the target token may be corrected by a sum, a weighted sum, an average, or a weighted average of the ASR score and the LHS of the target token. For example, a corrected ASR score ASRT′[i] of the i-th token may be determined as in Equation 3 below.
In Equation 3, ASRT′[i] denotes an ASR score of the i-th token, LT[i] denotes an LHS of the i-th token, and λ denotes a weight of ASRT′[i]. ASRT′[i] denotes a corrected ASR score of the i-th token.
The text data corresponding to the speech data may be determined based on the corrected ASR score. Among candidates of the i-th token, a candidate of the i-th token having the highest corrected ASR score may be determined to be the text data corresponding to the speech data. For example, referring to
The multilingual speech recognition method described above with reference to
Referring to
The pieces of candidate text data may be obtained as the speech recognition result of the input speech data. At least some tokens of the pieces of candidate text data may be different from each other. For example, first candidate text data of “Please recommend me a magazine for men” and second candidate text data of “Please recommend me a magasin for men” may be obtained as a speech recognition result of speech data in which “Please recommend me a magazine for men” is uttered. For example, the pieces of candidate text data may correspond to the speech recognition result of the speech data determined based on pronunciation in an ASR module of a multilingual speech recognition model.
According to one or more embodiments, the multilingual speech recognition method may include operation 520 of determining LHSs of the pieces of candidate text data based on a similarity between language identification results of a plurality of tokens included in the pieces of candidate text data.
According to one or more embodiments, operation 520 of determining the LHSs of the pieces of candidate text data may include determining LHSs of each of the plurality of tokens included in the pieces of candidate text data and determining the LHSs of the pieces of candidate text data based on the sum of the LHSs of each of the plurality of tokens. For example, the LHSs of the pieces of candidate text data may be determined as a sum, a weighted sum, an average, or a weighted average of the LHSs of each of the plurality of tokens included in the pieces of candidate text data. For example, the LHSs of the pieces of candidate text data may be determined to be a value obtained by converting the sum or the weighted sum of the LHSs of each of the plurality of tokens included in the pieces of candidate text data.
According to one or more embodiments, the determining of the LHSs of each of the plurality of tokens may include determining an LHS of a token based on a similarity between a language identification result of the token included in the plurality of tokens and a language identification result of a token sequence prior to the token. For example, the LHS of the token may be determined by Equation 1.
According to one or more embodiments, the determining of the LHSs of each of the plurality of tokens may include determining the LHS of the token based on a degree of similarity between the language identification result of the token included in the plurality of tokens and a language identification result of speech data corresponding to the plurality of tokens.
For example, an LHS LS of the candidate text data may be determined as in Equation 4 below.
In Equation 4, I denotes the length of a token of the candidate text data or the number of tokens included in the candidate text data. For example, LT[i] denotes an LHS of an i-th token included in the candidate text data determined by Equation 1. For example, LT[i] may be determined based on a similarity between a language identification result of speech data corresponding to data (e.g., a one-hot vector) indicating a language corresponding to the speech data and a language identification result of the i-th token. That is, the LHS LS of the candidate text data may correspond to the average of the LHSs of the tokens included in the candidate text data.
According to one or more embodiments, operation 520 of determining the LHSs of the pieces of candidate text data may include determining a parameter regarding a proportion of the LHSs of the pieces of candidate text data, based on a language change probability determined corresponding to each of the tokens included in the pieces of candidate text data, and correcting the LHSs of the pieces of candidate text data based on the parameter.
The parameter regarding a proportion of the LHSs of the pieces of candidate text data may be determined to be a lesser value as the language change probability of the tokens included in the pieces of candidate text data increases. For example, the parameter regarding a proportion of the LHSs of the pieces of candidate text data may be determined as a sum, a weighted sum, a product, an average, or an inverse number of a weighted average of the language change probabilities of each of the tokens included in the pieces of candidate text data. For example, the parameter regarding a proportion of the LHSs of the pieces of candidate text data may be determined to be a value obtained by converting a product, a sum, or an inverse number of a weighted sum of the language change probabilities of each of the tokens included in the pieces of candidate text data.
For example, the LHS LS of the candidate text data may be determined as in Equation 5 below.
In Equation 5, I denotes the length of a token of the candidate text data or the number of tokens included in the candidate text data, LT[i] denotes the LHS of the i-th token included in the candidate text data, and β denotes the parameter regarding the proportion of the LHS of the candidate text data. For example, β may be determined as a product of parameters regarding the proportion of the LHSs of the tokens included in the candidate text data.
According to one or more embodiments, the multilingual speech recognition method may include operation 530 of determining, among the pieces of candidate text data, text data corresponding to the speech data, based on the LHSs of the pieces of candidate text data.
According to one or more embodiments, operation 530 of identifying the text data corresponding to the speech data may include determining, among the pieces of candidate text data, the text data corresponding to the speech data, based on the probability that the pieces of candidate text data correspond to the speech data and LHSs of the pieces of candidate text data. The probability that the pieces of candidate text data correspond to the speech data may be the probability that the pieces of candidate text data, which are determined based on pronunciation in the ASR module of a multilingual speech recognition model, correspond to the speech recognition result of the speech data. For example, the probability that the pieces of candidate text data correspond to the speech data may be determined based on the probability that each token included in the pieces of candidate text data corresponds to the speech data.
According to one or more embodiments, operation 530 of identifying the text data corresponding to the speech data may include determining ASR scores of the pieces of candidate text data based on the probability that the pieces of candidate text data correspond to the speech data, correcting the ASR scores of the pieces of candidate text data based on the LHSs of the pieces of candidate text data, and identifing the text data corresponding to the speech data, based on the ASR scores of the pieces of candidate text data.
The ASR scores of the pieces of candidate text data may be determined based on the probability of which text the pieces of candidate text data correspond to, based on the pronunciation of the speech data. The ASR scores may be determined by the speech recognition result output from the ASR module of the multilingual speech recognition model. For example, the ASR scores of the pieces of candidate text data may be determined based on the probability that each token included in the pieces of candidate text data corresponds to the speech data or the ASR score of each token included in the pieces of candidate text data. For example, the ASR scores of the pieces of candidate text data may be corrected by a sum, a weighted sum, an average, or a weighted average of the ASR scores of the pieces of candidate text data and the LHSs of the pieces of candidate text data.
The text data corresponding to the speech data may be determined based on the corrected ASR scores of the pieces of candidate text data.
For example, referring to
According to one or more embodiments, the candidate text data corresponding to the speech data may be determined based on an ASR score ASRS. For example, the candidate text data corresponding to the first node 610, the candidate text data corresponding to the second node 620, and the candidate text data corresponding to the third node 630 may be data belonging to the top n (n is any natural number) or m % (m is any positive real number) having the high ASR score ASRS among possible combinations of tokens. For example, the candidate text data corresponding to the first node 610, the candidate text data corresponding to the second node 620, and the candidate text data corresponding to the third node 630 may be data having the ASR score ASRS that is greater than or equal to a predetermined threshold value.
The ASR score ASRS of the candidate text data corresponding to the first node 610 may be determined to be 0.4, and an LHS LS may be determined to be 1. Pieces of candidate text data corresponding to the first node 610 may be similar to each other as data indicating that the language identification results of tokens included in the path from the root to the first node 610, which is the last token, are all most likely to correspond to English, and accordingly, the LHS LS may be determined to be 1.
The ASR score ASRS of the candidate text data corresponding to the third node 630 may be determined to be 0.6, and the LHS LS may be determined to be 0.75. The candidate text data corresponding to the third node 630 may be data indicating that the language identification results of most of the tokens included in the path from the root to the first node 610, which is the last token, are most likely to correspond to English, while referring to a fourth node 640, the language identification results of some tokens may be data indicating that the language identification results are most likely to correspond to French. Accordingly, the LHS LS may be determined to be a value that is less than the LHS LS of the candidate text data corresponding to the first node 610, which is data indicating that the language identification results of all tokens are most likely to correspond to English.
For example, the ASR score ASRS of the candidate text data may be corrected to a value that is the sum of the ASR score ASRS and the LHS LS. The candidate text data corresponding to the first node 610 having the highest corrected ASR score ASRS′ may be determined to be the text data corresponding to the speech data. The candidate text data corresponding to the first node 610 may be output as the speech recognition result of the speech data.
According to one or more embodiments, a multilingual speech recognition apparatus 700 may be an apparatus that performs the multilingual speech recognition method described above with reference to
Referring to
The ASR module 710 may be a module that determines an ASR score of a token corresponding to the speech data 701 or a token sequence. As described above, the ASR score may be determined based on the probability of which text a target token corresponds to, based on pronunciation of the speech data 701.
The LHS module 720 may be a module that determines an LHS of the token corresponding to the speech data 701 or the token sequence. As described above, the LHS may be determined based on a similarity between language identification results of tokens.
The score integration module 730 may be a module that determines the text data 702 corresponding to the speech data 701, based on the ASR score determined by the ASR module 710 and the LHS determined by the LHS module 720. As described above, the ASR score may be corrected by a sum, a weighted sum, an average, or a weighted average of the ASR score and the LHS. The score integration module 730 may obtain the corrected ASR score by calculating the sum, the weighted sum, the average, or the weighted average of the ASR score and the LHS. The text data 702 with the maximum corrected ASR score may be output as a speech recognition result of the speech data 701.
An LHS module 800 shown in
Referring to
The language identification module 820 may be a module that determines a language identification result M of the token corresponding to the input speech data or the token sequence. The language identification result M may include probability data that the token or the token sequence corresponds to each recognizable language.
The code switch module 810 may be a module that determines a language change probability of the token corresponding to the input speech data or the token sequence. For example, the code switch module 810 may include a model (e.g., an n-gram model and a training model) that estimates the language change probability of the token. The code switch module 810 may output the language change probability of the token or the token sequence using the model that estimates the language change probability of the token. The LHS module 800 may or may not include the code switch module 810.
The LHS calculation module 830 may be a module that determines an LHS of the token or the token sequence based on a language identification result of the token or the token sequence obtained by the language identification module 820. According to one or more embodiments, the LHS calculation module 830 may also determine the LHS of the token or the token sequence based on the language identification result of the token or the token sequence obtained by the language identification module 820 and the language change probability of the token or the token sequence obtained by the code switch module 810.
Referring to
According to one or more embodiments, the processor 901 may perform at least one operation of the multilingual speech recognition method described above with reference to
For example, the processor 901 may perform at least one of determining an LHS of a target token based on a language identification result of the target token obtained as a speech recognition result of input speech data and determining text data corresponding to the input speech data, based on the LHS of the target token and a probability that the target token corresponds to the input speech data.
For example, the processor 901 may perform at least one of obtaining pieces of candidate text data corresponding to the input speech data, based on a speech recognition result of the input speech data, determining LHSs of the pieces of candidate text data based on a similarity between language identification results of a plurality of tokens included in the pieces of candidate text data, and determining, among the pieces of candidate text data, the text data corresponding to the input speech data, based on the LHSs of the pieces of candidate text data.
According to one or more embodiments, the memory 903 may be a volatile memory or a non-volatile memory and may store data related to the multilingual speech recognition method described above with reference to
According to one or more embodiments, the memory 903 may store a program in which the multilingual speech recognition method described above with reference to
According to one or more embodiments, the multilingual speech recognition apparatus 900 may be connected to an external device (e.g., a personal computer (PC) or a network) through the I/O device 905 and exchange data therewith. For example, the multilingual speech recognition apparatus 900 may receive speech data through the I/O device 905 and output text data as a speech recognition result of the speech data.
According to one or more embodiments, the multilingual speech recognition apparatus 900 may further include other components not shown in the drawings.
For example, the multilingual speech recognition apparatus 900 may include a communication module. The communication module may provide a function for the multilingual speech recognition apparatus 900 to communicate with other electronic devices or other servers through a network. According to one or more embodiments, the memory 903 may not be a component of the multilingual speech recognition apparatus 900 but may be included in an external device accessible by the multilingual speech recognition apparatus 900. In this case, the multilingual speech recognition apparatus 900 may receive data stored in the memory 903 included in the external device through the communication module and may transmit data to be stored in the memory 903.
In another example, the multilingual speech recognition apparatus 900 may further include other components such as a transceiver, various sensors, a database, etc.
The embodiments described herein may be implemented using a hardware component, a software component, and/or a combination thereof. A processing device may be implemented using one or more general-purpose or special-purpose computers, such as, for example, a processor, a controller and an arithmetic logic unit (ALU), a DSP, a microcomputer, an FPGA, a programmable logic unit (PLU), a microprocessor or any other device capable of responding to and executing instructions in a defined manner. The processing device may run an operating system (OS) and one or more software applications that run on the OS. The processing device also may access, store, manipulate, process, and create data in response to execution of the software. For purpose of simplicity, the description of a processing device is used as singular; however, one skilled in the art will appreciate that a processing device may include multiple processing elements and/or multiple types of processing elements. For example, the processing device may include a plurality of processors, or a single processor and a single controller. In addition, different processing configurations are possible, such as parallel processors.
The software may include a computer program, a piece of code, an instruction, or some combination thereof, to independently or uniformly instruct or configure the processing device to operate as desired. Software and data may be stored in any type of machine, component, physical or virtual equipment, or computer storage medium or device capable of providing instructions or data to or being interpreted by the processing device. The software may also be distributed over network-coupled computer systems so that the software is stored and executed in a distributed fashion. The software and data may be stored by one or more non-transitory computer-readable recording mediums.
The methods according to the above-described embodiments may be recorded in non-transitory computer-readable media including program instructions to implement various operations of the above-described embodiments. The media may also include, alone or in combination with the program instructions, data files, data structures, and the like. The program instructions recorded on the media may be those specially designed and constructed for the purposes of embodiments, or they may be of the kind well-known and available to those having skill in the computer software arts. Examples of non-transitory computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROM discs and/or DVDs; magneto-optical media such as optical discs; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory (ROM), random access memory (RAM), flash memory, and the like. Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher-level code that may be executed by the computer using an interpreter.
The above-described hardware devices may be configured to act as one or more software modules in order to perform the operations of the above-described embodiments, or vice versa.
As described above, although the embodiments have been described with reference to the limited drawings, a person skilled in the art may apply various technical modifications and variations based thereon. For example, suitable results may be achieved if the described techniques are performed in a different order, and/or if components in a described system, architecture, device, or circuit are combined in a different manner, or replaced or supplemented by other components or their equivalents.
Accordingly, other implementations are within the scope of the following claims.
Number | Date | Country | Kind |
---|---|---|---|
10-2024-0011061 | Jan 2024 | KR | national |
10-2024-0084463 | Jun 2024 | JP | national |
This application is a bypass-continuation of International Application No. PCT/KR2024/016938, filed on Oct. 31, 2024, which is based on and claims priority to Korean Patent Application No. 10-2024-0011061, filed in the Korean Intellectual Property Office on Jan. 24, 2024, and Korean Patent Application No. 10-2024-0084463, filed in the Korean Intellectual Property Office on Jun. 27, 2024, the disclosures of which are incorporated herein by reference in their entireties.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/KR2024/016938 | Oct 2024 | WO |
Child | 19008332 | US |