Speech analyzing systems match a received speech signal to a stored database of speech patterns. One system, a speech recognizer, interprets the speech patterns, or sequences of speech patterns to produce text. Another system, a vocoder, is a speech analyzer and synthesizer which digitally encodes an audio signal for transmission. The audio signal received by either of these devices often includes environmental noise. The noise acts to mask the speech signal, and can degrade the quality of the output speech of a vocoder or decrease the probability of correct recognition by a speech recognizer. It would be desirable to filter out the environmental noise to improve the performance of a vocoder or speech recognizer.
The invention, in various embodiments, addresses the issue of encoding speech in an audio signal containing noise. According to one aspect, this is achieved through the use of an adaptive noise codebook.
In one approach, the invention provides systems and methods for identifying speech sounds in an audio signal. According to one aspect, this is achieved through the use of a speech codebook and an adaptive noise codebook. The speech codebook includes a plurality of speech codebook entries corresponding to speech sounds, and the noise codebook includes a plurality of noise codebook entries corresponding to noise sounds. According to one feature, upon receipt of an audio signal including noise sounds, a noise codebook entry is updated based on the noise sounds. According to another feature, upon receipt of an audio signal including speech, speech sounds in the audio signal are identified using the speech codebook and noise codebook.
According to one implementation, a speech sound is identified by creating one or more noisy speech templates as a function of the speech codebook entries and the noise codebook entries, and matching a segment of the audio signal to a noisy speech template. In one embodiment, a noisy speech template is created by combining a speech codebook entry with a noisy codebook entry. According to one feature, the speech codebook entry and the noise codebook entry are each characterized by a vector comprising a gain parameter and at least one line spectral frequency parameter. The speech codebook entry may be combined with the noise codebook entry by converting the vectors into autocorrelation parameter vectors and adding the autocorrelation parameter vectors. According to a further embodiment, the matching of a segment of the audio signal to a noisy speech template is terminated upon identification of an acceptable match.
According to one configuration, the identified speech sounds and pauses are digitally encoded for transmission. According to one feature, the identified speech sounds and pauses are encoded at a bit rate of less than 2500 bps. According to another feature, the identified speech sounds and pauses are encoded at a variable bit rate. The bit rate for encoding an identified speech sound or a pause may be a function of the characteristics of the speech sound or pause.
In one embodiment, a noisy speech codebook is used in identifying the speech sounds. The noisy speech codebook includes a plurality of noisy speech templates. A template may be created as a function of a speech codebook entry and a noise codebook entry. The noisy speech codebook may include a plurality of templates created as a function of the plurality of speech codebook entries and a first noise codebook entry. In this embodiment, the noisy speech codebook may be updated based on a second plurality of templates created as a function of the plurality of speech codebook entries and a second noise codebook entry. According to an alternative embodiment, the noisy speech codebook includes a plurality of templates created as a function of the plurality of speech codebook entries and the plurality of noise codebook entries.
According to one feature, a noise codebook entry similar to the noise component of the audio signal is selected. This noise codebook entry may be used to create noisy speech templates.
According to one implementation, the audio signal is processed to detect the presence of speech in the audio signal. Segments of the audio signal absent of speech may be used to generate noise codebook entries. According to one embodiment, the audio signal is temporally parsed into segments. The segments may be processed to detect the presence of speech in a segment.
According to one feature, a speech codebook entry includes characteristics of a speech sound, and a noise codebook entry includes characteristics of a noise sound. The characteristics may include frequency, duration, amplitude, and gain characteristics. The characteristics may include parameters describing a frequency versus amplitude spectrum corresponding to a speech sound, or parameters describing a frequency versus amplitude spectrum corresponding to a noise sound. The characteristics may include parameters describing a frequency versus amplitude spectrum corresponding to a segment of the audio signal. These parameters may include at least one of gain and predictor coefficients, gain and reflection coefficients, gain and line spectral frequencies, and autocorrelation coefficients. In one embodiment, the characteristics of a segment of the audio signal are matched with the characteristics of a speech codebook entry and the characteristics of a noise codebook entry.
According to one feature, an entry in the noise codebook may be updated. The noise codebook entry may be updated by adapting the entry based on a segment of the audio signal absent of speech. The noise codebook entry may be updated by replacing the entry with a new entry generated from a segment of the input audio signal absent of speech. According to one implementation, the noise codebook entry updated is the least frequently used noise codebook entry. According to another implementation, the noise codebook entry updated is the least recently used noise codebook entry. According to a further implementation, the noise codebook entry updated is the least recently updated noise codebook entry. According to still another implementation, the noise codebook entry updated is the entry that is most similar to a new entry generated from a segment of the audio signal absent of speech.
According to one embodiment, the adaptive noise codebook includes at least 10 entries. According to another embodiment, the adaptive noise codebook includes at least 20 entries. According to another embodiment, the adaptive noise codebook includes at least 50 entries.
According to various embodiments, a voicing codebook is used in identifying speech sounds. The voicing codebook may include a plurality of voicing codebook entries. According to one feature, a voicing codebook entry corresponds to a voicing characteristic. The voicing characteristic my include any of presence of voicing, absence of voicing, pitch, and pitch trajectory. According to another feature, a voicing codebook entry represents at least one frame from an audio signal.
In one configuration, the audio signal is temporally divided into frames. A frame may have a predetermined duration, or a frame may have a variable duration. In one embodiment, the audio signal is temporally divided into segments. According to one feature, a segment comprises at least one of a frame, a plurality of sequential frames, and a plurality of overlapping frames.
According to another aspect, the invention provides a device including a receiver for receiving an audio signal, a first database for storing a plurality of speech codebook entries representative of a plurality of speech sounds, a second database for storing a plurality of noise codebook entries representative of a plurality of noise sounds, a first processor for updating an entry in the noise codebook based on noise sounds in the audio signal, and a second processor for identifying speech sounds in the audio signal as a function of the speech codebook entries and the noise codebook entries.
In one embodiment, the second processor of the device includes a template processor, which combines a noise codebook entry and a speech codebook entry and creates at least one template. Additionally, the second processor includes a match processor, which matches a segment of the audio signal with one of the at least one templates.
According to one implementation, the second processor creates a template by combining at least one speech codebook entry with at least one noise codebook entry. According to one feature, the at least one speech codebook entry and the at least one noise codebook entry each include a vector having a gain parameter and at least one line spectral frequency parameter. According to a further embodiment, the second processor further includes a transform processor for transforming the speech codebook entry vector and the noise codebook entry vector into autocorrelation parameter vectors, and adding the autocorrelation parameter vectors.
According to one embodiment, the device further includes a third processor for digitally encoding the identified speech sounds and pauses for transmission. The third processor may include an encoder that encodes the identified speech sounds and pauses at a bit rate of less than 2500 bits per second. The third processor may include an encoder that encodes the identified speech sounds and pauses at a variable bit rate. The third processor may determine the bit rate for encoding one of an identified speech sound and a pause as a function of the characteristics of one of the identified speech sound and pause.
In one configuration, the device further includes a third database for storing a noisy speech codebook including a plurality of templates. According to one implementation, the second processor creates a template as a function of at least one speech codebook entry and at least one noise codebook entry. The second processor may create the plurality of templates as a function of the plurality of speech codebook entries and a first noise codebook entry. According to one feature, the second processor includes a noisy speech codebook updater for updating the noisy speech codebook based on a second plurality of templates created as a function of the plurality of speech codebook entries and a second noise codebook entry. According to another feature, the second processor creates the plurality of templates as a function of the plurality of speech codebook entries and the plurality of noise codebook entries.
In one embodiment, the device includes a fourth processor for detecting speech in a segment of the audio signal. According to one feature, the first processor generates noise codebook entries from segments of the audio signal absent of speech.
According to one feature, the second processor selects a noise codebook entry as a function of a noise component of the audio signal.
According to one embodiment, the device includes a third processor for temporally parsing the audio signal into segments. The third processor may detect the presence of speech in a segment of the audio signal.
According to one feature, a speech codebook entry includes characteristics of a speech sound, and a noise codebook entry includes characteristics of a noise sound. In one embodiment, the device includes a third processor for matching the characteristics of at least one of a speech codebook entry and a noise codebook entry with characteristics of a segment of the audio signal. The characteristics may include at least one of frequency, duration, amplitude, and gain characteristics. The characteristics may include parameters describing a frequency versus amplitude spectrum corresponding to at least one of a speech sound and a noise sound. Additionally, the characteristics may include parameters describing a frequency versus amplitude spectrum corresponding to a segment of the audio signal. In this embodiment, the parameters may include at least one of gain and predictor coefficients, gain and reflection coefficients, gain and line spectral frequencies, and autocorrelation coefficients.
According to one embodiment, the first processor updates an entry in the noise codebook by adapting the entry based on a segment of the audio signal absent of speech. According to an alternative embodiment, the first processor updates an entry in the noise codebook by replacing the entry with a new entry generated from a segment of the audio signal absent of speech.
In one configuration, the device includes a third database for storing a voicing codebook including a plurality of voicing codebook entries. A voicing codebook entry may correspond to at least one voicing characteristic. According to one feature, the at least one voicing characteristic includes at least one of presence of voicing, absence of voicing, pitch, and pitch trajectory.
The foregoing and other objects and advantages of the invention will be appreciated more fully from the following further description thereof, with reference to the accompanying drawings. These depicted embodiments are to be understood as illustrative of the invention and not as limiting in any way.
To provide an overall understanding of the invention, certain illustrative embodiments will now be described, including systems, methods and devices for providing improved analysis of speech, particularly in noisy environments. However, it will be understood by one of ordinary skill in the art that the systems and methods described herein can be adapted and modified for other suitable applications and that such other additions and modifications will not depart from the scope hereof.
The matcher 112, which identifies the speech sounds in an audio signal, may include a processor 114 and at least one database 118. The database 118 stores a speech codebook 120, a noise codebook 122, and, optionally, a noisy speech codebook 124. According to alternative embodiments, the codebooks 120, 122, and 124 may be stored in separate databases. The processor 114 creates the noisy speech codebook 124 as a function of the speech codebook 120 and the noise codebook 122, as described in greater detail with respect to
An enlargement of one exemplary noise entry, noise entry 212b, is also shown in
According to various embodiments, the noise codebook 202 may contain 2, 4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, or 16384 noise entries 212. Additionally, the codebook may contain any integer number of noise entries. According to a preferred embodiment, the noise codebook 202 contains 20 noise entries 212. According to an alternative embodiment, each noise codebook entry represents a plurality of frames of noise.
Additionally, each noise entry 212 includes a usage data counter 218. In one implementation, the usage data counter 218 counts how many times the corresponding noise entry 212 has been adapted. According to one embodiment, the usage data counters 218 of noise entries 212 that have never been adapted or replaced store a value of zero, and every time a noise entry 212 is adapted, the usage data counter 218 is incremented by one. When a noise entry 212 is replaced, the corresponding usage data counter 218 is reset to one. In another embodiment, when a noise entry 212 is replaced, the corresponding usage data counter 218 is reset to zero. In an alternative embodiment, the usage data counters 218 track how many times the noise entries 212 have been selected.
The illustrative voicing codebook 204 includes voicing entries 220a-220d corresponding to four sound voicing patterns. Each voicing entry 220a-220d corresponds to a two frame voicing pattern. Entry 220a, a “voiced-voiced” voicing entry, corresponds to two frames of a voiced signal. Entry 220b, a “voiced-unvoiced” voicing entry, corresponds to a first frame of a voiced signal followed by a second frame of an unvoiced signal. Entry 220c, an “unvoiced-voiced” voicing entry, corresponds to a first frame of an unvoiced signal followed by a second frame of a voiced signal. Entry 220d, an “unvoiced-unvoiced” voicing entry, corresponds to two frames of an unvoiced signal. According to one feature, the “unvoiced-unvoiced” voicing entry may represent two frames of unvoiced speech, two frames of speech-absent environmental noise, or one frame of unvoiced speech and one frame of speech-absent noise. According to one embodiment, two consecutive frames of the input signal are matched with one of the four entries 220a-220d.
The voicing codebook 204 also contains pitch entries 222a-222c corresponding to pitch and pitch trajectories. Pitch entries 222a contain possible pitch values for the first frame, corresponding to the “voiced-unvoiced” voicing entry 220b. Pitch entries 222b contain possible pitch values for the second frame, corresponding to the “unvoiced-voiced” voicing entry 220c. Pitch entries 222c contain pitch values and pitch trajectories for the first and second frames, corresponding to the “voiced-voiced” voicing entry 220d. The pitch trajectory information includes how the pitch is changing over time (for example, if the pitch is rising or falling). According to one embodiment, pitch entries 222a include 199 entries, pitch entries 222b include 199 entries, and pitch entries 222c include 15,985 entries. However, according to alternative embodiments, the pitch entries 222a, 222b, and 222c may include 50, 100, 150, 250, 500, 1000, 2500, 5000, 7500, 10000, 12500, 15000, 17500, 20000, 25000, or 50000 entries.
An enlarged representation of each of the stage-codebooks 232, 234, and 236 is shown in
According to one embodiment, each speech codebook entry 240, 244, and 248 represents a plurality of frames of speech. According to this embodiment, each speech codebook entry 240, 244, and 248 represents a spectral trajectory, wherein a spectral trajectory is the sequence of spectra that model the plurality of frames. In various embodiments, each speech codebook entry 240, 244, and 248 represents 2, 4, 8, 10, 15, 20, 30, 40, or 50 frames of speech. In a preferred embodiment, each codebook entry 240, 244, and 248 represents four frames of speech.
Each entry in the stage-2 speech codebook 234 represents a possible perturbation of any entry 240 in the stage-1 speech codebook 232. According to one implementation, in which each entry 240 and 244 represents a spectral trajectory, a selected stage-1 codebook entry, e.g. stage-1 codebook entry 240m, is combined with a selected stage-2 codebook entry, e.g. stage-2 codebook entry 244n, by combining the corresponding spectra of the entries 240m and 244n. For example, if g1(θ) is the spectrum of the kth frame from stage-1 codebook entry 240m and g2(θ) is the spectrum of the kth frame from stage-2 codebook entry 244n, their product, g1(θ)*g2(θ), for each k, provides the combined speech spectral trajectory.
In one implementation, the spectra of a spectral trajectory are represented using 257 samples of the log-spectrum:
gp=log g(2*π*p/256) for p=0, 1, . . . , 256
where the samples are taken at equally spaced frequencies θ=2*π*p/256 from p=0 to p=256. Thus, for a spectral trajectory including three frames, the stage-codebook entry 240, 244, or 248 is a vector of 3*257 values representing a sequence of 3 log-spectra. By storing these log-values in each stage-codebook 232, 234, and 236, a vector from the stage-1 codebook 232 may be summed with a vector from the stage-2 codebook to create a vector of 3*257 values representing a sequence of 3 log-spectra. The sequence of spectra can be obtained from these log-spectra by exponentiation; this yields a vector of 3*257 nonnegative values. Each group of 257 nonnegative values can be converted into a sequence of autocorrelation values, as described further in relation to
This process may be repeated with the stage-3 codebook entries 248. The vector from the stage-1 codebook entry 240m may be summed with the vector from the stage-2 codebook entry 244n and the vector from the stage-3 codebook entry 248p to yield a vector of 3*257 values representing a sequence of three log-spectra.
As described in greater detail with respect to
To take into account noise obscuring the speech sounds in the input signal, the matcher 112 compares segments of the audio signal with noisy speech templates instead of comparing segments to the speech stage-codebooks 232, 234, and 236 directly. To create a noisy speech template, the parameter vector 214 of a noise codebook entry 212 and the parameter vector of the combined stage-1 codebook entry 240, stage-2 codebook entry 244, and stage-3 codebook entry 248, are converted to autocorrelation parameter vectors, as described in further detail with respect to
According to one embodiment, a plurality of noisy speech templates are generated and stored in a noisy speech codebook.
According to one embodiment, a noisy speech template is created for each stage-codebook entry 240, 244, and 248. According to the illustrative embodiment, the noisy speech codebook 300 is generated by combining the autocorrelation vectors of a selected sequence of noise codebook entries with the autocorrelation vectors of each frame of a speech codebook entry. However, according to alternative embodiments, the speech encoding system 100 maintains separate noisy speech codebooks for each noise entry. These noisy speech codebooks may be updated by selecting a second noise codebook entry, and replacing each noisy speech codebook entry with a template generated by combining the second noise codebook entry with each speech codebook entry. As shown in
Referring back to
An autocorrelation parameter vector is generated from a speech parameter vector. The nth autocorrelation value rn of an autocorrelation parameter vector G, may be calculated as a function of the spectrum g(θ) representing a frame of a speech codebook entry using the following formula:
The autocorrelation parameter vector G has a length N, where N is the number of samples in the frame represented by g(θ). Similarly, for a noise codebook entry 212, the nth autocorrelation value qn of an autocorrelation parameter vector M, may be calculated as a function of the spectrum μ(θ) representing the frame of the noise-codebook entry 212, using the following formula:
The autocorrelation parameter vector M also has a length N, where N is the number of samples in the frame represented by μ(θ).
According to one implementation, a frame of a noisy-speech template autocorrelation parameter vector S is the sum of a speech entry autocorrelation parameter vector G and a noise entry autocorrelation parameter vector M:
S=G+M
According to a further embodiment, the spectrum s(θ) representing a frame of a noisy-speech template may be calculated as the sum of the spectrum g(θ) representing a frame of a speech-codebook entry and the spectrum μ(θ) representing the frame of a noise codebook entry.
s(θ)=g(θ)+μ(θ)
Optionally, the noisy speech templates may be aggregated to form a noisy speech codebook (step 504), as described in relation to
Next, a processor matches a segment of the audio signal containing speech to a noisy speech template (step 508), thereby identifying the speech sound.
Referring to
Next, the stage 2 stage-codebook 234 is used. The matcher 112 combines each stage 2 entry 244 with the selected stage 1 entry 240m, creates noisy speech templates from this combination and the selected noise entries 212, and matches the noisy speech templates to the segment. The matcher 112 identifies and selects the noisy speech template used in forming the best match, e.g. the template derived from the combination of stage 1 entry 240m, stage 2 entry 244n, and the selected noise entries 212.
Last, the stage 3 stage-codebook 236 is used. The matcher 112 combines each stage 3 entry 248 with the selected stage 1 entry 240m and stage 2 entry 244n, creates noisy speech templates from this combination and the noise entries 212, and matches the noisy speech templates to the segment. The matcher 112 identifies and selects the noisy speech template, used in forming the best match, e.g. the template derived from stage 1 entry 240m, stage 2 entry 244n, stage 3 entry 248p, and the selected noise entries 212. According to other embodiments, the matcher 112 may select a plurality of noisy speech templates derived from the entries from each stage-codebook 232, 234, and 236, combining the selected entries from one stage with each entry in the subsequent stage. Selecting multiple templates from each stage increases the pool of templates to choose from, improving accuracy at the expense of increased computational cost.
According to one embodiment, to match a segment of the audio signal to an entry in the speech codebook 208 (step 508), the matcher 112 uses stage-codebooks 232, 234, and 236 sequentially, along with the noise codebook 202, to derive the best noisy speech template match. According to this embodiment, each stage-codebook entry 240, 244, and 248 represents a plurality of frames, thus representing a spectral trajectory. Each noise codebook entry 212 represents a single frame, and thus a single spectrum. Therefore, a noise codebook entry spectrum is identified and selected for each frame of a stage-codebook entry.
The matcher 112 begins with a first stage-1 codebook entry, e.g. stage-1 codebook entry 240a, which represents a four-spectrum (i.e. four frame) spectral trajectory. For the first speech spectrum in the stage-1 codebook entry 240a, the matcher 112 creates a set of noisy speech spectra by combining the first speech spectrum with the noise spectrum of each noise entry 212 in the noise codebook 202. The matcher 112 compares each of these noisy speech spectra to the first frame in the audio signal segment, and computes a frame-log-likelihood value (such as the frame log-likelihood value, discussed below) for each noisy speech spectrum. The frame-log-likelihood value represents the probability that the computed noisy speech spectrum matches the first frame of the segment. Based on the frame-log-likelihood values, the matcher 112 determines which noise spectrum yields the highest frame-log-likelihood value for the first frame of the first speech codebook entry 240a.
The matcher 112 repeats this process for each frame in the spectral trajectory of the first stage-1 codebook entry 240a and each corresponding frame of the input audio signal segment, determining which noise spectrum yields the highest frame-log-likelihood value for each frame. The matcher 112 sums the highest frame-log-likelihood value of each frame of the first stage-1 codebook entry 240a to yield the segment-log-likelihood value. The first stage-1 codebook entry 240a segment-log-likelihood value corresponds to the probability that the audio segment matches the combination of the speech spectral trajectory of the first stage-1 codebook entry 240a and the selected noise spectral trajectory that maximizes the segment-log-likelihood.
The matcher 112 repeats this process for each stage-1 codebook entry 240, generating a segment-log-likelihood value and a corresponding noise spectral trajectory for each stage-1 codebook entry 240. The matcher 112 selects the stage-1 codebook entry 240 noise spectral trajectory pairing having the highest segment-log-likelihood value.
After selecting a stage-1 codebook entry-noise spectral trajectory pairing, the matcher 112 proceeds to the stage-2 speech codebook 234. The matcher 112 calculates new spectral trajectories by combining the selected stage-1 codebook entries with each of the stage-2 codebook entries. Using the noise spectral trajectory selected above, the matcher 112 calculates a segment-log-likelihood value for each of the combined spectral trajectories, and selects the stage-2 codebook entry 244 that yields the combined spectral trajectory having the highest segment-log-likelihood value. This represents the “best” combination of stage-1 codebook 232 and stage-2 codebook 234 spectral trajectories. The matcher 112 repeats this process for the stage-3 codebook 236, combining each stage-3 codebook entry 248 with the combination of the selected stage-1 entry 240, stage-2 entry 244, and noise trajectory entries. The received speech sounds can be uniquely identified by the selected stage-1, stage-2, and stage-3 codebooks, the noise codebook entries 212 corresponding to the selected noise trajectory, and the voicing codebook entries 220, which, when combined together, create a noisy speech template.
As mentioned above, the matcher 112 can calculate the likelihood that a noisy speech spectrum matches a frame of an audio signal by employing a Hybrid Log-Likelihood Function (Lh) (step 508). This function is a combination of the Exact Log-Likelihood Function (Le) and the Asymptotic Log-Likelihood Function (La). The Exact function is computationally expensive, while the alternative Asymptotic function is computationally cheaper, but yields less exact results. The Exact function is:
where R is a Symmetric Positive-Definite (SPD) covariance matrix and has a block-Toeplitz structure, x is the frame of noisy speech data samples, and s is the hypothesized speech-plus-noise spectrum. The function includes a first part, before the second minus-sign, and a second part, after the second minus-sign. According to one embodiment including a single input signal, R may be a Toeplitz matrix. According to alternative embodiments including a plurality of input signals, R is a block-Toeplitz matrix as described above. The Asymptotic function is:
According to one embodiment, including a single input signal, the term “tr[f(θ)s(θ)−1]” is replaced with the term “f(θ)s(θ)−1”. According to one feature, the Asymptotic function shown above is used in embodiments including a plurality of input signals. The Asymptotic function also includes two parts: a first part before the plus-sign, and second part after the plus-sign. The part of the Asymptotic function before the plus corresponds to the first part of the Exact function. Similarly, the part of the Asymptotic function after the plus corresponds to the second part of the Exact function. Combining the first part of the Exact function, for which a known algorithm (the Preconditioned Conjugate Gradient algorithm) reduces the computation cost, with the second part of the Asymptotic function (which can be evaluated using a Fast Fourier Transform) yields the Hybrid Log-Likelihood Function Lh:
This hybrid of the two algorithms is less expensive computationally, without yielding significant loss in performance.
After the matcher has matched a segment of the audio signal to a template, the identified speech sound is digitally encoded for transmission (step 510). According to one implementation, only the index of the speech codebook entry, or of each stage-codebook entry 240, 244, and 248, correlated to the selected noisy speech template, as described above, is transmitted. Additionally, the index of the voicing codebook entry of the selected template may be transmitted. Thus, the noise codebook entry information may not be transmitted. Segments of the audio signal absent of voiced speech may represent pauses in the speech signal or could include unvoiced speech. According to one embodiment, these segments are also digitally encoded for transmission.
The noise spectrum parameter vector is then compared with the parameter vectors 214 of one or more of the noise codebook entries 212 (step 606). According to one embodiment, the comparison includes calculating the spectral distance between the noise spectrum parameter vector of the analyzed segment and each noise codebook entry 212.
Based on this comparison, the processor determines whether a noise codebook entry will be adapted or replaced (step 608). According to one embodiment, the processor compares the smallest spectral distance found in the comparison to a predetermined threshold value. If the smallest distance is below the threshold, the noise codebook entry corresponding to this distance is adapted as described below. If the smallest distance is greater than the threshold, a noise codebook entry parameter vector is replaced by the noise spectrum parameter vector.
If a noise codebook entry 212 will be adapted, the processor finds the best noise codebook entry match (step 610), e.g. the noise codebook entry 212 with the smallest spectral distance from the current noise spectrum. The best noise codebook entry match is combined with the noise spectrum parameter vector (step 612) to result in a modified noise codebook entry. According to one embodiment, autocorrelation vectors are generated for the best noise codebook entry match and the noise spectrum parameter vector. The modified codebook entry is created by combining 90% of the autocorrelation vector for best noise codebook entry match and 10% of the autocorrelation vector for the noise spectrum parameter vector. However, any relative proportion of the autocorrelation vectors may be used. The modified noise codebook entry replaces the best noise codebook entry match, and the codebook is updated (614).
Alternatively, a noise codebook entry parameter vector may be replaced by the noise spectrum parameter vector (step 608). According to another embodiment, the noise codebook entry is updated (step 614) by replacing the least frequently used noise codebook entry 212. According to a further embodiment, the noise codebook entry is updated (step 614) by replacing the least recently used noise codebook entry. According still another embodiment, the noise codebook entry is updated by replacing the least recently updated noise codebook entry.
According to one implementation, the signal may be encoded at a variable bit-rate. For example, a first segment may be encoded at 600 bps, as described above, and a second segment may be encoded at 300 bps, as described above. According to one configuration based on fixed duration segments composed of two frames, the encoding of each segment is determined as a function of the voicing properties of the frames. If it is determined that both frames of the segment are unvoiced and likely to be speech absent, a 2-bit code is transmitted together with a 13-bit speech codebook entry index. If it is determined that both frames are unvoiced and either frame is likely to have speech present, a different 2-bit code is transmitted together with a 39-bit speech codebook entry index. If at least one of the two frames is determined to be voiced, a 1-bit code is transmitted together with a 39-bit speech codebook entry index and a 14-bit voicing codebook entry index.
This encoding corresponds to one implementation of a variable-bit-rate vocoder which has been tested using 22.5 ms frames and yields an average bit rate of less than 969 bps. According to this implementation, about 20% of segments were classified as “unvoiced-unvoiced” and likely speech-absent, about 20% of segments were classified as “unvoiced-unvoiced” and likely speech-present, and about 60% of segments were classified as “voiced-unvoiced,” “unvoiced-voiced,” or “voiced-voiced.” Using the bit rates described above, and calculating the average occurrence of each type of segment, this results in an average of 3+8.2+32.4=43.6 bits per 45 ms segment, or less than 969 bps.
Those skilled in the art will know or be able to ascertain using no more than routine experimentation, many equivalents to the embodiments and practices described herein. Accordingly, it will be understood that the invention is not to be limited to the embodiments disclosed herein, but is to be understood from the following claims, which are to be interpreted as broadly as allowed under the law.
This application claims priority under 35 U.S.C. § 119(e) to U.S. Provisional Application No. 60/652,931 titled “Noise Robust Vocoder: Advanced Speech Encoding” filed Feb. 15, 2005, and U.S. Provisional Application No. 60/658,316 titled “Methods and Apparatus for Noise Robust Vocoder” filed Mar. 2, 2005, the entirety of which are hereby incorporated by reference.
The U.S. Government has a paid-up license in this invention and the right in limited circumstances to require the patent owner to license others on reasonable terms as provided for by the terms of Contract No. N00024-03-C-6314 awarded by the Naval Sea Systems Command.
Number | Date | Country | |
---|---|---|---|
60652931 | Feb 2005 | US | |
60658316 | Mar 2005 | US |