Method and device for detecting voice activity

Information

  • Patent Grant
  • 7003452
  • Patent Number
    7,003,452
  • Date Filed
    Wednesday, August 2, 2000
    24 years ago
  • Date Issued
    Tuesday, February 21, 2006
    18 years ago
Abstract
The invention concerns a method for detecting voice activity in a digital speech signal, in at least a frequency band, for example by means of a detecting automaton whereof the status is controlled on the basis of an energy analysis of the signal. The control of said automaton, or more generally the determination of voice activity, comprises a comparison, in the frequency band, of two different versions of the speech signal one of which at least is a noise-corrected version.
Description
BACKGROUND OF THE INVENTION

The present invention relates to digital techniques for processing speech signals. It relates more particularly to the techniques utilizing voice activity detection so as to perform different processings depending on whether the signal does or does not carry voice activity.


The digital techniques in question come under varied domains: coding of speech for transmission or storage, speech recognition, noise reduction, echo cancellation, etc.


The main difficulty with processes for detecting voice activity is that of distinguishing between voice activity and the noise which accompanies the speech signal.


The document WO99/14737 describes a method of detecting voice activity in a digital speech signal processed on the basis of successive frames and in which an a priori denoising of the speech signal of each frame is carried out on the basis of noise estimates obtained during the processing of one or more previous frames, and the variations in the energy of the a priori denoised signal are analyzed so as to detect a degree of voice activity of the frame. By carrying out the detection of voice activity on the basis of an a priori denoised signal, the performance of this detection is substantially improved when the surrounding noise is relatively strong.


In the methods customarily used to detect voice activity, the energy variations of the (direct or denoised) signal are analyzed with respect to a long-term average of the energy of this signal, a relative increase in the instantaneous energy suggesting the appearance of voice activity.


An aim of the present invention is to propose another type of analysis allowing voice activity detection which is robust to the noise which may accompany the speech signal.


SUMMARY OF THE INVENTION

According to the invention, there is proposed a method for detecting voice activity in a digital speech signal in at least one frequency band, whereby the voice activity is detected on the basis of an analysis comprising a comparison, in the said frequency band, of two different versions of the speech signal, one at least of which is a denoised version obtained by taking account of estimates of the noise included in the signal.


This method can be executed over the entire frequency band of the signal, or on a subband basis, as a function of the requirements of the application using voice activity detection.


Voice activity can be detected in a binary manner for each band, or measured by a continuously varying parameter which may result from the comparison between the two different versions of the speech signal.


The comparison typically pertains to respective energies, evaluated in the said frequency band, of the two different versions of the speech signal, or to a monotonic function of these energies.


Another aspect of the present invention relates to a device for detecting voice activity in a speech signal, comprising signal processing means designed to implement a method as defined hereinabove.


The invention further relates to a computer program, loadable into a memory associated with a processor, and comprising portions of code for implementing a method as defined hereinabove upon the execution of the said program by the processor, as well as to a computer medium, on which such a program is recorded.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a schematic diagram of a signal processing chain using a voice activity detector according to the invention;



FIG. 2 is a schematic diagram of an exemplary voice activity detector according to the invention;



FIGS. 3 and 4 are flow charts of signal processing operations performed in the detector of FIG. 2;



FIG. 5 is a graphic showing an exemplary profile of energies calculated in the detector of FIG. 2 and illustrating the principle of voice activity detection;



FIG. 6 is a diagram of a detection automaton implemented in the detector of FIG. 2;



FIG. 7 is a schematic diagram of another embodiment of a voice activity detector according to the invention;



FIG. 8 is a flow chart of signal processing operations performed in the detector of FIG. 7;



FIG. 9 is a graphic of a function used in the operations of FIG. 8.





DETAILED DESCRIPTION

The device of FIG. 1 processes a digital speech signal s. The signal processing chain represented produces voice activity decisions δn,j which are usable in a manner known per se by application units, not represented, affording functions such as speech coding, speech recognition, noise reduction, echo cancellation, etc. The decisions δn,j can comprise a frequency resolution (index j), this making it possible to enhance applications operating in the frequency domain.


A windowing module 10 puts the signal s into the form of successive windows or frames of index n, each consisting of a number N of samples of digital signal. In a conventional manner, these frames may exhibit mutual overlaps. In the remainder of the present description, the frames will be regarded, without this being in any way limiting, as consisting of N=256 samples at a sampling frequency Fe of 8 kHz, with a Hamming weighting in each window, and overlaps of 50% between consecutive windows.


The signal frame is transformed into the frequency domain by a module 11 applying a conventional fast Fourier transform algorithm (FFT) for calculating the modulus of the spectrum of the signal. The module 11 then delivers a set of N=256 frequency components of the speech signal, which are denoted Sn,f, where n designates the current frame number, and f a frequency of the discrete spectrum. Owing to the properties of digital signals in the frequency domain, only the first N/2=128 samples are used.


To calculate the estimates of the noise contained in the signal s, we do not use the frequency resolution available at the output of the fast Fourier transform, but a lower resolution, determined by a number I of frequency subbands covering the [0,Fe/2] band of the signal. Each subband i (1≦i≦I) extends between a lower frequency f(i−1) and an upper frequency f(i), with f(0)=0, and f(I)=Fe/2. This chopping into subbands can be uniform (f(i)−f(i−1)=Fe/2I). It may also be non-uniform (for example according to a barks scale). A module 12 calculates the respective averages of the spectral components Sn,f of the speech signal on a subband basis, for example through a uniform weighting such as:
Sn,i=1f(i)-f(i-1)f[f(i-1),f(i)]Sn,f


This averaging reduces the fluctuations between the subbands by averaging the contributions of the noise in these subbands, and this will reduce the variance of the noise estimator. Furthermore, this averaging makes it possible to reduce the complexity of the system.


The averaged spectral components Sn,i are addressed to a voice activity detection module 15 and to a noise estimation module 16. {circumflex over (B)}n,i denotes the long-term estimate of the noise component produced by the module 16 in relation to frame n and to subband i.


These long-term estimates {circumflex over (B)}n,i may for example be obtained in the manner described in WO99/14737. It is also possible to use simple smoothing by means of an exponential window defined by a forget factor λB:

{circumflex over (B)}n,iB.{circumflex over (B)}n−1,i+(1−λB).Sn,i

with λB equal to 1 if the voice activity detector 15 indicates that subband i bears voice activity, and equal to a value lying between 0 and 1 otherwise.


Of course, it is possible to use other long-term estimates representative of the noise component included in the speech signal, these estimates may represent a long-term average, or else a minimum of the component Sn,j over a sufficiently long sliding window.



FIGS. 2 to 6 illustrate a first embodiment of the voice activity detector 15. A denoising module 18 executes, for each frame n and each subband i, the operations corresponding to steps 180 to 187 of FIG. 3, so as to produce two denoised versions {circumflex over (E)}p1,n,i, {circumflex over (E)}p2,n,i of the speech signal. This denoising is done by non-linear spectral subtraction. The first version {circumflex over (E)}p1,n,i, is denoised in such a way as not to be less, in the spectral domain, than a fraction β1i of the long-term estimate {circumflex over (B)}n−τ,i. The second version {circumflex over (E)}p2,n,i is denoised in such a way as not to be less, in the spectral domain, than a fraction β2j of the long-term estimate {circumflex over (B)}n−τ1,i. The quantity τ1 is a delay expressed as a number of frames, which may be fixed (for example τ1=1) or variable. The more confident one is in the voice activity detection, the smaller the delay will be. The fractions β1i and β2i (such that β1i2i) may be dependent on or independent of subband i. Preferred values correspond for β1i to an attenuation of 10 dB, and for β2i to an attenuation of 60 dB, i.e. β1i≈0.3 and β2i≈0.001.


In step 180, the module 18 calculates, with the resolution of the subbands i, the frequency response Hpn,i of the a priori denoising-filter, according to:
Hpn,j=Sn,i-αn-τ1,i·B^n-τ1,iSn-τ2,i

where τ2 is a positive or zero integer delay and α′n,i is a noise overestimation coefficient. This overestimation coefficient α′n,i may be dependent on or independent of the frame index n and/or the subband index i. In a preferred embodiment, it depends both on n and i, and it is determined as described in document WO99/14737. A first denoising is performed in step 181: {circumflex over (E)}pn,i=Hpn,i.Sn,i. In steps 182 to 184, the spectral components {circumflex over (E)}p1,n,i are calculated according {circumflex over (E)}p1,n,i=max ({circumflex over (E)}pn,i:β1i.{circumflex over (B)}n−τ1,i), and in steps 185 to 187, the spectral components {circumflex over (E)}p2,n,i are calculated according to {circumflex over (E)}p2,n,i=max({circumflex over (E)}pn,i:β2i.{circumflex over (B)}n−τ1,i).


The voice activity detector 15 of FIG. 2 comprises a module 19 which calculates energies of the denoised versions of the signal {circumflex over (E)}p1,n,i and {circumflex over (E)}p2,n,i respectively lying in m frequency bands designated by the index j (1≦j≦m, m≧1). This resolution may be the same as that of the subbands defined by the module 12 (index i), or a finer resolution of possibly as much as the whole of the useful band [0, Fe/2] of the signal (case m=1). By way of example, the module 12 can define I=16 uniform subbands of the band [0, Fe/2], and the module 19 can retain m=3 wider bands, each band of index j covering the subbands of index i ranging from imin(j) to imax(j), with imin(1)=1, imin(j+1)=imax(j)+1 for 1≦j<m, and imax(m)=1. In step 190 (FIG. 3), the module 19 calculates the energies per band:
El,n,j=i=imin(j)imax(j)[f(i)-f(i-1)]·E^pl,n,i2E2,n,j=i=imin(j)imax(j)[f(i)-f(i-1)]·E^p2,n,i2


A module 20 of the voice activity detector 15 performs a temporal smoothing of the energies E1,n,j and E2,n,j for each of the bands of index j, this corresponding to steps 200 to 205 for FIG. 4. The smoothing of these two energies is performed by means of a determined smoothing window by comparing the energy E2,n,j of the most denoised version with its previously calculated smoothed energy Ē2,n−1,j, or with a value of the order of this smoothed energy Ē2,n−1,j, (tests 200 and 201). This smoothing window can be an exponential window defined by a forget factor λ lying between 0 and 1. This forget factor λ can take three values: the one λr very close to 0 (for example λr=0) chosen in step 202 if E2,n,j≦Ē2,n−1,j; the second λq very close to 1 (for example λq=0.99999) chosen in step 203 if E2,n,j>ΔĒ2,n−1,j, Δ being a coefficient bigger than 1; and the third λp lying between 0 and λq (for example λp=0.98) chosen in step 204 if Ē2,n−1,j<E2,n−1,j≦ΔĒ2,n−1,j. The exponential smoothing with the forget factor λ is then performed conventionally in step 205 according to:

Ē1,n,j=λ.Ē1,n−1,j+(1−λ).E1,n,j
Ē2,n,j=λ.Ē2,n−1,j+(1−λ).E2,n,j


An exemplary variation over time of the energies E1,n,j and E2,n,j and of the smoothed energies Ē1,n,j, and Ē2,n,j is represented in FIG. 5. It may be seen that good tracking of the smoothed energies is achieved when the forget factor is determined on the basis of the variations in the energy E2,n,j corresponding to the most denoised version of the signal. The forget factor λp makes it possible to take into account the increases in the level of the background noise, the energy reductions being tracked by the forget factor λr. The forget factor λq very close to 1 means that the smoothed energies do not track the abrupt energy increases due to speech. However, the factor λq remains slightly less than 1 so as to avoid errors caused by an increase in the background noise which may arise during a fairly long period of speech.


The voice activity detection automaton is controlled in particular by a parameter resulting from a comparison of the energies E1,n,j and E2,n,j. This parameter can in particular be the ratio dn,j=E1,n,j/E2,n,j. It may be seen in FIG. 5 that this ratio dn,j allows proper detection of the speech phases (represented by hatching).


The control of the detection automaton can also use other parameters, such as a parameter related to the signal-to-noise ratio: snrn,j=E1,n,j1,n,j, this amounting to taking into account a comparison between the energies E1,n,j and Ē1,n,j. The module 21 for controlling the automata relating to the various bands of index j calculates the parameters dn,j and snrn,j in step 210, then determines the state of the automata. The new state δn,j of the automaton relating to band j depends on the previous state δn−1,j, on dn,j and on snrn,j, for example as indicated in the diagram of FIG. 6.


Four states are possible: δj=0 detects silence, or absence of speech; δj=2 detects the presence of voice activity; and the states δj=1 and δj=3 are intermediate states of ascent and descent. When the automaton is in the silence state (δn−1,j=0), it remains there if dn,j exceeds a first threshold α1j, and if it switches to the ascent state in the converse case. In the ascent state (δn−1,j=1), it returns to the silence state if dn,j exceeds a second threshold α2j; and it switches to the speech state in the converse case. When the automaton is in the speech state (δn−1,j=2), it remains there if snrn,j exceeds a third threshold α3j, and it switches to the descent state in the converse case. In the descent state (δn−1,j=3), the automaton returns to the speech state if snrn,j exceeds a fourth threshold α4j, and it returns to the silence state in the converse case. The thresholds α1j, α2j, α3j, and α4j may be optimized separately for each of the frequency bands j.


It is also possible for the automata relating to the various bands to be made to interact by the module 21.


In particular, it may force each of the automata relating to each of the subbands to the speech state as soon as one among them is in the speech state. In this case, the output of the voice activity detector 15 relates to the whole of the signal band.


The two appendices to the present description show a source code in the C++ language, with a fixed-point data representation corresponding to an implementation of the exemplary voice activity detection method described hereinabove. To embody the detector, one possibility is to translate this source code into executable code, to record it in a program memory associated with an appropriate signal processor, and to have it executed by this processor on the input signals of the detector. The function a_priori_signal_power presented in appendix 1 corresponds to the operations incumbent on the modules 18 and 19 of the voice activity detector 15 of FIG. 2. The function voice_activity_detector presented in appendix 2 corresponds to the operations incumbent on the modules 20 and 21 of this detector.


In the particular example of the appendices, the following parameters have been employed: τ1=1; τ2=0; β1i=0.3; β2i=0.001; m=3; Δ=4.953; λp=0.98; λq=0.99999; λr=0; α1j=α2j=α4j=1.221; α3j=1.649. Table 1 hereinbelow gives the correspondences between the notation employed in the above description and in the drawings and that employed in the appendix.












TABLE I









subband
I



E[subband]
Sn,i



module
Êpn,i or Êp1,n,i or Êp2,n,i



param.beta_a_priori1
β1j



param.beta_a_priori2
β2j



vad
j-1



param.vad_number
m



P1[vad]
E1,n,j−1



P1s[vad]
Ē1,n,j−1



P2[vad]
E2,n,j−1



P2s[vad]
Ē2,n,j−1



DELTA_P
Log(Δ)



d
Log(dn,j)



snr
Log(snrn,j)



NOISE
silence state



ASCENT
ascent state



SIGNAL
speech state



DESCENT
descent state



D_NOISE
Log(α1j)



D_SIGNAL
Log(α2j)



SNR_SIGNAL
LOG(α3j)



SNR_NOISE
Log(α4j)










In the variant embodiment illustrated by FIG. 7, the denoising module 25 of the voice activity detector 15 delivers a single denoised version {circumflex over (E)}pn,i of the speech signal, so that the module 26 calculates its energy E2,n,j for each band j. The other version, in which the module 26 calculates the energy, is represented directly by the non-denoised samples Sn,i.


As before, various denoising processes may be applied by the module 25. In the example illustrated by steps 250 to 256 of FIG. 8, the denoising is done by nonlinear spectral subtraction with a noise overestimation coefficient dependent on a quantity ρ related to the signal-to-noise ratio. In steps 250 to 252, a preliminary denoising is performed for each subband of index i according to:

S′n,i=max(Sn,i−α.{circumflex over (B)}n−1,i;β.{circumflex over (B)}n−1,i)

the preliminary overestimation coefficient being for example α=2, and the fraction β possibly corresponding to a noise attenuation of the order of 10 dB.


The quantity ρ is taken equal to the ratio S′n,i/Sn,i in step 253. The overestimation factor f(ρ) varies in a nonlinear manner with the quantity ρ, for example as represented in FIG. 9. For the values of ρ closest to 0 (ρ<ρ1), the signal-to-noise ratio is low, and it is possible to take an overestimation factor f(ρ)=2. For the highest values of ρ (ρ2≦ρ≦1), the noise is weak and need not be overestimated (f(ρ)=1). Between ρ1 and ρ2, f(ρ) decreases from 2 to 1, for example linearly. The denoising proper, providing the version {circumflex over (E)}pn,i is performed in steps 254 to 256:

Êpn,i=max(Sn,i−f(ρ).{circumflex over (B)}n−1,i;β.{circumflex over (B)}n−1,i)


The voice activity detector 15 considered with reference to FIG. 7 uses, in each frequency band of index j (and/or in full band), a detection automaton having two states, silence or speech. The energies E1,n,j and E2,n,j calculated by the module 26 are respectively those contained in the components Sn,i of the speech signal and those contained in the denoised components {circumflex over (E)}pn,i calculated over the various bands as indicated in step 260 of FIG. 8. The comparison of the two different versions of the speech signal pertains to respective differences between the energies E1,n,j and E2,n,j and a lower bound of the energy E2,n,j of the denoised version.


This lower bound E2min,j can in particular correspond to a minimum value, over a sliding window, of the energy E2,n,j of the denoised version of the speech signal in the frequency band considered. In this case, a module 27 stores in a memory of the first-in first-out type (FIFO) the L most recent values of the energy E2,n,j of the denoised signal in each band j, over a sliding window representing for example of the order of 20 frames, and delivers the minimum energies
E2min,j=minE2,n-k,j0k<L

over this window (step 270 of FIG. 8). In each band, this minimum energy E2min,j serves as lower bound for the module 28 for controlling the detection automaton, which uses a measure Mj given by
Mj=E2,n,j-E2min,iE2,n,j-E2min,j.(step280)


The automaton can be a simple binary automaton using a threshold Aj, possibly dependent on the band considered: If Mj≧Aj, the output bit δn,j of the detector represents a silence state of the band j, and if Mj≦Aj, it represents a speech state. As a variant, the module 28 could deliver a nonbinary measure of the voice activity, represented by a decreasing function of Mj.


As a variant, the lower bound E2min,j used in step 280 could be calculated with the aid of an exponential window, with a forget factor. It could also be represented by the energy over band j of the quantity β.{circumflex over (B)}n−1,i serving as floor in the denoising by spectral subtraction.


In the foregoing, the analysis performed in order to decide on the presence or absence of voice activity pertains directly to energies of different versions of the speech signal. Of course, the comparisons could pertain to a monotonic function of these energies, for example a logarithm, or to a quantity having similar behavior to the energies according to voice activity (for example the power).












APPENDIX 1















/*******************************************************************


******


* description


* -----------


* NSS module:


* signal power before VAD


*


*******************************************************************


******/


/*------------------------------------------------------------------


-----*








*
included files







*------------------------------------------------------------------


-----*/


#include <assert.h>


#include “private.h”


/*------------------------------------------------------------------


-----*








*
private







*------------------------------------------------------------------


-----*/


Word32 power(Word16 module, Word16 beta, Word16 thd, Word16 val);


/*------------------------------------------------------------------


-----*








*
a_priori_signal_power







*------------------------------------------------------------------


-----*/


void a_priori_signal_power


(








/* IN */
Word16 *E, Word16 *internal_state, Word16 *max_noise,







Word16 *long_term_noise,









Word16 *frequential_scale,


/* IN&OUT */
Word16 *alpha,


/* OUT */
Word32 *P1, Word32 *P2


)


{









int vad;



for(vad = 0; vad < param.vad_number; vad++) {









int start = param.vads[vad].first_subband_for_power;



int stop = param.vads[vad].last_subband;



int subband;



int uniform_subband;



uniform_subband = 1;



for(subband = start; subband <= stop; subband++)









if(param.subband_size[subband] != param.subband_size[start]







)









uniform_subband = 0;









P1[vad] = 0; move32();



P2[vad] = 0; move32();



test(); if(sub(internal_state[vad], NOISE) == 0) {









for(subband = start; subband <= stop; subband++) {









Word32 pwr;



Word16 shift;



Word16 module;



Word16 alpha_long_term;



alpha_long_term = shr(max_noise[subband], 2); move16();



test(); test(); if(sub(alpha_long_term, long_term_noise







[subband]) >= 0) {









alpha[subband] = 0×7fff; move16();



alpha_long_term = long_term_noise[subband]; move16();









} else if(sub(max_noise[subband], long_term_noise[subban







d]) < 0) {









alpha[subband] = 0×2000; move16();



alpha_long_term = shr(long_term_noise[subband],2); move







16();









} else {









alpha[subband] = div_s(alpha_long_term, long_term_noise







[subband]); move16();









}



module = sub(E[subband], shl(alpha_long_term, 2)); move







16();









if(uniform_subband) {









shift = shl(frequential_scale[subband], 1); move16();









} else {









shift = add(param.subband_shift[subband], shl(frequen







tial_scale[subband], 1)); move16();









}



pwr = power(module, param.beta_a_priori1, long_term_noise







[subband], long_term_noise[subband]);









pwr = L_shr(pwr, shift);



P1[vad] = L_add(P1[vad], pwr); move32();



pwr = power(module, param.beta_a_priori2, long_term_noise







[subband], long_term_noise[subband]);









pwr = L_shr(pwr, shift);



P2[vad] = L_add(P2[vad], pwr); move32();









}









} else {









for(subband = start; subband <= stop; subband++) {









Word32 pwr;



Word16 shift;



Word16 module;



Word16 alpha_long_term;



alpha_long_term = mult(alpha[subband], long_term_noise







[subband]); move16();









module = sub{E[subband], shl(alpha_long_term, 2}); move







16();









if(uniform_subband) {









shift = sh1(frequential_scale[subband], 1); move16();









} else {









shift = add(param.subband_shift[subband], sh1(frequen







tial_scale[subband], 1)); move16();









}



pwr = power(module, param.beta_a_priori1, long_term_noise







[subband], E[subband]);









pwr = L_shr(pwr, shift);



P1[vad] = L_add(P1[vad], pwr); move32();



pwr = power(module, param.beta_a_priori2, long_term_noise







[subband], E[subband]);









pwr = L_shr(pwr, shift);



P2[vad] = L_add(P2[vad], pwr); move32();









}









}









}







}


/*------------------------------------------------------------------


-----*








*
power







*------------------------------------------------------------------


-----*/


Word32 power(Word16 module, Word16 beta, Word16 thd, Word16 val)


{









Word32 power;



test(); if(sub(module, mult(beta, thd)) <= 0) {









Word16 hi, lo;



power = L_mult(val, val); move32();



L_Extract(power, &hi, &lo);



power = Mpy_32_16(hi, lo, beta); move32();



L_Extract(power, &hi, &lo);



power = Mpy_32_16(hi, lo, beta); move32();









} else {









power = L_mult(module, module); move32();









}



return(power);







}



















APPENDIX 2















/*******************************************************************


******


* description


* -----------


* NSS module:


* VAD


*


*******************************************************************


******/


/*------------------------------------------------------------------


-----*








*
included files







*------------------------------------------------------------------


-----*/


#include <assert.h>


#include “private-h”


#include “simutool.h”


/*------------------------------------------------------------------


-----*








*
private







*------------------------------------------------------------------


-----*/








#define DELTA_P
(1.6 * 1024)


#define D_NOISE
(.2 * 1024)


#define D_SIGNAL
(.2 * 1024)


#define SNR_SIGNAL
(.5 * 1024)


#define SNR_NOISE
(.2 * 1024)







/*------------------------------------------------------------------


-----*








*
voice_activity_detector







/*------------------------------------------------------------------


-----*/


void voice_activity_detector


{








/* IN */
Word32 *P1, Word32 *P2, Word16 frame_counter,


/* IN&OUT */
Word32 *P1s, Word32 *P2s, Word16 *internal_state,


/* OUT */
Word16 *state







}


{









int vad;



int signal;



int noise;



signal = 0; move16();



noise = 1; move16();



for(vad = 0; vad < param.vad_number; vad++) {









Word16 snr, d;



Word16 logP1, logP1s;



Word16 logP2, logP2s;



logP2 = logfix(P2[vad]); move16();



logP2s = logfix(P2s[vad]); move16();



test(); if(L_sub(P2[vad], P2s[vad]) > 0) {









Word16 hi1, lo1;



Word16 hi2, lo2;



L_Extract(L_sub(P1[vad], P1s[vad]), &hi1, &lo1);



L_Extract(L_sub(P2[vad], P2s[vad]), &hi2, &lo2);



test(); if(sub(sub{logP2, logP2s}, DELTA_P) < 0) {









P1s[vad] = L_add(P1s[vad], L_shr(Mpy_32_16(hi1, lo1, 0×6







666), 4)); move32();









P2s[vad] = L_add(P2s[vad], L_shr(Mpy_32_16(hi2, lo2, 0×6







666), 4)); move32();









} else {









P1s[vad] = L_add(P1s[vad], L_shr(Mpy_32_16(hi1, lo1, 0×6







8db), 13)); move32();









P2s[vad] = L_add(P2s[vad], L_shr(Mpy_32_16(hi2, lo2, 0×6







8db), 13)); move32();









}









} else {









P1s[vad] = P1[vad]; move32();



P2s[vad] = P2[vad]; move32();









}



logP1 = logfix(P1[vad]); move16();



logP1s = logfix(P1s[vad]); move16();



d = sub(logP1, logP2); move16();



snr = sub(logP1, logP1s); move16();



ProbeFix16(“d”, &d, 1, 1.);



ProbeFix16(“_snr”, &snr, 1, 1.);







{









Word16 pp;



ProbeFix16(“p1”, &logP1, 1, 1.);



ProbeFix16(“p2”, &logP2, 1, 1.);



ProbeFix16(“p1s”, &logP1s, 1, 1.);



ProbeFix16(“p2s”, &logP2s, 1, 1.);



pp = logP2 − logP2s;



ProbeFix16(“dp”, &pp, 1, 1.);







}









test(); if(sub(internal_state[vad], NOISE) == 0)









goto LABEL_NOISE;









test(); if(sub(internal_state[vad], ASCENT) == 0)









goto LABEL_ASCENT;









test(); if(sub(internal_state[vad], SIGNAL) == 0)









goto LABEL_SIGNAL;









test(); if(sub(internal_state[vad], DESCENT) == 0)









goto LABEL_DESCENT;









LABEL_NOISE:



test(); if(sub(d, D_NOISE) < 0) {









internal_state[vad] = ASCENT; move16();









}



goto LABEL_END_VAD;



LABEL_ASCENT:



test(); if(sub(d, D_SIGNAL) < 0) {









internal_state[vad] = SIGNAL; move16();



signal = 1; move16();



noise = 0; move16();









} else {









internal_state[vad] = NOISE; move16();









}



goto LABEL_END_VAD;



LABEL_SIGNAL:



test(); if(sub(snr, SNR_SIGNAL) < 0) {









internal_state[vad] = DESCENT; move16();









} else {









signal = 1; move16();









}



noise = 0; move16();



goto LABEL_END_VAD;



LABEL_DESCENT:



test(); if(sub(snr, SNR_NOISE) < 0) {









internal_state[vad] = NOISE; move16();









} else {









internal_state[vad] = SIGNAL; move16();



signal = 1; move16();



noise = 0; move16();









}



goto LABEL_END_VAD;



LABEL_END_VAD:



;







}


*state = TRANSITION; move16();


test(); test(); if(signal != 0) {









test(); if(sub(frame_counter, param.init_frame_number) >= 0) {









for(vad = 0; vad < param.vad_number; vad++) {









internal_state[vad] = SIGNAL; move16();









}



*state = SIGNAL; move16();









}



} else if(noise != 0) {









*state = NOISE; move16();









}







}








Claims
  • 1. Method for detecting voice activity in a digital speech signal in at least one frequency band, wherein the voice activity is detected on the basis of an analysis comprising the step of comparing two different versions of the speech signal, wherein the two different versions of the speech signal are two versions denoised by non-linear spectral subtraction, wherein a first of the two versions is denoised in such a way as not to be less, in the spectral domain, than a first fraction of a long-term estimate representative of a noise component included in the speech signal, and the second of the two versions is denoised in such a way as not to be less, in the spectral domain, than a second fraction of said long-term estimate, smaller than said first fraction.
  • 2. Method according to claim 1, wherein said comparison is performed on respective energies, evaluated in said frequency band, of the two different versions of the speech signal, or to a monotonic function of said energies.
  • 3. Method according to claim 1, wherein said analysis further comprises a time smoothing of the energy of one of said versions of the speech signal, and a comparison between the energy of said version and the smoothed energy.
  • 4. Method according to claim 3, wherein the comparison between the energy of said version and the smoothed energy controls transitions of a voice activity detection automaton from a speech state to a silence state, and wherein the comparison of the two different versions of the speech signal controls transitions of the detection automaton from the silence state to the speech state.
  • 5. Method according to claim 1, wherein said analysis further comprises a time smoothing of the energy of each of the two versions of the speech signal, by means of a smoothing window determined by comparing the energy of the second of the two versions with the smoothed energy of the second of the two versions.
  • 6. Method according to claim 5, wherein the smoothing window is an exponential window defined by a forgetting factor.
  • 7. Method according to claim 6, comprising the step of allocating a substantially zero value to the forgetting factor when the energy of the second of the two versions is less than a value of the order of the smoothed energy of the second of the two versions.
  • 8. Method according to claim 7, comprising the step of allocating a first value substantially equal to 1 to the forgetting factor when the energy of the second of the two versions is greater than said value of the order of the smoothed energy multiplied by a coefficient bigger than 1, and allocating a second value lying between 0 and said first value to the forgetting factor when the energy of the second of the two versions is greater than said value of the order of the smoothed energy and less than said value of the order of the smoothed energy multiplied by said coefficient.
  • 9. Method according to claim 1, wherein the first and second fractions correspond substantially to attenuations of 10 dB and 60 dB, respectively.
  • 10. Method according to claim 1, wherein the comparison of the two different versions of the speech signal is performed on respective differences between the energies of said two versions in said frequency band and a lower bound of the energy of the denoised version of the speech signal in said frequency band.
  • 11. Device for detecting voice activity in a speech signal, comprising signal processing means for analyzing the speech signal in at least one frequency band, wherein the processing means comprise: first non-linear spectral subtraction means to provide a first version of the speech signal as a denoised version which is not less, in the spectral domain, than a first fraction of a long-term estimate representative of a noise component included in the speech signal;second non-linear spectral subtraction means to provide a second version of the speech signal as a denoised version which is not less, in the spectral domain, than a second fraction of said long-term estimate, said second fraction being smaller than said first fraction; andmeans for comparing the first and second versions of the speech signal.
  • 12. Device according to claim 11, wherein the processing means comprise means for evaluating, in said frequency band, energies of said first and second versions of the speech signal, whereby inputs of the comparison means comprise said energies or a monotonic function of said energies.
  • 13. Device according to claim 11, wherein the processing means further comprises means for performing a time smoothing of the energy of one of said first and second versions of the speech signal, and means for comparing the energy of said version and the smoothed energy.
  • 14. Device according to claim 13, wherein the processing means comprise a voice activity detection automaton having a plurality of states including a speech state and a silence state, means for controlling transitions of the voice activity detection automaton from the speech state to the silence state based on a comparison between the energy of said one of said first and second versions and the smoothed energy, and means for controlling transitions of the voice activity detection automaton from the silence state to the speech state based on a comparison of the first and second versions of the speech signal.
  • 15. Device according to claim 11, wherein the processing means further comprises means for performing a time smoothing of the energy of each of the first and second versions of the speech signal, by means of a smoothing window determined by comparing an energy of the second version with the smoothed energy of the second version.
  • 16. Device according to claim 15, wherein the smoothing window is an exponential window defined by a forgetting factor.
  • 17. Device according to claim 16, wherein the processing means further comprises means for allocating a substantially zero value to the forgetting factor when the energy of the second version is less than a value of the order of the smoothed energy of the second version.
  • 18. Device according to claim 17, wherein the processing means further comprises means for allocating a first value substantially equal to 1 to the forgetting factor when the energy of the second version is greater than said value of the order of the smoothed energy multiplied by a coefficient bigger than 1, and for allocating a second value lying between 0 and said first value to the forgetting factor when the energy of the second version is greater than said value of the order of the smoothed energy and less than said value of the order of the smoothed energy multiplied by said coefficient.
  • 19. Device according to claim 11, wherein the first and second fractions correspond substantially to attenuations of 10 dB and 60 dB, respectively.
  • 20. Device according to claim 11, wherein the comparison of the first and second versions of the speech signal is performed on respective differences between the energies of said first and second versions in said frequency band and a lower bound of the energy of the denoised version of the speech signal in said frequency band.
  • 21. A computer program product, loadable into a memory associated with a processor, and comprising portions of code for execution by the processor to detect voice activity in an input digital speech signal in at least one frequency band, whereby the voice activity is detected on the basis of an analysis comprising the step of comparing two different versions of the speech signal, wherein the two different versions of the speech signal are two versions denoised by non-linear spectral subtraction, wherein a first of the two versions is denoised in such a way as not to be less, in the spectral domain, than a first fraction of a long-term estimate representative of a noise component included in the speech signal, and the second of the two versions is denoised in such a way as not to be less, in the spectral domain, than a second fraction of said long-term estimate, smaller than said first fraction.
  • 22. A computer program product according to claim 21, wherein said comparison is performed on respective energies, evaluated in said frequency band, of the two different versions of the speech signal, or to a monotonic function of said energies.
  • 23. A computer program product according to claim 21, wherein said analysis further comprises a time smoothing of the energy of one of said versions of the speech signal, and a comparison between the energy of said version and the smoothed energy.
  • 24. A computer program product according to claim 23, wherein the comparison between the energy of said version and the smoothed energy controls transitions of a voice activity detection automaton from a speech state to a silence state, and wherein the comparison of the two different versions of the speech signal controls transitions of the detection automaton from the silence state to the speech state.
  • 25. A computer program product according to claim 21, wherein said analysis further comprises a time smoothing of the energy of each of the two versions of the speech signal, by means of a smoothing window determined by comparing the energy of the second of the two versions with the smoothed energy of the second of the two versions.
  • 26. A computer program product according to claim 25, wherein the smoothing window is an exponential window defined by a forgetting factor.
  • 27. A computer program product according to claim 26, wherein said analysis further comprises the step of allocating a substantially zero value to the forgetting factor when the energy of the second of the two versions is less than a value of the order of the smoothed energy of the second of the two versions.
  • 28. A computer program product according to claim 27, wherein said analysis further comprises the steps of allocating a first value substantially equal to 1 to the forgetting factor when the energy of the second of the two versions is greater than said value of the order of the smoothed energy multiplied by a coefficient bigger than 1, and allocating a second value lying between 0 and said first value to the forgetting factor when the energy of the second of the two versions is greater than said value of the order of the smoothed energy and less than said value of the order of the smoothed energy multiplied by said coefficient.
  • 29. A computer program product according to claim 21, wherein the first and second fractions correspond substantially to attenuations of 10 dB and 60 dB, respectively.
  • 30. A computer program product according to claim 21, wherein the comparison of the two different versions of the speech signal is performed on respective differences between the energies of said two versions in said frequency band and a lower bound of the energy of the denoised version of the speech signal in said frequency band.
Priority Claims (1)
Number Date Country Kind
99 10128 Aug 1999 FR national
PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/FR00/02220 8/2/2000 WO 00 5/3/2001
Publishing Document Publishing Date Country Kind
WO01/11605 2/15/2001 WO A
US Referenced Citations (12)
Number Name Date Kind
4682361 Selbach et al. Jul 1987 A
4982341 Laurent Jan 1991 A
6427134 Garner et al. Jul 2002 B1
6453285 Anderson et al. Sep 2002 B1
6453289 Ertem et al. Sep 2002 B1
6453291 Ashley Sep 2002 B1
6477489 Lockwood et al. Nov 2002 B1
6549586 Gustafsson et al. Apr 2003 B1
6591234 Chandran et al. Jul 2003 B1
6604071 Cox et al. Aug 2003 B1
6658380 Lockwood et al. Dec 2003 B1
6717991 Gustafsson et al. Apr 2004 B1
Foreign Referenced Citations (1)
Number Date Country
WO 9914737 Mar 1999 WO