1. Field of the Invention
The invention relates to a filter comprising at least two inputs for receiving at least two signals, and an output for supplying an output signal.
The invention further relates to a signal processing device comprising such a filter.
The invention further relates to a teleconferencing system, a voice-controlled electronic device, and a noise cancellation system.
The invention further relates to a method for filtering at least two signals and for supplying an output signal.
Recent developments in audio and video systems require the use of multiple channel processing and reproduction with acoustic echo cancellers (AEC) and noise cancellers. For example in mini-group video conferencing systems, multiple channel transmission leads to a better “localization” of the diverse people in the room. This enhances the intelligibility and naturalness of the speech.
Further, multiple channel echo cancellation is needed in voice-controlled stereo audio and video equipment, such as television receivers, radio receivers, CD players, etc.
A multiple channel AEC can, in general, not be created by a simple combination of multiple single channel AEC's.
2. Description of the Related Art
U.S. Pat. No. 5,828,756 discloses a stereophonic communication system, such as a teleconferencing system, which involves selectively reducing the correlation between the individual channel signals of the stereophonic system. Herein, non-linearities are added to the input signals to reduce the correlation. However by adding these non-linearities, audible artifacts in the output signals are introduced. These non-linearities can (sometimes) be accepted in teleconferencing systems but are certainly not acceptable in other applications, such as supplying music, etc.
It is, inter alia, an object of the invention to provide a filter, which overcomes the objections of the prior art. To this end a first aspect of the invention provides a filter described in the opening paragraph, characterized in that the coefficient updates are determined in a transformed domain, and the filter comprises means for reducing the effect of correlation between the input signals on the coefficient updates.
Herewith, the performance of the adaptive filters is improved without a huge increase in computational complexity.
The invention and additional features, which may be optionally used to implement the invention to advantage, will be apparent from and elucidated with references to the examples described below, hereinafter and shown in the accompanying drawings, in which:
In the description, the equations, matrices, etc., are shown as described below. Signals are denoted by lower case characters, constants by upper case. Underlining is used for vectors, lower case for time domain, and upper case for frequency domain. Matrices are denoted by bold face upper case, e.g., I. The dimension is put in superscript (e.g., the B×Q matrix X is given by XB,Q, for a square matrix the second dimension is omitted). Diagonal matrices are denoted by a double underline, e.g., P, with its diagonal denoted as P=diag {P}. A subscript i, e.g., wi, denotes the i'th version. The k'th element of w is given by (w)k. Finally, appending [k] denotes the time index, (.)t denotes the transpose, (.)* denotes the complex conjugate, and (.)h denotes the Hermitain transpose (complex conjugate transpose).
A general multiple input adaptive FIR filter, depicted in
For S>a≧0, the FIR filter Wa performs the convolution of the signal xa[k] and the coefficients wa,0[k] . . . wa,N−1[k] of that filter. The output signal êa[k] of such a filter can be described as follows
With, for S>a≧0,
The output of the multiple input adaptive filter is given by
These filter parts of the separate (adaptive) filters W0 until WS−1 can be implemented efficiently in frequency domain with help of partitioning, block processing and Discrete Fourier Transforms (DFTs). A reduction in computational complexity is obtained since the convolutions per sample in the time domain transform to element-wise multiplications per block in the frequency domain. We use block processing with block length B and DFTs of length M, with M≧N+B−1. The transformation of the input signals can be described for S>a≧0 by
where FM is the M×M Fourier matrix. The (a,b)'th element (for 0≦a<M,0≦b<M) of the Fourier matrix is given by
where j=√{square root over (−1)}.
The filter can then be computed in the frequency domain, by
Note that the frequency domain filter coefficients are related to the time domain coefficients, for all S>a≧0 this can be denoted by
To obtain an efficient implementation, the block length B must be chosen in the same order as the filter length N, which results a large processing delay.
To reduce the processing delay, the filter can be partitioned into smaller pieces of length B, and with g=┌N/B┐, we get the implementation of
For the update part of the filter, one can use S separate update algorithms to improve convergence behavior the input signals can be decorrelated separately in the time domain by using RLS like algorithms, leading to a huge computational complexity. Complexity reduction can be obtained by implementation in the frequency domain with (partitioned) Block Frequency Domain Adaptive Filters, as described in G. P. M. Egelmeers, Real time realization of large adaptive filters, Ph.D. thesis, Eindhoven University of Technology, Eindhoven (The Netherlands), November 1995. When there is correlation between the input signals of the filters, this might still lead to very bad convergence behavior, due to the non-uniqueness problem.
In this application, it is proposed to use a partitioned algorithm in the frequency domain that reduces the effect of the cross-correlation between the input signals on the algorithm's convergence behavior. To reduce complexity, block processing, with block length A to compute the sum of A consecutive updates with each iteration, is used. For S>a≧0, the coefficient vectors waN[lA] are partitioned into gu parts of length Z with gu=┌N/Z┐ such that, for S>j≧0
with for S>j≧0 and gu>i≧0
A Fourier transform length L is used with L≧Z+A−1, we define the input signal Fourier transforms for S>a≧0 as
The diagonal matrices XaL[lA] contain the vector XaL[lA] as main diagonal, so for S>a≧0
XaL[lA]=diag{XaL[lA]}.
An overlap-save method is used to compute the correlation involved in the adaptation process in the frequency domain, the frequency domain transform of the residual signal vector equals
The set of update equations for the filter coefficients in the MFDAF (Multiple Input Frequency Domain Adaptive Filter) algorithm can now be defined for gu>i≧0 by
with
and the transformation matrix GS.Z,S.L is given by
with
GZ,L=(JZ0Z,L-Z)(FL)−1.
The input channel's power matrix PS.L [lA] is defined by
where
The expectation operator ε { } of the above equation has to be replaced by an estimation routine.
The power matrix PS.L [lA] can be estimated by
To reduce the number of multiplications, the step-size parameter α of equation is incorporated in the above power estimation routine by defining
so
2a(PSL[lA])−1=(PaSL[lA])−1 (12)
Estimation of the power matrix PaS.L[lA] can then be done by
Direct application of this algorithm leads to stability problems. When the input signal power in a certain frequency bin is very small, the power in that bin will decrease to a (very) small value. The inverse of the matrix will then have large values and will be inaccurate (due to numerical and estimation errors) In the ideal case, the eigenvalues of the power matrix estimate cancel the eigenvalues of the input signal power matrix. Due to estimation errors, this goal is only approximated, and the mismatch introduces a deviation from the ideal convergence behavior and might even lead to instability. Especially when some of the eigenvalues of the estimate of the inverse power matrix get large, and do not (exactly) cancel a (small) eigenvalue of the input signal power matrix, instability might occur. A lower limit to the eigenvalues of the estimate of the power matrix can solve this problem. In the single channel case (or when we forget the cross-terms), we can solve this problem by applying a lower limit to the lower values. We can do this because the eigenvalues of a diagonal matrix are equal to the elements of the diagonal, so we actually limit the eigenvalues. In the multiple channel case we also have to limit the eigenvalues to assure stability, but these no longer equal the elements on the diagonal.
We know, however, that all eigenvalues of the power matrix are positive. We can now create a lower limit on the eigenvalues by shifting them by the suggested minimum. We know that for all eigenvalues λ of a matrix A, the determinant of A+(Pmin−λ′)·I must be zero. So for all λ′, eigenvalue of A+Pmin·I, there must be a λ, eigenvalue of A, such that λ′=λ+Pmin (and the other way around). This means that by adding a constant Pmin to the main diagonal of a matrix, all eigenvalues of that matrix are shifted by Pmin). So we define:
which results in
This is shown in
The effect of this shifting of the eigenvalues on the (theoretical ideal) convergence behavior of the algorithm will be very small, and in practice the algorithm is much more stable.
Although Pa,limS.L[lA] is a sparse matrix, computing its inverse would still require L inversions of S×S matrices, which takes on the order of L·S3 operations. As we, however, do not need the inverse itself, but only its matrix-vector product with the input signals, we can also look at it as solving the system
which requires on the order of L·S2 operations. Another option is to estimate the inverse of PS.L[lA] directly, which also results in a number of operations proportional to L·S2. However, also in this case, we have to limit the eigenvalues to assure stability.
A simple algorithm is given by:
with
We can incorporate α, which results in
The above algorithm does not guarantee a matrix (PaS·L[lA])−1 with positive eigenvalues, and, therefore, introduces a lot of stability problems. In the single channel case, we are able to stabilize the algorithm by using a lower limit on the estimate, which automatically results in positive eigenvalues because the matrix is diagonal, but that is not possible in the multiple channel case.
Positive Eigenvalues
An exact transformation of an algorithm for estimating PaS·L[lA] with positive eigenvalues will lead to an estimation algorithm for the inverse with positive eigenvalues. This can be done by using the matrix inversion lemma. When there is matrix A such that
A=B+C·D·E (15)
then the inverse matrix (A)−1 of A can be expressed by
(A)−1=(B)−1−(B)−1·C·((D)−1+E·(B)−1·C)−1·E·(B)−1. (16)
By choosing
and
we obtain by using Equation (14)
with
Equation (19) involves no matrix inversion, and only L/2+1 divisions, as the matrix DL[lA] of Equation (20) is a real valued diagonal matrix.
Limits on the Eigenvalues
An operation on the inverse power matrix that is equivalent to adding a constant to the diagonal of the (non-inverse) power matrix, would solve the problem. Adding a full (S·L)×(S·L) identity matrix and trying to find an equivalent operation on the inverse power matrix with the matrix inversion lemma, results in an algorithm that requires the matrix inversion we would like to avoid, so we try
with
(XlimS·L,L[lA])*·(XlimS·L,L[lA])t=(XS·L,L[lA])*·(XS·L,L[lA])t+Pmin·2αL·IS·L. (22)
As the matrix I has rank S·L and the product matrices (XlimS·L.L[lA])*·(XlimS·L,L[lA])t and (XS·L,L[lA])*·(XS·L,L[lA])t both have a rank of (at most) L, this is not possible for S>1. As we need the average of (XS·L,L[lA])*·(XS·L,L[lA])t, we can find a solution by taking the average over S consecutive updates. We will try to find XlimS·L,L[lA] such that
with for i=l mod S
with
For S=1, we get
U1=√{square root over (Pmin·2αL)} (26)
For S>1, there are an infinite number of solutions. If we try to keep the maximum distortion (the largest matrix element) as small as possible, we have to choose, for all S>j≧0 and S>i≧0
uj,i=±√{square root over (Pmin·2αL·)} (27)
If there is a real symmetric matrix UL for S=L, then a real symmetric matrix U2L for S=2L is given by
Using the above equation, we can construct all U2i with i>0. If S+1 is not a power of two, then we will use the matrix U2i where 2i>S+1>2i−1, and use the last S rows. In Table 1, the power matrix estimation algorithm using a direct inverse estimation with limits on the eigenvalues is summarized.
Initialization
1. Su=2┌log
U1=√{square root over (Pmin)}·2αL
for i=1 to log2(Su)do
2. begin
end
3. Initialize power matrix.
Iteration
4. Calculate (DL[lA])−1
Table 1: Direct inverse power update with limits.
Note that the inverse of PS·L [lA] is also a sparse matrix with the same structure, and we define
where for 0≦i<S and 0≦j<S
Ti,jL[lA]=diag{Ti,jL[lA]}.
Further, also the output of the programmable filters is supplied to a dynamic echo suppressor DES6, which is coupled with an output to the output of the stereo echo canceller. Dynamic echo suppressors are described in International Patent Application No. WO-97-45995, corresponding to U.S. Pat. No. 6,546,099.
A full stereo communication requires four stereo AECs, two on the near end side and two on the far end side. In
Also in the multiple input noise canceller, we can apply a DES (which is, in fact, not suppressing an echo, but is similar to the DES in the AEC's case) and programmable filters to improve performance, as shown in
Above some examples of application of a stereo echo canceller and of a noise canceller are described. It is to be noted that the invention can be used in different applications and is not restricted to the described applications.
Number | Date | Country | Kind |
---|---|---|---|
99202026 | Jun 1999 | EP | regional |
00201142 | Mar 2000 | EP | regional |
Number | Name | Date | Kind |
---|---|---|---|
4355368 | Zeidler et al. | Oct 1982 | A |
4903247 | Van Gerwen et al. | Feb 1990 | A |
4939685 | Feintuch | Jul 1990 | A |
5828756 | Benesty et al. | Oct 1998 | A |
5987143 | Okuno et al. | Nov 1999 | A |
6246760 | Makino et al. | Jun 2001 | B1 |
6546099 | Janse | Apr 2003 | B1 |
Number | Date | Country |
---|---|---|
WO9745995 | Dec 1997 | WO |