The following invention is in the field of acoustic analysis, spatial sound recording, microphone array signal processing, and spatial filtering. Some embodiments of the present invention relate to a method that can be used to determine the filter coefficients of a diffuse sound filter, i.e., a filter for extracting diffuse sound (reverberant sound) from the recordings with a microphone array. Some embodiments relate to a corresponding computer program. Some embodiments relate to an apparatus that can be used to determine the filter coefficients of a diffuse sound filter.
Sound acquisition with microphone arrays in reverberant environments typically aims at capturing the direct sound of the sound sources while attenuating noise and reverberation. For many applications it would be beneficial if we were able to extract also the reverberant sound while suppressing the direct sound and noise. For instance in spatial sound reproduction [Pulkki2007, Thiergart2013, Kowalczyk2013], the reverberation present at the recording side needs to be reproduced at the reproduction side to recreate the desired spatial impression. Moreover, given an estimate of the reverberant sound, we can compute parameters such as the signal-to-reverberation ratio or reverberant sound power, which represent crucial information for various other applications.
While the estimation of direct sound components (e.g., using source separation, dereverberation, or noise reduction) is well addressed in literature, only few approaches exist for extracting reverberant sound. Usually, reverberation is modeled as a (time-varying) diffuse sound field. To extract the diffuse sound, single-channel filters have been used recently (e.g., in [Pulkki2007, Thiergart2013]), which yield poor performance when multiple sources are active or for transient-like signals. A better performance can be achieved with multi-channel filters (e.g., [Kowalczyk2013, Thiergart2013b]). Unfortunately, currently existing multi-channel filters are not optimal and do not yield a suitable directivity pattern for capturing diffuse sound.
It would therefore be desirable to provide a diffuse sound filter having improved performance in terms of diffuse sound extraction and/or direct sound suppression. It may also be desirable that the diffuse sound filter has a directional response that is highly omnidirectional, with the exception of directions of arrival of direct sound components. A highly omnidirectional directional response is desired since the diffuse sound arrives from all directions at the microphone array.
According to an embodiment, a method may have the steps of: defining a linear constraint for filter coefficients of a diffuse sound filter, the linear constraint being based on a spatial coherence between a first diffuse sound portion in a first microphone signal and a second diffuse sound portion in a second microphone signal, the first microphone signal being captured by a first microphone and the second microphone signal being captured by a second microphone spaced apart from the first microphone in a known manner; calculating at least one of a direction of arrival of at least one direct sound, signal statistics over the first and second microphone signals, and noise statistics over the first and second microphone signals; and determining the filter coefficients of the diffuse sound filter by solving an optimization problem concerning at least one of the direction of arrival of the at least one direct sound, the signal statistics, and the noise statistics while considering the linear constraint for the filter coefficients.
According to another embodiment, a non-transitory digital storage medium may have stored thereon a computer program for performing a method comprising: defining a linear constraint for filter coefficients of a diffuse sound filter, the linear constraint being based on a spatial coherence between a first diffuse sound portion in a first microphone signal and a second diffuse sound portion in a second microphone signal, the first microphone signal being captured by a first microphone and the second microphone signal being captured by a second microphone spaced apart from the first microphone in a known manner; calculating at least one of a direction of arrival of at least one direct sound, signal statistics over the first and second microphone signals, and noise statistics over the first and second microphone signals; and determining the filter coefficients of the diffuse sound filter by solving an optimization problem concerning at least one of the direction of arrival of the at least one direct sound, the signal statistics, and the noise statistics while considering the linear constraint for the filter coefficients, when said computer program is run by a computer.
According to another embodiment, an apparatus may have: a linear constraint calculator configured to define a linear constraint for filter coefficients of a diffuse sound filter, the linear constraint being based on a spatial coherence between a first diffuse sound portion in a first microphone signal and a second diffuse sound portion in a second microphone signal, the first microphone signal being captured by a first microphone and the second microphone signal being captured by a second microphone spaced apart from the first microphone in a known manner; a calculator configured to calculate at least one of a direction of arrival of at least one direct sound, signal statistics over the first and second microphone signals, and noise statistics over the first and second microphone signals; and a filter coefficients calculator configured to determine the filter coefficients of the diffuse sound filter by solving an optimization problem concerning at least one of the direction of arrival of the at least one direct sound, the signal statistics, and the noise statistics while considering the linear constraint for the filter coefficients.
A method is provided which comprises defining a linear constraint for filter coefficients of a diffuse sound filter. The linear constraint is based on a spatial coherence between a first diffuse sound portion in a first microphone signal and a second diffuse sound portion in a second microphone signal. The first microphone signal is captured by a first microphone and the second microphone signal is captured by a second microphone spaced apart from the first microphone in a known manner. The method also comprises calculating at least one of a direction of arrival of at least one direct sound, signal statistics over the first and second microphone signals, and noise statistics over the first and second microphone signals. The method further comprises determining the filter coefficients of the diffuse sound filter by solving an optimization problem concerning at least one of the direction of arrival of the at least one direct sound, the signal statistics, and the noise statistics while considering the linear constraint for the filter coefficients.
Embodiments provide a computer program for implementing the above-described method when being executed on a computer or signal processor is provided.
Further embodiments provide an apparatus comprising a linear constraint calculator configured to define a linear constraint for filter coefficients of a diffuse sound filter. The linear constraint is based on a spatial coherence between a first diffuse sound portion in a first microphone signal and a second diffuse sound portion in a second microphone signal. The first microphone signal is or has been captured by a first microphone and the second microphone signal is or has been captured by a second microphone spaced apart from the first microphone in a known manner. The apparatus also comprises a statistics calculator configured to calculate at least one of a direction of arrival of at least one direct sound, signal statistics over the first and second microphone signals, and noise statistics over the first and second microphone signals and the second microphone signal. The apparatus further comprises a filter coefficients calculator configured to determine the filter coefficients of the diffuse sound filter by solving an optimization problem concerning at least one of the direction of arrival of the at least one direct sound, the signal statistics, and the noise statistics while considering the linear constraint for the filter coefficients.
Embodiments are based on the insight that a diffuse sound filter may be determined while taking into account at least one linear constraint that is related to the diffuse sound portions of the microphone signals.
Embodiments of the present invention will be detailed subsequently referring to the appended drawings, in which:
The terms “direct sound” and “diffuse sound” are defined as follows.
Direct sound: sounds that arrive from mainly a specific prominent direction at the microphones. The direct sound can represent for instance the sound travelling directly from the sound source to the microphone or a distinct room reflection. Direct sounds can be for instance plane waves or spherical waves with a specific direction of arrival. When the direction of arrival of a direct sound is known, one can compute the relative transfer function of the direct sound between the microphones given that the microphone geometry is known.
Diffuse sound: sound that arrives at the microphones from all directions. The diffuse sound can represent for instance the later reverberation present in a room. Typically, no prominent direction-of-arrival can be associated with a diffuse sound (isotropic sound field), i.e., the sound is arriving with equal mean power from all directions. Moreover, the relative transfer functions of the diffuse sound between the microphones should be assumed random and unobservable. However, the mean relative transfer functions of the diffuse sound between the microphones are usually known for specific microphone setups and diffuse field assumptions or can be measured.
The following subsections summarize existing approaches to extract diffuse sound (or reverberant sound) from a microphone recording. In the following, M denotes the number of microphones used. We assume that all microphone signals have been transformed into the time-frequency domain where k is the frequency index and n is the time index (note that the filters can typically be applied in the time-domain as well). The microphones capture L plane waves (referred to as direct sound) propagating in a diffuse field. The DOA of the l-th plane wave is represented by the unit-norm vector nl(k,n). In the time-frequency domain, the signal of the m-th (omnidirectional) microphone can be written as
Here, Xl(k,n) is the sound pressure of the l-th plane wave, Xd(k,n,dm) is the diffuse sound, Xn(k,n,dm) is a stationary noise (e.g., self-noise or background noise), and dm is a vector describing the microphone position (of the m-th microphone) in a given coordinate system.
The aim of this invention is to estimate Xd (k,n,dm) at position dm.
Single-channel filters extract the diffuse sound from a single microphone signal (=1). Such filters are used for example in Directional Audio Coding [Pulkki2007] or in the Virtual Microphone [Thiergart2013].
An estimate of the diffuse sound is found by multiplying one of the microphone signals, for example the microphone signal of the first microphone X1 (k,n), with a filter H(k,n), e.g.,
{circumflex over (X)}d(k,n,dm)=X1(k,n)H(k,n)
Usually, the filter H(k,n) is a Wiener filter, which is given by
where ϕd is the power of the diffuse sound and ϕu is the power of the plane waves and the stationary noise. In some applications, the square-root Wiener filter (i.e., the square-root of H) is used instead of the Wiener filter. Note that in order to compute H(k,n), one has to estimate the power ϕd and ϕu. For this purpose, we can consider for instance the signal-to-diffuse ratio (SDR), which can be estimated as explained in [Thiergart2012]. Alternatively, H(k,n) can be found by estimating the so-called diffuseness, as described in [Pulkki2007, Thiergart2013]. Estimating the SDR or diffuseness typically demands more than one microphone. Nevertheless, the diffuse sound is finally obtained by filtering a single microphone signal.
An example system for extracting the diffuse sound with a single-channel filter is illustrated in
Multi-channel filters consider M>1 microphones. Such filters have been used for instance in [Thiergart2013b, Kowalczyk2013]. For the following derivations, let us represent the M microphone signals by a vector x(k,n)=[X1(k,n), X2(k,n), . . . , XM(k,n)]T. The diffuse sound at the m-th microphone is estimated via a linear combination of the M microphone signals, i.e.,
{circumflex over (X)}d(k,n,dm)=wmH(k,n)x(k,n)
where wm is a complex weight-vector of length M. The weights of wm need to be computed such that an accurate estimate of the diffuse sound is obtained.
The straight-forward way to find an appropriate filter is to compute the weights wm such that the L plane waves are suppressed while the stationary noise Xn(k,n,dm), which is contained in the microphone signals, is minimized. Expressed mathematically, the filter weights are given by
subject to the linear constraints
wHal(k,n)=0∀l
Here, Φn is the PSD matrix (power spectral density matrix) of the stationary noise, i.e., Φn=E{xnxnH}, which can be estimated with well-known approaches for instance when no diffuse sound or direct sound is present. Moreover, al is the so-called propagation vector. Its elements are the relative transfer function of the l-th plane wave from the m-th microphone to the other microphones. Hence, al is a column vector with length M (remember: only the diffuse sound at the m-th microphone is estimated by the wm-weighted linear combination of the M microphone signals; the diffuse sound at the other microphones is substantially redundant, as these signals are related via relative transfer functions from the m-th microphone to the other microphones and could be calculated in this manner, if needed). The elements of al depend on the DOA of the l-th plane wave. This means al is a function of the DOA of the l-th plane wave, i.e., al=ƒ(nl). Since al depends on the direct sound (i.e., plane waves), it is referred to as direct sound constraint in the following. With this spatial filter we essentially create a beamformer, which has a pick-up pattern with nulls towards the directions of the L plane waves. As a result, all plane waves are suppressed. Unfortunately, solving this minimization problem above leads to zero weights wm since we only have null constraints, i.e., the diffuse sound cannot be extracted.
To overcome this problem and to avoid zero filter weights, [Thiergart2013b, Kowalczyk2013]proposes to use the same filter but with an additional constraint, given by
wHa0(k,n)=1
where a0 is a propagation vector that corresponds to a specific DOA n0, from which no plane wave arrives. With this constraint one avoids zero filter-weights, but still does not capture the undesired direct sound. As a result, with this filter only diffuse sound and some noise is captured but all plane waves are attenuated. In [Thiergart2013b], the DOA n0, to which the vector a0 corresponds, is found by choosing the direction which has the largest angular distance to all DOAs nl(k,n) of the plane waves. For instance if a single plane wave is arriving from 0 degree, then nl(k,n) would correspond to 180 degree. Unfortunately, the DOA n0 does not guarantee that we obtain a diffuse sound estimate with as little noise as possible. Moreover, the resulting pick-up pattern is not very optimal for capturing diffuse sound, since it becomes highly directive at higher frequencies. This is a drawback when aiming at capturing diffuse sound from all directions.
An example of a resulting pick-up pattern is depicted in
A closed-form solution to compute the filter weights wm given the constraints above can be found in [VanTrees2002]. In order to compute the spatial filter, one needs to know the DOA of the L plane waves, namely to compute the direct sound constraints al and a0. This DOA information can be determined with well-known narrowband DOA estimators, such as Root MUSIC or ESPRIT. Note further that the elements of a0 are typically complex and a0 typically needs to be recomputed for each k and n, since the DOAs of the plane waves should be assumed highly time-varying. The highly fluctuating a0 can lead to audible artifacts.
An example system for extracting the diffuse sound with the presented multi-channel filter is illustrated in
In this invention, we propose a novel multichannel filter for extracting reverberant sound that overcomes the limitations of the aforementioned filters. The proposed spatial filter is characterized by a directivity pattern, which tends to an omnidirectional pattern, except for the directions-of-arrival (DOAs) of the direct sound for which it exhibits spatial nulls. This represents a highly desired property for capturing diffuse sound from all directions with low distortion.
In the following, we propose a multi-channel filter to estimate the diffuse sound Xd (k,n,dm) at position dm with M>1 microphones. As for the multi-channel filters described above, the diffuse sound pressure at the m-th microphone is estimated by performing a linear combination of the microphone signals, i.e.,
{circumflex over (X)}d(k,n,dm)=wmH(k,n)x(k,n).
The weight-vector wm, which is proposed in the following, minimizes a specific cost function and is linearly constrained similarly to the multi-channel filters described above.
However, in contrast to the multi-channel filters described above, we propose to use a linear constraint which does not depend on the direct sound (i.e., on the L plane waves). More precisely, the proposed novel constraint is not a function of the DOAs of the plane waves or the corresponding relative transfer functions of the plane waves between the microphones, respectively.
In contrast, the proposed novel constraint depends on statistical information on the diffuse sound, i.e., the proposed novel constraint depends on the relative transfer functions of the diffuse sound between the microphones. We will show in the following that the proposed novel constraint is a function of the coherence or correlation of the diffuse sound between the microphones. This coherence corresponds to the mean relative transfer function of the diffuse sound between the microphones.
The proposed spatial filter is obtained by minimizing a specific cost function while satisfying a distortionless constraint for the diffuse sound. This constraint corresponds to the relative transfer function of the diffuse sound between the microphones. Expressed mathematically, the filter is computed as subject to the linear constraint
Here, J is the cost function to be minimized by the filter. The cost function can be for instance the stationary noise power at the filter output, the interfering energy at the filter output, or the quadratic error of the estimated diffuse sound. Examples for J will be provided in the embodiments. The constraint vector bm is given by bm(k,n)=[B1,m(k,n), B2,m(k,n), . . . , BM,m(k,n)]T. The m′-th element Bm′,m is the relative transfer function of the diffuse sound between microphone m and m′. This relative transfer function is given by
Note that the m-th element of bm is equal to 1. With this constraint we capture the diffuse sound without distortion. In fact, let xd(k,n)=[Xd (k,n,d1), Xd (k,n,d2), . . . , Xd (k,n,dM)]T be a vector containing the recorded diffuse sound. With the equations above, this vector can be written as
xd(k,n)=bm(k,n)Xd(k,n,dm).
The diffuse sound at the output of the filter is given by wH(k,n)xd(k,n), which is identical to Xd (k,n,dm) since wHbm(k,n)=1. Therefore, this filter captures the diffuse sound without distortion. The relative transfer functions in bm typically cannot be estimated in practice since it is basically random, i.e., we have a different realization of the transfer function for each k and n. Thus, in practice, Bm′,m is computed as the mean relative transfer function between microphone m and m′, i.e.,
Bm′,m(k,n)=γm′,m(k,n).
This mean relative transfer function γm′,m corresponds to the so-called spatial coherence of the diffuse sound between microphone m and m′, which is defined as
where (⋅)* denotes complex conjugate. This spatial coherence describes the correlation of the diffuse sound between microphone m and m′ in the frequency domain. This coherence depends on the specific diffuse sound field. The coherence can be measured in advance for a given room. Alternatively, the coherence is known from theory for specific diffuse sound fields [Elko2001]. For instance for a spherically isotropic diffuse sound field, which can often be assumed in practice, we have
where sine denotes the sine function, ƒ is the acoustical frequency for the given frequency band k, and c is the speed of sound. Moreover, rm′,m is the distance between microphone m and m′. When using the spatial coherence as the linear constraint Bm′,m, which represents the mean relative transfer function of the diffuse sound between the microphones, then the obtained filter is equivalent to the sum of many linearly constrained spatial filters, where each of these filters captures a different realization of the random diffuse sound without distortion.
With the diffuse sound constraint introduced above, we obtain a spatial filter that captures the diffuse sound equally well from all directions. This is in contrast to the multi-channel filters described above, which captures the sound mainly from one direction, namely the direction to which the chosen propagation vector a0 corresponds.
Note that the diffuse sound constraint bm is conceptually very different from the direct sound constraints al and a0. Therefore, the novel filter proposed in this section is conceptually very different compared to the multi-channel filters described above.
A block scheme of the proposed invention is depicted in
Minimizing the Output Power Satisfying a Diffuse Sound Constraint
In this embodiment, we define a spatial filter that minimizes the entire output of the filter subject to the diffuse sound constraint. The diffuse sound constraint ensures that the diffuse sound is preserved by the spatial filter while the remaining signal parts (undesired stationary noise and plane waves) are minimized. The filter weights wm are computed as
subject to the linear constraint
wHbm(k,n)=1.
A close-form solution to this filter is given by [VanTrees2002]
Here, Φx is the PSD matrix of the microphone signals, which can be computed as
Φx(k,n)=E{x(k,n)xH(k,n)},
where x(k,n) is the vector containing the microphone signals. In practice, the expectation is approximated for instance by a temporal averaging. Moreover, the elements of the constraint vector bm (k,n)=[B1,m(k,n), B2,m(k,n), . . . , BM,m(k,n)]T correspond to the spatial coherence of the diffuse sound between microphone m and m′, i.e.,
Bm′,m(k,n)=γm′,m(k,n).
Actually, the spatial coherence Bm′,m does not need to depend on time (that is, Bm′,m (k,n)=Bm′,m (k)), hence the spatial coherence can be estimated in advance or assume a theoretical value. The spatial coherence may be either estimated from the microphone signals (during periods where only the diffuse sound is present) using
or given as a priori information assuming a specific diffuse sound field. In the latter case, we use for instance the spatial coherence for a spherically isotropic diffuse sound field, i.e.,
Note that the sinc function might be replaced by other functions depending on the assumed sound field. For different diffuse sound fields there exist different coherence functions that are known a priori. Examples can be found in [Elko2001].
A block scheme of this embodiment is shown in
The filter computed in this embodiment has the following advantages compared to other spatial filter (e.g., the filters described in the background art):
Linearly Constrained Minimum Variance Filter
This embodiment represents a combination of the novel approach and the state-of-the-art approach of multi-channel filters described above in connection with
subject to the linear constraints
wHbm(k,n)=1
and
wHaL(k,n)=0∀l
Clearly, the filter minimizes only the stationary noise at the output. The undesired plane waves are suppressed with the second linear constraints (as explained above for the multi-channel filters,
wHm(k,n)=gH(k,n)[CH(k,n)Φn−1(k)C(k,n)]−1CH(k,n)Φn−1(k)
Here, vector C=[bm, a1, a2, . . . , aL] is the constraint matrix containing the linear constraints defined above and g=[1,O]T (O being a zero-vector of length L) are the corresponding responses. As for the multi-channel filter shown in
A block scheme of this embodiment is shown in
An example of a resulting pick-up pattern for this filter is depicted in
The filter computed in this embodiment has the following advantages compared to other spatial filter (e.g., the filters described in the background art):
Combined Approach
The spatial filters shown in
The approach proposed in the following represents a so-called parametric multi-channel Wiener filter (PMWF) that can be scaled between a so-called minimum mean square error (MMSE) spatial filter and the spatial filter in
The weight-vector of the proposed PMWF is computed as
subject to
E{|Xd(k,n,dm)−{circumflex over (X)}d(k,n,dm)|2}<σ2
where σ2 is the maximum squared absolute error of the estimated diffuse sound. Solving this optimization problem leads to
where we define
β=α(ϕdbmHΦx−1bm)
Here, αϵ[0,1] is a user-defined control parameter. For α=0, we obtain the MMSE spatial filter which minimizes the mean-squared error of the estimated diffuse sound. For α=1, we obtain the spatial filter proposed in
or with a decision directed approach as explained in [Kowalczyk2013]. In the preceding formula, Γd is the M×M spatial coherence matrix for the diffuse sound. The (m,m′)-th element of Γd is the spatial coherence γm′,m between microphone m and m′. This spatial coherence γm′,m was already defined above.
A block scheme of this embodiment is shown in
An individual microphone signal, for example for the m-th microphone, Xm(k,n) is a combination of the L direct sound portions X1=l to X1=L, the diffuse sound portion Xd, and noise Xn, i.e.
Relative transfer functions B1,m, B2,m, . . . , Bm′,m, . . . , BM,m for the diffuse sound between the other microphones to the m-th microphone are schematically illustrated in
The following list provides a brief overview of some of the aspects that have been described above:
Although some aspects have been described in the context of an apparatus, it is clear that these aspects also represent a description of the corresponding method, where a block or device corresponds to a method step or a feature of a method step. Analogously, aspects described in the context of a method step also represent a description of a corresponding block or item or feature of a corresponding apparatus.
Depending on certain implementation requirements, embodiments of the invention can be implemented in hardware or in software. The implementation can be performed using a digital storage medium, for example a floppy disk, a DVD, a CD, a ROM, a PROM, an EPROM, an EEPROM or a FLASH memory, having electronically readable control signals stored thereon, which cooperate (or are capable of cooperating) with a programmable computer system such that the respective method is performed.
Some embodiments according to the invention comprise a non-transitory data carrier having electronically readable control signals, which are capable of cooperating with a programmable computer system, such that one of the methods described herein is performed.
Generally, embodiments of the present invention can be implemented as a computer pro-gram product with a program code, the program code being operative for performing one of the methods when the computer program product runs on a computer. The program code may for example be stored on a machine readable carrier.
Other embodiments comprise the computer program for performing one of the methods described herein, stored on a machine readable carrier.
In other words, an embodiment of the inventive method is, therefore, a computer program having a program code for performing one of the methods described herein, when the computer program runs on a computer.
A further embodiment of the inventive methods is, therefore, a data carrier (or a digital storage medium, or a computer-readable medium) comprising, recorded thereon, the computer program for performing one of the methods described herein.
A further embodiment of the inventive method is, therefore, a data stream or a sequence of signals representing the computer program for performing one of the methods described herein. The data stream or the sequence of signals may for example be configured to be transferred via a data communication connection, for example via the Internet.
A further embodiment comprises a processing means, for example a computer, or a programmable logic device, configured to or adapted to perform one of the methods described herein.
A further embodiment comprises a computer having installed thereon the computer pro-gram for performing one of the methods described herein.
In some embodiments, a programmable logic device (for example a field programmable gate array) may be used to perform some or all of the functionalities of the methods described herein. In some embodiments, a field programmable gate array may cooperate with a microprocessor in order to perform one of the methods described herein. Generally, the methods may be performed by any hardware apparatus.
While this invention has been described in terms of several embodiments, there are alterations, permutations, and equivalents which fall within the scope of this invention. It should also be noted that there are many alternative ways of implementing the methods and compositions of the present invention. It is therefore intended that the following appended claims be interpreted as including all such alterations, permutations and equivalents as fall within the true spirit and scope of the present invention.
Number | Date | Country | Kind |
---|---|---|---|
13196672 | Dec 2013 | EP | regional |
14156014 | Feb 2014 | EP | regional |
This application is a continuation of copending International Application No. PCT/EP2014/076252, filed Dec. 2, 2014, which is incorporated herein in its entirety by this reference thereto, and which claims priority from European Applications Nos. EP 13 196 672.3, filed Dec. 11, 2013, and from EP 14 156 014.4, filed Feb. 20, 2014, which are each incorporated herein in its entirety by this reference thereto.
Number | Name | Date | Kind |
---|---|---|---|
20040175006 | Kim | Sep 2004 | A1 |
20050201204 | Dedieu | Sep 2005 | A1 |
20050232441 | Beaucoup | Oct 2005 | A1 |
20080069366 | Soulodre et al. | Mar 2008 | A1 |
20080089522 | Baba | Apr 2008 | A1 |
20090252355 | Mao et al. | Oct 2009 | A1 |
20090299742 | Toman et al. | Dec 2009 | A1 |
20100169103 | Pulkki et al. | Jul 2010 | A1 |
20120063608 | Soulodre et al. | Mar 2012 | A1 |
20120099732 | Visser | Apr 2012 | A1 |
20120128160 | Kim et al. | May 2012 | A1 |
20130022206 | Thiergart | Jan 2013 | A1 |
20130066628 | Takahashi et al. | Mar 2013 | A1 |
20130216047 | Kuech | Aug 2013 | A1 |
20130230184 | Kuech | Sep 2013 | A1 |
20130272548 | Visser | Oct 2013 | A1 |
20150016622 | Kawaguchi et al. | Jan 2015 | A1 |
Number | Date | Country |
---|---|---|
2013061421 | Apr 2013 | JP |
2013171076 | Sep 2013 | JP |
2013179388 | Sep 2013 | JP |
2343562 | Jan 2009 | RU |
2008034221 | Mar 2008 | WO |
Entry |
---|
Taseska et al, “MMSE-based blind source extraction in diffuse noise fields using a complex coherence-based a priori SAP estimator.” Sep. 4, 2012-Sep. 6, 2012. pp. 1-4. |
Souden et al, “A study of the LCMV and MVDR noise reduction filters.” Sep. 2010. pp. 1-11. |
Elko, “Spatial coherence functions for differential microphones in isotropic noise fields.” 2001. pp. 1-25. |
Elko, G.W., “Spatial Coherence Functions for Differential Microphones in Isotropic Noise Fields”, Microphone Arrays: Signal Processing Techniques and Applications; edited by M. Brandstein and D. Ward; Chapter 4; Springer, 2001, pp. 61-85. |
Kowalczyk, K. et al., “Sound Acquisition in Noisy and Reverberant Environments Using Virtual Microphones”, 2013 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA, Oct. 2013, 4 pages. |
Pulkki, V, “Spatial Sound Reproduction with Directional Audio Coding”, Journal of the AES. vol. 55, No. 6. New York, NY, USA., Jun. 2007, pp. 503-516. |
Thiergart, O. et al., “An Informed LCMV Filter Based on Multiple Instantaneous Direction-of-Arrival Estimates”, 2013 IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), Vancouver, BC, Canada., May 2013, pp. 659-663. |
Thiergart, O. et al., “Geometry-based Spatial Sound Acquisition Using Distributed Microphone Arrays”, IEEE Transactions on Audio, Speech, and Language Processing; vol. 21, No. 12, Dec. 2013, pp. 2583-2594. |
Thiergart, O. et al., “On the Spatial Coherence in Mixed Sound Fields and its Application to Signal-to-Diffuse Ratio Estimation”, The Journal of the Acoustical Society of America; vol. 132; No. 4, Oct. 2012, pp. 2337-2346. |
Van Trees, H. L., “Detection, Estimation, and Modulation Theory: Part IV: Array Processing”, John Wiley & Sons, vol. 1, Apr. 2002, 1470 pages. |
Number | Date | Country | |
---|---|---|---|
20160293179 A1 | Oct 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/EP2014/076252 | Dec 2014 | US |
Child | 15178530 | US |