The invention relates to an interpolation circuit in accordance with the preamble of claim 1. As defined therein, this interpolation circuit includes a first branch provided with a circuit for power-specific summation of the first and second microphone signals. A possible embodiment of such a circuit for power-specific summation is known from WO2011/057922A1. In the context of the present invention, a circuit for power-specific summation is to be understood as a circuit deriving an output signal based on two input signals, with the proviso that the power of the output signal is mainly equal to the sum of the power quantities of the two input signals.
Each interpolation method is based on a weighted summation of two signals. The summation signal can, however, only be interpolated correctly up to a particular frequency or wavelength at which the sampling theorem is still satisfied. Thus, a signal can only be calculated correctly if the distance between the microphones to be interpolated is not greater than half the wavelength. Beyond this, the phase can not be determined in a defined manner any more, resulting in comb filters and corresponding sound colorations.
The latter are prevented through power-specific summation in the interpolation method, as is described in WO2011/057922A1. As a result it is possible to simulate a virtual microphone in the desired location without any sound losses.
The invention intends to further improve the interpolation circuit. To this end, the interpolation circuit defined in the preamble of the main claim is characterized as specified in accordance with the features of the characterizing portion of the main claim. Preferred practical examples of the interpolation circuit of the invention are defined in the subclaims. The invention is based on the following inventive concept.
The localized perception of sound waves is substantially determined by the delay periods of the sound paths of low-frequency sound components. As these delay periods are represented in the phase of the corresponding low-frequency signal components, a correct phase of the virtual microphone signal is crucial for an unimpaired localized perception. The phase of the virtual microphone signal is a function of the location variable determining the position of the virtual microphone.
The correct delay period values, or phase values, of a virtual microphone are mapped with adequate accuracy for sufficiently low-frequency signal components by a traditional interpolation of real microphone signals; such an interpolation shall in the following be referred to as phase-specific interpolation.
The acoustic perception of sound sources is substantially determined by the ratios of the acoustic power of sound components of different frequencies, however is independent of whether or not the phase of the signal is correct.
With the exception of low-frequency signal components, the traditional interpolation is not suited due to infraction of the sampling condition because it falsifies the power ratios of different frequencies while also not providing a correct phase of the virtual microphone signal.
It is a property of frequency-dependent, approximately constant-power interpolation, hereinafter referred to as power-specific interpolation, that it does not substantially alter the power ratios of different frequencies and therefore results in a sound perception of the virtual microphone which approximately corresponds to the one of a real microphone in the corresponding position.
Inasmuch as a power-specific interpolation is not necessarily also phase-specific, an improvement of the localized perception is achieved by restricting the power-specific interpolation to high-frequency signal components and combining it with a phase-specific interpolation for the remaining, low-frequency signal components. This in turn is achieved in that processing is distributed to two different branches.
Further details also result from the following further reflections.
Power-specific interpolation is realized by the application of power-related weighting factors to the input signals of a power-specific summer, wherein the summation as in WO2011/057922A1 is employed for the power-specific summer, and the weighting factors are power-related in that the sum of their squared values is 1.
Processing of the microphone signals in the frequency range, which serves the purpose of power-specific interpolation, is advantageously employed concurrently for a separation between low-frequency and high-frequency signal components.
Combining of the two interpolation types is executed by weighted mixing of the signals of the two processing branches in dependence on the frequency parameter, wherein the weighting factors are a continuous function of the frequency. This largely prevents the generation of discontinuities in the frequency spectrum of the combined signal which would otherwise result in audible interferences for some signals.
If the calculation of the interpolated signal value of the corresponding frequency and the corresponding interpolation type is omitted for those frequencies and the one processing branch where the weighting factor of mixing is zero, this brings about the advantage of saving a part of the processing expenditure.
The selection of a summer for the power-specific interpolation, the phases of which are a smooth function of the weighted input signals, has the effect that interfering disruptions in sound perception are not produced during a continuous change of the control signal of the virtual microphone. The summation as in WO2011/057922A1 meets this requirement and is therefore utilized.
Both in a traditional interpolation and a power-specific interpolation, the phase function of the location variable of the virtual microphone in most cases deviates from the phase function of a real microphone placed in the position of the virtual microphone. The phase values of the virtual microphone are mapped with improved accuracy in that the location variable is converted to a control signal of the interpolation by an antidistortion calculation. Approximating calculations are sufficient. The antidistortion function typically maps the value 0 to 0 and the value 1 to 1, and the development in between typically is symmetrical. The most simple approximation is the proportionality function.
A further improvement of the phase values of the virtual microphone is achieved by adapting the phase function of the power-specific interpolation to the phase function of the traditional interpolation. This prevents interfering amplitude errors during transition between the two interpolation types in the frequency range of changeover between the signal contributions of the processing branches, and is achieved by employing separate, different antidistortion calculations for the control signals of the two interpolations. A typical, sufficiently accurate antidistortion function for the control signal of the traditional interpolation is the proportionality function. A typical, sufficiently accurate antidistortion function for the control signal of the power-specific interpolation is the squared sine function.
The invention is explained in more depth by making reference of the description of the figures, wherein
The first circuit branch 104 is provided with a means 108 for power-specific summation of the signals supplied at the first 105 and second 106 inputs of the first circuit branch and for outputting a power-specific summation signal at the output 107 of the first circuit branch 104.
The first circuit branch 104 is further provided with a multiplication circuit 124 coupled between the first input 105 of the first circuit branch and a first input 126 of the means 108 for power-specific summation. The circuit branch 104 is furthermore provided with a multiplication circuit 125 coupled between the second input 106 of the first circuit branch and a second input 127 of the means for power-specific summation. The multiplication circuits 124, 125 are each provided with a control input that is coupled to the control input 103 of the interpolation circuit via a control signal conversion circuit 131.
The second circuit branch 109 is provided with a first multiplication circuit 120 and a second multiplication circuit 121 having inputs coupled to the first 110 and the second input 111, respectively, of the second circuit branch, and outputs coupled to respective inputs of a second signal combination circuit 122, the output of which is coupled to the output 112 of the second circuit branch 109. The first and second multiplication circuits 120, 121 are each provided with a control input that is coupled to the control input 103 of the interpolation circuit via a control signal conversion circuit 130.
The respective outputs 107, 112 of the first and second circuit branches 104 and 109 are coupled to respective inputs 115, 118 of a signal combination circuit 116 via respective multiplication circuits 113 and 114. An output 119 of the signal combination circuit 116 is coupled to the output 102 of the interpolation circuit.
Interpolation is preferably carried out in the frequency range. In this case transformation circuits 133 and 134 are provided which convert the microphone signals from the time range into the frequency range, e.g. by means of fast Fourier transform, and having a transformation circuit 135 which converts the output signal of the signal combination circuit 116 from the frequency range into the time range, e.g. by means of inverse fast Fourier transform.
The multiplication circuits 120, 121 are adapted to multiply the signals supplied to them by first and second multiplication factors (1−f,f), wherein first and second multiplication factors are dependent on the control signal (r). In a preferred manner,
f=r
B, (Eq. 1),
wherein B is a constant that is greater than zero, preferably equal to 1.
The multiplication circuits 124,125 are adapted to multiply the signals supplied to them by third and fourth multiplication factors that are equal to (1−g)1/2 and g1/2, wherein third and fourth multiplication factors are dependent on the control signal (r). The factor g may be dependent on r in various ways. One possibility is
g=r
C (Eq. 2),
wherein C is a constant that is greater than zero, preferably equal to 1. In this case it is achieved that the signal at the output 107 of the first branch 104 is adapted to the signal at the output 112 of the second branch 109 in the amplitude as well as in simple approximation of the phase. Or, g=sinD(r*π/2), wherein D is a constant that is greater than zero, preferably equal to 2. In this case the same conditions apply as in the case g=rC wherein, however, the accuracy of approximation of the phase is additionally improved. The multiplication circuits 113 and 114 are adapted to multiply the signals supplied to them by respective frequency-dependent multiplication factors 1−c(k) and c(k), wherein k is a frequency parameter. In a preferred embodiment a condition for c(k) is that for k=0 it is a constant E1 that is preferably equal to 1 and decreases for increasing values of k until c(k) is equal to a constant E0, preferably equal to 0, for higher values of k. Conversely, it is thus true for the multiplication factor 1−c(k) that it is 1−E1 for k=0 and increases for increasing values of k until it becomes 1−E0 for higher values of k. This means that the contribution of the second branch 109 is mainly in the low frequency range, however that this contribution decreases for higher frequencies and is taken over by the contribution of the first branch 104.
The means 108 for power-specific summation as shown in
With regard to a practical example where an interpolated microphone signal is derived from two microphone signals of two juxtaposed microphones of the microphone arrangement in
r=A*(φ−φm)/(φm+1−φm) (Eq. 3)
wherein A is a constant that is preferably equal to 1, and
wherein φm and φm+1 are the corner positions of the two microphones 301 and 302 on the circle and φ is a corner variable indicating the corner position where a virtual microphone between the two microphones is assumed to be arranged on the circle, and wherein the interpolated microphone signal at the output of the interpolation circuit is assumed to be the output signal of this virtual microphone.
The operation of the interpolation circuit according to
It shall be assumed that the position of the virtual microphone may be described through a parametric interpolation of location along a suitably devised connecting line between the positions of the adjacent real microphones 301, 302, that the parameter of this interpolation of location is scaled by an appropriately defined scaling function so that the scaling yields 0 at the position of the microphone 301 and 1 at the position of the microphone 302, and that the scaling result is adopted as the control signal r of the circuit in
For example, in the arrangement in
The circuit in
All of such branching and recombination is carried out with signals transformed into the frequency range, and the operations in the branches relate to spectral values. The spectral values of the input signals are each generated from the respective input signal by a spectral transformation unit in the input signal path, and the output signal is generated from the spectral values of the output signal by an inverse spectral transformation unit in the output signal path. This spectral processing enables power-specific summation and the transition of the interpolation types, which shall be elucidated further below.
Spectral values should be understood to be vector variables having a frequency as an index, and each vector element is processed in the same manner. In difference from this, an improved example realization for a vector element only carries out the operations of a branch if the weighting factor of the branch in question and of the frequency index in question is not 0 upon recombination of the branches. The weighting factors of the recombination shall be explained in more detail further below.
The interpolations are each composed of an application of weighting factors to the input spectral values and of a summation, wherein the weighting factors of the interpolation are controlled by a control variable.
The power-specific signal interpolation meets the condition that the output power should be approximately equal to the sum of the input power, in that both the involved summation meets this condition (power-specific summation), and furthermore in weighting the sum of the output powers is equal to the sum of the input powers. In weighting this condition is met due to the fact that the squared weighting factors add up to 1.
The operation of a power-specific summation will be described further below in the explanations for
The phase-specific interpolation is a linear interpolation which operates in a manner that is known per se.
In order for each interpolation type to obtain a frequency-dependent proportion of its effect, frequency-dependent weighting factors are applied to the spectral values upon recombination of the signal branches. The weighting factors of the recombination expediently add up to 1.
The transition range of the interpolation types is realized through the frequency-dependent weighting of the recombination. The curve of the frequency dependency is preferably smooth, whereby audible interferences in the resultant signal are prevented.
The location of the transition range with regard to the frequency is advantageously selected such that the power ratios of different frequencies are not yet altered strongly by the phase-specific interpolation for frequencies below the transition range. This approximately comes about for a frequency in an order where the distance of the adjacent real microphones is one quarter of the wavelength of a sound wave propagating in the direction of the connecting line.
The antidistortion calculation for the control variable of the interpolation that is provided for the improvement of the phase values of the virtual microphone at frequencies in the transition range of the interpolation types is carried out separately for the two branches by respective control signal conversion circuits 130 and 131. The antidistortion function is realized through an antidistortion curve which is selected to compensate the phase characteristics of the signal interpolation such as to approximate it to the phase characteristics of the interpolation of location. For example, the antidistortion curve is determined in advance through comparisons of phase measurements or phase estimates with a real microphone and phase measurements or phase estimates with the aid of the present circuit. The expression “phase characteristics” refers to the dependency of the phase of an interpolated spectral value on the control variables of the interpolation and on the respective spectral values to be interpolated. The antidistortion can only compensate the dependency on the control variables, not the dependency on the two spectral values to be interpolated. For determining the antidistortion curve it is therefore expedient to consider only those case where the influence of the spectral values to be interpolated is small, and an average or typical case is assumed. Those are the cases in which the difference of the phases of the spectral values to be interpolated is small, which is true for the typical acoustic applications at sufficiently low frequencies and thus also for the intended transition range of the interpolation types.
Identifying the inputs 201, 200 of the means 108 for power-specific summation as 127 or 126 in
In summary it may be said that the operation of the partial circuits of the two signal branches differs in the following points:
Altogether, the comportment of the circuit with regard to the phase may be described as follows: For signal components in the range of high frequencies only the first branch takes effect, in which the phase resulting from ensuring the correct power of the interpolation is not taken into account. For signal components in the range of low frequencies only the second branch takes effect, which ensures the correct phase of the interpolation. In a transition range at medium frequencies a combination of both branches takes effect in with the branches change over continually and exhibit only a small difference, if any, in their phase.
The circuit in
The given arrangement results in a calculated k-th complex output spectral value Y(k) of the signal at the output 213 of the means 108 as
Y(k)=m(k)·Z1(k)+Z2(k). (Eq. 4)
In analogy with the method of WO2011/057922A1, the multiplication factor m(k) is calculated as follows:
eZ
1(k)=Real(Z1(k))·Real(Z1(k))+Imag(Z1(k))·Imag(Z1(k)) (Eq. 5.1)
eZ
2(k)=Real(Z2(k))·Real(Z2(k))+Imag(Z2(k))·Imag(Z2(k)) (Eq. 5.2)
x(k)=Real(Z1(k))·Real(Z2(k))+Imag(Z1(k))·Imag(Z2(k)) (Eq. 5.3)
w(k)=x(k)/(eZ1(k)+L·eZ2(k)) (Eq. 5.4)
m(k)=(w(k)2+1)1/2−w(k) (Eq. 5.5)
wherein
The degree L of limitation of the comb filter compensation is a numerical value which determines the degree in which the probability of the occurrence of artefacts perceived to be interfering is reduced. This probability is given when the amplitude of the spectral values of the signal at the input 203 of the calculation unit is small compared with that of the spectral value of the signal at the input 202 of the calculation unit. At a condition of L>=0, L typically is constant and L<1. If L=0, a reduction of the probability of artefacts does not ensue. The greater L, the lower is the probability of artefacts, however this equally has the effect of partially reducing the compensation of sound colorations due to comb filter effects that is aimed at by the circuit. L is selected such that artefacts just about are not perceived any more in accordance with experience.
It will now be shown that the power ratios of different frequencies between the inputs and the output of the means 108 for power-specific summation are not altered substantially.
To this end the sum of the input spectral powers is compared to the output spectral power for a frequency index k.
The respective spectral power values eZ1(k) and eZ2(k) for the complex input spectral values Z1(k) and Z2(k) were already indicated in (Eq. 5.1) and (Eq. 5.2), and in the same way there results for the k-th spectral power value eY(k) of the signal at the output 213 of the means 108
eY(k)=Real(Y(k))·Real(Y(k))+Imag(Y(k))·Imag(Y(k)).
When L=0 is assumed and substituted in the equation (Eq. 5.4) given above, the equation is simplified to
w
0(k)=x(k)/eZ1(k),
and with w0(k) instead of w(k) and with corresponding substitutions
m
0(k)=(w0(k)2+1)1/2−w0(k)
and
Y
0(k)=m0(k)·Z1(k)+Z2(k)
it is possible by well-known mathematical processes to thereby solve an equation
eY
0(k)=eZ1(k)+eZ2(k),
which shows the accurate equality of output power and sum of the input powers at L=0.
The application of the parameter L with L>0 results in a deviation from the accurate equality of power for the single frequency index k, with the ensuing restriction of:
eY(k)≈eZ1(k)+eZ2(k),
whereas L>0 has the advantageous effect of the probability of the occurrence of artefacts perceived as interfering being reduced.
These artefacts may come about with the named w0(k) because a zero crossing of Z1(k), even if it is continuous, results in a non-continuous polarity reversal of Y0(k), and they may be perceived as interfering if the contribution of the spectral proportion thereby effected to the overall signal is sufficiently great. The discontinuity is eliminated by L>0.
The interpolation circuit of
As was already mentioned, this circuit generates an interpolated signal at the output 102 for a virtual microphone assumed to be arranged in the position 401 on the circle in
For φ=φm+1, it may be derived from Formula (Eq. 3) that r=1. Accordingly, due to Formula (Eq. 1) there also follows f=1, and due to Formula (Eq. 3) there also follows g=1. It is thus evident from
For φ situated between φm and φ=φm+1, the Formulae (Eq. 1), (Eq. 2), (Eq. 3) and (Eq. 4) are to be applied. The k-th complex spectral value S[k] of the output signal s of the virtual microphone in the location φ as a function of φ, c(k), Am[k] and Am+1[k] then has the following form:
S[k]=(((((Real(((r(φ))C)1/2·Am+1[k])·Real((1−(r(φ))C)1/2·Am[k])+Imag(((r(φ))C)1/2·Am+1[k])·Imag((1−(r(φ))C)1/2·Am[k]))/((Real(((r(φ))C)1/2·Am+1[k])·Real(((r(φ))C)1/2·Am+1[k])+Imag(((r(φ))C)1/2·Am+1[k])·Imag(((r(φ))C)1/2·Am+1[k]))+L·(Real((1−(r(φ))C)1/2·Am[k])·Real((1−(r(φ))C)1/2·Am[k])+Imag((1−(r(φ))C)1/2·Am[k])·Imag((1−(r(φ))C)1/2·Am[k]))))2+1)1/2−((Real(((r(φ))C)1/2·Am+1[k])·Real((1−(r(φ))C)1/2·Am[k])+Imag(((r(φ))C)1/2·Am+1[k])·Imag((1−(r(φ))C)1/2·Am[k]))/((Real(((r(φ))C)1/2·Am+1[k])·Real(((r(φ))C)1/2·Am+1[k])+Imag(((r(φ))C)1/2·Am+1[k])·Imag(((r(φ))C)1/2·Am+1[k]))+L·(Real((1·(r(φ))C)1/2·Am[k])·Real((1−(r(φ))C)1/2·Am[k])+Imag((1−(r(φ))C)1/2·Am[k])·Imag((1−(r(φ))C)1/2·Am[k])))))·(((r(φ))C)1/2·Am+1[k])+((1−(r(φ))C)1/2·Am[k]))·(1−c(k))+(((r(φ))B·Am+1[k])+((1−(r(φ))B)·Am[k]))·c(k),
with
r(φ)=A·(φ−φm)/(φm+1−φm). (Eq. 6)
Or, when expressed in the form of single calculation steps:
r=A·(φ−φm)/(φm+1−φm) (Eq. 6.1)
U
1(k)=(r)B·Am+1[k] (Eq. 6.2)
U
2(k)=(1−(r)B)·Am[k] (Eq. 6.3)
U(k)=(U1(k))+(U2(k)) (Eq. 6.4)
Z
1(k)=((r)C)1/2·Am+1[k] (Eq. 6.5)
Z
2(k)=(1−(r)C)1/2·Am[k] (Eq. 6.6)
eZ
1(k)=Real(Z1(k))·Real(Z1(k))+Imag(Z1(k))·Imag(Z1(k)) (Eq. 6.7)
eZ
2(k)=Real(Z2(k))·Real(Z2(k))+Imag(Z2(k))·Imag(Z2(k)) (Eq. 6.8)
x(k)=Real(Z1(k))·Real(Z2(k))+Imag(Z1(k))·Imag(Z2(k)) (Eq. 6.9)
w(k)=(x(k))/((eZ1(k))+L·(eZ2(k))) (Eq. 6.10)
m(k)=((w(k))2+1)1/2−(w(k)) (Eq. 6.11)
Y(k)=(m(k))·(Z1(k))+(Z2(k)) (Eq. 6.12)
S[k]=(Y(k))·(1−c(k))+(U(k))·c(k) (Eq. 6.13)
Now it shall be explained by referring to
The following is now true for r.
r=A*(l−lm)/(lm+1−lm) (Eq. 7)
wherein A is a constant, preferably equal to 1, and
wherein lm and lm+1 indicate the positions of the two microphones 502 and 503 on the straight line 505 and L is the distance variable indicating the position of the virtual microphone between the two microphones 502 and 503 on the straight line 505. The interpolated microphone signal at the output of the interpolation circuit is then assumed to be the output signal of this virtual microphone 506.
The operation is analogous to the operation already described in the foregoing.
The interpolation circuit may just as well be applied to other microphone arrangements where the microphones are arranged along a curve and not on a straight or circle line.
The operation of the circuit in
The multiplication factor in this case is termed mS and is calculated as follows:
eZ
1(k)=Real(Z1(k))·Real(Z1(k))+Imag(Z1(k))·Imag(Z1(k)) (Eq. 8.1)
eZ
2(k)=Real(Z2(k))·Real(Z2(k))+Imag(Z2(k))·Imag(Z2(k)) (Eq. 8.2)
x(k)=Real(Z1(k))·Real(Z2(k))+Imag(Z1(k))·Imag(Z2(k)) (Eq. 8.3)
m
S(k)=((eZ1(k)+eZ2(k))/(eZ1(k)+eZ2(k)+2·x(k)))1/2 (Eq. 8.4)
wherein
Similar to the case of the circuit in
Y(k)=(Z1(k)+Z2(k))·mS(k) (Eq. 9)
is now equal to the sum of the input powers, i.e.:
eY(k)=eZ1(k)+eZ2(k).
In difference from the circuit in
The means 108″ contains a calculation unit 710, two multiplication circuits 720 and 740, and a signal combination unit 730. The inputs 701 (127 in
The input 700 of the means 108″ is coupled to a second input of the multiplication circuit 740. The input 701 of the means 108″ is coupled to a second input of the multiplication circuit 720. The outputs of the multiplication circuits 720 and 740 are coupled to respective inputs of the signal combination unit 730. An output of the signal combination unit 730 is coupled to the output 713 of the means 108″ which has its output 713 coupled to the output 107 of the first circuit branch 104. The calculation unit 710 is adapted to derive multiplication factors m1(k) and m2(k) in dependence on the signals at the inputs 702 and 703 of the calculation unit 710, and to supply these multiplication factors to the respective outputs 711 and 712.
The practical example in
The case differentiation criterion is the sign of x(k), wherein x(k) is defined in accordance with the previously named formulae. The sign differentiates correlated (+) spectral components from anti-correlated (−) spectral components of the input signals, or 0 indicates non-correlated spectral components. The differentiation has the effect of these various spectral components being treated differently.
For correlated spectral components (with x(k)>0) the multiplication factors as in
The multiplication factors m1(k) and m2(k) are accordingly calculated as follows:
eZ
1(k)=Real(Z1(k))·Real(Z1(k))+Imag(Z1(k))·Imag(Z1(k)) (Eq. 10.1)
eZ
2(k)=Real(Z2(k))·Real(Z2(k))+Imag(Z2(k))·Imag(Z2(k)) (Eq. 10.2)
x(k)=Real(Z1(k))·Real(Z2(k))+Imag(Z1(k))·Imag(Z2(k)) (Eq. 10.3)
w(k)=x(k)/(eZ1(k)+L·eZ2(k)) (Eq. 10.4)
m(k)=(w(k)2+1)1/2−w(k) (Eq. 10.5)
m
S(k)=((eZ1(k)+eZ2(k))/(eZ1(k)+eZ2(k)+2·x(k)))1/2 (Eq. 10.6)
m
1(k)=m(k)|x(k)<=0 (Eq. 10.7.1)
m
1(k)=mS(k)|x(k)>0 (Eq. 10.7.2)
m
2(k)=1|x(k)<=0 (Eq. 10.8.1)
m
2(k)=mS(k)|x(k)>0 (Eq. 10.8.2)
wherein
The k-th complex output spectral value Y(k) of the signal at the output 713 of the means 108″ is therefore:
Y(k)=m1(k)·Z1(k)+m2(k)·Z2(k). (Eq. 11)
The explanation of the further operation is entirely along the lines of the explanations for
Number | Date | Country | Kind |
---|---|---|---|
TO2011A000890 | Oct 2011 | IT | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/EP2012/069799 | 10/5/2012 | WO | 00 | 4/3/2014 |