The present invention relates to a method for the non-linear estimation of wireless signals from several sources, the time/frequency representation of which shows an unknown non-zero proportion of zero components, using an array made up of P>2 antennas, when the directional vectors U and V of the sources emitting these signals are additionally known or estimated.
It is commonly necessary to estimate wireless signals (soundproofing) originating from radars, communication systems, or acoustic signals (audio or sonar), and received by a listening system made up of an antenna array.
The received signal results from a temporal and spectral mixture of no more than 2 sources, the directional vectors of which are assumed to be known, since they are estimated beforehand.
The criterion traditionally used is the maximum likelihood (ML), leading to processing by spatial linear filtering, which improves the signal-to-noise ratio by a factor equal to the number of sensors in the single-source case.
In the scenarios being considered (known directional vectors), this processing leads to a nonbiased linear estimate with minimal signal variance for which no a priori knowledge is available.
Other methods that are more complicated to carry out can be used, such as Capon filtering, when the directional vectors are imperfectly or not known (“Robust Adaptive Beamforming”, eds P. Stoica and J. Li, Wiley, 2006).
None of these methods exploit any a priori on the signal, and in particular none make it possible to correctly delimit the temporal and/or spectral supports of the signal, since a linear (ML) or pseudo-linear (Capon) processing always provides an output signal, even if at the input, the measurement is only made up of noise.
The problem is to access finer knowledge of the signal.
The invention aims to propose a method enabling a finer determination of the components of the signal.
To that end, the invention relates to a method including the following steps:
The method thus makes it possible to answer the following questions: for which of the signals present (their number being assumed to be limited to 2 locally), what are the temporal and spectral supports of the supposed signal described by the components obtained using a time/frequency analysis? And, what is the value of each component when it is not zero? The answer to these questions makes it possible to improve the knowledge of the signal.
According to specific embodiments, the method includes one or more of the following features:
The invention will be better understood upon reading the following description, done in reference to the appended drawings:
The device 8 for estimating a mixture of signals coming from several sources 12 according to the invention illustrated in
The invention is suitable both for monopolar antenna arrays and bipolar antenna arrays.
The device further includes computing modules. In different alternative embodiments of the device for estimating signals according to the invention, the computer modules can be arranged with different architectures; in particular, each step of the method can be carried out by a separate module, or on the contrary, all of the steps can be grouped together within a single information processing unit 14.
The computing unit(s) are connected to the sensors by any means suitable for transmitting the received signals.
The computing unit(s) include information storage means, as well as computing means making it possible to carry out the algorithms of
Advantageously, the reception is done on a spatial diversity array (interferometric array) and the demodulation of the signal allowing the “baseband lowering” is done by the same local oscillator for all of the antennas in order to ensure coherence. The received signal is sampled in real or complex (double demodulation and quadrature or any other method) on each channel.
The received signal, filtered in a band typically of several hundred MHz, is modeled by: s0(t)ei
This signal is sampled at a rhythm Te such that 1/Te>>2×band of the wanted signal.
The weighted and overlapping Discrete Fourier Transforms (DFT) of this signal are calculated over NTDF points. The weighting serves to reduce the secondary lobes. Since this weighting causes of variation in the contribution of the data to the DFT (the data at the center of the temporal support of the DFT being assigned a much higher weight than the data on the edges of the support), which can go as far as a loss of the short signals, the hypothesis is used of overlapping of the temporal supports.
The collected measurements are therefore the results of the DFTs. They constitute a time-frequency grid, the boxes of which are called time-frequency boxes. Each box of the grid contains a complex vector that is the result of the Discrete Fourier Transforms for all of the channels, for a given time interval and frequency interval.
For the frequencies of the signals and the distances involved here, the wave front is considered to be planar. The antennas therefore receive the signal with a phase difference depending on the two angles between the wave plane and the plane of the antennas.
In the mono-phase case, the measurements collected on the complete network are therefore written as follows:
X
n
=s
n
U+W
n
In the case of a monopolar interferometric array, U is written as follows:
where g is a complex scalar dependent on the polarization of the incident wave and its arrival direction, and where ui are modulus 1 complex numbers representing the geometric phase shifts associated with the direction of the incident wave. It is possible to choose one of the antennas of the receiver as phase reference.
In the case of a interferometric array, U is written as follows:
U=hH+vV,
where h and v are complex scalars such as |h|2+|v|2=1 that express the polarization of the incident wave, and where H (V, resp.) is the response from the array to a horizontally (vertically, resp.) polarized wave. H and V only depend on the direction and frequency of the incident wave.
In all cases, U is of dimension P, where P is the number of antennas used.
It is possible to consider that U is standardized, and that sn carries the power of the signal and the mean gain of the array.
U is kept by the DFT, which is a linear transform, and is found on the signal output by the DFT.
In the general scenario, it is possible to have a mixture of K signals (K optionally being greater than the resolution of the array). The signal is then written:
The fundamental hypothesis is that when the set of N time/frequency boxes is restricted to a rectangular zone or window with index j, the complexity of the environment is such that in such a window, the mixture of the signals is limited to two signals.
The model then becomes:
X
jn
=s
j1(n)Uj1+sj2(n)Uj2+Wn Equation 2 Expression of a signal mixture in a window
Windows of predetermined size are defined to cover the time-frequency grid. All of these windows form a division of the grid. The size of the windows is chosen such that no more than two signals from two sources are present in each window.
The vector U or the vectors U and V designating the unitary directional vector(s) formed by the incident signal(s) relative to the array are next extracted (or estimated), using any suitable known means.
A loop is formed to travel all of the windows defining the division of the grid.
For each window, a step for estimating the signal(s) is carried out using the conditional expectation and approximations that can be made thereto for the specific model of the signals.
The final processing is broken down into:
The estimate of the signal is done over a window where the situation is mono-source or dual-source.
In a mono-source situation, the vectorial signal measured for the box with index n is
X
n
=s
n
U+W
n
, n=1, 2, . . . Equation 3 Expression of the measured signal (mono-source scenario)
In a dual-source situation, one has:
X
n
=s
n
U+c
n
V+W
n
, n=1, 2 . . . Equation 4 Expression of the measured signal (dual-source scenario)
In the writings of Equation 3 and Equation 4, U and V are unitary directional vectors of a monopolar or bipolar array, sn and cn are complex signals that must be estimated; Wn is the noise, which is spatially white and in n, Gaussian, centered and with covariance: E(WnWn*)=2σ2Ip where Ip is the unit matrix of Cp (P being the number of receiving channels). * designates the conjugated transpose.
Hereinafter, it will be assumed that U and V are estimated by Û and {circumflex over (V)}, for example using a method of the MUSIC type, and that this estimate having been done over a number of boxes N>>1, it is possible to make the approximation U=Û, V={circumflex over (V)}.
To account for the fact that sn (or cn) can be zero for certain n, without knowing the modulation of the signal in advance, sn (or cn) is modeled as independent samples in n of a random variable whereof the probability density is a mixture (q, 1−q, 1<1) of two Gaussians centered on respective variances 2σj2 (j=1 for sn and j=2 for cn), and 2τ2, with τ2<σ2<σj2.
2σj2 is the power of the wanted signal if it is present when τ2 is neglected in the expression (1−q)2τ2+q2σj2 of the mean power.
τ is a regularization parameter of the model that makes it possible to apply the Bayes formula for laws of probability allowing probability densities relative to the Lebesgue measurement; however, the physical reality is that there is no signal when this is modeled using the variance-centered Gaussian 2τ2 (power 2τ2). That is why, at the end of the calculations, only the limit of the expressions is retained when τ→0.
Sn and cn are considered to be independent.
In both cases, mono-source or dual-source, the measured signal is written in the single form:
X
n
=MS
n
+W
n
, n=1, 2, . . . N Equation 5 Metrical form of the measured signal
where M=U and Sn=sn in mono-source, and M=(U V), Sn(sn cn)T in dual-source, with M known and σ2 known.
In the rest of the document, to make the notations lighter, with the understanding that the processing is done on each box with index n independently, the index n is omitted and X denotes the measurement in a time/frequency box, s the mono-source signal and S=(s c)T the dual-source signal in a time/frequency box.
The a priori knowledge of S is given by a probability density. In the mono-source case:
In the dual-source case:
where q1=q2,q2=q3=q(1−q),q4=(1−q)2 if the sources are considered to be independent and equally probable (hereinafter, we generalize to a distribution q1,q2,q3,q4 of 4 situations not connected by the above expressions),
And
are the covariance matrices of S for the four possible cases.
S is estimated using the conditional expectation by using the mono-source and dual-source models (Equation 5, Equation 6 and Equation 5, Equation 7).
The Conditional Expectation (CE) is the estimator S that minimizes the mean quadratic deviation E(∥S−Ŝ∥2). It is also nonbiased, and provides an explicit solution for Ŝ. It is built as follows:
Let X be the measurement; its probability density, which depends on the parameter to be estimated S, is interpreted as the conditional probability density of X knowing S. One therefore has p(X/S) and p(S) derived from the a piori knowledge of S.
Ŝ is given by the explicit formula:
p(S/X), the conditional probability of S knowing X, is obtained by the Bayes formula.
In the case where X=MS+V and p(S) is Gaussian, covariance-centered C, it is possible to find S analytically, which is generally not the case.
This is a linear function of X. In fact (in dimensions 2):
Let us take
By completing the “square” in S, we have:
where K2 is a constant (=1/π44σ4) in dimension 2. Σ is a Hermitian matrix defined positive. One deduces from this:
The complete expressions of Equation 12 will be used to find the desired estimator for our problem.
In the case where S is a Gaussian sample, Equation 9, Equation 10, Equation 11, Equation 12 and Equation 13 yield:
which can also be written:
{circumflex over (S)}=(2σ2Σ−1)−1M*X=(M*M+2σ2C−1)−1M*X
One concludes from this that if 2σ2C−1<<M*M, Ŝ is reduced to the maximum likelihood estimator of S using the model of Equation 5 with no a priori knowledge of S: ŜMV=(M*M)−1M*X.
The condition 2σ2C−1<<M*M as matrices is also expressed by C>>2σ2(M*M)−1, which means that the a priori on S that is defined by C does not provide real information on S.
The estimate of the signal in the mono-source case is as indicated below.
In the one-dimensional case for S=s (mono-source), M=U, and therefore M*M=1;
The matrix C is reduced to the constant c;
is deduced from this
This then yields:
The conditional expectation is obtained, in the Gaussian case for s, by the quotient of Equation 15 by Equation 16:
The estimate of the signal in the dual-source case is as indicated below.
The conditional expectation estimator is obtained using Equation 15 and Equation 16 for the mixture density of S given by Equation 7.
After simplifying by the common factor
in all of the terms in the numerator and the denominator, one obtains:
Let us take Γj=2σ2Cj−1 (without dimension),
Q
j=(M*M+2σ2Cj−1)−1=(M*M+Γj)−1=Σj/2σ2
We have:
One deduces from this:
The expression of the estimate of the signal is subject to an approximation to allow it to be estimated as indicated below in the dual-source case.
The products of determinants in Equation 19 are respectively equal to the following expressions, an approximation of which is provided for a good signal-to-noise ratio (σ12/σ2>>1 and σ22/σ2>>1) and for τ→0.
detQ
1
detΓ
1=[(1+σ2/σ12)(1+σ2/σ22)−|U*V|2]−1σ4/σ12σ22≈(1−|U*V|2)−1σ4/σ12σ22
detQ
2
detΓ
2=[(1+σ2/σ12)(1+σ2/τ2)−|U*V|2]−1σ4/σ12τ2≈(τ2/σ2)(1+σ2/σ12)−1·(σ4/σ12τ2)≈σ2/σ12
detQ
3
detΓ
3=[(1+σ2/τ2)(1+σ2/σ22)−|U*V|2]−1σ4/σ22τ2≈σ2/σ22
detQ
4
det Γ
4=[(1+σ2/τ2)2−|U*V|2]−1σ4/τ4≈1 Equation 21 Det Qi det Γi
Likewise, one finds, for Qj when τ→0:
Q1 is unchanged.
One can see that the products det Qi det Γi have a finite limit in each of the four situations, as do the matrices Qj, which is a satisfactory behavior.
One has thus obtained a first expression of the estimator.
In reality, only one of the terms in Equation 19 is preponderant for each box, which leads to a first simplification. The new estimating processing of the Ŝ is deduced from this. Ŝ=Qj0M*X where
Which is simplified as:
M*X is given by:
The det Qj·det Γj are given by Equation 21.
The Qj are given by Equation 22.
The qj are for example given by
if there is independence of the 4 situations and equal probability for s≠0,c≠0.
F(j) is given by:
F(j1)=ln(q1detQ1detΓ1)+X*MQ1M*X/2σ2=ln(q2(1−|U*V|2)σ2/σ12σ22)+X*MQ1M*X/2σ2
F(j2)=ln(q2detQ2detΓ2)+X*MQ2M*X/2σ2=ln(q(1−q)σ2/σ12)+X*UU*X/2σ2
F(j3)=ln(q3detQ3detΓ3)+X*MQ3M*X/2σ2=ln(q(1−q)σ2/σ22)+X*VV*X/2σ2
F(j4)=ln(q4detQ4detΓ4)+X*MQ4M*X/2σ2=ln((1−q)2)
For j0=1, the maximum likelihood estimator is found. Indeed, in this case,
and Q1=(M*M)−1 such that
For j0=2,ŜT=(U*X,0) (filtering by the directional vector of source 1).
For j0=3,ŜT=(0,V*X) (filtering by the directional vector of source 2).
For j0=4,ŜT=(0,0)
Where the symbol T designates the transpose.
One has thus “linearized” the optimal processing, since one has obtained four linear filters, controlled by the decision on the type of situation for each box: both sources are present/source 1 is present/source 2 is present/neither of the sources is present. The obtained estimator is called “Conditional Expectation with 4 Linear Filters”.
In this way, we have simplified the optimal estimator, by breaking it down into two steps:
It is satisfactory to see that the estimator is independent of τ, which is the expected behavior, since τ is not a physical parameter, but an artifice making it possible to modify the “absence of signal” situation by a very tight Gaussians.
This estimator requires the calculation of three quadratic forms and a test. The difficulty remains calculating the unknown parameters qj,j=1 . . . 4, σ12, σ22 (the power of the noise 2σ2 is presumed to be known).
In the specific case where the 2 sources are independent, and have the same power and the same presence rate, the parameters can be estimated by calculating the empirical moments of order 2 and 4.
If we call 2σ′2 the shared value of the variance of the Gaussian representing each source, and q the probability shared by the 2 sources, everything happens as if we were in a mono-source situation, with a single source of variance 2σM′2=2σ′2 and presence probability qM=2q.
σM′2 and qM are then given by the following equations:
In the general case (independent sources), there are 4 parameters of the model: q1, q2, σ12, σ22. One skilled in the art knows how to generalize the method of moments above to higher orders to obtain the estimates of these parameters.
An alternative of the estimating processing consists of simplifying the decision step previously described, as follows:
The conditional expectation considers all four possible situations:
s≠0,c≠0;s≠0,c=0;s=0,c≠0;s=0,c=0
It is normally necessary to address a decision problem with four hypotheses.
To simplify, we propose to test s≠0 against s=0 independently of c on the one hand, and to test c≠0 against c=0 independently of s on the other hand. We therefore perform two tests with two hypotheses instead of one test with four hypotheses.
These tests will be done from preprocessed measurements U*X, V*X.
T
the 2 hypotheses for s
where the (.) indicate the scalar x scalar product
and where u=U*W,v=V*W:(u,v) is therefore Gaussian, centered and with covariance:
and where c is an unknown parameter.
It is a problem not varying by the group of sector translations (U*V 1)T and a linear hypothesis problem (see “Testing Statistical Hypothesis”, 3rd edition, E. L. Lehmann, J. P. Domano, Springer, 2005). It may be processed by first performing a projection on the orthogonal of (U*V 1)T to eliminate c, then testing the presence of s using the chi2 test. The projection is written:
The test to be performed therefore pertains to the measurement |U*X−(U*V)V*X|2:
|U*X−(U*V)V*X|2>ou<λ Equation 26 Simplified hypothesis test on s
Which amounts to the same thing as performing the test:
where ŝMV is the estimate within the meaning of the maximum likelihood of s.
T
In the same way for c, M*X is projected on the orthogonal of (1 V*U)T in order to eliminate the terms in s, and the new measurement to be considered is obtained:
One obtains the following test:
|(V*U)U*X−V*X|2>ou<μ Equation 27 Simplified hypothesis test on c
which can be written
where ĉMV is the estimate of c within the meaning of the maximum likelihood of c.
In dual-sources, the proposed estimator consists of performing the following operations:
As illustrated in
Then, depending on the situation, spatial filtering is applied in steps 331 to 334 under the following conditions, which makes it possible to obtain the so-called Conditional Expectation with Independent Decisions (CEID) estimator:
If |ŝMV|>seuil and |ĉMV|>seuil:ŜECDI=(M*M)−1M*X (step 331)
If |ŝMV|>seuil and |ĉMV|>seuil:ŝECDI=U*X,ĉECDI=0 (step 332)
If |ŝMV|>seuil and |ĉMV|>seuil:ŝECDI=0,ĉECDI=V*X (step 333)
If |ŝMV|>seuil and |ĉMV|>seuil:ŜECDI=0 (step 334)
In mono-source, the proposed estimator consists of performing the following operations, as illustrated in
From the calculation of the dual-source maximum likelihood estimator ŝMV=U*X done in step 210, thresholding of the modulus of ŝMV is done in step 350, then, depending on the situation, spatial filtering is applied in steps 361 or 362 under the following conditions:
If |ŝMV|>seuilŝECDI=U*X (step 361)
If |ŝMV|<seuilŝECDI=0 (step 362)
Advantageously, the threshold for steps 320, 350 is determined as follows:
Pfa refers to the probability of deciding s≠0 whereas s=0, and Pd is the probability of deciding s≠0 whereas s≠0.
For example and non-limitingly, it is proposed to use the Neyman-Pearson criterion, which consists of determining the Pfa (for example, several percent), and in return, maximizing Pd, which makes it possible to obtain a threshold on λ′ and μ′. For example and non-limitingly, it is also possible to adjust the value of λ′ (μ′, resp.) such that 1−Pd=Pfa around a set RSB.
A close alternative of the Maximum Likelihood called Thresholded Maximum Likelihood (TML) consists of performing the following operations:
Calculating the maximum likelihood estimator ŜMV=(M*M)−1M*X
Thresholding the modulus of the components of ŜMV
Depending on the situation, applying spatial filtering
If |ŝMV|<seuil:ŝMVS=0, otherwise ŝMVS=ŝMV
If |ĉMV|<seuil:ĉMVS=0, otherwise ĉMVS=ĉMV
Another alternative consists of using one of the two previous estimators to obtain an initialization of the unknown parameters qj,j=1 . . . 4, σ12, σ22, then applying the Conditional Expectation estimator or the Conditional Expectation with 4 Linear Filters estimator.
Number | Date | Country | Kind |
---|---|---|---|
1402981 | Dec 2014 | FR | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2015/081213 | 12/23/2015 | WO | 00 |