1. Field of the Invention
The present invention relates generally to sensor network and, particularly, to wireless sensor networks in which a plurality of wireless sensors are spread over a geographical location. More particularly, the invention relates to estimation in such a network utilizing a noise-constrained diffusion least mean square method for estimation in the adaptive network.
2. Description of the Related Art
In reference to wireless sensor networks, the term “diffusion” is used to identify the type of cooperation between sensor nodes in the wireless sensor network. That data that is to be shared by any sensor is diffused into the network in order to be captured by its respective neighbors that are involved in cooperation.
Wireless sensor networks include a plurality of wireless sensors spread over a geographic area. The sensors take readings of some specific data and, if they have the capability, perform some signal processing tasks before the data is collected from the sensors for more detailed thorough processing.
A “fusion-center based” wireless network has sensors transmitting all the data to a fixed center, where all the processing takes place. An “ad hoc” network is devoid of such a center and the processing is performed at the sensors themselves, with some cooperation between nearby neighbors of the respective sensor nodes.
Recently, several algorithms have been developed to exploit this nature of the sensor nodes and cooperation schemes have been formalized to improve estimation in sensor networks.
Least mean squares (LMS) algorithms are a class of adaptive filters used to mimic a desired filter by finding the filter coefficients that relate to producing the least mean squares of the error signal (i.e., the difference between the desired and the actual signal). The LMS algorithm is a stochastic gradient descent method, in that the filter is only adapted based on the error at the current time.
dk(i)=uk,iw0+vk. (1)
where uk,i is a known regressor row vector of length M, w0 is an unknown column vector of length M and vk(i) represents additive noise. The output and regressor data are used to produce an estimate of the unknown vector, given by ψk,i.
Diffusion LMS (DLMS) techniques use the data to find an estimate for the unknown vector. There are two strategies that may be employed. In the first strategy, each node combines its own estimate with the estimates of its neighbors using some combination technique, and then the combined estimate is used for updating the node estimate. This method is referred to as Combine-then-Adapt (CIA) diffusion. It is also possible to first update the estimate using the estimate from the previous iteration and then combine the updates from all neighbors to form the final estimate for the iteration. This method is known as Adapt-then-Combine (ATC) diffusion. Simulation results show that ATC diffusion outperforms CTA diffusion.
Using LMS, the CTA diffusion algorithm is given by:
where {clk}IεN
The ATC algorithm is achieved by reversing the order of the equations of equation set (2):
Equation (4) is a global cost function. The local cost function at each node is given by:
Jk(w)=E└|dk−ukw|2┘. (5)
Completing the squares and using the notation E[uk*uk]=Ru,k, equation (5) can be rewritten as:
Jk(w)=∥w−wk∥R
The global cost function now becomes:
The above model assumes that each node has access to the entire network. In a practical system, however, a node has access only to its closest neighbors. Thus, the cost function has to be approximated with only data from neighbors being shared at each node. As a result, the weighting matrix for the second term does not remain Ru,l but, instead, has to be replaced by a constant weighting factor blk. The value of wl is also replaced by its intermediate estimate from node l, i.e., ψl. Eventually, the cost function becomes:
Defining Jk1(w)=E└|dk−ukw|2┘ yields:
Conventional LMS algorithms are not equipped to efficiently provide such estimates, particularly in complex and large adaptive networks. The diffusion scheme based on conventional LMS algorithms lacks knowledge of noise power in the system. Thus, it cannot perform well in the presence of noise. Thus, a noise-constrained diffusion least mean square method for estimation in adaptive networks solving the aforementioned problems is desired.
The noise-constrained diffusion least mean square method for estimation in adaptive networks is based on the Least Mean Squares (LMS) algorithm. The method provides for a variable step size in which the step-size variation rule results directly from the noise constraint. In a first embodiment, the noise-constrained diffusion least mean square method for estimation in adaptive networks is given by the following steps: (a) establishing an adaptive network having N nodes, where N is an integer greater than one, and for each node k, a number of neighbors of node k is given by Nk, including the node k, where k is an integer between one and N; (b) establishing an integer i and initially setting i=1; (c) calculating an output of the adaptive network at each node k as dk(i)=uk,iw0+vk(i), where uk,i represents a known regressor row vector of length M, w0 represents an unknown column vector of length M and vk(i) represents noise in the adaptive network, where M is an integer; (d) calculating an error value ek(i) at each node k as ek(i)=dk(i)−uk,iwk,i-1, where wk,i represents an estimate of an output vector for node k at iteration i; (e) calculating a local estimate for each node neighboring node k, ψk,i, for each node k as ψk,i=wk,i-1+μk(1+γλk,i-1)uk,i*ek(i), where
β and γ being selectable unitless parameters, μk represents a node step size for each node k, and where σv,k2 represents an additive noise variance; (f) calculating an estimate of an output vector wk,i for each node k as
where l is an integer, clk represents a combination weight for node k, which is fixed; (g) if ek(i) is greater than a selected error threshold, then setting i=i+1 and returning to step (c); otherwise, (h) defining a set of output vectors wk for each node k, where wk=wk,i; and (i) storing the set of output vectors in computer readable memory.
These and other features of the present invention will become readily apparent upon further review of the following specification.
Similar reference characters denote corresponding features consistently throughout the attached drawings.
The noise-constrained diffusion least mean square method for estimation in adaptive networks is based on the Least Mean Squares (LMS) algorithm. The method uses a variable step size in which the step-size variation rule results directly from the noise constraint.
Assuming knowledge of the additive noise variance σv,k2, the cost function of equation (8) may be modified as follows. The second term in equation (9) goes to zero asymptotically. This means that the problem reduces to minimizing Jk1(w) with respect to w based on the constraint Jk1(w)=σv,k2. The Lagrangian for this problem is given by:
where the last term is added as a correction term to avoid any spurious behavior. Using the Robbins-Munro algorithm, the adaptive solution now becomes:
Solving the first partial derivative yields:
Similarly, solving the second partial derivative yields:
which reduces to:
λk,i=λk,i-1+βγ(E└|dk−ukw|2┘−σv,k2)−2βγλk,i-1. (15)
Replacing βγ by β/2 and incorporating the partial derivatives into the algorithm yields the steepest descent solution:
Equation (16) can be broken into a two-step process, namely
Combining equations (17) and (18) with equation (16) produces the steepest descent solution to the noise-constrained problem.
In order to obtain the adaptive solution, Ru,k, Rdu,k and E└|dk−ukw|2┘ are replaced by their instantaneous values. Noting that ek(i)=dk(i)−uk,iwk,i-1, the replacement provides:
Thus, equations (20)-(22) form the noise-constrained diffusion LMS (NCDLMS) method using the ATC approach. The ATC noise-constrained diffusion least mean square method for estimation in adaptive networks is given by the following steps: (a) establishing an adaptive network having N nodes, where N is an integer greater than one, and for each node k, a number of neighbors of node k is given by Nk, including the node k, where k is an integer between one and N; (b) establishing an integer i and initially setting i=1; (c) calculating an output of the adaptive network at each node k as dk(i)=uk,iw0+vk(i), where uk,i represents a known regressor row vector of length M, w0 represents an unknown column vector of length M and vk(i) represents noise in the adaptive network, where M is an integer; (d) calculating an error value ek(i) at each node k as ek(i)=dk(i)−uk,iwk,i-1, where wk,i represents an estimate of an output vector for node k at iteration i; (e) calculating a local estimate for each node neighboring node k, ψk,i, for each node k as φk,i=wk,i-1+μk(1+γλk,i-1)uk,i*ek(i), where
β and γ being selectable unitless parameters, μk represents a node step size for each node k, and where σv,k2 represents an additive noise variance; (f) calculating an estimate of an output vector wk,i for each node k as
where l is an integer, clk represents a combination weight for node k, which is fixed; (g) if ek(i) is greater than a selected error threshold, then setting i=i+1 and returning to step (c); otherwise, (h) defining a set of output vectors wk for each node k, where wk=wk,i; and (i) storing the set of output vectors in computer readable memory.
In order to perform mean transient analysis on the ATC scheme, we use the following conditions:
w(o)Δ=INwo, diΔ=col{d1(i), . . . ,dN(i)},
UiΔ=diag{u1,i, . . . uN,i}, viΔ=col{v1(i), . . . vN(i)},
where is the Kronecker product and the N×1 vector di is given by di=Uiw(o)+vi. Defining D=diag{μ1IM, . . . , μNIM} and Li-1=diag{(1+γλ1,i-1)IM, . . . , (1+γλN,i-1)IM}, then the combiner weight-matrix may be defined as CΔ=[c1, . . . , cN], which allows us to write GΔ=CIM. Next, defining the weight-error vector as wiΔ=wo−wi and noting the fact that wiΔ=col{wi, . . . , wN}, yields:
wiΔ=w(o)−wi. (26)
Rearranging the above produces:
Taking expectation value and solving the above yields:
E[wi]=G(INM−DE[Li-1]RU)E[wi-1]. (28)
Letting E[Bi-1]=(INM−DE[Li-1]RU), it can be seen that the stability of the system depends on
∥GE[Bi-1]∥2≦∥G∥2·∥E[Bi-1]∥2. (29)
Since we already have G=CIM and we know that Bi-1 and RU are Hermitian and block diagonal, equation (29) reduces to:
|
A combiner rule is picked such that ∥C∥2≦1, so that the cooperative scheme provides robustness over the non-cooperative scheme. For a symmetric matrix C, we have ∥C∥2=1 so that:
|
This means that the cooperative system will always be stable as long as the non-cooperative system is stable. For the non-cooperative system to be stable, the step-size should be such that |
Expanding the above and solving gives the step-size range for node k:
where λk,∞ is the steady-state value for the Lagrangian and should ideally reduce to zero, resulting in the familiar expression for the step-size:
Simulation results are provided below to illustrate the performance of the present NCDLMS method. Comparisons are shown against the Diffusion LMS (DLMS) method. Simulations have been performed for an exemplary network containing N=15 nodes connected as shown in
The NCDLMS algorithm is compared with DLMS for both the CTA as well as the ATC variations. Also plotted is the case where no cooperation takes place between nodes; i.e., simple LMS adaptation without combining. Step-sizes for the DLMS and no cooperation case are all set to μk=0.01. For the NCDLMS algorithm, the initial step-size is set at μk=0.025, whereas β=0.01 and γ=141. For all three diffusion cases, the Metropolis rule is used for combining.
The processor 12 may be associated with, or incorporated into, any suitable type of computing device, for example, a personal computer or a programmable logic controller. The display 16, the processor 12, the memory 14, the user interface 18, network interface 20 and any associated computer readable media are in communication with one another by any suitable type of data bus, as is well known in the art. Additionally, other standard components, such as a printer or the like, may interface with system 10 via any suitable type of interface.
Examples of computer readable media include a magnetic recording apparatus, an optical disk, a magneto-optical disk, and/or a semiconductor memory (for example, RAM, ROM, etc.). Examples of magnetic recording apparatus that may be used in addition to memory 14, or in place of memory 14, include a hard disk device (HDD), a flexible disk (FD), and a magnetic tape (MT). Examples of the optical disk include a DVD (Digital Versatile Disc), a DVD-RAM, a CD-ROM (Compact Disc-Read Only Memory), and a CD-R (Recordable)/RW.
It is to be understood that the present invention is not limited to the embodiments described above, but encompasses any and all embodiments within the scope of the following claims.
Number | Name | Date | Kind |
---|---|---|---|
6026121 | Sadjadpour | Feb 2000 | A |
6377611 | Hwang | Apr 2002 | B1 |
6741707 | Ray et al. | May 2004 | B2 |
7298774 | Miyata et al. | Nov 2007 | B2 |
20020041678 | Basburg-Ertem et al. | Apr 2002 | A1 |
20050201457 | Allred et al. | Sep 2005 | A1 |
20080136645 | Lin et al. | Jun 2008 | A1 |
20080260141 | Givens | Oct 2008 | A1 |
20100033764 | Guo et al. | Feb 2010 | A1 |
20120106357 | Zummo et al. | May 2012 | A1 |
Number | Date | Country |
---|---|---|
101252559 | Aug 2008 | CN |
20050001888 | Jan 2005 | KR |
Entry |
---|
Saeed et al. “Noise-Constrained Diffusion Least Mean Squares Over Adaptive networks” 2010, IEEE, pp. 288-292. |
Lopes et al. “Diffusion Least Mean Squares Over Adaptive Networks: Formulation and Performance,” IEEE Trans. Signal Processing, vol. 56, No. 7, pp. 3122-3136, Jul. 2008. |
M.O. Bin Saeed, A. Zerguine and S.A. Zummo, “Variable Step-Size Least Mean Square Algorithms Over Adaptive Networks”, 10th International Conference on Information Science, Signal Processing and their Applications (ISSPA 2010), May 2010. |
Number | Date | Country | |
---|---|---|---|
20120135691 A1 | May 2012 | US |