Recently proposed distributed adaptive estimation algorithms for wireless sensor networks (WSNs) do not consider errors due to noisy links, which occur during the transmission of local estimates between sensors. In this paper, we study the effect of noisy links on the performance of distributed incremental least-mean-square (DILMS) algorithm for the case of Gaussian regressors. More specifically, we derive theoretical relations which explain how steady-state performance of DILMS algorithm (in terms of mean-square deviation (MSD), excess mean-square error (EMSE), and mean-square error (MSE)) is affected by noisy links. In our analysis, we use a spatial-temporal energy conservation argument to evaluate the steady-state performance of the individual nodes across the entire network. Our simulation results show that there is a good match between simulations and derived theoretical expressions. However, the important result is that unlike the ideal links case, the steady-state MSD, EMSE and MSE are not monotonically increasing functions of step size parameter when links are noisy. In addition, the optimal step size is found in a closed form for a special case which minimizes the steady-state values of MSD, EMSE, and MSE in each node.
1. Introduction
In many wireless sensor network applications, multiple displaced sensors are used to estimate and track an unknown parameter, for example, average temperature, level of water contaminants, or a target position [1, 2]. In general, parameter estimation in WSN can be solved by either a centralized approach or a decentralized approach [3]. In a centralized approach, the spatially distributed sensor send their locally processed data to a fusion center to form the final estimate [4–6]. As the number of nodes increases, centralized processing becomes computationally prohibitive, since it would require communications over longer range which leads to reduced battery life. On the other hand, in decentralized estimation, spatially displaced sensors provide local estimates by collaborating with other nodes in the network [7–9].
In some WSN applications, the parameter estimation task must be done whereas the statistical model for the underlying processes of interest is not available, or it changes over time. This issue motivated the development of special class of decentralized approaches known as distributed adaptive estimation schemes [10–14]. In these schemes, every node is equipped with local computing ability to derive a local estimate and share it with its predefined neighbors. Using cooperative processing in conjunction with adaptive filtering per node enables the entire network and also each individual node to track not only the variations of the environment but also the topology of the network.
Depending on the manner by which the nodes communicate with each other, they may be referred to as incremental algorithms or diffusion algorithms. In the incremental mode, a cyclic path through the network is required, and nodes communicate with neighbors within this path [10–12]. The given algorithms in [10–12] use different adaptive filter in their structure, such as LMS, recursive least-squares (RLS), and affine projection. When more communication and energy resources are available, a diffusion cooperative scheme can be applied where nodes communicate with all of their neighbors, and no cyclic path is required. Both LMS-based and RLS-based diffusion algorithms have been considered in the literature [13, 14]. In [15], we have considered the quantization effects on the steady-state performance of DILMS algorithm.
1.1. Problem Description
Analysis of distributed estimation algorithms in the presence of noisy links is an important physical layer issue which has been considered for different algorithms in the literature [16–21]. Nevertheless, this issue has not been considered in distributed adaptive estimation algorithms. Hence, in this paper, we study the performance of DILMS algorithm in a WNS with noisy links between sensors. The importance of such a study arises from the following facts:
The WSN with noisy links is more realistic assumption than WSN with ideal links.
The performance of distributed adaptive estimation algorithm (e.g., DILMS) can vary significantly when they are used in noisy links WSN, which makes it vital to analyze their performance.
Our aim in this paper is to derive some theoretical expressions that describe the steady-state performance of DILMS algorithm with noisy links. It must be noted that analyzing the DILMS algorithm in a WSN with noisy links is challenging task, since an adaptive network comprises a system of systems that processes data cooperatively in both time and space in presence of noisy links. To be more specific, the contributions of this paper are listed below.
1.2. Contributions
We show that the performance of distributed adaptive estimation algorithms drastically decreases when links are noisy.
We show that unlike the ideal link case [10], the steady-state MSD, EMSE, and MSE curves are not monotonically increasing functions of step size when links are noisy. This result is very important in real-world DILMS implementation and highlights the importance of our work.
The optimal step size is found in a closed form for a special case which minimizes the steady-state values of MSD, EMSE, and MSE in each sensor.
1.3. Paper Organization
The remainder of this paper is organized as follows: in Section 2, we introduce the DILMS algorithm. In Section 3, we present our analysis and explain some results of the derived theoretical results in Section 4. Simulation results are given in Section 5, and Finally, the conclusions are drawn in Section 6.
1.4. Notation
Throughout the paper, we adopt boldface letters for random quantities and normal font for nonrandom (deterministic) quantities. The “*” symbol is used for both complex conjugation for scalars and Hermitian transpose for matrices. The main symbols used in this paper are listed in Table 1.
Symbols and their descriptions.
symbol
description
transposition
absolute squared operation
real value of a
statistical expectation of a
column vector with entries a and b
weighted norm for a column vector x,
trace of matrix A
μ
step size parameter in DILMS algorithm
identity matrix
2. The DILMS Algorithm
Suppose that a WSN is deployed to estimate an unknown vector from measurements collected at N nodes in a network. Each node k has access to time realizations of zero-mean spatial data , where each is a scalar measurement and each is a row regression vector. We collect the regression and measurement data into global matrices as
then the estimation problem is formulated as
In the appendix, we introduce a motivating application where an estimation problem (2) arises. The optimal solution is given by [10]
where and . Note that the cost function (2) can be decomposed as [10]
Using this property, in [10], a distributed incremental LMS strategy with a cyclic estimation structure is proposed, as follows:
where indicates the local estimate at the node k and time i and indicates the overall estimate at iteration i. For each time i, each node utilizes the local data and received from the node to obtain . At the end of this cycle, is employed as both the global estimate and the initial condition for the next time instant. Note that to implement the DILMS, the time realizations are used. The update equation for DILMS in noisy link condition changes to
where the vector is the channel noise term between sensor k and which is assumed to be additive with zero mean and covariance matrix . No distributional assumptions are required on the noise sequence. To show the effect of noisy links on the performance of DILMS algorithms, we consider a network with nodes. The observation noise has a variance of , and also we assume for channel noise. The curves are obtained by averaging over 300 experiments with (see Figure 1). As it is clear from Figure 1, the performance of DILMS algorithms drastically decreases when links are noisy.
The effect of noisy links on the performance of DILMS algorithm.
3. Performance Analysis
3.1. Data Model and Assumptions
In order to pursue the performance analysis, we will rely on the energy conservation approach of [10, 22]. Also, to carry out the performance analysis, we first need to assume a model for the data as is commonly done in the literature of adaptive algorithms. In the subsequent analysis, the following assumptions will be considered.
The desired unknown vector relates to the via
where is white noise term with variance and is independent of for all .
is independent of for .
is independent of for .
is independent of for all .
3.2. Weighted Energy Conservation Relation
In steady-state analysis, we are interested in evaluating the MSD, EMSE, and MSE for every node k which are defined as
where , , and . Note that due to incremental cooperation (cyclic path), for we use in (8). We further define the weighted a priori and a posteriori local errors for each node k as follows:
Multiplying the previous equation from left by and using the definitions in (9), we have
By replacing the from (11) and equating the weighted norm of both sides of the resultant equation, we arrive to the following relation:
We find from (13) that the cross terms are canceled out. Equation (13) is a space-time version of the weighted energy conservation relation used in [10] in the context of regular adaptive implementations.
3.3. Weighted Variance Relation
In this section, we use the energy conservation relation to evaluate the steady-state performance of the DILMS algorithm in every node when the links between nodes are noisy. To this aim, we need to have a recursive equation for . To obtain such a recursion we replace from (12) into (13) to get
We have dropped the time index i for compactness of notation. Now, we can relate the to via
If we replace (15) into (14), take expectation of both sides, and use assumptions (A.1)–(A.4), we obtain
By considering the error definitions and , we can obtain the following relations:
Using the property and (17), we can expand (16) as
Recursion (21) is a variance relation that can be used to infer the steady-state performance at every node k. Note that is solely regressors dependent and, therefore, decoupled from the weight error vector. For simplicity, in this work, we consider the following assumption.
The regressors arise from a source with circular Gaussian distribution with covariance matrix .
We introduce the eigndecomposition , where is a diagonal matrix with the eigenvalues of and is unitary, that is, . Define the transformed quantities
Using the above definitions, (21) and (22) can be rewritten in the equivalent forms
To proceed, we need to evaluate the moments in (24) as follows:
where and for circular complex data and for real data. Replacing these moments, (24) can be written as
Note from (27) that choosing to be diagonal, will be diagonal as well, suggesting a more compact notation. Thus, we introduce the column vectors
where the notation will be used in two ways: is a column vector containing the main diagonal of , and is a diagonal matrix whose entries are those of the vector λ. Therefore, using the diagonal notation, we obtain the following linear relation between the corresponding vectors :
where is a matrix that includes statistics of local data and given by
Using (30), can be rewritten in a more compact form as
For the sake of clarity, we reintroduce the time index i but drop the notation from the subscripts in (31) for compactness. Expression (31) becomes
We replaced by in order to indicate that the weighting matrix can be node dependent.
3.4. Steady-State Behavior
By comparing (34) with the similar equation for DILMS algorithm with ideal links (i.e., equation (55) in [10]), we can conclude that the desired steady-state MSD, EMSE, and MSE for DILMS algorithm with noisy links at node k can be expressed as
where
Note that in (38), the subscripts are all . It is evident that the effect of channel noise is addition of term to in the case of ideal link.
4. Discussion on Derived Theoretical Results
An important result is that unlike the ideal links case, in the presence of noisy links, the MSD, EMSE, and MSE curves are not monotonically increasing functions of step size. To show this result more clearly and to make (35)–(37) analytically more tractable, we assume that
, , .
We further assume that μ is small enough so that can be approximated as
So, is now a diagonal matrix and as a result, matrix will be diagonal as well. For small μ, we have , then Π can be approximated as
so that
Similarly, using the assumptions in (A.6), we have
so that becomes
Now, replacing (43) and (45) into (35) and using , we obtain
similarly, we can find the following approximations for EMSE and MSE as
We can easily conclude from (46)-(47) that the MSD, EMSE, and MSE curves are not monotonically increasing function of step size parameter. In Figure 2, we have shown the MSD as a function of μ when , , , and for different values of . As it is clear from Figure 2, for (i.e., noiseless links), the MSD curve is a monotonically increasing function of μ.
The steady-state MSD (in dB) curve as a function of μ and for different values of .
Remark 1.
To explain this behavior, we consider again the update (6). For small μ, the channel noise term say is dominant term in update equation, so as the steady-state performance deteriorates. As μ increases, the effect of channel noise term decreases, and finally as μ becomes larger, the steady-state performance deteriorates again like any adaptive algorithm.
Remark 2.
The optimal step size for MSD is given by
It must be noted that (48) is also the optimal step size for EMSE and MSE curves.
Remark 3.
Note that according to the given results in [10], as step size becomes smaller , steady-state values of MSD, EMSE, and MSE in each node should be more smaller too, but this is not the case in the presence of noisy links. In fact, μ must be chosen more carefully in real world.
5. Simulation
In this section, we provide computer simulations to compare the theoretical expressions with simulation results. To conduct our simulation results, we consider the following steps:
consider a network with noisy link and generate the measurement and regression data,
select a parameter (which is known for us but unknown for DILMS algorithm),
let the DILMS algorithm estimate in WSN with noisy links and data (generated in step 1),
obtain the MSD, EMSE, and MSE simulation results,
we apply the data (generated in step 1) to our derived theoretical results,
finally, we compare the resultant simulation results with derived theoretical results.
To this aim, we consider a distributed network with nodes, and choose and .
5.1. Regressors with Shift Structure
Although the analysis relied on the independence assumptions, simulations presented in this subsection were carried out using regressors with shift structure to cope with realistic scenarios. The regressors are generated at each node k according to the following recursion:
The expression above describes a first-order autoregressive (AR) process with a pole at , is a white, zero-mean, Gaussian random sequence with unity variance or a uniform random sequence with unity variance, and . In this way, the covariance matrix of the regressors is Toeplitz matrix with entries , with . The MSD, EMSE, and MSE are obtained by averaging the last 200 samples. Each curve is obtained by averaging over 100 independent experiments. The steady-state curves are generated by running the network learning process for 2000 iterations. We consider real data ().
We assume that for node k, covariance matrix is a diagonal matrix and has different values at the diagonal so
In fact, at each node k, the noise term is generated to result the required (assumed) covariance matrix . The statistical profiles for the mentioned parameters are illustrated in Figure 3.
Node profile and channel noise information.
In Figures 4–6, the steady-state of MSD, EMSE, and MSE for are plotted, respectively. It is clear from Figures 4–6 that there is a good match between simulation and theory. Note also that despite the diverse statistical profile, the MSD in Figure 4 is roughly even over the network. On the other hand, the EMSE and the MSE are more sensitive to local statistics, as depicted in Figures 5 and 6.
Steady-state MSD versus node, .
Steady-state EMSE versus node, .
Steady-state MSE versus node, .
In Figures 7 and 8, the MSD and EMSE for different values of μ for node are plotted. We note that unlike the ideal link case [10], here, the steady-state MSD and EMSE (and also MSE) are not a monotonically increasing function of μ.
Steady-state MSD versus μ, node .
Steady-state EMSE versus μ, node .
5.2. Independent Regressors
In this case, we assume that the regressors data arise from independent Gaussian, where their eigenvalue spread is . We assume for channel noise. The observation noise variance and are shown in Figure 9. In Figures 10 and 11, the steady-state of MSD and EMSE for are plotted. It is clear from Figures 10 and 11 that there is a good match between simulation and theory.
The Observation noise power profile, (a) and (b) .
Steady-state MSD versus node, .
Steady-state EMSE versus node, .
6. Conclusions and Future Work
In this paper, we derived theoretical relations to predict the performance of incremental distributed least-mean square algorithm (DILMS) when the links between nodes were noisy. Starting with the weighted energy conservation relation, we derived a variance relation for our problem. However, The important result is that unlike the ideal link case, the steady-state MSD, EMSE, and MSE are not a monotonically increasing function of step size when links are noisy. The simulation results revealed that there is good match between the derived closed-form expressions for MSD, EMSE, and MSE for every node in the network and the simulation results. Note that although in this paper we focused on the LMS adaptive filter, we can extend it for other adaptive filters. In our future work, we will consider more sophisticated cooperation modes (rather than incremental mode), such as diffusion mode.
Footnotes
Appendix
Consider that a network with N sensors is deployed to observe a physical phenomenon such as temperature, humidity, or any other events in a specified environment. At time i, the kth node collects a measurement (a time-realization data). By assuming an autoregressive (AR) model to represent these measurements, we have
where is additive zero-mean noise and the coefficients are parameters of the underlying model. If we define the parameter vector
and the regression vector
then (A.1) at each node k can be rewritten as an equivalent linear measurement model
The objective becomes to estimate the model parameter vector from the measurements and over the network and thereby has the form of a system identification problem.
Acknowledgment
This research is partially supported by Iran Telecommunications Research Center (ITRC) which is appreciated.
References
1.
EstrinD.GirodL.PottieG.SrivastavaM.Instrumenting the world with wireless sensor networksProceedings of IEEE Interntional Conference on Acoustics, Speech, and Signal ProcessingMay 2001Salt Lake City, Utah, USA203320362-s2.0-0034842559
2.
AkyildizI. F.SuW.SankarasubramaniamY.CayirciE.Wireless sensor networks: a surveyComputer Networks20023843934222-s2.0-003708689010.1016/S1389-1286(01)00302-4
3.
XiaoJ. J.RibeiroA.LuoZ. Q.GiannakisG. B.Distributed compression-estimation using wireless sensor networksIEEE Signal Processing Magazine200623427412-s2.0-3374633887810.1109/MSP.2006.1657815
4.
LiJ.AlRegibG.Distributed estimation in energy-constrained wireless sensor networksIEEE Transactions on Signal Processing20095710374637582-s2.0-7034964057510.1109/TSP.2009.2022874
5.
RibeiroA.GiannakisG. B.Bandwidth-constrained distributed estimation for wireless sensor networks—part I: Gaussian caseIEEE Transactions on Signal Processing2006543113111432-s2.0-3324449122910.1109/TSP.2005.863009
6.
RibeiroA.GiannakisG. B.Bandwidth-constrained distributed estimation for wireless sensor networks—part II: Unknown probability density functionIEEE Transactions on Signal Processing2006547278427962-s2.0-3374569903510.1109/TSP.2006.874366
7.
LuoZ. Q.Universal decentralized estimation in a bandwidth constrained sensor networkIEEE Transactions on Information Theory2005516221022192-s2.0-1714441978310.1109/TIT.2005.847692
8.
XiaoJ. J.LuoZ. Q.Decentralized estimation in an inhomogeneous sensing environmentIEEE Transactions on Information Theory20055110356435752-s2.0-2684454086810.1109/TIT.2005.855580
9.
XiaoL.BoydS.LallS.A spacetime diffusion scheme for peer-topeer least-squares estimationProceedings of the International Conference on Information Processing in Sensor Networks2006Nashville, Tenn, USA168176
10.
LopesC. G.SayedA. H.Incremental adaptive strategies over distributed networksIEEE Transactions on Signal Processing2007558406440772-s2.0-3454787014410.1109/TSP.2007.896034
11.
SayedA. H.LopesC. G.Distributed recursive least-squares strategies over adaptive networksProceedings of the 40th Asilomar Conference on Signals, Systems, and Computers (ACSSC '06)October-November 2006Pacific Grove, Calif, USA2332372-s2.0-4704910457010.1109/ACSSC.2006.356622
12.
LiL.ChambersJ. A.LopesC. G.SayedA. H.Distributed estimation over an adaptive incremental network based on the affine projection algorithmIEEE Transactions on Signal Processing20105811511642-s2.0-7294910852410.1109/TSP.2009.20250745071198
13.
LopesC. G.SayedA. H.Diffusion least-mean squares over adaptive networks: formulation and performance analysisIEEE Transactions on Signal Processing2008567312231362-s2.0-4664910833010.1109/TSP.2008.917383
14.
CattivelliF. S.LopesC. G.SayedA. H.Diffusion recursive least-squares for distributed estimation over adaptive networksIEEE Transactions on Signal Processing2008565186518772-s2.0-4664911523910.1109/TSP.2007.913164
15.
RastegarniaA.TinatiM. A.KhaliliA.Performance analysis of quantized incremental LMS algorithm for distributed adaptive estimationSignal Processing2010908262126272-s2.0-7795120827210.1016/j.sigpro.2010.02.019
16.
SchizasI. D.RibeiroA.GiannakisG. B.Consensus in ad hoc WSNs with noisy links—part I: distributed estimation of deterministic signalsIEEE Transactions on Signal Processing20085613503642-s2.0-3774901551810.1109/TSP.2007.906734
17.
SchizasI. D.GiannakisG. B.RoumeliotisS. I.RibeiroA.Consensus in Ad hoc WSNs with noisy links—part II: distributed estimation and smoothing of random signalsIEEE Transactions on Signal Processing2008564165016662-s2.0-4184914380310.1109/TSP.2007.908943
18.
AysalT. C.BarnerK. E.Constrained decentralized estimation over noisy channels for sensor networksIEEE Transactions on Signal Processing2008564139814102-s2.0-4184913434010.1109/TSP.2007.909006
19.
AysalT. C.BarnerK. E.Blind decentralized estimation for bandwidth constrained wireless sensor networksIEEE Transactions on Wireless Communications200875146614712-s2.0-4714909300710.1109/TWC.2008.0606874524301
20.
KarS.MouraJ. M. F.Distributed consensus algorithms in sensor networks with imperfect communication: link failures and channel noiseIEEE Transactions on Signal Processing20095713553692-s2.0-5864910642910.1109/TSP.2008.2007111
21.
ChamberlandJ. -F.VeeravalliV. V.The impact of fading on decentralized detection in power constrained wireless sensor networks3Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '04)May 2004Montreal, Canada837840
22.
SayedA. H.Fundamentals of Adaptive Filtering2003Hoboken, NJ, USAWiley