A distributed source coding approach is proposed for robust data communications in sensor networks. When sensor measurements are quantized, possible correlations between the measurements can be exploited to reduce the overall rate of communication required to report these measurements. Robust distributed source coding (RDSC) approaches differentiate themselves from other works in that the reconstruction error of all sources will not exceed a given upper bound, even if only a subset of the multiple descriptions of the distributed source code are received. We deal with practical aspects of RDSC in the context of scalar quantization of two correlated sources. As a benchmark to evaluate the performance of the proposed scheme, we derive theoretically achievable distortion-rate performances of an RDSC for two jointly Gaussian sources by applying known results on the classical multiple description source coding.
1. Introduction and Motivation
Sensor networks are usually collections of remote sensors that gather temporal and spatial information about a source distributed in time and space. Each sensor reports its measurements to a central unit which extracts meaningful information about the distributed source being measured. The spatial and temporal correlations between sensor measurements can be exploited to reduce the amount of communication required between the sensors and the central unit to achieve a desired reconstruction quality of the distributed source. We call this problem the distributed source coding (DSC) problem.
For low-cost sensors, remote communication links to the central server are often unreliable. Therefore, at each given point in time, only a subset of all measurements might be available at the decoding unit. Even though the overall communication rate can be minimized by a distributed coding scheme that completely removes the redundancies between the sensor measurements, the reconstruction quality of the sources when a subset of the measurements are available might be unreasonably poor. Thus, robust distributed source coding is a generalization of the well studied multiple description (MD) coding in which all descriptions are describing a single source.
A robust distributed source coding (RDSC) scheme, on the other hand, minimizes the overall communication rate while satisfying a given bound on the maximum reconstruction error when only a subset of the measurements are available. Theoretical aspects of RDSC problem are addressed in [1, 2]. Practical DSC schemes often consist of a quantizer followed by lossless discrete source coders. The correlations between the quantized values of two sensors are used to reduce the overall rate required to communicate them to the server. This reduction is still possible even if each sensor has only access to its own readings, in compliance with Slepian-Wolf's theorem, as pursued practically in [3]. Moreover, the quantizers can also be designed to optimally exploit this correlation in conjunction with the lossless coders following them. Such schemes are reported in [2, 4–7] mainly for two sources with or without side information. Scalability of distributed coding of correlated sources is considered in [5]. Similarly, [7] demonstrates that for binary symmetric and additive white Gaussian noise channel the correlation between sources can be useful in reducing quantization distortion and protecting data when transmitted over noisy channels. Discussion on distortion performance gain and quantizer design using intersensor communication is reported in [8].
This work is a practical attempt to address the RDSC problem in the context of optimally designing scalar quantizers. We also introduce a paradigm shift in the problem formulations, which we believe is more appropriate for a distributed measurement context. The goal of conventional DSC and RDSC problems is the reconstruction of a single source X from the encoded versions of a number of noisy measurements of X (one for each sensor). We, on the other hand, assume that each sensor i is measuring a separate random variable , with 's being possibly dependent. The objective is to minimize the expected distortion of the reconstructions of all sources.
The rest of this paper is organized as follows. In Section 2, we formulate the problem and present a graph theoretic, iterative algorithm to find a locally minimal solution of the problem in Section 3. To provide a benchmark against which the designs in Section 3 can be compared, a theoretically achievable distortion-rate performance of RDSC of two jointly Gaussian sources is derived in Section 4. Experimental results are reported in Section 5 and the paper is concluded in Section 6.
2. Problem Formulation
Our overall communications model is depicted in Figure 1. In this model, the outputs of two scalar sources, , , are quantized with two quantizers and . The quantizer cell boundaries are denoted by , where represents the real line. Let and be the discrete random variables as a result of quantizing and . and are then fed into their respective discrete lossless encoders. The encoders might or might not take into account the correlations between the two symbols in order to reduce the overall communication rates. Based on this, a number of different coding cases are possible as suggested in [4], which determine the communication rates used in further optimization processes. As an example, for variable length coding of quantized source symbols, if all dependencies between the symbols in the two encoders are ignored, the communication rate would simply be , the sum of the entropy rates of marginal quantized symbols. On the other hand, if the dependencies of the symbols are fully exploited to achieve the Slepian-Wolf bound, then the communication rate would be .
Joint and marginal reconstructions.
The reconstructions of input vector source given only or are called the marginal reconstructions and are denoted by and , respectively. The reconstructed vector source using both and is denoted by . The marginal and joint distortions are therefore
where and the expectations are over joint distribution of and . The quantizer arguments in and are dropped when the dependence on the quantized symbols and is understood.
The problem can now be formulated as a feasibility problem or the problem of characterizing all the achievable tuples . A suitable optimization problem corresponding to this feasibility problem is the following:
for a given triplet , assuming the optimization problem is feasible. This formulation is identical to multiple description coding (MDC) problem when . The optimal quantizer design for MDC has been considered in [9, 10]. A similar approach to that in [9] is adopted in this paper to deal with the more general RDSC problem. This involves minimization of a Lagrangian similar to those in [4, 9] corresponding to the constrained optimization problem in (2) as given below:
where , and λ are proper Lagrangian coefficients.
3. An Iterative Design Algorithm
The joint optimization of and seems to be computationally prohibitive due to the interdependence of the conditional PDFs of each source given the quantized value of the other. We instead adopt an iterative approach in which each quantizer is optimized assuming the other one is fixed and known. Given two initial quantizers and , a series of successively improved quantizers , , , can be constructed as follows:
The iteration stops after a predetermined number of steps or when the reduction in the value of becomes negligible in consecutive iterations.
The actual minimization procedures in (4) are performed with a generalization of the approach in [9]. The optimization problem is cast into the shortest path problem as follows. First, the domain of is uniformly prequantized into N cells of interval length Δ each, N being sufficiently large. Next, we construct an acyclic directed graph G, called quantizer graph, by associating each prequantization cell boundary j, , with a node of the graph G. We introduce a directed edge from node j to node , , to represent the interval , where is the smallest value in the support of that is taken into account. By including in G the edges of all pairs of j and , G becomes a complete directed acyclic graph.
When designing the quantizer in the nth iteration with fixed as , we assign to the edge from node j to node the following weight:
where and represent the transmission loss probabilities of each of the two descriptions.
If the interval is a cell of the quantizer , then the quantity is the contribution of the cell to the Lagrangian cost function . We use the conditional distribution of given to compute the expected marginal and joint distortions and the rate. Figure 2 illustrates how to calculate the statistics of the reconstruction of for fixed and given . To be more specific, for the squared error distortion measure, the weights are explicitly calculated in the Appendix.
Given the quantized value of , the distribution of can be used to find the distortion for any quantization of .
Now, one can easily see that the problem of designing optimal for fixed is equivalent to finding the shortest path from node to node N of graph G for an appropriate λ value. Given the value of λ, this shortest path problem can be solved in time. The targeted transmission rate r can be met through a binary search on λ.
It should be noted that this procedure will only produce quantizers of connected cells. For a DSC problem where only the performance of the system with both descriptions available is important, limiting the design to those quantizers with connected cells might significantly reduce the performance [4]. On the other hand, when acceptable marginal constructions are required, each individual quantizer will also be close to an optimal quantizer marginally designed for the source, which necessarily requires quantizers of connected cells.
4. Achievable RDSC Performance for Jointly Gaussian Sources
The problem of RDSC for symmetric Gaussian sources is considered in [1, 2]. Given n sensors that observe noisy versions of a single source X, an achievable distortion-rate performance in reconstruction of X is derived when any subset of the sensors data are available. We adopt a simple theoretical formulation more suited to our framework in studying achievable distortion-rate performance. In formulations of [1, 2], the goal is the reconstruction of a specific source X when the sensors observe noisy versions of X. However, in our case, the sensors are to sense different but correlated sources. This necessitates a slight paradigm shift in our formulation. We use known results on classical multiple description source coding of Gaussian sources in [11] to derive achievable distortion-rate performances for a two-dimensional jointly Gaussian RDSC problem. This distortion-rate analysis is useful in evaluating the performance of scalar quantization results in Section 3. We report our results for a symmetric case only, although a general derivation is possible. Also, these theoretical results would only apply to the case where both sensor measurements are known to both encoders.
The region of all achievable distortion-rate pairs for MDC of a Gaussian source is known [11]. For symmetric descriptions of a Gaussian source of variance , the result of [11] can be written as
where if and is zero otherwise.
Now, consider two jointly Gaussian random variables and , each of which has zero mean and variance . The joint distribution of these two sources is completely determined by their correlation coefficient . These two sources can be written as the sum of two uncorrelated and hence independent Gaussian sources each of zero mean as and , where T and U are the DC and AC components of and . Also, , . To describe or with R bits, one can equivalently describe their DC and AC components. Applying the results in [11] to T and U independently, we arrive at the following achievable distortion-rate performance:
Note that d and are the average distortions of reconstructing both sources and . It should once again be emphasized that the coding assumes the values of both sources and are known at both encoders. A closed form achievable bound on d can be found at high rates. We derive one such bound by assuming the marginal distortions to be small but still much larger than the smallest marginal distortion promised by their distortion-rate functions. More precisely, we assume . This assumption is usually satisfied for practically meaningful designs considered in the next section. Physically, the assumption promises a small distortion given any of the descriptions alone, while allowing for a large improvement when both descriptions are available. Under this assumption, the minimum achievable joint distortion from (7) is
for any given marginal rate R and distortion d. The joint distortion in (7) can then be written as
Minimizing under the constraints and requires and . Equation (8) follows from inserting these relations back into (9).
5. Experimental Results
In this section, we investigate the performance of our proposed system and report improvements over a naive scheme. When the two descriptions are equally important, that is, , the minimization in (3) can be written in the following equivalent form:
In this new form, the parameter p has a precise meaning as follows. Assume that the two descriptions are transmitted through two independent channels. The probability that each of the two transmissions fails is p. Then, in (10) is the expected distortion of the reconstructed sources and . The relative importance of the marginal and joint distortions is reflected in parameter p. For relatively large failure rates p (poor channels), one would expect the two quantizers and to be close to an optimal quantizer for a single Gaussian. The quantizer cells of and are however somewhat interleaved to reduce the distortion when both quantization outputs are available at the decoder. An example of a pair of optimized quantizers for , , and is shown in Figure 3. While other choices are possible as discussed in Section 3, to calculate the overall communication rate throughout this section, the Slepian-Wolf bound is assumed to be achieved; that is, we assume .
Quantizers optimally interleaved to reduce the distortion when both quantized values are present, while each of them is close to an optimal quantizer for a single Gaussian source.
The joint and marginal distortions for , , and different marginal rates are depicted in Figure 4. The distortion of the optimal quantizer for a single Gaussian source and the same rate, obtained from the Lloyd-Max algorithm, is also shown in the same figure for comparison. Evidently, the difference between the two is extremely small. In other words, the penalty in the performance of marginal reconstruction is negligible, when jointly designing the quantizers for two sources versus designing the quantizers for each source independently. The distortion-rate performance of the proposed distributed coding scheme, consisting of scalar quantizers and entropy coders, is compared against theoretically achievable performances given by (8). The comparison is carried out in the following manner. For fixed p and ρ, the quantizers are optimally designed for a target marginal distortion. This design results in a marginal bit rate and a joint distortion, which is then plotted as a function of this marginal rate and is compared to what is theoretically achievable for the same marginal rate and distortion.
The joint and marginal distortions as a function of marginal rate R.
Figure 4 also demonstrates the joint distortion achieved with a naive quantization scheme. This quantization scheme is represented as “Lloyd-joint” and is based on independently designing an optimal quantizer for and having the same number of cells as and using the Lloyd-Max algorithm. Our proposed method exploits the correlation between the sources to improve the expected distortion. As is evident from the figure, the improvements at low rates are remarkable.
6. Concluding Remarks
This paper considered the practice of RDSC in the context of optimally designed scalar quantizers. A practically convenient coding scheme, consisting of a pair of quantizers and discrete lossless coders to robustly encode two correlated sensor measurements, is proposed in this paper. When the particular form of the lossless coders is chosen, we devised an optimization algorithm for designing the quantizers. It is then shown that a proper choice of the quantizers can significantly increase the overall system performance.
Sensors are assumed to be measuring some random field; therefore, aside from temporal correlations between the measurements of a sensor, a great number of correlations can exist between the measurements of nearby sensors too. Such correlations can be exploited to reduce the amount of communication required to report the sensor measurements to a server.
Footnotes
Appendix
Competing Interests
The authors declare that they have no competing interests.
References
1.
ChenJ.BergerT.Robust distributed source codingIEEE Transactions on Information Theory20085483385339810.1109/tit.2008.926389MR24510022-s2.0-48849097837
2.
IshwarP.PuriR.PradhanS. S.RamchandranK.On compression for robust estimation in sensor networksProceedings of the IEEE International Symposium on Information TheoryJune-July 200319310.1109/ISIT.2003.1228207
3.
AaronA.GirodB.Compression with side information using turbo codesProceedings of the Data Compression Conference (DCC '02)2002Snowbird, Utah, USA25226110.1109/DCC.2002.999963
4.
Rebollo-MonederoD.Quantization and transforms for distributed source coding [Ph.D. thesis]2007Stanford University
5.
SaxenaA.RoseK.On scalable distributed coding of correlated sourcesIEEE Transactions on Signal Processing20105852875288310.1109/TSP.2010.2042488MR27894302-s2.0-77951189318
6.
WagnerA. B.KellyB. G.AltugY.Distributed rate-distortion with common componentsIEEE Transactions on Information Theory20115774035405710.1109/tit.2011.2145570MR28404412-s2.0-79959539582
7.
WernerssonN.KarlssonJ.SkoglundM.Distributed quantization over noisy channelsIEEE Transactions on Communications20095761693170010.1109/TCOMM.2009.06.0704822-s2.0-67650675055
8.
SunJ. Z.GoyalV. K.Intersensor collaboration in distributed quantization networksIEEE Transactions on Communications20136193931394210.1109/TCOMM.2013.071813.1208332-s2.0-84884901817
9.
DumitrescuS.WuX.Lagrangian optimization of two-description scalar quantizersIEEE Transaction on Information Theory200753113990401210.1109/tit.2007.907498MR24465512-s2.0-36348940577
10.
VaishampayanV. A.Design of multiple description scalar quantizersIEEE Transactions on Information Theory199339382183410.1109/18.256491ZBL0784.940092-s2.0-0027597173
11.
OzarowL. H.On a source-coding problem with two channels and three receiversThe Bell System Technical Journal198059101909192110.1002/j.1538-7305.1980.tb03344.xMR596429