Asymptotic Noise Analysis of High Dimensional
Consensus
Usman A. Khan, Soummya Kar and José M. F. Moura
Department of Electrical and Computer Engineering
Carnegie Mellon University, Pittsburgh, PA 15213 USA
Email: ukhan@ece.cmu.edu, soummyak@andrew.cmu.edu, moura@ece.cmu.edu
Abstract—The paper studies the effect of noise on the
asymptotic properties of high dimensional consensus (HDC).
HDC offers a unified framework to study a broad class of
distributed algorithms with applications to average consensus,
leader-follower dynamics in multi-agent networks and distributed
sensor localization. We show that under a broad range of perturbations, including inter-sensor communication noise, random
data packet dropouts and algorithmic parameter uncertainty, a
modified version of the HDC converges almost surely (a.s.) We
characterize the asymptotic mean squared error (m.s.e.) from
the desired agreement state of the sensors (which, in general,
vary from sensor to sensor) and show broad conditions on
the noise leading to zero asymptotic m.s.e. The convergence
proof of the modified HDC algorithm is based on stochastic
approximation arguments and offers a general framework to
study the convergence properties of distributed algorithms in the
presence of noise.
Index Terms—High Dimensional Consensus, Random Link
Failures, Communication Noise, Almost Sure Convergence,
Stochastic Approximation.
I. I NTRODUCTION
Distributed signal processing is an active research area
and offers a scalable, iterative and low complexity alternative to centralized data processing. Extensive research efforts
fostered the development of basic network utility schemes,
including distributed average consensus, distributed parameter
estimation and inference algorithms, distributed sensor network localization, distributed compressed sensing, just to name
a few. To envision extensive real time implementation and
deployment of such networks and schemes, it is of importance
to analyze and design robust versions of these algorithms.
Indeed, such networks operate under scarce resources like
limited bandwidth and power, unpredictable environments and
random operating conditions. In [1],[2] a generic class of
algorithms, called HDC (High Dimensional Consensus), was
presented which unified a large class of distributed network
algorithms and offered a common framework for their analysis
and design. In this paper, we continue the study of HDC, with a
focus on their robustness properties. Specifically, we consider
a modified version of the HDC, called M-HDC, designed to
This work was partially supported by the DARPA DSO Advanced Computing and Mathematics Program Integrated Sensing and Processing (ISP)
Initiative under ARO grant # DAAD 19-02-1-0180, by NSF under grants
# ECS-0225449 and # CNS-0428404, by the Office of Naval Research under
MURI N000140710747, and by an IBM Faculty Award.
978-1-4244-5827-1/09/$26.00 ©2009 IEEE
191
cope with the effects of environmental randomness on the
operation of the HDC. Under a broad range of random perturbations, including inter-sensor communication noise, random
data packet dropouts and uncertainty in model matrices (to
be explained later) we prove a.s. convergence of the M-HDC
algorithm and explicitly characterize the deviation from the
desired equilibrium state in terms of the residual m.s.e. The
M-HDC is a stochastic approximation based algorithm and
employs a temporally decreasing weight sequence for the
iterative scheme. As explained in the paper, the adaptively
chosen weight sequence controls the effect of environmental
noise leading to robust behavior. On the other hand, the incorporation of robustness in this way incurs a slower convergence
rate in comparison to the non-robust HDC version, which
converges at a geometric rate in the absence of noise.
We comment briefly on the organization of the rest of the
paper. Section II reviews prior work on HDC, whereas the
robust version M-HDC together with the noise assumptions
are presented in Section III. The main convergence results are
presented in Section IV and Section V discusses several applications of the robust M-HDC. Finally Section VI concludes
the paper.
II. P RIOR WORK
Consider a network of N sensor nodes communicating over
a directed graph, G = (Θ, A). The N network agents are
divided into two classes, namely anchors and sensors with
κ denoting the set of anchors and Ω the sensors, such that,
Θ = κ ∪ Ω. Let uk ∈ R1×m be the state associated to the kth
anchor, and let xl ∈ R1×m be the state associated to the lth
sensor. We are interested in studying linear iterative algorithms
of the form
uk (t + 1)
=
xl (t + 1)
=
uk (t),
j∈KΩ (l)∪{l}
k ∈ κ,
plj xj (t) +
(1)
blk uk (0),(2)
k∈Kκ (l)
for l ∈ Ω, where: t ≥ 0 is the discrete-time iteration index; and
plj ’s and blk ’s are the state updating coefficients. We assume
that the updating coefficients are constant over the components
of the m-dimensional state. The study of such distributed
iterated algorithms was initiated in [2], [3], [4], where we
introduced the term Higher Dimensional Consensus (HDC)
to describe such algorithms (see [4] for a justification of the
nomenclature.) In particular, we note that eqns. 1,2 consider
Asilomar 2009
the HDC algorithm under no randomness in communication
and weight computation as treated in [4]. In this section we
review some results from [4] on the properties of HDC in
a noise-free environment and in later sections consider the
effect of random perturbations on the steady state behavior of
the same.
For the purpose of analysis, we write the HDC (1)–(2) in
matrix form. Define
T
U(t) = uT1 (t), . . . , uTK (t) ,
(3)
T
T
T
(4)
X(t) = xK+1 (t), . . . , xN (t) ,
P = {plj } ∈ RM ×M ,
B = {blk } ∈ RM ×K .
(5)
M ×m
K×m
Note that U(t) ∈ R
and X(t) ∈ R
. With the above
notation, we write (1)–(2) concisely as
U(t + 1)
I 0
U(t)
=
,
(6)
X(t + 1)
B P
X(t)
C(t + 1)
=
ΥC(t).
(7)
Υ
Note that the graph, G , associated to the N × N iteration
matrix, Υ, must be a subgraph of G. In other words, the
sparsity of Υ is dictated by the sparsity of the underlying
sensor network. In the iteration matrix, Υ: its submatrix, P,
collects the updating coefficients of the M sensors with respect
to the M sensors; and its submatrix, B, collects the updating
coefficients of the M sensors with respect to the K anchors.
From (6), the matrix form of the HDC in (2) is
X(t + 1)
=
PX(t) + BU(0),
t ≥ 0.
A. No anchors: B = 0
In this case, the HDC reduces to
X(t + 1)
As discussed in Section III, the HDC algorithm is implemented as (1)–(2), and its matrix representation is given
by (7). The study of the forward HDC problem (referred to
as the HDC for brevity in the sequel) can be divided into the
following two cases: (A) no anchors; and (B) with anchors.
We briefly review these two cases separately, their applications
being considered in Section V.
192
PX(t),
=
Pt+1 X(0).
(9)
An important problem covered by this case is averageconsensus. As well known, when
ρ(P) = 1,
(10)
and under some minimal assumptions on P and on the network
connectivity, (9) converges to the average of the initial sensors’
states. For more precise and general statements in this regard,
see for instance, [5], [6]. Average-consensus, thus, is a special
case of the HDC, when B = 0 and ρ(P) = 1. This problem
has been studied in great detail, a detailed set of references is
provided in [2].
The rest of this paper deals entirely with the case ρ(P) < 1
and the term HDC subsumes the ρ(P) < 1 case, unless explicitly noted. Note that, when B = 0, the HDC (with ρ(P) < 1)
leads to X∞ = 0, which is not interesting.
B. With anchors: B = 0
This extends the average-consensus to “higher dimensions”
(as explained in [4].) Lemma 1 establishes: (i) the conditions
under which the HDC converges; (ii) the limiting state of the
network; and (iii) the rate of convergence of the HDC.
Lemma 1 ([4]) Let B = 0 and U(0) ∈
/ N (B), where N (B)
is the null space of B. If
(8)
As with all distributed algorithms of interest, there are two
objectives behind the study of HDC algorithms. First, given
the sensor network and the weight matrices, P and B, it is
of interest to determine the asymptotic properties of the HDC
iterates, i.e., the limiting sensor states, provided they converge.
The second problem is the inverse or learning problem, where
it is desired to design the weight matrices P and B respecting
the sparsity of the given communication network, such that the
sensors converge to some desired state. In [4] we consider the
learning problem in detail, which leads to a multi-objective
design criteria exhibiting the trade-off between convergence
rate and steady state m.s.e. The current paper deals solely
with the forward or analysis problem, where we assume that
appropriate weight matrices P and B are given and the
objective is to analyze the asymptotic properties of the HDC
under random perturbations.
=
ρ(P) < 1,
(11)
then the limiting state of the sensors, X∞ , is given by
X∞ lim X(t + 1) = (I − P)
t→∞
−1
BU(0),
(12)
and the error, E(t) = X(t) − X∞ , decays exponentially to 0
with exponent ln(ρ(P)), i.e.,
1
lim sup lnE(t) ≤ ln(ρ(P)).
t→∞ t
(13)
The above lemma shows that the limit state of the sensors, X∞ , is independent of the sensors’ initial conditions
and is given by (12), for any X(0) ∈ RM ×m . It is also
straightforward to show that if ρ(P) ≥ 1, then the HDC
algorithm (8) diverges for all U(0) ∈
/ N (B), where N (B)
is the null space of B. Clearly, the case U(0) ∈ N (B) is not
interesting as it leads to X∞ = 0.
The following section considers a modified version of the
HDC algorithm presented in [4] to cope with the effect of environmental randomness including inter-sensor communication
noise, channel effects and incorrect weight design.
III. P ROBLEM FORMULATION
In this section we present M-HDC, a modified version of the
HDC algorithm to cope with the effect of noise. Before stating
the assumptions on the random operating conditions formally,
we consider a relaxed version of the HDC, eqns. 1,2:
Relaxed-HDC: In vector form, the relaxed HDC iterations
take the form:
U(t + 1) = U(0),
t≥0
(14)
X(t + 1) = (1 − α)X(t) + α [PX(t) + BU(0)] , t ≥ 0 (15)
where the relaxation parameter 0 ≤ α ≤ 1 is a design
constant. Under appropriate choice of α it can be shown
that the convergence properties of the (unrelaxed) HDC are
retained. Since the iterations above can be separated over the
columns of U(t) and X(t), we an alternative representation
of the HDC takes the form:
j
j
U (t + 1) = U (0),
t≥0
(16)
j
j
j
X (t + 1) = (1 − α)X (t) + α PX (t) + BU (0) , t ≥ 0
(17)
Here, Uj (t), Xj (t), denote the j-th column of the state
matrices, U(t), X(t) respectively.
Thus, to implement the sequence of iterations in (17)
perfectly, the n-th sensor at iteration t needs the corresponding
rows of the matrices P and B, and, in addition, the current
states, Ulj (t), Xlj (t), l ∈ Θn (j-th component of l-th sensor
coordinates), of its neighbors. The computation of the matrices
P and B may require online sensing or measurement (for
example, in the case of distributed sensor localization [1], the
weight matrices P, B consist of local barycentric coordinates
computed from inter-sensor distance measurements, which
are susceptible to random ranging errors) and hence can
be estimated to a limited degree of accuracy only. Also,
because of imperfect communication, each sensor receives
only noisy versions of its neighbors current state. Hence,
in a random environment, the iteration sequence in (17)
needs to be modified accordingly, because, each sensor has
only partial imperfect information about the system. In the
following, we identify the sources of randomness, which may
be introduced in (17), because of imperfect communication
and noisy distance computation, and state them formally as
assumptions. Also, note that the choice of α = 1 reduces
eqn. (15) to the update eqn. (8), i.e., the unrelaxed version of
the HDC as considered in Section II. The key point to note
here, that if α < 1, unlike the unrelaxed case, the relaxed
HDC attaches a non-trivial weight (constant over time) to
the previous sensor states. As will be formulated later, this
relaxation is an important step to make HDC resilient to noisy
perturbations, where the relaxation parameter α is chosen to
be a time-varying sequence, so that the cumulative effect of
noise is controlled as the algorithm progresses.
We present the key technical assumptions on the different
types of random perturbations on the HDC before presenting
the M-HDC algorithm.
(C1) Randomness in system matrices: At each iteration,
each sensor needs the corresponding row of the system
matrices B and P, which can be, possibly, random due to the
effect of imperfect online sensing. We assume that at each
j
193
n (t) and
iteration, the n-th sensor can only get estimates, B
Pn (t), of the corresponding rows of the B and P matrices
respectively. In the generic imperfect communication case,
we have
B (t)
B(t)
= B + SB + S
(18)
B (t)}t≥0 is an independent sequence of random
where, {S
matrices with,
B (t)] = 0, ∀t,
E[S
B (t)2 = kB < ∞. (19)
sup E S
t≥0
Here, SB is the mean measurement error. Similarly, for the
case of P, we have
P (t),
P(t)
= P + SP + S
(20)
P (t)}t≥0 is an independent sequence of random
where {S
matrices with,
P (t)] = 0, ∀t,
E[S
P (t)2 = kP < ∞. (21)
sup E S
t≥0
Note that this way of writing B(t),
P(t) does not require
the noise model to be additive. It only says that any random
object may be written as the sum of a deterministic mean part
and the corresponding zero mean random part. The moment
assumptions in eqns. (19,21) are very weak and in particular,
is satisfied if the sequences {B(t)}
t≥0 and {P(t)}t≥0 are
i.i.d.
(C2) Random Link Failure: We assume that the intersensor communication links fail randomly. This happens,
for example, in wireless sensor network applications, where
occasionally data packets are dropped. To this end, if the
sensors n and l share a communication link (or, l ∈ Θn ), we
assume that the link fails with some probability 1 − qnl at
each iteration, where 0 < qnl ≤ 1. We associate with each
such potential network link, a binary random variable, enl (t),
where enl (t) = 1 indicates that the corresponding network
link is active at time t, whereas enl (t) = 0 indicates a link
failure.
(C3) Additive Channel Noise: Define the family of indepenj
dent zero mean random variables, {vnl
(t)}n,l,j,t , such that
j
sup E[vnl
(t)]2 = kv < ∞.
(22)
n,l,j,t
We assume that at the t-th iteration, if the network link (n, l)
j
is active, sensor n receives only a corrupt version, ynl
(t), of
j
sensor l’s state, cl (t), given by
j
j
ynl
(t) = cjl (t) + vnl
(t).
(23)
This models the channel noise. It is to be noted, that the
moment assumption in eqn. (22) is very weak and holds, in
particular, if the channel noise is i.i.d.
(C4) Independence: We assume that the sequences,
B (t), S
P (t)}t≥0 , {enl (t)}n,l,t , and {v j (t)}n,l,j,t are mu{S
nl
tually independent. Note that, in the above, we do not put
restrictions on the distributional form of the random errors,
but, assume they obey some weak moment conditions.
Clearly, under the random environment model (as detailed in
Assumptions (C1)-(C4), the sensors cannot update their states
according to the iterations, given in (17). We now consider the
following state update recursion for the random environment
case:
Algorithm M-HDC:
xjn (t
+ 1)
=
(1 −
α (t)) xjn (t)
j
ujl + vnl
(t)
xjl (t)
+
j
vnl
(t)
+ α(t)
+ α(t)
l∈κ∩Θn
l∈Ω∩Θn
nl (t)
enl (t)B
qnl
nl (t) = P
nl (t)
P
(24)
enl (t)
−1
qnl
(25)
enl (t)
−1 .
qnl
(26)
Clearly, by (C4), the matrices B(t)
∈ RM ×(m+1) and P(t)
∈
M ×(m+1)
R
are zero mean. Also, by the bounded moment
assumptions in (C1), we have
2
sup E B(t)
=
kB < ∞,
t≥0
2
=
kP < ∞.
sup E P(t)
t≥0
(27)
Hence, the iterations in (24) can be written in vector form as
xj (t + 1)
=
(1 − α(t)) xj (t) + α(t) P(t)
+ P(t) xj (t)
+ B(t)
+ B(t) uj + η j (t) ,
(28)
where, the nth element of the vector, η j (t), is given by
nl (t) + P
nl (t) v j (t)
ηnj (t) =
P
nl
l=n
+
l=n
nl (t) v j (t).
nl (t) + B
B
nl
α2 (t) < ∞.
(32)
t≥0
This condition, commonly assumed in the adaptive control
and adaptive signal processing literature, assumes that the
weights decay to zero, but not too fast.
(D2) Low Error Bias: We assume the small bias condition:
(33)
in addition to ρ(P) < 1. We note that this condition ensures
that the matrix (I − P − SP ) is invertible.
We refer to the algorithm in (24) as M-HDC (Distributed
Localization in Random Environment) in the sequel.
To write the algorithm M-HDC in a compact form, we
introduce some notation. Define the random matrices, B(t)
∈
M ×(m+1)
M ×(m+1)
R
and P(t) ∈ R
, as
nl (t) = B
nl (t)
B
t≥0
ρ (P + SP ) < 1.
nl (t)
enl (t)P
qnl
, n ∈ Ω, 1 ≤ j ≤ m
(D1) Persistence Condition:
α(t) > 0,
α(t) = ∞,
(29)
By (C1)-(C4), the sequence, {η j (t)}t≥0 , is zero mean independent with
sup E η j (t)2 = kη < ∞.
(30)
t
From (C1), the iteration sequence in (28) can be written as
xj (t + 1) = xj (t) − α(t) (I − P − SP ) xj (t)
P (t) + P(t)
xj (t)
− (B + SB ) uj − S
B (t) + B(t)
uj − η j (t) .
(31)
− S
We now make two additional design assumptions on the
algorithm:
194
In the following section, we study convergence properties
of the M-HDC algorithm, under the assumptions (C1)-(C4),
(D1)-(D2).
IV. C ONVERGENCE AND RELATED RESULTS
We have the following main result regarding the limiting
state of the M-HDC algorithm:
Theorem 2 Let {X(t)}t≥0 , 1 ≤ j ≤ m, be the state sequence
generated by the iterations, given by (24), under the assumptions (C1)-(C4), (D1)-(D2). Then,
P lim X(t) = (I − P − SP )
t→∞
−1
(B + SB ) U(0), ∀j = 1.
(34)
The proof relies on stochastic approximation based arguments
([7]) and follows similar reasoning as adopted in [8],[1] for
proving convergence of distributed network algorithms and is
omitted. We discuss the consequences of Theorem 2.
Remark 3 Theorem 2 shows that the M-HDC converges a.s.
to a deterministic state under the broad noise assumptions discussed above. However, in the steady state there is a residual
error from the desired convergence state (I − P)−1 BU(0).
This is due to the the presence of non-zero error bias terms
SP , SB in the estimation of the weight matrices P and B
respectively. In particular, if the error biases SP , SB are
zero, then irrespective of the randomness in communication,
packet dropouts, the M-HDC converges a.s. to the desired
state (I − P)−1 BU(0) where the HDC (noise free environment) converges. However, it is important to note here, that
the decreasing weight sequence α(t) lowers the convergence
rate of the M-HDC algorithm as compared to the geometric convergence rate attained by the HDC. In other words,
robustness can be achieved by sacrificing convergence rate.
Typical stochastic approximation based arguments [9] suggest
a
that a weight sequence {α(t)} of the form α(t) = t+1
(satisfying the persistence condition) leads to a convergence
rate of the order of √1t , i.e., the transient error X(t) −
−1
(I − P − SP ) (B + SB ) U(0) decreases at a rate of √1t as
t → ∞.
V. A PPLICATIONS
We briefly comment on applications of HDC where robustness to random environmental perturbations is desired.
A very important and canonical application is robust distributed average consensus where random inter-sensor data
packet dropouts, quantized transmission, additive channel
noise require the designing of robust consensus algorithms.
An extensive treatment in this line can be found in [8],[10].
As mentioned earlier, another application which, in addition
requires robustness to random weight matrices, comes from
distributed sensor localization [1], where the matrices P and B
are local inter-sensor barycentric coordinates computed from
incorrect distance estimates.
VI. C ONCLUSIONS
The paper presents a modified version of the HDC algorithm, M-HDC, which adds robustness to HDC in the
presence of random environmental perturbations. We show
that the M-HDC converges a.s. to a deterministic state and
characterizes the residual m.s.e. from the desired convergence
state. In particular, the residual m.s.e. is zero if the error biases
resulting from computation of the weight matrices P and B
are zero. Finally, we demonstrate the usefulness of the MHDC algorithm as a unified framework for studying several
important distributed applications in sensor networks including
average consensus and distributed sensor localizations.
R EFERENCES
[1] U. A. Khan, S. Kar, and J. M. F. Moura, “Distributed sensor localization
in random environments using minimal number of anchor nodes,” IEEE
Transactions on Signal Processing, vol. 57, no. 5, May. 2009, also, in
arXiv: http://arxiv.org/abs/0802.3563.
[2] U. Khan, S. Kar, and J. M. F. Moura, “Distributed algorithms in sensor
networks,” in Handbook on Sensor and Array Processing, S. Haykin
and K. J. R. Liu, Eds. New York, NY: Wiley-Interscience, 2009, to
appear, 33 pages.
[3] U. A. Khan, S. Kar, and J. M. F. Moura, “Higher dimensional consensus
algorithms in sensor networks,” in IEEE 34th International Conference
on Acoustics, Speech, and Signal Processing, Taipei, Taiwan, Apr. 2009.
[4] U. Khan, S. Kar, and J. M. F. Moura, “Higher dimensional consensus:
learning in large-scale networks,” April 2009, accepted for publication
in the IEEE Transactions on Signal Processing. [Online]. Available:
http://arxiv.org/abs/0904.1840
[5] L. Xiao and S. Boyd, “Fast linear iterations for distributed averaging,”
Systems and Controls Letters, vol. 53, no. 1, pp. 65–78, Apr. 2004.
[6] R. Olfati-Saber, J. A. Fax, and R. M. Murray, “Consensus and cooperation in networked multi-agent systems,” Proceedings of the IEEE,
vol. 95, no. 1, pp. 215–233, Jan. 2007.
[7] M. B. Nevelson and R. Z. Hasminskii, Stochastic Approximation and
Recursive Estimation. Providence, Rhode Island: American Mathematical Society, 1973.
[8] S. Kar and J. M. F. Moura, “Distributed consensus algorithms in
sensor networks: Quantized data,” November 2007, submitted for
publication, 30 pages, see http://arxiv.org/abs/0712.1609. [Online].
Available: http://arxiv.org/abs/0712.1609
[9] S. Kar, J. M. F. Moura, and K. Ramanan, “Distributed
parameter estimation in sensor networks: Nonlinear observation
models and imperfect communication,” Aug. 2008, submitted for
publication, see also http://arxiv.org/abs/0809.0009. [Online]. Available:
http://arxiv.org/abs/0809.0009
[10] S. Kar and J. M. F. Moura, “Distributed consensus algorithms in sensor
networks: Link failures and channel noise,” IEEE Transactions on
Signal Processing, 2008, accepted for publication, 30 pages. [Online].
Available: http://arxiv.org/abs/0711.3915
195