Step 1

advertisement
Consensus-Based Distributed
Least-Mean Square Algorithm
Using Wireless Ad Hoc Networks
Gonzalo Mateos, Ioannis Schizas and Georgios B. Giannakis
ECE Department, University of Minnesota
Acknowledgment: ARL/CTA grant no. DAAD19-01-2-0011
1
Motivation
 Estimation using ad hoc WSNs raises exciting challenges
 Communication constraints
Single-hop communications
 Limited power budget
 Lack of hierarchy / decentralized processing
Consensus
 Unique features
 Environment is constantly changing (e.g., WSN topology)
 Lack of statistical information at sensor-level
 Bottom line: algorithms are required to be
 Resource efficient
 Simple and flexible
 Adaptive and robust to changes
2
Prior Works
 Single-shot distributed estimation algorithms
 Consensus averaging [Xiao-Boyd ’05, Tsitsiklis-Bertsekas ’86, ’97]
 Incremental strategies [Rabbat-Nowak etal ’05]
 Deterministic and random parameter estimation [Schizas etal ’06]
 Consensus-based Kalman tracking using ad hoc WSNs
 MSE optimal filtering and smoothing [Schizas etal ’07]
 Suboptimal approaches [Olfati-Saber ’05], [Spanos etal ’05]
 Distributed adaptive estimation and filtering
 LMS and RLS learning rules [Lopes-Sayed ’06 ’07]
3
Problem Statement
 Ad hoc WSN with
sensors
 Single-hop communications only. Sensor ‘s neighborhood
 Connectivity information captured in
 Zero-mean additive (e.g., Rx, quantization) noise
 Each sensor
, at time instant
 Acquires a regressor
and scalar observation
 Both zero-mean w.l.o.g and spatially uncorrelated
 Least-mean squares (LMS) estimation problem of interest
4
Centralized Approaches
 If
,
jointly stationary
Wiener solution
 If global (cross-) covariance matrices
,
available
Steepest-descent converges avoiding matrix inversion
 If (cross-) covariance info. not available or time-varying
Low complexity suggests (C-) LMS adaptation
Goal: develop a distributed (D-) LMS algorithm for ad hoc WSNs
5
A Useful Reformulation

Introduce the bridge sensor subset
1)
2)

For all sensors
,
For
, there must
such that
such that
Consider the convex, constrained optimization
Proposition [Schizas etal’06]: For
WSN is connected, then
satisfying 1)-2) and the
6
Algorithm Construction

Problem of interest

Two key steps in deriving D-LMS
1)
Resort to the alternating-direction method of multipliers
Gain desired degree of parallelization
2)
Apply stochastic approximation ideas
Cope with unavailability of statistical information
7
Derivation of Recursions

Associated augmented Lagrangian

Alternating-direction method of Lagrange multipliers
Three-step iterative update process
Step 1: Multipliers
Step 2: Local estimates
Step 3: Bridge variables
Dual iteration
Minimize
Minimize
w.r.t.
w.r.t.
8
Multiplier Updates

Recall the constraints

Use standard method of multipliers type of update

Requires
from the bridge neighborhood
9
Local Estimate Updates

Given by the local optimization

First order optimality condition

Proposed recursion inspired by Robbins-Monro algorithm
1)
2)

is the local prior error
is a constant step-size
Requires


Already acquired bridge variables
Updated local multipliers
10
Bridge Variable Updates

Similarly,

Requires


from the neighborhood
from the neighborhood in a startup phase
11
D-LMS Recap and Operation

In the presence of communication noise, for
Step 1:
Step 2:
Step 3:
Steps 1,2:
Rx
from
Sensor

Step 3:
Tx
Rx
to
from
Tx
to
Bridge sensor
Simple, fully distributed, only single-hop exchanges needed
12
Further Insights
 Manipulating the recursions for
and
yields
 Introduce the instantaneous consensus error at sensor
 The update of
becomes
 Superposition of two learning mechanisms
 Purely local LMS-type of adaptation
 PI consesus loop
tracks the consensus set-point
13
D-LMS Processor
Sensor j
Local LMS
Algorithm
Consensus Loop
PI Regulator
To
 Network-wide information enters through the set-point
 Expect increased performance with
Flexibility
14
Mean Analysis

Independence setting
(As1)
signal assumptions for
is a zero-mean white random vector
, with spectral radius
(As2) Observations obey a linear model
where
is a zero-mean white noise
(As3)

Define
and
are statistically independent
and
Goal: derive sufficient conditions under which
15
Dynamics of the Mean
Lemma: Under (As1)-(As3), consider the D-LMS algorithm
initialized with
Then for
,
.
is given by the second-order recursion
with
and
, where
 Equivalent first-order system by state concatenation
16
First-Order Stability Result
Proposition: Under (As1)-(As3), the D-LMS algorithm whose
positive step-sizes
such that
sense i.e.,
and relevant parameters are chosen
, achieves consensus in the mean
 Step-size selection based on local information only
 Local regressor statistics
 Bridge neighborhood size
17
Simulations
node WSN,
Regressors:
Observations:
D-LMS:
True time-varying weight:
i.i.d.
,
18
Loop Tuning
 Adequately selecting
actually does make a difference
 Compared figures of merit:
 MSE (Learning curve):
 MSD (Normalized estimation error):
19
Concluding Summary
 Developed a distributed LMS algorithm for general ad hoc WSNs
 Intuitive sensor-level processing
 Local LMS adaptation
 Tunable PI loop driving local estimate to consensus
 Mean analysis under independence assumptions
step-size selection rules based on local information
 Simulations validate mss convergence and tracking capabilities
 Ongoing research
 Stability and performance analysis under general settings
 Optimality: selection of bridge sensors,
 D-RLS. Estimation/Learning performance Vs complexity tradeoff
20
Download