ESTIMATION-BASED ADAPTIVE FILTERING AND CONTROL

advertisement
ESTIMATION-BASED ADAPTIVE
FILTERING AND CONTROL
a dissertation
submitted to the department of electrical engineering
and the committee on graduate studies
of stanford university
in partial fulfillment of the requirements
for the degree of
doctor of philosophy
Bijan Sayyar-Rodsari
July 1999
c Copyright by Bijan Sayyar-Rodsari 1999
All Rights Reserved
ii
I certify that I have read this dissertation and that in my opinion it is fully
adequate, in scope and quality, as a dissertation for the degree of Doctor of
Philosophy.
Professor Jonathan How
(Principal Adviser)
I certify that I have read this dissertation and that in my opinion it is fully
adequate, in scope and quality, as a dissertation for the degree of Doctor of
Philosophy.
Professor Thomas Kailath
I certify that I have read this dissertation and that in my opinion it is fully
adequate, in scope and quality, as a dissertation for the degree of Doctor of
Philosophy.
Dr. Babak Hassibi
I certify that I have read this dissertation and that in my opinion it is fully
adequate, in scope and quality, as a dissertation for the degree of Doctor of
Philosophy.
Professor Carlo Tomasi
Approved for the University Committee on Graduate Studies:
iii
Abstract
Adaptive systems have been used in a wide range of applications for almost four
decades. Examples include adaptive equalization, adaptive noise-cancellation, adaptive vibration isolation, adaptive system identification, and adaptive beam-forming.
It is generally known that the design of an adaptive filter (controller) is a difficult nonlinear problem for which good systematic synthesis procedures are still lacking.
Most existing design methods (e.g. FxLMS, Normalized-FxLMS, and FuLMS) are adhoc in nature and do not provide a guaranteed performance level. Systematic analysis
of the existing adaptive algorithms is also found to be difficult. In most cases, addressing even the fundamental question of stability requires simplifying assumptions
(such as slow adaptation, or the negligible contribution of the nonlinear/time-varying
components of signals) which at the very least limit the scope of the analysis to the
particular problem at hand.
This thesis presents a new estimation-based synthesis and analysis procedure for
adaptive “Filtered” LMS problems. This new approach formulates the adaptive filtering (control) problem as an H∞ estimation problem, and updates the adaptive weight
vector according to the state estimates provided by an H∞ estimator. This estimator
is proved to be always feasible. Furthermore, the special structure of the problem
is used to reduce the usual Riccati recursion for state estimate update to a simpler
Lyapunov recursion. The new adaptive algorithm (referred to as estimation-based
adaptive filtering (EBAF) algorithm) has provable performance, follows a simple update rule, and unlike previous methods readily extends to multi-channel systems
and problems with feedback contamination. A clear connection between the limiting behavior of the EBAF algorithm and the classical FxLMS (Normalized-FxLMS)
iv
algorithm is also established in this thesis.
Applications of the proposed adaptive design method are demonstrated in an Active Noise Cancellation (ANC) context. First, experimental results are presented for
narrow-band and broad-band noise cancellation in a one-dimensional acoustic duct.
In comparison to other conventional adaptive noise-cancellation methods (FxLMS
in the FIR case and FuLMS in the IIR case), the proposed method shows much
faster convergence and improved steady-state performance. Moreover, the proposed
method is shown to be robust to feedback contamination while conventional methods
can go unstable. As a second application, the proposed adaptive method was used
for vibration isolation in a 3-input/3-output Vibration Isolation Platform. Simulation results demonstrate improved performance over a multi-channel implementation
of the FxLMS algorithm. These results indicate that the approach works well in
practice. Furthermore, the theoretical results in this thesis are quite general and can
be applied to many other applications including adaptive equalization and adaptive
identification.
v
Acknowledgements
This thesis has greatly benefited from the efforts and support of many people whom
I would like to thank. First, I would like to thank my principle advisor Professor
Jonathan How. This research would not have been possible without Professor How’s
insights, enthusiasm and constant support throughout the project. I appreciate his
attention to detail and the clarity that he brought to our presentations and writings.
I would also like to acknowledge the help and support of Dr. Alain Carrier from Lockheed Martin’s Advanced Technology Center. His careful reading of all the manuscripts
and reports, his provocative questions, and his dedication to meaningful research has
greatly influenced this work. I would like to gratefully acknowledge members of my
defense and reading committee, Professor Thomas Kailath, Professor Carlo Tomasi,
and Dr. Babak Hassibi. It was from a class instructed by Professor Kailath and Dr.
Hassibi that the main concept of this thesis originated, and it was their research that
this thesis is based on. It is impossible to exaggerate the importance of Dr. Hassibi’s
contributions to this thesis. He has been a great friend and advisor throughout this
work for which I am truly thankful.
My thanks also goes to Professor Robert Cannon and Professor Steve Rock for giving me the opportunity to interact with wonderful friends in the Aerospace Robotics
Laboratory. The help from ARL graduates, Gordon Hunt, Steve Ims, Stef Sonck,
Howard Wang, and Kurt Zimmerman was crucial in the early stages of the research
at Lockheed. I have also benefited from interesting discussions with fellow ARL students Andreas Huster, Kortney Leabourne, Andrew Robertson, Heidi Schubert, and
Bruce Woodley, on both technical and non-technical issues. I am forever thankful for
their invaluable friendship and support. I also acknowledge the camaraderie of more
vi
recent ARL members, Tobe Corazzini, Steve Fleischer, Eric Frew, Gokhan Inalhan,
Hank Jones, Bob Kindel, Ed LeMaster, Mel Ni, Eric Prigge, and Luis Rodrigues.
I discussed all aspects of this thesis in great detail with Arash Hassibi. He helped
me more than I can thank him for. Lin Xiao and Hong S. Bae set up the hardware for
noise cancellation and helped me in all experiments. I appreciate all their assistance.
Thomas Pare, Haitham Hindi, and Miguel Lobo provided helpful comments about the
research. I also acknowledge the assistance from fellow ISL students, Alper Erdogan,
Maryam Fazel, and Ardavan Maleki. I would like to also name two old friends, Khalil
Ahmadpour and Mehdi Asheghi, whose friendship I gratefully value.
I owe an immeasurable amount of gratitude to my parents, Hossein and Salehe, my
sister, Mojgan, and my brother, Bahman, for their support throughout the numerous
ups and downs that I have experienced. Finally, my sincere thanks goes to my wife,
Samaneh, for her gracious patience and strength. I am sure they agree with me in
dedicating this thesis to Khalil.
vii
Contents
Abstract
iv
Acknowledgements
vi
List of Figures
xii
1 Introduction
1
1.1
Motivation
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
1
1.2
Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4
1.3
An Overview of Adaptive Filtering (Control) Algorithms . . . . . . .
6
1.4
Contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
10
1.5
Thesis Outline
12
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2 Estimation-Based adaptive FIR Filter Design
14
2.1
Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
15
2.2
EBAF Algorithm - Main Concept
. . . . . . . . . . . . . . . . . . .
16
2.3
Problem Formulation
. . . . . . . . . . . . . . . . . . . . . . . . . .
18
2.4
2.3.1
H2 Optimal Estimation
. . . . . . . . . . . . . . . . . . . . .
19
2.3.2
H∞ Optimal Estimation . . . . . . . . . . . . . . . . . . . . .
20
H∞ -Optimal Solution
. . . . . . . . . . . . . . . . . . . . . . . . . .
21
2.4.1
γ-Suboptimal Finite Horizon Filtering Solution . . . . . . . .
21
2.4.2
γ-Suboptimal Finite Horizon Prediction Solution . . . . . . .
22
2.4.3
The Optimal Value of γ . . . . . . . . . . . . . . . . . . . . .
23
2.4.3.1
23
Filtering Case
. . . . . . . . . . . . . . . . . . . . .
viii
2.4.3.2
2.4.4
Prediction Case . . . . . . . . . . . . . . . . . . . .
27
Simplified Solution Due to γ = 1 . . . . . . . . . . . . . . . .
29
2.4.4.1
Filtering Case: . . . . . . . . . . . . . . . . . . . . .
29
2.4.4.2
Prediction Case: . . . . . . . . . . . . . . . . . . . .
30
2.5
Important Remarks
. . . . . . . . . . . . . . . . . . . . . . . . . . .
30
2.6
Implementation Scheme for EBAF Algorithm . . . . . . . . . . . . .
32
2.7
Error Analysis
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
35
2.7.1
Effect of Initial Condition . . . . . . . . . . . . . . . . . . . .
35
2.7.2
Effect of Practical Limitation in Setting y(k) to ŝ(k|k) (ŝ(k))
36
2.8
Relationship to the Normalized-FxLMS/FxLMS Algorithms . . . . .
2.8.1
Prediction Solution and its Connection to the FxLMS Algorithm
2.8.2
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
38
Filtering Solution and its Connection to the Normalized-FxLMS
Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.9
38
Experimental Data & Simulation Results
40
. . . . . . . . . . . . . . .
41
2.10 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
57
3 Estimation-Based adaptive IIR Filter Design
58
3.1
Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
59
3.2
Problem Formulation
. . . . . . . . . . . . . . . . . . . . . . . . . .
61
Estimation Problem . . . . . . . . . . . . . . . . . . . . . . .
63
Approximate Solution . . . . . . . . . . . . . . . . . . . . . . . . . .
65
3.2.1
3.3
3.3.1
γ-Suboptimal Finite Horizon Filtering Solution to the Linearized
Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.3.2
γ-Suboptimal Finite Horizon Prediction Solution to the Linearized Problem
3.3.3
66
. . . . . . . . . . . . . . . . . . . . . . . . .
Important Remarks
66
. . . . . . . . . . . . . . . . . . . . . . .
66
3.4
Implementation Scheme for the EBAF Algorithm in IIR Case . . . .
67
3.5
Error Analysis
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
69
3.6
Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . .
70
3.7
Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
77
ix
4 Multi-Channel Estimation-Based Adaptive Filtering
4.1
4.2
Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
79
4.1.1
79
4.4
Multi-Channel FxLMS Algorithm
. . . . . . . . . . . . . . .
Estimation-Based Adaptive Algorithm for Multi Channel Case
. . .
81
. . . . . . . . . . . . . . . . . . . . . .
85
Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . .
85
4.3.1
Active Vibration Isolation . . . . . . . . . . . . . . . . . . . .
86
4.3.2
Active Noise Cancellation . . . . . . . . . . . . . . . . . . . .
89
4.2.1
4.3
78
H∞ -Optimal Solution
Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
5 Adaptive Filtering via Linear Matrix Inequalities
104
5.1
Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
5.2
LMI Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
5.2.1
Including H2 Constraints
. . . . . . . . . . . . . . . . . . . . 110
5.3
Adaptation Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . 111
5.4
Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
5.5
Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120
6 Conclusion
121
6.1
Summary of the Results and Conclusions
6.2
Future Work
. . . . . . . . . . . . . . . 121
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
A Algebraic Proof of Feasibility
A.1 Feasibility of γf = 1
126
. . . . . . . . . . . . . . . . . . . . . . . . . . . 126
B Feedback Contamination Problem
128
C System Identification for Vibration Isolation Platform
132
C.1 Introduction
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
C.2 Identified Model
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
C.2.1 Data Collection Process . . . . . . . . . . . . . . . . . . . . . 133
C.2.2 Consistency of the Measurements . . . . . . . . . . . . . . . . 134
C.2.3 System Identification
. . . . . . . . . . . . . . . . . . . . . . 137
x
C.2.4 Control design model analysis . . . . . . . . . . . . . . . . . . 140
C.3 FORSE algorithm
. . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
Bibliography
155
xi
List of Figures
1.1
General block diagram for an FIR Filterm . . . . . . . . . . . . . . . . . .
13
1.2
General block diagram for an IIR Filter
. . . . . . . . . . . . . . . . . . .
13
2.1
General block diagram for an Active Noise Cancellation (ANC) problem . . . .
46
2.2
A standard implementation of FxLMS algorithm . . . . . . . . . . . . . . .
47
2.3
Pictorial representation of the estimation interpretation of the adaptive control
problem: Primary path is replaced by its approximate model
. . . . . . . . .
47
2.4
Block diagram for the approximate model of the primary path
. . . . . . . .
48
2.5
Schematic diagram of one-dimensional air duct . . . . . . . . . . . . . . . .
48
2.6
Transfer functions plot from Speakers #1 & #2 to Microphone #1
. . . . . .
49
2.7
Transfer functions plot from Speakers #1 & #2 to Microphone #2
. . . . . .
49
2.8
Validation of simulation results against experimental data for the noise cancellation problem with a single-tone primary disturbance at 150 Hz. The primary
disturbance is known to the adaptive algorithm. The controller is turned on at
t ≈ 3 seconds.
2.9
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
50
Experimental data for the EBAF algorithm of length 4, when a noisy measurement
of the primary disturbance (a single-tone at 150 Hz) is available to the adaptive
algorithm (SNR=3). The controller is turned on at t ≈ 5 seconds.
. . . . . .
51
2.10 Experimental data for the EBAF algorithm of length 8, when a noisy measurement
of the primary disturbance (a multi-tone at 150 and 180 Hz) is available to the
adaptive algorithm (SNR=4.5). The controller is turned on at t ≈ 6 seconds.
.
52
2.11 Experimental data for the EBAF algorithm of length 16, when a noisy measurement of the primary disturbance (a band limited white noise) is available to the
adaptive algorithm (SNR=4.5). The controller is turned on at t ≈ 5 seconds.
xii
.
53
2.12 Simulation results for the performance comparison of the EBAF and (N)FxLMS
algorithms. For 0 ≤ t ≤ 5 seconds, the controller is off. For 5 < t ≤ 20 seconds
both adaptive algorithms have full access to the primary disturbance (a singletone at 150 Hz). For t ≥ 20 seconds the measurement of Microphone #1 is used
as the reference signal (hence feedback contamination problem). The length of
the FIR filter is 24.
. . . . . . . . . . . . . . . . . . . . . . . . . . . .
54
2.13 Simulation results for the performance comparison of the EBAF and (N)FxLMS
algorithms. For 0 ≤ t ≤ 5 seconds, the controller is off. For 5 < t ≤ 40 seconds
both adaptive algorithms have full access to the primary disturbance (a band
limited white noise). For t ≥ 40 seconds the measurement of Microphone #1 is
used as the reference signal (hence feedback contamination problem). The length
of the FIR filter is 32.
. . . . . . . . . . . . . . . . . . . . . . . . . . .
55
2.14 Closed-loop transfer function based on the steady state performance of the EBAF
and (N)FxLMS algorithms in the noise cancellation problem of Figure 2.13.
3.1
. .
56
General block diagram for the adaptive filtering problem of interest (with Feedback
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
72
3.2
Basic Block Diagram for the Feedback Neutralization Scheme . . . . . . . . .
72
3.3
Basic Block Diagram for the Classical Adaptive IIR Filter Design . . . . . . .
73
3.4
Estimation Interpretation of the IIR Adaptive Filter Design
. . . . . . . . .
73
3.5
Approximate Model For the Unknown Primary Path . . . . . . . . . . . . .
74
3.6
Performance Comparison for EBAF and FuLMS Adaptive IIR Filters for Single-
Contamination)
Tone Noise Cancellation. The controller is switched on at t = 1 second. For
1 ≤ t ≤ 6 seconds adaptive algorithm has full access to the primary disturbance.
For t ≥ 6 the output of Microphone #1 is used as the reference signal (hence
feedback contamination problem). . . . . . . . . . . . . . . . . . . . . . .
3.7
75
Performance Comparison for EBAF and FuLMS Adaptive IIR Filters for MultiTone Noise Cancellation. The controller is switched on at t = 1 second. For
1 ≤ t ≤ 6 seconds adaptive algorithm has full access to the primary disturbance.
For t ≥ 6 the output of Microphone #1 is used as the reference signal (hence
feedback contamination problem). . . . . . . . . . . . . . . . . . . . . . .
xiii
76
4.1
General block diagram for a multi-channel Active Noise Cancellation (ANC) problem
4.2
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
91
Pictorial representation of the estimation interpretation of the adaptive control
. . . . . . . . .
91
4.3
Approximate Model for Primary Path . . . . . . . . . . . . . . . . . . . .
92
4.4
Vibration Isolation Platform (VIP)
. . . . . . . . . . . . . . . . . . . . .
92
4.5
A detailed drawing of the main components in the Vibration Isolation Platform
problem: Primary path is replaced by its approximate model
(VIP). Of particular importance are: (a) the platform supporting the middle mass
(labeled as component #5), (b) the middle mass that houses all six actuators (of
which only two, one control actuator and one disturbance actuator) are shown
(labeled as component #11), and (c) the suspension springs to counter the gravity (labeled as component #12). Note that the actuation point for the control
actuator (located on the left of the middle mass) is colocated with the load cell
(marked as LC1). The disturbance actuator (located on the right of the middle
. . . . . . . . . . . . . . . . . .
93
4.6
SVD of the MIMO transfer function . . . . . . . . . . . . . . . . . . . . .
94
4.7
Performance of a multi-channel implementation of EBAF algorithm when distur-
mass) actuates against the inertial frame.
bance actuators are driven by out of phase sinusoids at 4 Hz. The reference signal
available to the adaptive algorithm is contaminated with band limited white noise
(SNR=3). The control signal is applied for t ≥ 30 seconds.
4.8
95
Performance of a multi-channel implementation of FxLMS algorithm when simulation scenario is identical to that in Figure 4.7.
4.9
. . . . . . . . . .
. . . . . . . . . . . . . . .
96
Performance of a multi-channel implementation of EBAF algorithm when disturbance actuators are driven by out of phase multi-tone sinusoids at 4 and 15 Hz.
The reference signal available to the adaptive algorithm is contaminated with band
limited white noise (SNR=4.5). The control signal is applied for t ≥ 30 seconds.
97
4.10 Performance of a multi-channel implementation of FxLMS algorithm when simulation scenario is identical to that in Figure 4.9.
. . . . . . . . . . . . . . .
98
4.11 Performance of a Multi-Channel implementation of the EBAF for vibration isolation when the reference signals are load cell outputs (i.e. feedback contamination
exists). The control signal is applied for t ≥ 30 seconds.
xiv
. . . . . . . . . . .
99
4.12 Performance of the Multi-Channel noise cancellation in acoustic duct for a multitone primary disturbance at 150 and 200 Hz. The control signal is applied for
t ≥ 2 seconds.
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
4.13 Performance of the Multi-Channel noise cancellation in acoustic duct when the
primary disturbance is a band limited white noise. The control signal is applied
for t ≥ 2 seconds.
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
4.14 Closed-loop vs. open-loop transfer functions for the steady state performance of
the EBAF algorithm for the simulation scenario shown in Figure 4.13.
. . . . 102
5.1
General block diagram for an Active Noise Cancellation (ANC) problem . . . .
5.2
Cancellation Error at Microphone #1 for a Single-Tone Primary Disturbance
5.3
Typical Elements of Adaptive Filter Weight Vector for Noise Cancellation Problem
in Fig. 5.2
. 116
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
5.4
Cancellation Error at Microphone #1 for a Multi-Tone Primary Disturbance
5.5
Typical Elements of Adaptive Filter Weight Vector for Noise Cancellation Problem
in Fig. 5.4
B.1
. 118
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
Block diagram of the approximate model for the primary path in the presence of
the feedback path
C.1
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131
Magnitude of the scaling factor relating load cell’s reading of the effect of control
actuators to that of the scoring sensor . . . . . . . . . . . . . . . . . . . .
C.2
146
Magnitude of the scaling factor relating load cell’s reading of the effect of disturbance actuators to that of the scoring sensor after diagonalization . . . . . . .
C.5
145
Magnitude of the scaling factor relating load cell’s reading of the effect of control
actuators to that of the scoring sensor after diagonalization . . . . . . . . . .
C.4
144
Magnitude of the scaling factor relating load cell’s reading of the effect of disturbance actuators to that of the scoring sensor . . . . . . . . . . . . . . . . .
C.3
115
147
Comparison of SVD plots for the transfer function to the scaled/double-integrated
load cell data
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
C.6
Comparison of SVD plots for the transfer function to the actual load cell data .
C.7
Comparison of SVD plots for the transfer function to the scoring sensors
C.8
Comparison of SVD plots for the transfer function to the position sensors colocated
with the control actuators
148
. . . 149
. . . . . . . . . . . . . . . . . . . . . . . . . 149
xv
C.9
Comparison of SVD plots for the transfer function to the position sensors colocated
with the disturbance actuators
. . . . . . . . . . . . . . . . . . . . . . . 150
C.10 The identified model for the system beyond the frequency range for which measurements are available . . . . . . . . . . . . . . . . . . . . . . . . . . .
151
C.11 The final model for the system beyond the frequency range for which measurements are available . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
152
C.12 The comparison of the closed loop and open loop singular value plots when the
controller is used to close the loop on the identified model
. . . . . . . . . . 153
C.13 The comparison of the closed loop and open loop singular value plots when the
controller is used to close the loop on the real measured data
xvi
. . . . . . . . . 154
Chapter 1
Introduction
This dissertation presents a new estimation-based procedure for the systematic synthesis and analysis of adaptive filters (controllers) in “Filtered” LMS problems. This
new approach uses an estimation interpretation of the adaptive filtering (control)
problem to formulate an equivalent estimation problem. The adaptation criterion for
the adaptive weight vector is extracted from the H∞ -solution to this estimation problem. The new algorithm, referred to as Estimation-Based Adaptive Filtering (EBAF),
applies to both Finite Impulse Response (FIR) and Infinite Impulse Response (IIR)
adaptive filters.
1.1
Motivation
Least-Mean Squares (LMS) adaptive algorithm [51] has been the centerpiece of a wide
variety of adaptive filtering techniques for almost four decades. The straightforward
derivation, and the simplicity of its implementation (especially at the time of limited
computational power) encouraged experiments with the algorithm in a diverse range
of applications (e.g. see [51,33]). In some applications however, the simple implementation of the LMS algorithm was found to be inadequate. Subsequent attempts
to overcome its shortcomings have produced a large number of innovative solutions
that have been successful in practice. Commonly used algorithms such as normalized
1
1.1. MOTIVATION
2
LMS, correlation LMS [47], leaky LMS [21], variable-step-size LMS [25], and FilteredX LMS [35] are the outcome of such efforts. These algorithms use the instantaneous
squared error to estimate the mean-square error, and often assume slow adaptation
to allow for the necessary linear operations in their derivation (see Chapters 2 and 3
in [33] for instance). As Reference [2] points out:
“Many of the algorithms and approaches used are of an ad hoc nature;
the tools are gathered from a wide range of fields; and good systematic
approaches are still lacking.”
Introducing a systematic procedure for the synthesis of adaptive filters is one of the
main goals of this thesis.
Parallel to the efforts on the practical application of the LMS-based adaptive
schemes, there has been a concerted effort to analyze these algorithms. Of pioneering
importance are the results in Refs. [50] and [23]. Reference [50] considers the adaptation with LMS on stationary stochastic processes, and finds the optimal solution
to which the expected value of the weight vector converges. For sinusoidal inputs
however, the discussion in [50] does not apply. In [23] it is shown that for sinusoidal
inputs, when time-varying component of the adaptive filter output is small compared
to its time-invariant component (see [23], page 486), the adaptive LMS filter can be
approximated by a linear time-invariant transfer function. Reference [13] extends the
approach in [23] to derive an equivalent transfer function for the Filtered-X LMS
adaptive algorithm (provided the conditions required in [23] still apply). The equivalent transfer function is then used to analytically derive an expression for the optimum
convergence coefficients. A frequency domain model of the so-called filtered LMS algorithm (i.e. an algorithm in which the input or the output of the adaptive filter or
the feedback error signal is linearly filtered prior to use in the adaptive algorithm)
is discussed in [17]. The frequency domain model in [17] decouples the inputs into
disjoint frequency bins and places a single frequency adaptive noise canceler on each
bin. The analysis in their work utilizes the frequency domain LMS algorithm [11]
and assumes a time invariant linear behavior for the filter. Other important aspects
1.1. MOTIVATION
3
of the adaptive filters have also been extensively studied. The effect of the modeling error on the convergence and performance properties of the LMS-based adaptive
algorithms (e.g. [17,7]), and tracking behavior of the LMS adaptive algorithm when
the adaptive filter is tuned to follow a linear chirp signal buried in white noise [5,6],
are examples of these studies∗ . In summary, existing analysis techniques are often
suitable for analyzing only one particular aspect of the behavior of an adaptive filter
(e.g. its steady-state behavior). Furthermore, the validity of the analysis relies on
certain assumptions (e.g. slow convergence, and/or the negligible contribution of the
nonlinear/time-varying component of the adaptive filter output) that can be quite
restrictive. Providing a solid framework for the systematic analysis of adaptive filters
is another main goal of this thesis.
The reason for the difficulty experienced in both synthesis and analysis of adaptive
algorithms is best explained in Reference [37]:
“It is now generally realized that adaptive systems are special classes of
nonlinear systems . . . general methods for the analysis and synthesis of
nonlinear systems do not exist since conditions for their stability can be
established only on a system by system basis.”
This thesis introduces a new framework for the synthesis and analysis of adaptive
filters (controllers) by providing an estimation interpretation of the above mentioned
“nonlinear” adaptive filtering (control) problem. The estimation interpretation replaces the original adaptive filtering (control) synthesis with an equivalent estimation
problem, the solution of which is used to update the weight vector in the adaptive
filter (and hence the name estimation-based adaptive filtering). This approach is
applicable (due to its systematic nature) to both FIR and IIR adaptive filters (controllers). In the FIR case the equivalent estimation problem is linear, and hence exact
solutions are available. Stability, performance bounds, transient behavior of adaptive
FIR filters are thus precisely addressed in this framework. In the IIR case, however,
only an approximate solution to the equivalent estimation problem is available, and
∗
The survey here is intended to provide a flavor of the type of the problems that have captured
the attention of researchers in the field. The shear volume of the literature makes subjective selection
of the references unavoidable.
1.2. BACKGROUND
4
hence the proposed estimation-based framework serves as a reasonable heuristic for
the systematic design of adaptive IIR filters. This approximate solution however, is
based on realistic assumptions, and the adaptive algorithm maintains its systematic
structure. Furthermore, the treatment of feedback contamination (see Chapter 3 for a
precise definition), is virtually identical to that of adaptive IIR filters. The proposed
estimation-based approach is particularly appealing if one considers the difficulty with
the existing design techniques for adaptive IIR filters, and the complexity of available
solutions to feedback contamination (e.g. see [33]).
1.2
Background
The development of the new estimation-based framework is based on recent results
in robust estimation. Following the pioneering work in [52], the H∞ approach to
robust control theory produced solutions [12,24] that were designed to meet some
performance criterion in the face of the limited knowledge of the exogenous disturbances and imperfect system models. Further work in robust control and estimation
(see [32,46] and the references therein) produced straightforward solutions that allowed in-depth studies of the properties of the robust controllers/estimators. The
main idea in H∞ estimation is to design an estimator that bounds (in the optimum
case, minimizes) the maximum energy gain from the disturbances to the estimation
errors. Such a solution guarantees that for disturbances with bounded energy, the
energy of the estimation error will be bounded as well. In the case of an optimal
solution, an H∞ -optimal estimator will guarantee that the energy of the estimation
error for the worst case disturbance is indeed minimized [28].
Of crucial importance for the work in this thesis, is the result in [26] where the H∞ optimality of the LMS algorithm was established. Note that despite a long history
of successful applications, prior to the work in [26], the LMS algorithm was regarded
as an approximate recursive solution to the least-squares minimization problem. The
work in [26] showed that instead of being an approximate solution to an H2 minimization, the LMS algorithm is the exact solution to a minmax estimation problem. More
1.2. BACKGROUND
5
specifically, Ref. [26] proved that the LMS adaptive filter is the central a priori H∞ optimal filter. This result established a fundamental connection between an adaptive
control algorithm (LMS algorithm in this case), and a robust estimation problem.
Inspired by the analysis in [26], this thesis introduces an estimation interpretation of
a far more general adaptive filtering problem, and develops a systematic procedure for
the synthesis of adaptive filters based on this interpretation. The class of problems
addressed in this thesis, commonly known as “Filtered” LMS [17], encompass a wide
range of adaptive filtering/control applications [51,33], and have been the subject of
extensive research over the past four decades. Nevertheless, the viewpoint provided
in this thesis not only provides a systematic alternative to some widely used adaptive
filtering (control) algorithms (such as FxLMS and FuLMS) with superior transient
and steady-state behavior, but it also presents a new framework for their analysis.
More specifically, this thesis proves that the fundamental connection between adaptive filtering (control) algorithms and robust estimation extends to the more general
setting of adaptive filtering (control) problems, and shows that the convergence, stability, and performance of these classical adaptive algorithms can be systematically
analyzed as robust estimation questions.
The systematic nature of the proposed estimation-based approach enables an alternative formulation for the adaptive filtering (control) problem using Linear Matrix
Inequalities (LMIs), the ramifications of which will be discussed in Chapter 5. Several
researchers (see [18] and references therein) in the past few years have shown that
elementary manipulations of linear matrix inequalities can be used to derive less restrictive alternatives to the now classical state-space Riccati-based solution to the H∞
control problem [12]. Even though the computational complexity of the LMI-based
solution remains higher than that of solving the Riccati equation, there are three main
reasons that justify such a formulation [19]: (a) a variety of design specifications and
constraints can be expressed as LMIs, (b) problems formulated as LMIs can be solved
exactly by efficient convex optimization techniques, and (c) for the cases that lack
analytical solutions such as mixed H2 /H∞ design objectives (see [4], [32] and [45] and
references therein), the LMI formulation of the problem remains tractable (i.e. LMIsolvers are viable alternatives to analytical solutions in such cases). As will be seen
1.3. AN OVERVIEW OF ADAPTIVE FILTERING (CONTROL) ALGORITHMS
6
in Chapter 5, the LMI framework provides the machinery required for the synthesis
of a robust adaptive filter in the presence of modeling uncertainty.
1.3
An Overview of Adaptive Filtering (Control)
Algorithms
To put this thesis in perspective, this section provides a brief overview of the vast
literature on adaptive filtering (control). Reference [36] recognizes 1957 as the year
for the formal introduction of the term “adaptive system” into the control literature.
By then, the interest in filtering and control theory had shifted towards increasingly
more complex systems with poorly characterized (possibly time varying) models for
system dynamics and disturbances, and the concept of “adaptation” (borrowed from
living systems) seemed to carry the potential for solving the increasingly more complex control problems. The exact definition of “adaptation” and its distinction from
“feedback”, however, is the subject of long standing discussions (e.g. see [2,36,29]).
Qualitatively speaking, an adaptive system is a system that can modify its behavior
in response to changes in the dynamics of the system or disturbances through some
recursive algorithm. As a direct consequence of this recursive algorithm (in which
the parameters of the adaptive system are adjusted using input/output data), an
adaptive system is a “nonlinear” device.
The development of adaptive algorithms has been pursued from a variety of view
points. Different classifications of adaptive algorithms (such as direct versus indirect
adaptive control, model reference versus self-tuning adaptation) in the literature reflect this diversity [2,51,29]. For the purpose of this thesis, two distinct approaches for
deriving recursive adaptive algorithms can be identified: (a) stochastic gradient approaches that include LMS and LMS-Based adaptive algorithms, and (b) least-squares
estimation approaches that include adaptive recursive least-squares (RLS) algorithm.
The central idea in the former approach, is to define an appropriate cost function
that captures the success of the adaptation process, and then change the adaptive
1.3. AN OVERVIEW OF ADAPTIVE FILTERING (CONTROL) ALGORITHMS
7
filter parameters to reduce the cost function according to the method of steepest descent. This requires the use of a gradient vector (hence the name), which in practice
is approximated using instantaneous data. Chapter 2 provides a detailed description
of this approach for the problem of interest in this Thesis. The latter approach to
the design of adaptive filters is based on the method of least squares. This approach
closely corresponds to Kalman filtering. Ref. [44] provides a unifying state-space approach to adaptive RLS filtering. The main focus in this thesis however, is on the
LMS-based adaptive algorithms.
Since adaptive algorithms can successfully operate in a poorly known environment,
they have been used in a diverse field of applications that include communication
(e.g. [34,41]), process control (e.g. [2]), seismology (e.g. [42]), biomedical engineering
(e.g. [51]). Despite the diversity of the applications, different implementations of
adaptive filtering (control) share one basic common feature [29]: “an input vector and
a desired response are used to compute an estimation error, which is in turn used to
control the values of a set of adjustable filter coefficients.” Reference [29] distinguishes
four main classes of adaptive filtering applications based on the way the desired
signal is defined in the formulation of the problem: (a) identification: in this class of
applications an adaptive filter is used to provide a linear model for an unknown plant.
The plant and the adaptive filter are driven by the same input, and the output of the
plant is the desired response that adaptive filter tries to match. (b) inverse modeling:
here the adaptive filter is placed in series with an unknown (perhaps noisy) plant, and
the desired signal is simply a delayed version of the plant input. Ideally, the adaptive
filter converges to the inverse of the unknown plant. Adaptive equalization (e.g. [40])
is an important application in this class. (c) prediction: the desired signal in this case
is the current value of a random signal, while past values of the random signal provide
the input to the adaptive filter. Signal detection is an important application in this
class. (d) interference canceling: here adaptive filter uses a reference signal (provided
as input to the adaptive filter) to cancel unknown interference contained in a primary
signal. Adaptive noise cancellation, echo cancellation, and adaptive beam-forming
are applications that fall in this last class. The estimation-based adaptive filtering
algorithm in this thesis is presented in the context of adaptive noise cancellation, and
1.3. AN OVERVIEW OF ADAPTIVE FILTERING (CONTROL) ALGORITHMS
8
therefore a detailed discussion of the fourth class of adaptive filtering problems is
provided in Chapter 2.
There are several main structures for the implementation of adaptive filters (controllers). The structure of the adaptive filter is known to affect its performance,
computational complexity, and convergence. In this thesis, the two most commonly
used structures for adaptive filters (controllers) are considered. The finite impulse
response (FIR) transversal filter (see Fig. 1.1) is the structure upon which the main
presentation of the estimation-based adaptive filtering algorithm is primarily presented. The transversal filter consists of three basic elements: (a) unit-delay element,
(b) multiplier, and (c) adder, and contains feed forwards paths only. The number
of unit-delays specify the length of the adaptive FIR filter. Multipliers weight the
delayed versions of some reference signal, which are then added in the adder(s). The
frequency response for this filter is of finite length (hence the name), and contains
only zeros (all poles are at the origin in the z-plane). Therefore, there is no question
of stability for the open-loop behavior of the FIR filter. The infinite-duration impulse
response (IIR) structure is shown in Figure 1.2. The feature that distinguishes the
IIR filter from an FIR filter is the inclusion of the feedback path in the structure of
the adaptive filter.
As mentioned earlier, for an FIR filter all poles are at the origin, and a good
approximation of the behavior of a pole, in general, can only be achieved if the length
of the FIR filter is sufficiently long. An IIR filter, ideally at least, can provide a
perfect match for a pole with only a limited number of parameters. This means
that for a desired dynamic behavior (such as resonance frequency, damping, or cutoff
frequency), the number of parameters in an adaptive IIR filter can be far fewer than
that in its FIR counterpart. The computational complexity per sample for adaptive
IIR filter design can therefore be significantly lower than that in FIR filter design.
The limited use of adaptive IIR filters (compared to the vast number of applications for the FIR filters) suggests that the above mentioned advantages come at
a certain cost. In particular, adaptive IIR filters are only conditionally stable, and
therefore some provisions are required to assure stability of the filter at each iteration.
There are solutions such as Schur-Cohn algorithm ([29] pages 271-273) that monitor
1.3. AN OVERVIEW OF ADAPTIVE FILTERING (CONTROL) ALGORITHMS
9
the stability of the IIR filter (by determining whether all roots of the denominator of
the IIR filter transfer function are inside the unit circle). This however requires intensive on-line calculations. Alternative implementations of adaptive IIR filters (such
as parallel implementation [48], and lattice implementation [38]) have been suggested
that provide simpler stability monitoring capabilities. The monitoring process is independent of the adaptation process here. In other words, the adaptation criteria
do not inherently reject de-stabilizing values for filter weights. The monitoring process detects these de-stabilizing values and prevents their implementation. Another
significant problem with adaptive IIR filter design stems from the fact that the performance surface (see [33], Chapter 3) for adaptive IIR filters is generally non-quadratic
(see [33] pages 91-94 for instance) and often contains multiple local minima. Therefore, the weight vector may converge to a local minimum only (hence non-optimal
cost). Furthermore, it is noted that the adaptation rate for adaptive IIR filters can
be slow when compared to the FIR adaptive filters [33,31]. Early works in adaptive
IIR filtering (e.g. [16]) are for the most part extensions to Widrow’s LMS algorithm
of adaptive FIR filtering [51]. More recent works include modifications to recursive
LMS algorithm (e.g. [15]) that are devised for specific applications. In other words,
existing design techniques for adaptive IIR filters are application-specific and rely on
certain restrictive assumptions in their derivation. Our description of the Filtered-U
recursive LMS algorithm in Chapter 3 will further clarify this point. Furthermore,
as [33] points out: “The properties of an adaptive IIR filter are considerably more
complex than those of the conventional adaptive FIR filter, and consequently it is
more difficult to predict their behavior.” Thus, a framework that allows a unified
approach to the synthesis and analysis of adaptive IIR filters, and does not require
restrictive assumptions for its derivation would be extremely useful. As mentioned
earlier, this thesis provides such a framework.
Finally, for a wide variety of applications such as equalization in wireless communication channels, and active control of sound and vibration in an environment
where the effect of a number of primary sources should be canceled by a number of
control (secondary) sources, the use of a multi-channel adaptive algorithm is well justified. In general, however, variations of the LMS algorithm are not easy to extend to
1.4. CONTRIBUTIONS
10
multi-channel systems. Furthermore, the analysis of the performance and properties
of such multi-channel algorithms is complicated [33]. As Ref. [33] points out, in the
context of active noise cancellation, the successful implementation of multi-channel
adaptive algorithms has so far been limited to cases involving repetitive noise with a
few harmonics [39,43,49,13]). For the approach presented in this thesis, the syntheses
of single-channel and multi-channel adaptive algorithms are virtually identical. This
similarity is a direct result of the way the synthesis problem is formulated (see 4).
1.4
Contributions
In meeting the goals of this research, the following contributions have been made to
adaptive filtering and control:
1. An estimation-interpretation for adaptive “Filtered” LMS filtering (control)
problems is developed. This interpretation allows an equivalent estimation formulation for the adaptive filtering (control) problem. The adaptation criterion
for adaptive filter weight vector is extracted from the solution to this equivalent estimation problem. This constitutes a systematic synthesis procedure for
adaptive filters in filtered LMS problems. The new synthesis procedure is called
Estimation-Based Adaptive Filtering (EBAF).
2. Using an H∞ criterion to formulate the “equivalent” estimation problem, this
thesis develops a new framework for the systematic analysis of Filtered LMS
adaptive algorithms. In particular, the results in this thesis extend the fundamental connection between the LMS adaptive algorithm and robust estimation
(i.e. H∞ optimality of the LMS algorithm [26]) to the more general setting of
filtered LMS adaptive problems.
3. For the EBAF algorithm in the FIR case:
(a) It is shown that the adaptive weight vector update can be based on the
central filtering (prediction) solution to a linear H∞ estimation problem,
the existence of which is guaranteed. It is also shown that the maximum
1.4. CONTRIBUTIONS
11
energy gain in this case can be minimized. Furthermore, the optimal energy gain is proved to be unity, and the conditions under which this bound
is achievable are derived.
(b) The adaptive algorithm is shown to be implementable in real-time. The
update rule requires a simple Lyapunov recursion that leads to a computational complexity comparable to that of filtered LMS adaptive algorithms
(e.g. FxLMS). The experimental data, along with extensive simulations
are presented to demonstrate the improved steady-state performance of
the EBAF algorithm (over FxLMS and Normalized-FxLMS algorithms),
as well as a faster transient response.
(c) A clear connection between the limiting behavior of the EBAF algorithm
and the existing FxLMS and Normalized-FxLMS adaptive algorithms has
been established.
4. For the EBAF algorithm in the IIR case, it is shown that the equivalent estimation problem is nonlinear. A linearizing approximation is then employed
that makes systematic synthesis of adaptive IIR filter tractable. The performance of the EBAF algorithm in this case is compared to the performance of
the Filtered-U LMS (FuLMS) adaptive algorithm, demonstrating the improved
performance in the EBAF case.
5. The treatment of feedback contamination problem is shown to be identical to
the IIR adaptive filter design in the new estimation-based framework.
6. A multi-channel extension of the EBAF algorithm demonstrates that the treatment of the single-channel and multi-channel adaptive filtering (control) problems in the new estimation based framework is virtually the same. Simulation
results for the problem of vibration isolation in a 3-input/3-output vibration isolation platform (VIP) prove feasibility of the EBAF algorithm in multi-channel
problems.
7. The new estimation-based framework is shown to be amenable to a Linear Matrix Inequality (LMI) formulation. The LMI formulation is used to explicitly
1.5. THESIS OUTLINE
12
address the stability of the overall system under adaptive algorithm by producing a Lyapunov function. It is also shown to be an appropriate framework to
address the robustness of the adaptive algorithm to modeling error or parameter uncertainty. Augmentation of an H2 performance constraint to the H∞
disturbance rejection criterion is also discussed.
1.5
Thesis Outline
The organization of this thesis is as follows. In Chapter 2, the fundamental concepts
of the estimation-based adaptive filtering (EBAF) algorithm are introduced. The
application of the EBAF approach in the case of adaptive FIR filter design is also
presented in this chapter. In Chapter 3, the extension of the EBAF approach to the
adaptive IIR filter design is discussed. A multi-channel implementation of the EBAF
algorithm is presented in Chapter 4. An LMI formulation for the EBAF algorithm is
derived in Chapter 5. Chapter 6 concludes this dissertation with a summary of the
main results, and the suggestions for future work. This dissertation contains three
appendices. An algebraic proof for the feasibility of the unity energy gain in the
estimation problem associated with adaptive FIR filter design (in Chapter 2) is discussed in Appendix A. The problem of feedback contamination is formally addressed
in Appendix B. A detailed discussion of the identification process is presented in Appendix C. The identified model for the Vibration Isolation Platform (VIP), used as a
test-bed for multi-channel implementation of the EBAF algorithm, is also presented
in this appendix.
1.5. THESIS OUTLINE
13
x(k − 1)
x(k)
z −1
W0
x(k − 2)
x(k − N )
z −1
W1
z −1
WN −1
W2
WN
+
u(k)
Fig. 1.1: General block diagram for an FIR Filterm
x(k)
u(k)
r(k)
a0
+
+
z −1
a1
b1
z −1
a2
b2
r(k − 2)
z −1
bN
aN
Fig. 1.2: General block diagram for an IIR Filter
Chapter 2
Estimation-Based adaptive FIR
Filter Design
This chapter presents a systematic synthesis procedure for H∞ -optimal adaptive FIR
filters in the context of an Active Noise Cancellation (ANC) problem. An estimation
interpretation of the adaptive control problem is introduced first. Based on this interpretation, an H∞ estimation problem is formulated, and its finite horizon prediction
(filtering) solutions are discussed. The solution minimizes the maximum energy gain
from the disturbances to the predicted (filtered) estimation error, and serves as the
adaptation criterion for the weight vector in the adaptive FIR filter. This thesis refers
to the new adaptation scheme as Estimation-Based Adaptive Filtering (EBAF). It
is shown in this chapter that the steady-state gain vectors in the EBAF algorithm
approach those of the classical Filtered-X LMS (Normalized Filtered-X LMS) algorithm. The error terms, however, are shown to be different, thus demonstrating that
the classical algorithms can be thought of as an approximation to the new EBAF
adaptive algorithm.
The proposed EBAF algorithm is applied to an active noise cancellation problem
(both narrow-band and broad-band cases) in a one-dimensional acoustic duct. Experimental data as well as simulations are presented to examine the performance of
the new adaptive algorithm. Comparisons to the results from a conventional FxLMS
algorithm show faster convergence without compromising steady-state performance
14
2.1. BACKGROUND
15
and/or robustness of the algorithm to feedback contamination of the reference signal.
2.1
Background
This section introduces the context in which the new estimation-based adaptive filtering (EBAF) algorithm will be presented. It defines the adaptive filtering problem
of interest and describes the terminology that is used in this chapter. A conventional
solution to the problem based on the FxLMS algorithm is also outlined in this section. The discussion of key concepts of the EBAF algorithm and the mathematical
formulation of the algorithm are left to Sections 2.2 and 2.3, respectively.
Referring to Fig. 2.1, the objective in this adaptive filtering problem is to adjust
the weight vector in the adaptive FIR filter, W (k) = [w0 (k) w1 (k) ... wN (k)]T (k is
the discrete time index), such that the cancellation error, d(k) −y(k), is small in some
appropriate measure. Note that d(k) and y(k) are outputs of the primary path P (z)
and the secondary path S(z), respectively. Moreover,
1. n(k) is the input to the primary path,
2. x(k) is a properly selected reference signal with a non-zero correlation with the
primary input,
4
3. u(k) is the control signal applied to the secondary path (generated as u(k) =
[x(k) x(k − 1) · · · x(k − N)] W (k)),
4. e(k) is the measured residual error available to the adaptation scheme.
Note that in a typical practice, x(k) is obtained via some measurement of the primary
input. The quality of this measurement will impact the correlation between the
reference signal and the primary input. Similar to the conventional development of
the FxLMS algorithm however, this chapter assumes perfect correlation between the
two.
The Filtered-X LMS (FxLMS) solution to this problem is shown in Figure 2.2
where perfect correlation between the primary disturbance n(k) and the reference
signal x(k) is assumed [51,33]. Minimizing the instantaneous squared error, e2 (k), as
2.2. EBAF ALGORITHM - MAIN CONCEPT
16
an approximation to the mean-square error, FxLMS follows the LMS update criterion
(i.e. to recursively adapt the weight vector in the negative gradient direction)
µ 2
∇e (k)
2
e(k) = d(k) − y(k) = d(k) − S(k) ⊕ u(k)
W (k + 1) = W (k) −
where µ is the adaptation rate, S(k) is the impulse response of the secondary path,
and “⊕” indicates convolution. Assuming slow adaptation, the FxLMS algorithm
then approximates the instantaneous gradient in the weight vector update with
4
T
∇e2 (k) ∼
= −2 [x0 (k) x0 (k − 1) · · · x0 (k − N)] e(k) = −2h0 (k)e(k)
(2.1)
4
where x0 (k) = S(k) ⊕ x(k) represents a filtered version of the reference signal which
is available to the LMS adaptation (and hence the name (Normalized) Filtered-X
LMS). This yields the following adaptation criterion for the FxLMS algorithm
W (k + 1) = W (k) + µh0 (k)e(k)
(2.2)
A closely related adaptive algorithm is the one in which the adaptation rate is
normalized with the estimate of the power of the reference vector, i.e.
W (k + 1) = W (k) + µ
h0 (k)
e(k)
1 + µh∗ 0 (k)h0 (k)
(2.3)
where ∗ indicates complex conjugate. This algorithm is known as the NormalizedFxLMS algorithm.
In practice, however, only an approximate model of the secondary path (obtained
via some identification scheme) is known, and it is this approximate model that is
used to filter the reference signal. For further discussion on the derivation and analysis
of the FxLMS algorithm please refer to [33,7].
2.2
EBAF Algorithm - Main Concept
The principal goal of this section is to introduce the underlying concepts of the new
EBAF algorithm. For the developments in this section, perfect correlation between
2.2. EBAF ALGORITHM - MAIN CONCEPT
17
n(k) and x(k) in Fig. 2.1 is assumed (i.e. x(k) = n(k) for all k). This is the same
condition under which the FxLMS algorithm was developed. The dynamics of the
secondary path are assumed known (e.g. by system identification). No explicit model
for the primary path is needed.
As stated before, the objective in the adaptive filtering problem of Fig. 2.1 is to
generate a control signal, u(k), such that the output of the secondary path, y(k), is
“close” to the output of the primary path, d(k). To achieve this goal, for the given
reference signal x(k), the series connection of the FIR filter and the secondary path
must constitute an appropriate model for the unknown primary path. In other words,
with the adaptive FIR filter properly adjusted, the path from x(k) to d(k) must be
equivalent to the path from x(k) to y(k). Based on this observation, in Fig. 2.3 the
structure of the path from x(k) to y(k) is used to model the primary path. The
modeling error is included to account for the imperfect cancellation.
The above mentioned observation forms the basis for an estimation interpretation of the adaptive control problem. The following outlines the main steps for this
interpretation:
1. Introduce an approximate model for the primary path based on the architecture
of the adaptive path from x(k) to y(k) (as shown in Fig. 2.3). There is an
optimal value for the weight vector in the approximate model’s FIR filter for
which the modeling error is the smallest. This optimal weight vector, however,
is not known. State-space models are used for both FIR filter and the secondary
path.
2. In the approximate model for the primary path, use the available information to
formulate an estimation problem that recursively estimates this optimal weight
vector.
3. Adjust the weight vector of the adaptive FIR filter to the best available estimate
of the optimal weight vector.
Before formalizing this estimation-based approach, a closer look at the signals
(i.e. information) involved in Fig. 2.1 is provided. Note that e(k) = d(k) − y(k) +
Vm (k), where
2.3. PROBLEM FORMULATION
18
a. e(k) is the available measurement.
b. Vm (k) is the exogenous disturbance that captures the effect of measurement
noise, modeling error, and the initial condition uncertainty in error measurements.
c. y(k) is the output of the secondary path.
d. d(k) is the output of the primary path.
Note that unlike e(k), the signals y(k) and d(k) are not directly measurable. With
u(k) fully known, however, the assumption of a known initial condition for the secondary path leads to the exact knowledge of y(k). This assumption is relaxed later
in this chapter, where the effect of an “inexact” initial condition in the performance
of the adaptive filter is studied (Section 2.7).
The derived measured quantity that will be used in the estimation process can
now be introduced as
4
m(k) = e(k) + y(k) = d(k) + Vm (k)
2.3
(2.4)
Problem Formulation
Figure 2.4 shows a block diagram representation of the approximate model to the
primary path. A state space model, [ As (k), Bs (k), Cs (k), Ds (k) ], for the secondary
path is assumed. Note that both primary and secondary paths are assumed stable.
The weight vector, W (k) = [ w0 (k) w1 (k) · · · wN (k) ]T , is treated as the state vector
capturing the trivial dynamics, W (k + 1) = W (k), that is assumed
for the FIR filter.
With θ(k) the state variable for the secondary path, then ξ T =
W T (k) θT (k)
is
the state vector for the overall system.
The state space representation of the system is then
#"
#
"
#
"
0
W (k) 4
W (k + 1)
I(N +1)×(N +1)
= Fk ξk
=
Bs (k)h∗ (k) As (k)
θ(k)
θ(k + 1)
(2.5)
2.3. PROBLEM FORMULATION
19
where h(k) = [x(k) x(k − 1) · · · x(k − N)]T captures the effect of the reference input
x(·). For this system, the derived measured output defined in Eq. (2.4) is
#
"
h
i W (k)
4
m(k) = Ds (k)h∗ (k) Cs (k)
+ Vm (k) = Hk ξk + Vm (k)
θ(k)
(2.6)
A linear combination of the states is defined as the desired quantity to be estimated
"
#
h
i W (k)
4
s(k) = L1,k L2,k
= Lk ξk
(2.7)
θ(k)
For simplicity, the single-channel problem is considered here. Extension to the multichannel case is straight forward and is discussed in Chapter 4. Therefore, m(k) ∈
R1×1 , s(k) ∈ R1×1 , θ(k) ∈ RNs ×1 , and W (k) ∈ R(N +1)×1 . All matrices are then
of appropriate dimensions. There are several alternatives for selecting Lk and thus
the variable to be estimated, s(k). The end goal of the estimation based approach
however, is to set the weight vector in the adaptive FIR filter such that the output
of the secondary path, y(k) in Fig. 2.3, best matches d(k). So s(k) = d(k) is chosen,
i.e. Lk = Hk .
Any estimation algorithm can now be used to generate an estimate of the desired
quantity s(k). Two main estimation approaches are considered next.
2.3.1
H2 Optimal Estimation
Here stochastic interpretation of the estimation problem is possible. Assuming that
ξ0 (the initial condition for the system in Figure 2.4) and Vm (·) are zero mean uncorrelated random variables with known covariance matrices
"
#
#
"
h
i
ξ0
0
Π
0
∗
E
=
ξ0∗ Vm
(j)
0 Qk δkj
Vm (k)
(2.8)
4
ŝ(k|k) = F (m(0), · · · , m(k)), the causal linear least-mean-squares estimate of s(k), is
given by the Kalman filter recursions [27].
There are two primary difficulties with the H2 optimal solution: (a) The H2 solution is optimal only if the stochastic assumptions are valid. If the external disturbance
2.3. PROBLEM FORMULATION
20
is not Gaussian (for instance when there is a considerable modeling error that should
be treated as a component of the measurement disturbance) then pursuing an H2
filtering solution may yield undesirable performance; and (b) regardless of the choice
for Lk , the recursive H2 filtering solution does not simplify to the same extent as
the H∞ solution considered below. This can be of practical importance when the
real-time computational power is limited. Therefore, the H2 optimal solution is not
employed in this chapter.
2.3.2
H∞ Optimal Estimation
To avoid difficulties associated with the H2 estimation, we consider a minmax formulation of the estimation problem in this section. Here, the main objective is to limit
the worst case energy gain from the measurement disturbance and the initial condition uncertainty to the error in a causal (or strictly causal) estimate of s(k). More
specifically, the following two cases are of interest. Let ŝ(k|k) = Ff (m(0), · · · , m(k))
denote an estimate of s(k) given observations m(i) for time i = 0 up to and including
4
time i = k, and let ŝ(k) = ŝ(k|k − 1) = Fp (m(0), · · · , m(k − 1)) denote an estimate
of s(k) given m(i) for time i = 0 up to and including i = k − 1. Note that ŝ(k|k)
and ŝ(k) are known as filtering and prediction estimates of s(k), respectively. Two
estimation errors can now be defined: the filtered error
ef,k = ŝ(k|k) − s(k)
(2.9)
ep,k = ŝ(k) − s(k)
(2.10)
and the predicted error
Given a final time M, the objective of the filtering problem can now be formalized as
finding ŝ(k|k) such that for Π0 > 0
M
X
sup
Vm , ξ0
e∗f,k ef,k
k=0
ˆ
(ξ0 − ξˆ0 )∗ Π−1
0 (ξ0 − ξ0 ) +
M
X
k=0
≤ γ2
∗
Vm
(k)Vm (k)
(2.11)
2.4. H∞ -OPTIMAL SOLUTION
21
for a given scalar γ > 0. In a similar way, the objective of the prediction problem can
be formalized as finding ŝ(k) such that
M
X
sup
Vm , ξ0
e∗p,k ep,k
k=0
ˆ
(ξ0 − ξˆ0 )∗ Π−1
0 (ξ0 − ξ0 ) +
M
X
≤ γ2
(2.12)
∗
Vm
(k)Vm (k)
k=0
for a given scalar γ > 0. The question of optimality of the solution can be answered
by finding the infimum value among all feasible γ’s. Note that, for the H∞ optimal
estimation there is no statistical assumption regarding the measurement disturbance.
Therefore, the inclusion of the output of the modeling error block (see Fig. 2.3) in
the measurement disturbance is consistent with H∞ formulation of the problem. The
elimination of the “modeling error” block in the approximate model of primary path
in Fig. 2.4 is based on this characteristic of the disturbance in an H∞ formulation.
2.4
H∞ -Optimal Solution
For the remainder of this chapter, the case where Lk = Hk is considered. Referring
to Figure 2.4, this means that s(k) = d(k). To discuss the solution, from [27] the
solutions to the γ-suboptimal finite-horizon filtering problem of Eq. (2.11), and the
prediction problem of Eq. (2.12) are drawn. Finally, we find the optimal value of γ
and show how γ = γopt simplifies the solutions.
2.4.1
γ-Suboptimal Finite Horizon Filtering Solution
Theorem 2.1: [27]Consider the state space representation of the block diagram of
Figure 2.4, described by Equations (2.5)-(2.7). A level-γ H∞ filter that achieves
(2.11) exists if, and only if, the matrices
0
0
Hk
Ip
Ip
Rk =
and Re,k =
+
Pk Hk∗ L∗k
(2.13)
2
2
0 −γ Iq
0 −γ Iq
Lk
(here p and q are used to indicate the correct dimensions) have the same inertia for
all 0 ≤ k ≤ M, where P0 = Π0 > 0 satisfies the Riccati recursion
∗
Pk+1 = Fk Pk Fk∗ − Kf,k Re,k Kf,k
(2.14)
2.4. H∞ -OPTIMAL SOLUTION
22
where Kf,k =
Fk Pk
Hk∗ L∗k
−1
Re,k
If this is the case, then the central H∞ estimator is given by
ξˆk+1 = Fk ξˆk + Kf,k m(k) − Hk ξˆk , ξˆ0 = 0
−1
ŝ(k|k) = Lk ξˆk + (Lk Pk Hk∗ ) RHe,k
m(k) − Hk ξˆk
(2.15)
(2.16)
(2.17)
−1
with Kf,k = (Fk Pk Hk∗ ) RHe,k
and RHe,k = Ip + Hk Pk Hk∗ .
Proof: see [27].
2.4.2
γ-Suboptimal Finite Horizon Prediction Solution
Theorem 2.2: [27]For the system described by Equations (2.5)-(2.7), level-γ H∞
filter that achieves (2.12) exists if, and only if, all leading sub-matrices of
Lk
−γ 2 Ip 0
−γ 2 Ip 0
p
p
(2.18)
Rk =
and Re,k =
+
Pk L∗k Hk∗
0
Iq
0
Iq
Hk
have the same inertia for all 0 ≤ k < M. Note that Pk is updated according to Eq.
(2.14). If this is the case, then one possible level-γ H∞ filter is given by
ξˆk+1 = Fk ξˆk + Kp,k m(k) − Hk ξˆk , ξˆ0 = 0
(2.19)
ŝ(k) = Lk ξˆk
where
Kp,k =
Fk P̃k Hk∗
(2.20)
I+
Hk P̃k Hk∗
−1
(2.21)
and
P̃k = I − γ −2 Pk L∗k Lk
−1
Pk ,
(2.22)
Proof: see [27].
Note that the condition in Eq. (2.18) is equivalent to
I − γ −2 Pk L∗k Lk > 0,
for k = 0, · · · , M
(2.23)
and hence P̃k in Eq. (2.22) is well defined. P̃k can also be defined as
P̃k−1 = Pk−1 − γ −2 L∗k Lk ,
for k = 0, · · · , M
(2.24)
2.4. H∞ -OPTIMAL SOLUTION
23
which proves useful in rewriting the prediction coefficient, Kp,k in Eq. (2.21), as
follows. First, note that
−1
−1
Fk P̃k Hk∗ I + Hk P̃k Hk∗
= Fk P̃k−1 + Hk∗ Hk
Hk∗
(2.25)
and hence, replacing for P̃k−1 from Eq. (2.24)
Kp,k = Fk Pk−1 − γ −2 L∗k Lk + Hk∗ Hk
−1
Hk∗
(2.26)
Theorems 2.1 and 2.2 (Sections 2.4.1 and 2.4.2) provide the form of the filtering and
prediction estimators, respectively. The following section investigates the optimal
value of γ for both of these solutions, and outlines the simplifications that follow.
2.4.3
The Optimal Value of γ
The optimal value of γ for the filtering solution will be discussed first. The discussion
of the optimal prediction solution utilizes the results in the filtering case.
2.4.3.1
Filtering Case
2.4.3.1.1 γopt ≤ 1: First, it will be shown that for the filtering solution γopt ≤ 1.
Using Eq. (2.11), one can always pick ŝ(k|k) to be simply m(k). With this choice
ŝ(k|k) − s(k) = Vm (k), for all k
(2.27)
and Eq. (2.11) reduces to
M
X
sup
Vm ∈ L2 , ξ0
Vm (k)∗ Vm (k)
k=0
ˆ
(ξ0 − ξˆ0 )∗ Π−1
0 (ξ0 − ξ0 ) +
M
X
(2.28)
Vm (k)∗ Vm (k)
k=0
which can never exceed 1 (i.e. γopt ≤ 1). A feasible solution for the H∞ estimation
problem in Eq. (2.11) is therefore guaranteed when γ is chosen to be 1. Note that
it is possible to directly demonstrate the feasibility of γ = 1. Using simple matrix
2.4. H∞ -OPTIMAL SOLUTION
24
manipulation, it can be shown that for Lk = Hk and for γ = 1, Rk and Re,k have the
same inertia for all k.
2.4.3.1.2 γopt ≥ 1: To show that γopt is indeed 1, an admissible sequence of disturbances and a valid initial condition should be constructed such that γ could be made
arbitrarily close to 1 regardless of the filtering solution chosen. The necessary and
sufficient conditions for the optimality of γopt = 1 are developed in the course of
constructing this admissible
sequence
of disturbances.
T
T
T
Assume that ξˆ0 = Ŵ0 θ̂0
is the best estimate for the initial condition of the
system in the approximate model of the primary path (Fig. 2.4). Moreover, assume
that θ̂0 is indeed the actual initial condition for the secondary path in Fig. 2.4. The
actual initial condition for the weight vector of the FIR filter in this approximate
model is W0 . Then,
h
m(0) =
H0 ξˆ0 =
h
Ds (0)h∗ (0) Cs (0)
Ds (0)h∗ (0) Cs (0)
i
i
"
"
W0
θ̂0
Ŵ0
#
+ Vm (0)
(2.29)
#
θ̂0
where m(0) is the (derived) measurement at time k = 0. Now, if
∗
Vm (0) = Ds (0)h (0) Ŵ0 − W0 = KV (0) Ŵ0 − W0
(2.30)
(2.31)
then m(0) − H0 ξˆ0 = 0 and the estimate of the weight vector will not change. More
specifically, Eqs. (2.16) and (2.17) reduce to the following simple updates
ξˆ1 = F0 ξˆ0
(2.32)
ŝ(0|0) = L0 ξˆ0
(2.33)
which given L0 = H0 generates the estimation error
ef,0 = ŝ(0|0) − s(0)
= L0 ξˆ0 − L0 ξ0
= Ds (0)h∗ (0) Ŵ0 − W0
= Vm (0)
(2.34)
2.4. H∞ -OPTIMAL SOLUTION
25
Repeating a similar argument at k = 1 and 2, it is easy to see that if
Vm (1) = [Ds (1)h∗ (1) + Cs (1)Bs (0)h∗ (0)] Ŵ0 − W0 = KV (1) Ŵ0 − W0
(2.35)
and
∗
∗
∗
Vm (2) = [Ds (2)h (2) + Cs (2)Bs (1)h (1) + Cs (2)As (1)Bs (0)h (0)] Ŵ0 − W0
= KV (2) Ŵ0 − W0
(2.36)
then
m(k) − Hk ξˆk = 0,
for k = 1, 2
(2.37)
Note that when Eq. (2.37) holds, and with Lk = Hk , Eq. (2.17) reduces to
ŝ(k|k) = Lk ξˆk = Hk ξˆk
(2.38)
and hence
ef,k = ŝ(k|k) − s(k)
= ŝ(k|k) − [m(k) − Vm (k)]
= Hk ξˆk − [m(k) − Vm (k)]
h
i
= Hk ξˆk − m(k) + Vm (k)
= Vm (k)
for k = 1, 2
(2.39)
Continuing this process, KV (k), for 0 ≤ k ≤ M can be defined as

KV (0)


Ds (0)
0
0
0
···
0

h(0)


 


 KV (1)   Cs (1)Bs (0)
Ds (1)
0
0 ···
0   h(1) 

 



  C (2)A (1)B (0) C (2)B (1) D (2) 0 · · ·


0
(2)
K
h(2)
=
 V
s
s
s
s
s
  s







..
..
..
.. 






.
.
.

 
 . 
..
KV (M)
h(M)
.
···
Ds (M)
4
= ∆M ΛM
(2.40)
2.4. H∞ -OPTIMAL SOLUTION
26
such that Vm (k), ∀ k, is an admissible disturbance. In this case, Eq. (2.11) reduces
to
M
X
sup
ξ0
Vm (k)∗ Vm (k)
k=0
ˆ
(ξ0 − ξˆ0 )∗ Π−1
0 (ξ0 − ξ0 ) +
M
X
Vm (k)∗ Vm (k)
k=0
M
X
#
"
=
(Ŵ0 − W0 )∗
sup
ξ0
KV∗ (k)KV (k)
k=0
"
∗
ˆ
(ξ0 − ξˆ0 )∗ Π−1
0 (ξ0 − ξ0 ) + (Ŵ0 − W0 )
M
X
(Ŵ0 − W0 )
#
KV∗ (k)KV (k)
(Ŵ0 − W0 )
k=0
(2.41)
From Eq. (2.40), note that
M
X
KV∗ (k)KV (k) = Λ∗M ∆∗M ∆M ΛM = k ∆M ΛM k22
(2.42)
k=0
and hence the ratio in Eq. (2.41) can be made arbitrarily close to one if
lim k∆M ΛM k2 → ∞
M →∞
(2.43)
Eq. (2.43) will be referred to as the condition for optimality of γ = 1 for the
filtering solution.
Equation (2.43) can now be used to derive necessary and sufficient conditions for
optimality of γ = 1. First, note that a necessary condition for Eq. (2.43) is
lim kΛM k2 → ∞
M →∞
(2.44)
or equivalently
lim
M →∞
M
X
h∗ (k)h(k) → ∞
(2.45)
k=0
The h(k) that satisfies the condition in (2.45) is referred to as exciting [26]. Several
sufficient conditions can now be developed. Since
k∆M ΛM k2 ≥ σmin (∆M ) kΛM k2
(2.46)
2.4. H∞ -OPTIMAL SOLUTION
27
one sufficient condition is that
σmin (∆M ) > ,
∀ M, and > 0
(2.47)
Note that for LTI systems, the sufficient condition (2.47) is equivalent to the requirement that the system have no zeros on the unit circle. Another sufficient condition
is that h(k)’s be persistently exciting, that is
#
"
M
1 X
lim σmin
h(k)h∗ (k) > 0
M →∞
M k=0
(2.48)
which holds for most reasonable systems.
2.4.3.2
Prediction Case
The optimal value for γ can not be less than one in the prediction case. In the previous
section we showed that despite using all available measurements up to and including
time k, the sequence of the admissible disturbances, Vm (k) = KV (k) Ŵ0 − W0 for
k = 0, · · · , M (where KV (k) is given by Eq. (2.40)), prevented the filtering solution
from achieving γ < 1. The prediction solution that uses only the measurements up to
time k (not including k itself) can not improve over the filtering solution and therefore
the energy gain γ is at least one.
Next, it is shown that if the initial condition P0 is chosen appropriately (i.e. if it is
small enough), then γopt = 1 can be guaranteed. Referring to the Lyapunov recursion
of Eq. (2.65), the Riccati matrix at time k can be written as:
#
"
!
!∗
k−1
k−1
Y
Y
I
0
Pk =
Fj P0
Fj ,
Fj =
Bs (j)h∗ (j) As (j)
j=0
j=0
(2.49)
Defining
ΨjA = As (j)As (j − 1) · · · As (0)
(2.50)
Eq. (2.49) can be written as
# "
#∗
"
I
0
I
0
Pk = Pk−1 j
P0 Pk−1 j
(2.51)
∗
k
∗
k
j=0 ΨA Bs (j)h (k−1−j) ΨA
j=0 ΨA Bs (j)h (k−1−j) ΨA
2.4. H∞ -OPTIMAL SOLUTION
28
From Theorem 2.2, Section 2.4.2, the condition for the existence of a prediction
solution is (I − γ −2 Pk L∗k Lk ) > 0, or equivalently
(γ 2 − Lk Pk L∗k ) > 0
(2.52)
Note that Lk = [ Ds (k)h∗ (k) Cs (k) ], and therefore Eq. (2.52) can be re-written as
#
"
h
i
∗
h(k)D
(k)
s
γ 2 − Ds (k)h∗ (k) Cs (k) Pk
>0
(2.53)
∗
Cs (k)
Replacing for Pk from Eq. (2.51), and carrying out the matrix multiplications, Eq. (2.53)
is equivalent to
γ2
"
−
"
h(k)Ds∗ (k) +
Pk−1
∗j ∗
∗
j=0 h(k−1−j)Bs (j)ΨA Cs (k)
∗
Ψ∗k
A Cs (k)
h(k)Ds∗ (k) +
Pk−1
j=0
∗
h(k−1−j)Bs∗ (j)Ψ∗j
A Cs (k)
∗
Ψ∗k
A Cs (k)
#∗
× P0 ×
#
>0
(2.54)
Introducing
0∗
∗
h (k) = Ds h (k) +
k−1
X
Cs (k)ΨjA Bs (j)h∗ (k−1−j)
(2.55)
j=0
as the filtered version of the reference vector, h(k), Eq. (2.54) can be expressed as
#
"
i
h
0
h
(k)
γ 2 − h0 ∗ (k) Cs (k)ΨkA P0
>0
(2.56)
∗
Ψ∗k
C
(k)
A s
Selecting the initial value of the Riccati matrix, without loss of generality, as
#
"
µI 0
(2.57)
P0 =
0 αI
and the Eq. (2.56) reduces to
∗
∗
γ 2 − µh0 (k)h0 (k) − αCs (k)ΨkA Ψ∗k
A Cs (k) > 0
(2.58)
It is now clear that a prediction solution for γ = 1 exists if
∗
1 − αCs (k)ΨkA Ψ∗k
A Cs (k)
µ<
h0 ∗ (k)h0 (k)
(2.59)
Equation (2.59) is therefore the condition for optimality of γopt = 1 for the prediction
solution.
2.4. H∞ -OPTIMAL SOLUTION
29
2.4.4
Simplified Solution Due to γ = 1
2.4.4.1
Filtering Case:
The following shows that with Hk = Lk and γ = 1, the Riccati equation (2.14)
is considerably simplified. To this end, apply the matrix inversion lemma, (A +
BCD)−1 = A−1 − A−1 B[C −1 + DA−1 B]−1 DA−1 , to
"
# "
#
h
i
0
Ip
Hk
Re,k =
+
Pk Hk∗ Hk∗
0 −Iq
Hk
"
with A =
Ip
0
#
"
,B=
Hk
#
h
, C = I, and D = Pk
(2.60)
i
Hk∗ Hk∗ . It is easy to
0 −Iq
Hk
−1
verify that the term DA B is zero. Therefore
# "
#
"
h
i
Hk
0
Ip
−1
∗
∗
−
Pk Hk −Hk
Re,k =
−Hk
0 −Iq
(2.61)
In which case,
h
∗
Kf,k Re,k Kf,k
=
Fk Pk
Hk∗ Hk∗
i
"
−1
Re,k
Hk
#
Hk
!
Pk Fk∗
=0
for γ = 1 and for all k. Thus the Riccati recursion (2.14) reduces to the Lyapunov
recursion Pk+1 = Fk Pk Fk∗ with P0 = Π0 > 0.
Partitioning the Riccati matrix Pk in block matrices conformable with the block
matrix structure of Fk , (2.14) yields the following simple update


P11,k+1 = P11,k , P11,0 = Π11,0









 P
= P A (k) + P h(k)B ∗ (k), P
=Π
12,k+1
12,k
s
11,k
s
12,0
12,0





∗

P22,k+1 = Bs (k)h(k)∗ P11,k h(k)Bs∗ (k) + As (k)P12,k
h(k)Bs∗ (k)+





Bs (k)h∗ (k)P12,k A∗s (k) + As (k)P22,k A∗s (k), P22,0 = Π22,0
The filtering solution can now be summarized in the following theorem:
(2.62)
2.5. IMPORTANT REMARKS
30
Theorem 2.3: Consider the system described by Equations (2.5)-(2.7), with Lk =
Hk . If the optimality condition (2.43) is satisfied, the H∞ -optimal filtering solution
achieves γopt = 1, and the central H∞ -optimal filter is given by
ξˆk+1 = Fk ξˆk + Kf,k m(k) − Hk ξˆk , ξˆ0 = 0
(2.63)
−1
ŝ(k|k) = Lk ξˆk + (Lk Pk Hk∗ ) RHe,k
m(k) − Hk ξˆk
(2.64)
−1
with Kf,k = (Fk Pk Hk∗ ) RHe,k
and RHe,k = Ip + Hk Pk Hk∗ , where Pk satisfies the Lyapunov recursion
Pk+1 = Fk Pk Fk∗ , P0 = Π0 .
(2.65)
Proof: follows from the discussions above.
2.4.4.2
Prediction Case:
Referring to Eq. (2.26), it is clear that for γ = 1 and for Lk = Hk , the coefficient
Kp,k will reduce to Fk Pk Hk∗ . Therefore, the prediction solution can be summarized
as follows:
Theorem 2.4: Consider the system described by Equations (2.5)-(2.7), with Lk =
Hk . If the optimality conditions (2.43) and (2.59) are satisfied, and with P0 as defined
in Eq. (2.57), the H∞ -optimal prediction solution achieves γopt = 1, and the central
filter is given by
ˆ
ˆ
ˆ
ξk+1 = Fk ξk + Kp,k m(k) − Hk ξk , ξˆ0 = 0
(2.66)
ŝ(k) = Lk ξˆk
(2.67)
with Kp,k = Fk Pk Hk∗ where Pk satisfies the Lyapunov recursion (2.65).
Proof: follows from the discussions above.
2.5
Important Remarks
The main idea in the EBAF algorithm can be summarized as follows. At a given time
k, use the available information on; (a) measurement history, e(i) for 0 ≤ i ≤ k, (b)
control history, u(i) for 0 ≤ i < k, (c) reference signal history, x(i) for 0 ≤ i ≤ k, (d)
2.5. IMPORTANT REMARKS
31
the model of the secondary path and the estimate of its initial condition, and (e) the
pre-determined length of the adaptive FIR filter to produce the best estimate of the
actual output of the primary path, d(k). The key premise is that if d(k) is accurately
estimated, then the inputs u(k) can be generated such that d(k) is canceled. The
objective of the EBAF algorithm is to make y(k) match the optimal estimate of d(k)
(see Fig. 2.3). For the adaptive filtering problem in Fig. 2.1 , however, adaptation
algorithm only has direct access to the weight vector of the adaptive FIR filter.
Because of this practical constraint, the EBAF algorithm adapts the weight vector in
the adaptive FIR filter according to the estimate of the optimal weight vector given
by Eqs. (2.63)
or (2.66) (for
the filtering, or prediction solutions, respectively). Note
T
T
T
that ξˆ = Ŵ (k) θ̂ (k) . The error analysis for this adaptive algorithm is discussed
k
in Section 2.7. Now, main features of this algorithm can be described as follows:
1. The estimation-based adaptive filtering (EBAF) algorithm yields a solution that
only requires one Riccati recursion. The recursion propagates forward in time,
and does not require any information about the future of the system or the
reference signal (thus allowing the resulting adaptive algorithm to be real-time
implementable). This has come at the expense of restricting the controller to
an FIR structure in advance.
∗
2. With Kf,k Re,k Kf,k
= 0, Pk+1 = Fk Pk Fk∗ is the simplified Riccati equation,
which considerably reduces the computational complexity involved in propagating the Riccati matrix. Furthermore, this Riccati update always generates
a non-negative definite Pk , as long as P0 is selected to be positive definite (see
Eq. (2.65)).
3. In general, the solution to an H∞ filtering problem requires verification of the
fact that Rk and Re,k are of the same inertia at each step (see Eq. (2.13)). In a
p
similar way, the prediction solution requires that all sub-matrices of Rkp and Re,k
have the same inertia for all k (see Eq. (2.18)). This can be a computationally
expensive task. Moreover, it may lend to a breakdown in the solution if the
condition is not met at some time k. The formulation of the problem eliminates
2.6. IMPLEMENTATION SCHEME FOR EBAF ALGORITHM
32
the need for such checks, as well as the potential breakdown of the solution, by
providing a definitive answer to the feasibility and optimality of γ = 1.
4. When [ As (k), Bs (k), Cs (k), Ds (k) ] = [ 0, 0, 0, I ] for all k, (i.e. the output
of the FIR filter directly cancels d(k) in Figure 2.1), then the filtering/prediction
results reduce to the simple Normalized-LMS/LMS algorithms in Ref. [26] as
expected.
5. As mentioned earlier, there is no need to verify the solutions at each time step,
so the computational complexity of the estimation based approach is O(n3 )
(primarily for calculating Fk Pk FK∗ ), where
n = (N + 1) + Ns
(2.68)
where (N +1) is the length of the FIR filter, and Ns is the order of the secondary
path. The special structure of Fk however reduces the computational complexity
to O(Ns3 + Ns N), i.e. cubic in the order of the secondary path, and linear in the
length of the FIR filter (see Eq. (2.62)). This is often a substantial reduction
in the computation since Ns N. Note that the computational complexity for
FxLMS is quadratic in Ns and linear in N.
2.6
Implementation Scheme for EBAF Algorithm
Three sets of variables are used to describe the implementation scheme:
1. Best available estimate ofa variable: Referring
to Eqs. (2.16) and (2.19), and
noting the fact that ξˆT = Ŵ T (k) θ̂T (k) , Ŵ (k) can be defined as the estimate
k
of the weight vector, and θ̂(k) as the secondary path state estimate in the
approximate model of the primary path.
4
2. Actual value of a variable: Referring to Fig. 2.1, define u(k) = h∗ (k)Ŵ (k) as the
actual input to the secondary path, y(k) as the actual output of the secondary
path, and d(k) as the actual output of the primary path. Note that d(k) and
2.6. IMPLEMENTATION SCHEME FOR EBAF ALGORITHM
33
y(k) are not directly measurable, and that at each iteration the weight vector
in the adaptive FIR filter is set to Ŵ (k).
3. Adaptive algorithm’s internal copy of a variable: Recall that in Eq. (2.4), y(k)
is used to construct the derived measurement m(k). Since y(k) is not directly
available, the adaptive algorithm needs to generate an internal copy of this
variable. This internal copy (referred to as ycopy (k)) is constructed by applying
u(k) (the actual control signal) to a model of the secondary path inside the
adaptive algorithm. The initial condition for this model is θcopy (0). In other
words, the derived measurement is constructed as follows
m(k) = e(k) + ycopy (k)
(2.69)
θcopy (k + 1) = As (k)θcopy (k) + Bs (k)u(k)
(2.70)
ycopy (k) = Cs (k)θcopy (k) + Ds (k)u(k)
(2.71)
where
Given the identified model for the secondary path and its input u(k) = h∗ (k)Ŵ (k),
the adaptive algorithm’s copy of y(k) will be exact if the actual initial condition of the secondary path is known. Obviously, one can not expect to have the
exact knowledge of the actual initial condition of the secondary path. In the
next section, however, it is shown that when the secondary path is linear and
stable, the contribution of the initial condition to its output decreases to zero
as k increases. Therefore, the internal copy of y(k) will converge to the actual
value of y(k) over time.
Now, the implementation algorithm can be outlined as follows;
1. Start with Ŵ (0) = 0 and θ̂(0) = 0 as the initial guess for the state vector in the
approximate model of the primary path. Also assume that θcopy (0) = 0, and
h(0) = [ x(0) 0 · · · 0 ]T . The initial value for the Riccati matrix is P0 which
is chosen to be block diagonal. The role of P0 is similar to the learning rate in
LMS-based adaptive algorithms (see Section 5.3.2).
2.6. IMPLEMENTATION SCHEME FOR EBAF ALGORITHM
34
2. If 0 ≤ k ≤ M (finite horizon):
(a) Form the control signal
u(k) = h∗ (k)Ŵ (k)
(2.72)
to be applied to the secondary path. Note that applying u(k) to the
secondary path produces
y(k) = Cs (k)θ(k) + Ds (k)u(k)
(2.73)
at the output of the secondary path. This in turn leads to the following
error signal measured at time k:
e(k) = d(k) − y(k) + Vm (k)
(2.74)
which is available to the adaptive algorithm to perform the state update
at time k.
(b) Propagate the state estimate and the internal copy of the state of the
secondary path as follows
 "
# 
Ŵ (k + 1)



 =
θ̂(k + 1)


θcopy (k + 1)
i
h
"
# h
Kf,k Cs (k)
Fk + Kf,k [0 − Cs (k)]
Kf,k
Ŵ (k)




 θ̂(k) +



h
i
∗
θcopy (k)
As (k)
0
(Bs (k)h (k) 0
i

e(k)

(2.75)
where e(k) is the error sensor measurement at time k given by Eq. (2.74),
and Kf,k = Fk Pk Hk∗ (I + Hk Pk Hk∗ )−1 (see Theorem 2.3). Note that for the
prediction-based EBAF algorithm Kf,k should be replaced with Kp,k =
Fk Pk Hk∗ .
2.7. ERROR ANALYSIS
35
(c) update the Riccati matrix Pk using the Lyapunov recursion
"
#
P11
P12,k+1
=
∗
P12,k+1
P22,k+1
"
#"
#"
I
0
P11 P12,k
I
Bs (k)h∗ (k) As (k)
∗
P12,k
P22,k
0
#∗
Bs (k)h∗ (k) As (k)
(2.76)
Pk+1 will be used in (2.75) to update the state estimate.
3. Go to 2.
2.7
Error Analysis
In Section 2.6, it is pointed out that the proposed implementation scheme can deviate
from an H∞ -optimal solution for two main reasons:
1. The error in initial condition of the secondary path which can cause ycopy to be
different from y(k).
2. The additional error in the cancellation of d(k) due to the fact that y(k) can
not be set to ŝ(k|k) (or ŝ(k)). All one can do is to set the weight vector in the
adaptive FIR filter to be Ŵ (k).
Here, both errors are discussed in detail.
2.7.1
Effect of Initial Condition
As earlier discussions indicate, the secondary path in Fig. 2.1 is assumed to be linear.
For a linear system the output at any given time can be decomposed into two components: the zero-input component which is associated with the portion of the output
solely due to the initial condition of the system, and the zero-state component which
is the portion of the output solely due to the input to the system.
2.7. ERROR ANALYSIS
36
For a stable system, the zero-input component of the response will decay to zero
for large k. Therefore, any difference between ycopy (k) and y(k) (which with a known
input to the secondary path can only be due to the unknown initial condition) will
go to zero as k grows. In other words, exact knowledge of the initial condition of the
secondary path does not affect the performance of the proposed EBAF algorithm for
sufficiently large k.
2.7.2
Effect of Practical Limitation in Setting y(k) to ŝ(k|k)
(ŝ(k))
As pointed out earlier, the physical setting of the adaptive control problem in Fig. 2.1
only allows for the weight vector in the adaptive FIR filter to be adjusted to Ŵ (k).
In other words, the state of the secondary path can not be set to a desired value at
each step. Instead, θk evolves based on its initial condition and the control input,
u(k), that we provide. Assume that θ(k) is the actual state of the secondary path at
time k. The actual output of the secondary path is then
y(k) = Ds (k)h∗ (k)Ŵ (k) + Cs (k)θ(k)
(2.77)
which leads to the following cancellation error
d(k) − y(k) = d(k) − Ds (k)h∗ (k)Ŵ (k) + Cs (k)θ(k)
(2.78)
For the prediction solution of Theorem 2.4, adding the zero quantity ±Cs (k)θ̂(k) to
the right hand side of Equation (2.78), and taking the norm of both sides,
kd(k) − y(k)k = k d(k) − Ds (k)h∗ (k)Ŵ (k) + Cs (k)θ(k) ± Cs (k)θ̂(k) k
= k d(k) − Ds (k)h∗ (k)Ŵ (k) − Cs (k)θ̂(k) + Cs (k) θ̂(k) − θ(k) k
Therefore,
k Cs (k) θ̂(k) − θ(k) k
kd(k) − y(k)k
k d(k) − ŝ(k) k
≤
+
M
M
M
X
X
X
∗
∗ −1 ˜
∗
∗ −1 ˜
∗
˜0 +
˜
˜
V
(k)V
(k)
Π
+
V
(k)V
(k)
Π
+
Vm
(k)Vm (k)
ξ˜0∗ Π−1
ξ
ξ
ξ
ξ
ξ
m
0
m
0
0
m
0 0
m
0 0
k=0
k=0
k=0
(2.79)
2.7. ERROR ANALYSIS
37
where ξ˜0 = (ξ0 − ξˆ0 ) and ξk is defined in Eq. (2.5). Note that the first term in the right
hand side of Eq. (2.79) is the prediction error energy gain (see Eq. (2.12)). Therefore,
the energy gain of the cancellation error with the prediction-based EBAF exceeds the
error energy gain of the H∞ optimal prediction solution by the second term on the
right hand side of Eq. (2.79). It can be shown that when the primary inputs h(k)
are persistently exciting (see Eq. (2.48)), the dynamics for the state estimation error,
θ̂(k)−θ(k), are internally stable which implies that the second term on the right hand
side of Eq. (2.79) is bounded for all M, and in the limit when M → ∞∗ .
When Ds (k) = 0 for all k, an implementation of the filtering solution that utilizes
the most recent measurement, m(k), is feasible. In this case, the filtering solution in
Eqs. (2.16)-(2.17) can be written as follows:
ξˇk|k = ξˆk + Pk Hk∗ (Ip + Hk Pk Hk∗ )−1 m(k) − Hk ξˆk
(2.80)
ξˆk+1 = Fk ξˇk|k
(2.81)
ŝ(k|k) = Lk ξˇk|k
(2.82)
where the weight vector update in the adaptive FIR filter follows Eq. (2.80). With
a derivation identical to the one for prediction solution, it can be shown that the
performance bound in this case is
k Cs (k) θ̂(k|k) − θ(k) k
kd(k) − y(k)k
k d(k) − ŝ(k|k) k
≤
+
M
M
M
X
X
X
∗ −1 ˜
∗
∗ −1 ˜
∗
∗ −1 ˜
∗
˜
˜
˜
ξ0 Π0 ξ0 +
Vm (k)Vm (k) ξ0 Π0 ξ0 +
Vm (k)Vm (k) ξ0 Π0 ξ0 +
Vm
(k)Vm (k)
k=0
k=0
k=0
(2.83)
An argument similar to the prediction case shows that the second term on the right
hand side has a finite gain as well.
∗
Reference [24] shows that if the exogenous disturbance is assumed to be a zero mean white
noise process with unit intensity, and independent of the initial condition of the system ξ0 , then the
terminal state estimation error variance satisfies
E(ξk − ξ̂k )(ξk − ξˆk )∗ ≤ Pk
2.8. RELATIONSHIP TO THE NORMALIZED-FXLMS/FXLMS ALGORITHMS
2.8
38
Relationship to the Normalized-FxLMS/FxLMS
Algorithms
In this section, it will be shown that as k → ∞, the gain vector in the predictionbased EBAF algorithm converges to the gain vector in the classical Filtered-X LMS
(FxLMS) algorithm. Thus, FxLMS is an approximation to the steady-state EBAF.
The error terms in the two algorithms are shown to be different (compare Eqs. (2.89)
and (2.2)). Therefore it is expected that the prediction-based EBAF demonstrate
superior transient performance compared to the FxLMS algorithm. Simulation results in the next section agree with this expectation. The fact that the gain vectors
asymptotically coincide, agrees with the fact that the derivation of the FxLMS algorithm relies on the assumption that the adaptive filter and the secondary path are
interchangeable which can only be true in the steady state. Similar results are shown
for the connection between the filtering-based EBAF and the Normalized FxLMS
adaptive algorithms.
For the discussion in this section, the secondary path is assumed, for simplicity, to
be LTI, i.e. [ As , Bs , Cs , Ds ]. Note that for the LTI system, ΨkA in Eq. (2.50) reduces
to Aks . The Riccati matrix Pk in Eq. (2.51) can then be rewritten as
"
# "
#∗
I
0
I
0
Pk = Pk−1 j
P0 Pk−1 j
∗
k
∗
k
j=0 As Bs h (k−1−j) As
j=0 As Bs h (k−1−j) As
(2.84)
Eq. (2.84) will be used in establishing the proper connections between the filtered/predicted
solutions of Section 2.4 and the conventional Normalized-FxLMS/FxLMS algorithms.
2.8.1
Prediction Solution and its Connection to the FxLMS
Algorithm
To study the asymptotic behavior of the state estimate update, note that for an stable
secondary path Aks → 0 as k → ∞. Therefore, using Eq. (2.84)
# "
#∗
"
I
0
I
0
Pk → Pk−1 j
P0 Pk−1 j
as k → ∞(2.85)
∗
∗
j=0 As Bs h (k−1−j) 0
j=0 As Bs h (k−1−j) 0
2.8. RELATIONSHIP TO THE NORMALIZED-FXLMS/FXLMS ALGORITHMS
"
which for P0 =
"
Pk → Pk−1
P11 (0) P12 (0)
#
results in
P21 (0) P22 (0)
I
j
∗
j=0 As Bs h (k−1−j)
39
#
"
P11 (0) Pk−1
I
j
∗
j=0 As Bs h (k−1−j)
#∗
as k → ∞ (2.86)
Selecting P11 (0) = µI as in Eq. (2.57), and noting the fact that Kp,k = Fk Pk Hk∗
(Theorem 2.4), it is easy to see that as k → ∞
#
"
!∗
k−1
X
I
∗
j
∗
Kp,k → µ Pk
Ds h (k) +
Cs As Bs h (k−1−j)
j
∗
j=0 As Bs h (k−j)
j=0
#
"
I
h0 (k)
→ µ Pk
(2.87)
j
∗
A
B
h
(k−j)
s
j=0 s
and therefore the state estimate update in Theorem 2.4 becomes
# "
#
"
#"
Ŵ (k+1)
I
0
Ŵ (k)
=
+
θ̂(k+1)
Bs h∗ (k) As
θ̂(k)
"
#
h0 (k)
∗
µ Pk
m(k) − Ds h (k)Ŵ (k) − Cs θ̂(k) (2.88)
j
∗
0
j=0 As Bs h (k−j)h (k)
Thus,the following update for the weight vector is derived
∗
Ŵ (k+1) = Ŵ (k) + µh0 (k) m(k) − Ds h∗ (k)Ŵ (k) − Cs θ̂(k)
(2.89)
Note that m(k) = e(k) + ycopy (k) (see Eq. (2.69)), and hence the difference between
the limiting update rule of Eq. (2.89) (i.e. the prediction EBAF algorithm), and the
classical FxLMS algorithm of Eq. (2.2) will be the error term used by these algorithms.
More specifically, e(k) in the FxLMS algorithm is replaced with the following modified
error (using Eq. (2.71)):
e(k) + ycopy (k) − Ds h∗ (k)Ŵ (k) − Cs θ̂(k) = e(k) + Cs θcopy (k) − Cs θ̂(k).
Note that if y(k) is directly measurable, then the modified error will be
h
i
e(k) + y(k) − Ds h∗ (k)Ŵ (k) − Cs θ̂(k)
(2.90)
2.8. RELATIONSHIP TO THE NORMALIZED-FXLMS/FXLMS ALGORITHMS
40
The condition for optimality of γ = 1 in the prediction case (see Eq. (2.59)), can
also be simplified for stable LTI secondary path as k → ∞. Rewriting the optimality
condition for the prediction solution, Eq. (2.59), as
µ<
∗
1 − αCs Aks A∗k
s Cs
h0 ∗ (k)h0 (k)
(2.91)
for a stable secondary path, Aks → 0 as k → ∞, and hence
µ<
1
h0 ∗ (k)h0 (k)
as k → ∞
(2.92)
is the limiting condition for the optimality of γ = 1 in the prediction case. This is
essentially a filtered version of the well known LMS bound [26].
2.8.2
Filtering Solution and its Connection to the NormalizedFxLMS Algorithm
In the Filtered case the gain vector is Kf,k = Fk Pk Hk∗ (I + Hk Pk Hk∗ ). In Section 2.8.1
the limiting value for the quantity Fk Pk Hk∗ in Eq. (2.87) is computed. In a similar
way it can be shown that, with P11 (0) = µI, as K → ∞,
(I + Hk Pk Hk∗ ) → (1 + µh0∗ (k)h0 (k))
(2.93)
and hence the coefficient for the state estimate update in the filtering case becomes
!
h0 (k)
µ
Kf,k →
as k → ∞
(2.94)
Pk
j
∗
0
1 + µh0∗ (k)h0 (k)
j=0 As Bs h (k−j)h (k)
Thus the update rule for the weight vector in the filtering EBAF algorithm would be
Ŵ (k+1) = Ŵ (k) + µ
h0 ∗ (k)
∗
θ̂(k)
(2.95)
h
(k)
Ŵ
(k)
−
C
m(k)
−
D
s
s
(1 + µh0∗ (k)h0 (k))
which is similar to the Normalized-FxLMS algorithm (Eq. (2.3)) in which the error
signal is replaced with a modified error signal described by Eq. (2.90).
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
2.9
41
Experimental Data & Simulation Results
This section examines the performance of the proposed EBAF algorithm for the active noise cancellation (ANC) problem in a one dimensional acoustic duct. Figure 2.5
shows the schematic diagram of the one-dimensional air duct that is used in the experiments. The control objective is to attenuate (cancel in the ideal case) the disturbance
introduced into the duct by Speaker #1 (primary noise source) at the position of Microphone #2 (error sensor) by the control signal generated by Speaker #2 (secondary
source). Microphone #1 can be used to provide the reference signal for the adaptation
algorithm. Clearly, Microphone #1 measurements are affected by both primary and
secondary sources, and hence if these measurements are used as the reference signal
the problem, commonly known as feedback contamination, has to be addressed.
A dSPACE DS1102 DSP controller board (which includes TI’s C31 DSP processor
with 60 MHz clock rate, and 128k of 32-bit RAM), and its Matlab 5 interface are
used for real time implementation of the algorithm. A state space model (of order 10)
is identified for this one-dimensional acoustic system. Note that of the four identified
transfer functions, only the transfer function from Speaker #2 to Microphone #2 (i.e.
the secondary path) is required by the estimation-based adaptive algorithm. Figs.
2.6 and 2.7 show the identified transfer function for the one-dimensional duct. This
section will first provide experimental data that validate a corresponding simulation
result. More sophisticated experiments and simulations are then presented to study
various aspects of the EBAF algorithm.
Figure 2.8 shows the experimental data in a typical noise cancellation scenario,
along with corresponding plots from a simulation that is designed to mimic that
experiment. Here, the reading of Microphone #2 (i.e. the cancellation error) is
shown when an adaptive FIR filter of length 4 is used for noise cancellation. The
primary source is a sinusoidal tone at 150 Hz, which is also available to the adaptation
algorithm as the reference signal. A band-limited white noise (noise power = 0.008)
is used as the measurement noise for the simulation in Figure 2.8. The sampling
frequency is 1000 Hz for both experiment and simulation. P11,0 = 0.05I4×4 , P12,0 = 0,
and P22,0 = 0.005I10×10 are used to initialize the Riccati matrix in Eq. (2.14). The
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
42
experiment starts with adaptive controller off, and about 3 seconds later the controller
is turned on. The transient response of the adaptive FIR filter lasts for approximately
0.05 seconds. There is a 60 times reduction in the magnitude of the error. Therefore,
with the full access to the single tone primary disturbance, the EBAF algorithm
provides a fast and effective noise cancellation. The results from a corresponding
Matlab simulation (with the same filter length, and similar open loop error at 150
Hz) are also shown in Fig. 2.8. The transient behavior and the steady state response
in the simulation agree with the experimental data, thus assuring the validity of the
set up for the simulations presented in this chapter.
Figures 2.9, 2.10, and 2.11 show the results of more noise cancellation experiments
in the one dimensional acoustic duct. In all these three experiments, the reference
signal available to the adaptation scheme is formed such that a considerable level of
uncorrelated additive white noise corrupts the primary disturbance. This is done to
examine the robustness of the EBAF algorithm in the case where clean access to the
primary disturbance is not possible. In practice however, efforts are made to produce
as clean a reference signal as possible.
In Figure 2.9 the primary disturbance is a sinusoid of amplitude 0.3 volts at 150
Hz. The reference signal used by the EBAF algorithm is subject to a band limited
white Gaussian noise, and the signal to noise ratio is approximately 3. For this
experiment the length of the adaptive filter is 4, and the controller is turned on
at t = 4.7 seconds. A 35 times reduction in the magnitude of the disturbance at
Microphone #2 is measured. While there is a reduction of magnitude 3.5 in the
error at Microphone #1, this is achieved as a byproduct of the noise cancellation at
Microphone #2 (i.e. noise cancellation at Microphone #1 is not an objective of the
single-channel noise cancellation attempt in this experiment). Chapter 4 will address
the multi-channel noise cancellation in detail. Note that the magnitude of noise
cancellation in this case is lower than the cancellation achieved in Figure 2.8 where
the reference signal was fully known to the adaptive algorithm. This observation
confirms the well known fact that the quality of the reference signal (i.e. its degree
of correlation to the primary disturbance) profoundly affects the performance of the
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
43
adaptive algorithm [51,33]. Figure 2.10 shows the experimental results for multitone noise cancellation where a noisy reference signal is available to the adaptive
algorithm. Here the number of taps in the adaptive filter is 8, and the primary
disturbance consists of two sinusoids at 150 and 180 Hz. The signal to noise ratio for
the available reference signal in this case is approximately 4.5. A 16 times reduction
in the amplitude of the disturbances at Microphone #2 is recorded. Some reduction
in the amplitude of the noise at Microphone #1 is also recorded. Note that, the sole
objective of the EBAF algorithm, in this case, is to cancel the noise at the position
of Microphone #2, and therefore no attempt is made to reduce the disturbances at
Microphone #1.
Figure 2.11 shows the results of the EBAF algorithm in the case where the primary
disturbance is a band limited white noise. As in Figures 2.9 and 2.10, only a noisy
measurement of the primary disturbance is available to the adaptive algorithm. The
signal to noise ratio in this case is approximately 4.5. The length of the adaptive
filter in this case is 16, and a reduction of approximately 3 times in the measurements
of Microphone #2 is achieved. For a better performance, the number of taps for the
adaptive FIR filter should be increased. Hardware limitations, however, prevented
experiments with higher order FIR filters. The performance of the EBAF algorithm
with longer FIR filters is examined through simulations in the rest of this section.
In Figure 2.12, the effect of feedback contamination (i.e. the contamination of the
reference signal with the output of the adaptive FIR filter through some feedback
path) when the primary source is a single tone is studied in simulation. In [33] the
subject of feedback contamination is discussed in detail, where relevant references
to the conventional solutions to this problem are also listed. Here, however, the
objective is to show that the proposed EBAF algorithm maintains superior performance (compared to FxLMS and normalized-FxLMS (NFxLMS) algorithms) when
such a problem occurs and no additional information is furnished. Fig. 2.12 contains
a typical response to feedback contamination for EBAF, FxLMS and NFxLMS algorithms. For the first 5 seconds, the input to Speaker #2 is grounded (i.e. u(k) = 0
for k ≤ 5). Switching the controller on results in large transient behavior in the
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
44
case of FxLMS and NFxLMS while, for the EBAF algorithm, the transient behavior does not display the undesirable overshoot. Different operation scenarios (with
various filter lengths, and adaptation rates) were tested, and this observation holds
true in all cases. For the next 15 seconds, the primary source is directly available
to all adaptive algorithms and the steady-state performance (left plots) is virtually
the same. From k = 20 on (right plots), the output of Microphone #1 (which is
contaminated by the output of the FIR filter) is used as the reference signal. Once
again, Fig. 2.12 shows a typical result. Note that in the case of FxLMS and NFxLMS
the adaptation rate must be kept small enough to avoid unstable behavior when the
switch to contaminated reference signal takes place. The EBAF algorithm allows for
faster convergence in the face of feedback contamination. For the results in Fig. 2.12,
the length of the adaptive FIR filter (for all three algorithms) is 24. For the EBAF
algorithm P11,0 = 0.005I24×24 , P22,0 = 0.0005I10×10 , and P12,0 = 0. For FxLMS and
NFxLMS algorithms, the adaptation rates are 0.005 and 0.025, respectively.
Figure 2.13 considers the effect of feedback contamination in a wide-band (10−500
Hz) noise cancellation process. For the results in Fig. 2.13, the length of the adaptive
FIR filter is 32. For the EBAF algorithm P11,0 = 0.05I32×32 , P22,0 = 0.005I10×10 ,
and P12,0 = 0. For FxLMS and NFxLMS algorithms the adaptation rates are 0.0005
and 0.01, respectively. The FxLMS algorithm becomes unstable for faster adaptation
rates, hence forcing slow convergence (i.e. lower control bandwidth). For NFxLMS,
the normalization of the adaptation rate by the norm of the reference vector (a vector
of length 32 in this case) prevents unstable behavior. The response of the algorithm
under feedback contamination is however still slower than EBAF algorithm. Furthermore, the oscillations in cancellation error due to the switching between modes of
operation are significantly higher when compared to the oscillations in EBAF case.
Figure 2.14 shows the closed-loop (i.e. the transfer function from the primary disturbance source at Speaker #1 to Microphone #2 during steady-state operation of
the adaptive control algorithm) performance comparison for wide-band noise cancellation. The EBAF algorithm outperforms FxLMS and Normalized-FxLMS adaptive
algorithms, even though the same level of information is made available to all three
adaptation schemes. For the result presented here the length of the FIR filter (for
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
45
all three approaches) is 32, and the band-limited white noise which is used as the
primary source is available as the reference signal. Since the frequency response is
calculated based on the steady-state data, the adaptation rate of the algorithms is
not relevant. Measurement noise for all three simulations is a band-limited white
noise with power 0.008, which is found to result in a steady-state attenuation that is
consistent with the experiments.
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
46
Physical Plant
n(k)
Vm (k)
d(k)
Primary Path
(Unknown)
+
+
+
−
Adaptive FIR Filter
x(k)
...
z −1
W0
W1
Secondary
Path
(Known)
WN
+
u(k)
y(k)
Update Weight Vector
Adaptation
Algorithm
Digital Control System
Fig. 2.1: General block diagram for an Active Noise Cancellation (ANC) problem
e(k)
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
x(k)
47
Primary Path
(Unknown)
P (z)
d(k)
+
e(k)
+
Adaptive FIR
Filter
u(k)
Secondary Path
(known)
S(z)
−
y(k)
A Copy Of
(known)
Secondary Path
LMS
Adaptation
Algorithm
x0 (k)
Fig. 2.2: A standard implementation of FxLMS algorithm
Primary Path
Modeling Error
x(k)
A Copy Of
Secondary Path
An FIR Filter
Secondary Path
Adaptive FIR Filter
u(k)
Vm (k)
d(k)
+
+
+
−
+
e(k)
y(k)
Adaptation Algorithm
Fig. 2.3: Pictorial representation of the estimation interpretation of the adaptive control problem:
Primary path is replaced by its approximate model
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
x(k)
z −1
W0
x(k − 1)
W1
x(k − N)
WN
Replica of the FIR Filter
Vm (k)
Ds (k)
Bs (k)
u(k)
+
48
+
z −1
Cs (k)
+
+
m(k)
As (k)
Replica of the Secondary Path
Fig. 2.4: Block diagram for the approximate model of the primary path
Microphone #1
Speaker #1
Microphone #2
Speaker #2
Fig. 2.5: Schematic diagram of one-dimensional air duct
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
Speaker #1 → Microphone #1
Speaker #2 → Microphone #1
1
1
10
Magnitude
Magnitude
10
0
10
−1
10
−2
10
49
0
10
−1
10
−2
1
10
10
2
10
1
10
Frequency (Hz)
2
10
Frequency (Hz)
400
0
Phase (Deg.)
Phase (Deg.)
200
0
−200
−400
−600
−800
−1000
1
10
−500
−1000
−1500
−2000
1
10
2
10
Frequency (Hz)
2
10
Frequency (Hz)
Fig. 2.6: Transfer functions plot from Speakers #1 & #2 to Microphone #1
Speaker #1 → Microphone #2
Speaker #2 → Microphone #2
1
1
10
Magnitude
Magnitude
10
0
10
−1
10
−2
10
0
10
−1
10
−2
1
10
10
2
10
1
10
Frequency (Hz)
2
10
Frequency (Hz)
400
200
Phase (Deg.)
Phase (Deg.)
0
−500
−1000
−1500
−2000
1
10
2
10
Frequency (Hz)
0
−200
−400
−600
−800
−1000
1
10
2
10
Frequency (Hz)
Fig. 2.7: Transfer functions plot from Speakers #1 & #2 to Microphone #2
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
Cancellation Error
50
Steady-State Behavior
Transient Behavior
e(k) - Simulation
0.02
0.6
0.6
0.4
0.4
0.2
0.2
0
0
−0.2
−0.2
−0.4
−0.4
−0.6
−0.6
−0.8
2
3
4
−0.8
2.9
0.01
0
−0.01
2.95
3
3.05
−0.02
4
6
8
10
4
6
8
10
e(k) - Experiment
0.02
0.6
0.6
0.4
0.4
0.2
0.2
0
0
−0.2
−0.2
−0.4
−0.4
−0.6
−0.6
−0.8
2
3
Time (sec.)
4
−0.8
2.9
0.01
0
−0.01
2.95
3
3.05
Time (sec.)
−0.02
Time (sec.)
Fig. 2.8: Validation of simulation results against experimental data for the noise cancellation
problem with a single-tone primary disturbance at 150 Hz. The primary disturbance is known to
the adaptive algorithm. The controller is turned on at t ≈ 3 seconds.
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
51
Experimental Data for a Sinusoidal Primary Disturbance of 150 Hz
Error at Microphone #1
0.5
0
−0.5
0
2
4
6
8
10
12
14
16
18
20
0
2
4
6
8
10
12
14
16
18
20
Error at Microphone #2
0.5
0
−0.5
Time (sec.)
Fig. 2.9: Experimental data for the EBAF algorithm of length 4, when a noisy measurement of
the primary disturbance (a single-tone at 150 Hz) is available to the adaptive algorithm (SNR=3).
The controller is turned on at t ≈ 5 seconds.
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
52
Experimental Data for a Multi-Tone Sinusoidal Primary Disturbance of 150&180 Hz
Error at Microphone #1
0.5
0
−0.5
0
5
10
15
20
25
30
0
5
10
15
20
25
30
Error at Microphone #2
0.5
0
−0.5
Time (sec.)
Fig. 2.10: Experimental data for the EBAF algorithm of length 8, when a noisy measurement
of the primary disturbance (a multi-tone at 150 and 180 Hz) is available to the adaptive algorithm
(SNR=4.5). The controller is turned on at t ≈ 6 seconds.
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
53
Experimental Data for a Band Limited White Noise
0.5
0.4
0.3
Error at Microphone #2
0.2
0.1
0
−0.1
−0.2
−0.3
−0.4
−0.5
0
5
10
15
20
25
30
Time (sec.)
Fig. 2.11: Experimental data for the EBAF algorithm of length 16, when a noisy measurement
of the primary disturbance (a band limited white noise) is available to the adaptive algorithm
(SNR=4.5). The controller is turned on at t ≈ 5 seconds.
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
No Feedback Contamination
With Feedback Contamination
1
1
EBAF
0
−0.5
0
2
4
6
8
−1
20
10
FxLMS
e(k)
e(k)
24
26
28
30
FxLMS
0.5
0
−0.5
0
−0.5
0
2
4
6
8
−1
20
10
1
22
24
26
28
30
1
NFxLMS
NFxLMS
0.5
e(k)
0.5
e(k)
22
1
0.5
0
−0.5
−1
0
−0.5
1
−1
EBAF
0.5
e(k)
e(k)
0.5
−1
54
0
−0.5
0
2
4
6
Time (sec.)
8
10
−1
20
22
24
26
28
30
Time (sec.)
Fig. 2.12: Simulation results for the performance comparison of the EBAF and (N)FxLMS
algorithms. For 0 ≤ t ≤ 5 seconds, the controller is off. For 5 < t ≤ 20 seconds both adaptive
algorithms have full access to the primary disturbance (a single-tone at 150 Hz). For t ≥ 20 seconds
the measurement of Microphone #1 is used as the reference signal (hence feedback contamination
problem). The length of the FIR filter is 24.
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
55
Cancellation Error - Band-Limited-White-Noise Primary Source
2
EBAF
e(k)
1
0
−1
−2
0
10
20
30
40
50
60
70
80
2
100
FxLMS
1
e(k)
90
0
−1
−2
0
10
20
30
40
50
60
70
80
90
100
2
NFxLMS
e(k)
1
0
−1
−2
0
10
20
30
40
50
60
70
80
90
100
Time (sec.)
Fig. 2.13: Simulation results for the performance comparison of the EBAF and (N)FxLMS
algorithms. For 0 ≤ t ≤ 5 seconds, the controller is off. For 5 < t ≤ 40 seconds both adaptive
algorithms have full access to the primary disturbance (a band limited white noise). For t ≥
40 seconds the measurement of Microphone #1 is used as the reference signal (hence feedback
contamination problem). The length of the FIR filter is 32.
2.9. EXPERIMENTAL DATA & SIMULATION RESULTS
56
Transfer Function from Speaker #1 to Microphone #2
1
10
Open-Loop
EBAF
FxLMS
NFxLMS
0
Magnitude
10
−1
10
−2
10
1
10
2
10
Frequency (Hz)
Fig. 2.14: Closed-loop transfer function based on the steady state performance of the EBAF
and (N)FxLMS algorithms in the noise cancellation problem of Figure 2.13.
2.10. SUMMARY
2.10
57
Summary
The adaptive control problem has been approached from an estimation point of view.
More specifically, it has been shown that for a common formulation of the adaptive
control problem an equivalent estimation interpretation exists. Then, a standard H∞
estimation problem has been constructed that corresponds to the original adaptive
control problem, and have justified the choice of estimation criterion. The H∞ optimal filtering/prediction solutions have also been derived, and it has been proved
that the optimal energy gain is unity. The filtering/prediction solutions have been
simplified, and explained how these solutions form the foundation for an EstimationBased Adaptive Filtering (EBAF) algorithm. Meanwhile, the feasibility of the real
time implementation of the EBAF algorithm is justified.
An implementation scheme for the new algorithm has been outlined, and a corresponding performance bound has been derived. It is shown that the classical FxLMS
(Normalized-FxLMS) adaptive algorithms are approximations to the limiting behavior of the proposed EBAF algorithm. The EBAF algorithm is shown to display
improved performance when compared to commonly used FxLMS and NormalizedFxLMS algorithms. Simulations have been verified by conducting a noise cancellation
experiment, and showing that the experimental data reasonably match a corresponding simulation.
The systematic nature of the proposed EBAF algorithm can serve as the first step
towards methodical optimization of now pre-determined parameters of the FIR filter
(such as filter length, or adaptation rate). Furthermore, the analysis of the various
aspects of the algorithm directly benefits from the advances in robust estimation theory. Finally, more efficient implementation schemes can further reduce computational
complexity of the algorithm.
Chapter 3
Estimation-Based adaptive IIR
Filter Design
This chapter extends the “estimation-based” approach (introduced in Chapter 2) to
the synthesis of adaptive Infinite Impulse Response (IIR) filters (controllers). Systematic synthesis of adaptive IIR filters is proven to be difficult and existing design
practices are ad hoc in nature. The proposed approach in this chapter is based on an
estimation interpretation of the adaptive IIR filter (controller) design that replaces
the original adaptive filtering (control) problem with an equivalent estimation problem. Similar to the case with FIR filters, an H∞ criterion is chosen to formulate this
equivalent estimation problem. Unlike the FIR case however, the estimation problem
in this case is nonlinear in IIR filter parameters. At the present time, the nonlinear
robust estimation problem does not have an exact closed form solution. The proposed
solution in this chapter is therefore an “approximation” that uses the best available
estimate of the IIR filter parameters to locally linearize the nonlinear robust estimation problem at each time step. For the linearized problem, the estimation-based
adaptive algorithm is identical to the adaptive FIR filter design described in Chapter 2. The systematic nature of this new approach is particularly appealing given the
complexity of the existing design schemes for adaptive IIR filters (controllers).
Simulations for active noise cancellation in a one dimensional acoustic duct are
58
3.1. BACKGROUND
59
used to examine the main features of the proposed estimation-based adaptive filtering algorithm for adaptive IIR filters. The performance of the EBAF algorithm is
compared to that of a commonly used classical solution in adaptive control literature known as Filtered-U Recursive LMS algorithm. The comparison reveals faster
convergence, with improved steady-state behavior in the case of EBAF algorithm.
3.1
Background
This section defines the adaptive filtering problem of interest and describes the terminology that is used in the rest of this chapter. The description of feedback contamination problem, as well as an introductory discussion of the classical Filtered-U
recursive LMS algorithm are also included in this section.
Figure 3.1 shows the adaptive filtering problem of interest. Similar to the FIR
case (see Figure 2.1), the objective here is to adjust the weight vector in the adaptive
IIR filter such that the output of the secondary path provides an acceptable match to
the output of the primary path. Figure 3.1 includes the following signals: (a) ref (k):
the input to the primary path, which is the same as the reference signal available to
the adaptive filter when there is no feedback path, (b) u(k): the control signal applied
to the secondary path, (c) y(k): the output of the secondary path (i.e. the signal
that should match d(k)), and (d) e(k): the residual error which is used to update the
weight vector in the adaptive IIR filter. Note that, despite assuming full knowledge of
the primary path input in Figure 3.1, the reference signal available to the adaptation
algorithm will be affected by the adaptive filter output when a feedback path exists.
In practice, the input to the primary path is not always fully known, and a signal
with “enough” correlation to the primary input replaces ref (k) in Figure 3.1 [33].
The existence of the feedback path will effect the correlation between the reference
signal available to the adaptive algorithm (x(k) in Fig. 3.1) and the primary input to
the primary path (ref(k) in Fig. 3.1) and has a profound effect on the performance of
the adaptive filter. This phenomenon, known as feedback contamination, is extensively
studied in the adaptive control literature ([33], Chapter 3). A simple solution for this
feedback problem is to use a separate feedback (as a part of the control system) to
3.1. BACKGROUND
60
cancel the undesirable feedback signal. An implementation of this idea, known as
feedback neutralization, is shown in Fig. 3.2. Note that in this scheme, W (z) is the
adaptive FIR filter that generates the control signal. F̂ (z) is another adaptive FIR
filter whose output eliminates (in the ideal case) the effect of feedback contamination.
This approach, however, requires special care in the implementation to avoid the
cancellation of the reference signal all together (see [33] for details). A closer look at
Figure 3.2 indicates that even though feedback neutralization employs two FIR filters,
the overall adaptive controller is no longer a zero-only system (F̂ (z) is positioned in
a feedback path). Precisely speaking, feedback neutralization is indeed an adaptive
IIR filtering algorithm.
A more direct (and more general) approach to the design of adaptive IIR filters
in such circumstances is shown in Figure 3.3. In this approach the feedback path is
treated as a part of the overall plant for which the adaptive IIR filter is designed.
Of possible adaptive IIR algorithms, only the Filtered-U recursive LMS algorithm
(FuLMS) will be considered here [14]. This selection is justified by the fact that the
FuLMS adaptive algorithm exhibits the main features of a conventional adaptive IIR
filtering algorithm and has been used successfully in noise cancellation problems (see
Chapter 3 in [33] and the references therein). Referring to Figure 3.3, the residual
error is
e(k) = d(k) − s(k) ⊕ r(k) = d(k) − y(k)
(3.1)
where s(k) is the impulse response of the secondary path, and ⊕ indicates convolution.
Note that the conventional derivation of the Filtered-U algorithm does not include
the exogenous measurement disturbance Vm (k). The output of the IIR filter r(k) is
computed as
r(k) = aT (k)x(k) + bT (k)r(k − 1)
(3.2)
where a(k) = [ a0 (k) a1 (k) · · · aL−1 (k) ]T is the weight vector for A(z) at time k,
and b(k) = [ b1 (k) b2 (k) · · · bM (k) ]T is likewise defined for B(z). Moreover, x(k) =
[ x(k) x(k − 1) · · · x(k − L + 1) ]T and r(k−1) = [ r(k − 1) r(k − 2) · · · r(k − M) ]T
are reference signals for A(z) and B(z), respectively.
3.2. PROBLEM FORMULATION
61
T
Defining a new overall weight vector w(k) = aT (k) bT (k) , and a generalized
T
reference vector u(k) = xT (k) rT (k − 1) , Eq. (3.2) can be rewritten as
r(k) = wT (k)u(k)
(3.3)
which has the same format as the output of an ordinary FIR filter. Reference [33]
shows that the steepest-descent algorithm can be used to derive the following update
equation for the generalized weight vector
w(k + 1) = w(k) + µ [ŝ(k) ⊕ u(k)] e(k)
(3.4)
if the instantaneous squared error, e2 (k), is used to estimate the mean-square error,
E[ k e2 (k) k ]. Referring to Section 2.1, the error criterion here is the same as that
of the LMS algorithm. The algorithm is called the Filtered-U recursive LMS algorithm since it uses an estimate of the secondary path, ŝ(k), to filter the generalized
reference vector u(k). The derivation of this algorithm explicitly relies on a slow
convergence assumption and therefore, in general, the convergence rate µ should be
kept small. Slow adaptation also enables the derivation of an approximate instantaneous gradient vector that significantly reduces the computational complexity of
the algorithm (see [33], page 93). The ad hoc nature of the FuLMS, however, has
significantly complicated the analysis of the algorithm. As Reference [33] indicates,
the global convergence and stability of the algorithm have not been formally proved,
and the optimal solution in the case of large controller coefficients is found to be illconditioned. Nevertheless, successful implementations of this algorithm are reported,
and hence it will be used as the conventional design algorithm to which the EBAF
algorithm will be compared.
3.2
Problem Formulation
The underlying concept for estimation-based adaptive IIR filter design is essentially
the same as that of the estimation-based adaptive FIR filter design. For clarity,
however, the main steps in the estimation interpretation of the adaptive filtering
(control) problem are repeated here:
3.2. PROBLEM FORMULATION
62
1. Introduce an approximate model for the primary path based on the architecture
of the adaptive path from x(k) to y(k) (as shown in Fig. 3.1). The goal is to
find the optimal weight vector in the approximate model for which the modeling
error is the smallest. As in Chapter 2, state-space models are used for both the
adaptive filter (IIR in this case) and the secondary path.
2. In the approximate model for the primary path, use the available information to
formulate an estimation problem that recursively estimates this optimal weight
vector.
3. Adjust the weight vector of the adaptive IIR filter to the best available estimate
of the optimal weight vector.
For simplicity, the case without feedback contamination is considered first. The case
with feedback contamination is a straightforward extension, and will be discussed in
Appendix B. A model for the secondary path is assumed available (e.g. via identification). The primary path however, is completely unknown. Figure 3.4 provides
a pictorial presentation for the above mentioned estimation interpretation, which is
identical to Figure 2.3, except for the replacement of the FIR filter with an IIR one.
The main signals involved in Figure 3.4 are similar to those in Figure 2.2 and are
described here for easier reference. First, note that
e(k) = d(k) − y(k) + Vm (k)
(3.5)
where e(k) is the available error measurement, Vm (k) is the exogenous disturbance
that captures measurement noise, modeling error and initial condition uncertainty,
y(k) is the output of the secondary path, and d(k) is the output of the primary path.
Equation (3.5) can be rewritten as
e(k) + y(k) = d(k) + Vm (k)
(3.6)
where the left hand side is a noisy measurement of the output of the primary path
d(k). Since y(k) is not directly measurable (neither is d(k) of course), the trick is to
have the adaptive algorithm generate an internal copy of y(k), and then define the
3.2. PROBLEM FORMULATION
63
derived measured quantity as
4
m(k) = e(k) + ycopy (k) = d(k) + Vm (k)
(3.7)
which will be used in formulating the estimation problem. To generate the internal
copy of y(k), the adaptive algorithm uses the available model for the secondary path
and the known control input to the secondary path, u(k). See Section 2.7 for a
discussion of the impact of the initial condition error.
3.2.1
Estimation Problem
Figure 3.5 shows a block diagram representation of the approximate model to the
primary path. Here, a state space model, [ As (k), Bs (k), Cs (k), Ds (k) ], for the
secondary path is assumed. A second order IIR filter (with 5 parameters) models the
adaptive IIR filter. Define W (k) = [ a0 (k) b1 (k) · · · bN (k) a1 (k) · · · aN (k)]T (N = 2
in Fig. 3.5) to be the unknown optimal vector of the IIR filter parameters at time
T
is the state vector for the overall system. Note that
k. Then, ξ = W T (k) θT (k)
θ(k) captures the dynamics of the secondary path in the approximate model. The
state space representation of the system is
#
"
#
"
#"
I(2N +1)×(2N +1)
W (k + 1)
W (k)
0
=
θ(k + 1)
θ(k)
Bs (k)h∗ (k)
As (k)
4
ξk+1 = Fk ξk
(3.8)
where h(k) = [x(k) r(k − 1) · · · r(k − N) r(k − 1) · · · r(k − N)]T captures the effect of the reference input x(·). Note that
r(k) = x(k)a0 (k) + r(k − 1)b1 (k) + · · · + r(k − N)bN (k),
(3.9)
(where r(−1) = · · · = r(−N) = 0) and therefore, the system dynamics are nonlinear
in the IIR filter parameters. For this system, the derived measured output is
#
"
h
i W (k)
+ Vm (k)
m(k) =
Ds (k)h∗ (k) Cs (k)
θ(k)
4
= Hk ξk + Vm (k)
(3.10)
3.2. PROBLEM FORMULATION
64
where m(k) should be constructed at each step according to Equation (3.7). Note
that measurement equation is also nonlinear in the parameters of the IIR filter. Now,
define a generic linear combination of the states as the desired quantity to be estimated
"
#
h
i W (k)
s(k) =
L1,k L2,k
θ(k)
4
= Lk ξk
(3.11)
Here, θ(·) ∈ RNs ×1 , W (·) ∈ R(2N +1)×1 , m(·) ∈ R1×1 and s(·) ∈ R1×1 . All matrices are
then of appropriate dimensions. Different choices for Lk lead to different estimates of
W (k), and hence different adaptation criteria for the parameters in the IIR filter. As
in Chapter 2, an H∞ criterion will be used to directly estimate d(k) (hence Lk = Hk ).
The objective is to find an H∞ suboptimal filter ŝ(k|k) = F (m(0), · · · , m(k)) (or H∞
suboptimal predictor ŝ(k) = F (m(0), · · · , m(k − 1))), such that the worst case energy
gain from the measurement disturbance and the initial condition uncertainty to the
error in a causal estimate of s(k) = d(k) remains bounded. In other words, ŝ(k|k) is
sought such that for a given γf > 0 and Π0 > 0,
M
X
sup
Vm , ξ0
(s(k) − ŝ(k|k))∗ (s(k) − ŝ(k|k))
k=0
ξ0∗ Π−1
0 ξ0
+
M
X
≤ γf2
(3.12)
∗
Vm
(k)Vm (k)
k=0
Similarly ŝ(k) is desired such that for a given γp > 0 and Π0 > 0,
M
X
sup
Vm , ξ0
(s(k) − ŝ(k))∗ (s(k) − ŝ(k))
k=0
ξ0∗ Π−1
0 ξ0 +
M
X
≤ γp2
(3.13)
∗
Vm
(k)Vm (k)
k=0
Since Eqs. (3.8), (3.10), and (3.11) are nonlinear in IIR filter parameters, both of these
estimation problems are nonlinear. An exact closed form solution to the nonlinear
H∞ estimation problem is not yet available. One solution to this problem is to use
the following linearizing approximation; at each time step, we replace the IIR filter
3.3. APPROXIMATE SOLUTION
65
parameters, a0 , b1 , · · · bN , in Equation (3.9) with their best available estimate. This
reduces the estimation problems in Eqs. (3.12) and (3.13) into linear H∞ estimation
problems for which the solutions in Section 2.4 can be directly applied. A similar
linearizing approximation is commonly adopted in the optimal estimation problems
for nonlinear dynamic processes and is referred to as continuous relinearization in
extended Kalman filtering [9,20]. As Reference [9] indicates, in the context of optimal estimation for the nonlinear processes, “considerable success has been achieved
with this linearization approach”. Extensive simulations in this Thesis reveal equally
successful results in adaptive IIR filter design (see Section 3.6).
Remark: As Figure 3.5 suggests, feedback is an integral part of an IIR filter structure.
The nonlinearity of the robust estimation problem in Eqs. (3.12) and (3.13) is due
to the existence of this structural feedback loop. Obviously, a physical feedback path
outside the IIR filter (see Figure 3.1) will have the same effect. The treatment of
the nonlinearity in the case of an IIR filter, i.e. replacing the IIR filter parameters
in Equation (3.9) with their best available estimates, carries over to the case where
such a feedback path exists. Appendix B discusses the case with reference signal
contamination in more detail.
3.3
Approximate Solution
With the linearizing approximation of the previous section, at any given time k,
h(k) will be fully known. This eliminates the nonlinearity in Equations (3.8)-(3.11),
and allows the straightforward solutions in Section 2.4 to be used. Note that for the
linearized problem all the optimality arguments in Chapter 2 are valid and the simplifications to filtering (prediction) solutions also apply. The linearizing approximation,
however, prevents any claim on the optimality of the solution. Only the simplified
solutions to the linearized estimation problems of Eqs. (3.12) and (3.13) (i.e. the
central filtering solution corresponding to γf = 1 for the filtering solution, and the
central prediction solution corresponding to γp = 1 for the prediction solution) are
stated here.
3.3. APPROXIMATE SOLUTION
3.3.1
66
γ-Suboptimal Finite Horizon Filtering Solution to the
Linearized Problem
Theorem 3.1: Invoking Theorem 2.3 in Chapter 2, for the state space representation
of the block diagram of Figure 3.5, described by Equations (3.8)-(3.11), and for Lk =
Hk , the central H∞ -optimal filtering solution to the linearized problem is given by
ˆ
ˆ
ˆ
ξk+1 = Fk ξk + Kf,k m(k) − Hk ξk , ξˆ0 = 0
(3.14)
−1
ŝ(k|k) = Lk ξˆk + (Lk Pk Hk∗ ) RHe,k
m(k) − Hk ξˆk
(3.15)
with
−1
Kf,k = (Fk Pk Hk∗ ) RHe,k
and RHe,k = Ip + Hk Pk Hk∗
(3.16)
where Pk satisfies the Lyapunov recursion
Pk+1 = Fk Pk Fk∗ , P0 = Π0 .
3.3.2
(3.17)
γ-Suboptimal Finite Horizon Prediction Solution to the
Linearized Problem
Theorem 3.2: Invoking Theorem 2.4 in Chapter 2, for the state space representation
of the block diagram of Figure 3.5, described by Equations (3.8)-(3.11), and for Lk =
Hk , if (I − Pk L∗k Lk ) > 0, then the central H∞ -optimal prediction solution to the
linearized problem is given by
ˆ
ˆ
ˆ
ξk+1 = Fk ξk + Kp,k m(k) − Hk ξk , ξˆ0 = 0
(3.18)
ŝ(k) = Lk ξˆk
(3.19)
with Kp,k = Fk Pk Hk∗ where Pk satisfies the Lyapunov recursion (3.17).
3.3.3
Important Remarks
1. As indicated in Theorems 3.1 and 3.2, the solution to the linearized robust estimation problem (Eq. (3.12) for the filtering problem and Eq. (3.13) for the
prediction problem) requires the solution to only one Riccati equation. Furthermore, the Riccati solution propagates forward in time and does not involve any
3.4. IMPLEMENTATION SCHEME FOR THE EBAF ALGORITHM IN IIR CASE
67
information regarding the future of the system or the reference signal. Thus,
the resulting adaptive algorithm is real-time implementable.
2. Note that the Riccati update for the simplified solution to the linearized robust
estimation problems reduces to a Lyapunov recursion which always generates a
non-negative Pk as long as P0 > 0.
3. Based on the results in Theorem 3.1, the linearized filtering problem has a
guaranteed solution for γf = 1. This will prevent any breakdown in the solution
and allows real-time implementation of the algorithm.
4. Theorem 3.2 proves that the prediction solution to the linearized problem is
guaranteed to exist for γp = 1, as long as the condition (I − Pk L∗k Lk ) > 0
is satisfied. Furthermore, the discussion in Section 2.8 shows that this condition translates into an upper limit for the adaptation rate in the steady-state
operation of the adaptive system.
5. From features (2) and (3), there is no need to verify the solutions at each time
step, so the computational complexity of the estimation based approach is O(n3 )
(primarily for calculating Fk Pk FK∗ in Eq. (3.17)), where
n = (2N + 1) + Ns
(3.20)
where (2N + 1) is the total number of IIR filter parameters for an IIR filter of
order N, and Ns is the order of the secondary path. As in the FIR case, the
special structure of Fk reduces the computational complexity to O(Ns3 + Ns N),
i.e. cubic in the order of the secondary path, and linear in the length of the
FIR filter.
3.4
Implementation Scheme for the EBAF Algorithm in IIR Case
The implementation scheme parallels that of the adaptive FIR filter discussed in
Chapter 2. For easier reference, the main signals involved in the description of the
3.4. IMPLEMENTATION SCHEME FOR THE EBAF ALGORITHM IN IIR CASE
68
adaptive algorithm are briefly introduced here. For a more detailed description, see
Chapter 2. In what follows (a) Ŵ (k) is the estimate of the adaptive weight vector,
4
(b) θ̂(k) is the estimate of the state of the secondary path, (c) u(k) = h∗ (k)Ŵ (k)
is the actual control input to the secondary path, (d) y(k) and d(k) are the actual
outputs of the secondary and primary paths, respectively, (e) e(k) is the actual error
measurement, and (f) θcopy (k) and ycopy (k) are the adaptive algorithm’s internal copy
of the state and output of the secondary path which are used in constructing m(k)
according to Eq. (3.7). Now, the implementation algorithm can be outlined as follows:
1. Start with Ŵ (0) = 0 and θ̂(0) = 0 as the initial guess for the state vector in the
approximate model of the primary path. Also assume that θcopy (0) = 0, and
r(−1) = · · · = r(−N) = 0 (hence h(0) = [ x(0) 0 · · · 0 ]T ). The initial value
for the Riccati matrix is P0 which is chosen to be block diagonal.
2. If 0 ≤ k ≤ M (finite horizon):
(a) Linearize the nonlinear dynamics in Eqs. (3.8), (3.10), and (3.11) by substituting for a0 (k), b1 (k), · · ·, bN (k) with their best available estimates
(i.e. â0 (k) = Ŵ (1, k), b̂1 (k) = Ŵ (2, k), · · ·, b̂N (k) = Ŵ (N + 1, k)) in
Eq. (3.9).
(b) Form the control signal
u(k) = h∗ (k)Ŵ (k)
(3.21)
to be applied to the secondary path. Note that applying u(k) to the
secondary path produces
y(k) = Cs (k)θ(k) + Ds (k)u(k)
(3.22)
at the output of the secondary path. This in turn leads to the following
error signal measured at time k:
e(k) = d(k) − y(k) + Vm (k)
(3.23)
which is available to the adaptive algorithm to perform the state update
at time k.
3.5. ERROR ANALYSIS
69
(c) Propagate the state estimate and the internal copy of the state of the
secondary path as follows
 "
# 
Ŵ (k + 1)



 =
θ̂(k
+
1)


θcopy (k + 1)
i
h
"
# h
Kf,k Cs (k)
Fk + Kf,k [0 − Cs (k)]
Kf,k
Ŵ (k)




 θ̂(k) +



h
i
∗
θcopy (k)
As (k)
0
(Bs (k)h (k) 0
i

e(k)

(3.24)
where e(k) is the error sensor measurement at time k given by Eq. (3.23),
and Kf,k = Fk Pk Hk∗ (I + Hk Pk Hk∗ )−1 (see Theorem 3.1). Note that for
the prediction-based EBAF algorithm Kf,k must be replaced with Kp,k =
Fk Pk Hk∗ .
(d) update the Riccati matrix Pk using the Lyapunov recursion
"
#
P11
P12,k+1
=
∗
P12,k+1
P22,k+1
"
#"
#"
I
0
P11 P12,k
I
Bs (k)h∗ (k) As (k)
∗
P12,k
P22,k
0
#∗
Bs (k)h∗ (k) As (k)
(3.25)
Pk+1 will be used in (3.24) to update the state estimate.
3. Go to 2.
3.5
Error Analysis
As explained in Section 2.7, for a linear stable secondary path the contribution of
the initial condition at the output of the secondary path decays to zero for large k.
This means that ycopy (k) converges to y(k) for sufficiently large k. In other words,
3.6. SIMULATION RESULTS
70
exact knowledge of the initial condition of the secondary path does not affect the
performance of the proposed EBAF algorithm for sufficiently large k. Unlike the
discussions in Section 2.7 however, a performance bound for the EBAF algorithm in
the IIR case is not yet available.
3.6
Simulation Results
This section examines the performance of the proposed EBAF algorithm for the
active noise cancellation (ANC) problem in the one dimensional acoustic duct (see
Figure 2.5). The control objective is the same as that in Section 2.9, i.e. to attenuate
(cancel in the ideal case) the disturbance introduced into the duct by Speaker #1
(primary noise source) at the position of Microphone #2 (error sensor) by the control
signal generated by Speaker #2 (secondary source). Microphone #1 can be used to
provide the reference signal for the adaptation algorithm. Clearly, Microphone #1
measurements are affected by both primary and secondary sources, and if used as the
reference signal, feedback contamination exists.
A state space model (of order 10) is identified for this one-dimensional acoustic
system. Of four identified transfer functions (see Figs. 2.6 and 2.7), only the transfer
function from Speaker #2 to Microphone #2 (i.e. the secondary path) is used by the
estimation-based adaptive algorithm. The IIR filter in EBAF approach is of order 5
(i.e. a total of 11 parameters for adaptive IIR filter). Each FIR filter in the FuLMS
implementation is of order 6 (a total of 12 parameters in the adaptive filter). Note
that all the measurements in the simulations are subject to band-limited white noise
(with power 0.008), and the sampling frequency in all cases is 1 KHz.
Figure 3.6 compares the performance of an estimation-based adaptive IIR filter
with that of a FuLMS algorithm in a single-tone noise cancellation problem. The
frequency of the tone is 150 Hz. Referring to Fig. 3.6, the first second shows the open
loop measurement of the error sensor, Microphone #2. At t = 1 (sec) the adaptive
control is switched on. For the next 5 seconds both adaptive algorithms have access
to the primary disturbance (i.e. x(k) is fully known to the adaptation schemes). At
t = 6 the reference signal available to the adaptive algorithms is switched to the
3.6. SIMULATION RESULTS
71
measurements of Microphone #1. Note that in this case only a filtered version of the
primary disturbance which is also contaminated with the output of the adaptive filter
is available to the adaptive algorithms. The adaptation rate for FuLMS algorithm
is kept small (0.00003 in this case), to avoid unstable behavior when the reference
signal becomes contaminated. The EBAF algorithm converges faster than FuLMS
algorithm (EBAF algorithm reaches its steady-state behavior in approximately 1.5
seconds), while avoiding the unstable behavior due to feedback contamination. A 50
times reduction in the error amplitude (without feedback contamination) is recorded.
With feedback contamination however, only a 10 times reduction is achieved.
Figure 3.7, shows a similar scenario for the multi-tone case. Here the primary
disturbance consists of two sinusoidal signals at 150 and 140 Hz. A trend similar to
the single-tone case is observed here. EBAF algorithm is robust to feedback contamination and allows faster convergence rates. The performance however is not as good
as the single tone case. With uncontaminated reference signal only a reduction of
order 10 is achieved. With feedback contamination this performance is reduced to a
factor of 6. The results shown in Figures 3.6 and 3.7 capture the typical behavior of
the adaptive IIR filters under the EBAF and FuLMS algorithms.
3.6. SIMULATION RESULTS
ref (k)
72
Vm (k)
d(k)
Primary Path
(Unknown)
+
+
+
+
Feedback Path
(Known)
+
+
e(k)
−
Adaptive IIR
Filter
x(k)
u(k)
Secondary Path
(Known)
y(k)
Fig. 3.1: General block diagram for the adaptive filtering problem of interest (with Feedback
Contamination)
Primary Path
(Unknown)
P (z)
ref (k)
+
+
+
Vm (k)
d(k)
+
Feedback Path
F (z)
+
+
e(k)
−
+
+
−
Adaptive FIR Filter
F̂ (z)
u(k)
Secondary Path
(Known)
S(z)
y(k)
Adaptive FIR Filter
x(k)
W (z)
Fig. 3.2: Basic Block Diagram for the Feedback Neutralization Scheme
3.6. SIMULATION RESULTS
73
ref (k)
Vm (k)
d(k)
Primary Path
(Unknown)
+
+
Feedback Path
(Known)
+
+
e(k)
−
A(z)
+
Secondary Path
(Known)
+
x(k)
u(k)
y(k)
+
B(z)
Fig. 3.3: Basic Block Diagram for the Classical Adaptive IIR Filter Design
Primary Path
Modeling Error
x(k)
A Copy Of
Secondary Path
An IIR Filter
Adaptive IIR Filter
Secondary Path
u(k)
Vm (k)
+
d(k)
+
+
+
−
y(k)
Fig. 3.4: Estimation Interpretation of the IIR Adaptive Filter Design
e(k)
3.6. SIMULATION RESULTS
74
Approximate Model For Primary Path
x(k)
a0
+
r(k)
+
Vm (k)
Ds (k)
u(k)
Bs (k)
+
z −1
Cs (k)
z −1
As (k)
a1
b1
z −1
b2
a2
A Second Order IIR Filter
Fig. 3.5: Approximate Model For the Unknown Primary Path
+
+
d(k) m(k)
3.6. SIMULATION RESULTS
75
e(k) - EBAF Adaptive IIR
Single Tone Error Cancellation - EBAF vs. FuLMS
1
0.5
0
−0.5
e(k) - FuLMS Adaptive IIR
−1
0
1
2
3
4
5
6
7
8
9
10
0
1
2
3
4
5
6
7
8
9
10
1
0.5
0
−0.5
−1
Time (sec.)
Fig. 3.6: Performance Comparison for EBAF and FuLMS Adaptive IIR Filters for Single-Tone
Noise Cancellation. The controller is switched on at t = 1 second. For 1 ≤ t ≤ 6 seconds adaptive
algorithm has full access to the primary disturbance. For t ≥ 6 the output of Microphone #1 is
used as the reference signal (hence feedback contamination problem).
3.6. SIMULATION RESULTS
76
e(k) - EBAF Adaptive IIR
Multi Tone Error Cancellation - EBAF vs. FuLMS
1.5
1
0.5
0
−0.5
−1
e(k) - FuLMS Adaptive IIR
−1.5
0
1
2
3
4
5
6
7
8
9
10
0
1
2
3
4
5
6
7
8
9
10
1.5
1
0.5
0
−0.5
−1
−1.5
Time (sec.)
Fig. 3.7: Performance Comparison for EBAF and FuLMS Adaptive IIR Filters for Multi-Tone
Noise Cancellation. The controller is switched on at t = 1 second. For 1 ≤ t ≤ 6 seconds adaptive
algorithm has full access to the primary disturbance. For t ≥ 6 the output of Microphone #1 is
used as the reference signal (hence feedback contamination problem).
3.7. SUMMARY
3.7
77
Summary
A new framework for the synthesis and analysis of the IIR adaptive filters is introduced. First, an estimation interpretation of the adaptive filtering (control) is used
to formulate an equivalent nonlinear robust estimation problem. Then, an approximate solution for the equivalent estimation problem is provided. This approximate
solution is based on a linearizing approximation, from which the adaptation law for
the adaptive filter weight vector is extracted.
The proposed approach clearly indicates an inherent connection between the adaptive IIR filter design and a nonlinear robust estimation problem. This connection
brings the analysis and synthesis tools in robust estimation into the field of adaptive
IIR filtering (control). Simulation results demonstrate the feasibility of the proposed
EBAF algorithm.
Chapter 4
Multi-Channel Estimation-Based
Adaptive Filtering
This chapter extends the estimation-based adaptive filtering algorithm, discussed
in Chapter 2, to the multi-channel case where a number of adaptively controlled
secondary sources use multiple reference signals to cancel the effect of a number of
primary sources (i.e. disturbance sources) as seen by a number of error sensors. The
multi-channel estimation-based adaptive filtering algorithm is shown to maintain all
the main features of the single-channel solution, underlining the systematic nature of
the approach.
In addition to the noise cancellation problem in a one dimensional acoustic duct, a
structural vibration control problem is chosen to examine the performance of the proposed multi-channel adaptive algorithm. An identified model for a Vibration Isolation
Platform (VIP) is used for vibration control simulations in this chapter. The performance of the new multi-channel adaptive algorithm is compared to the performance
of a multi-channel implementation of the FxLMS algorithm.
78
4.1. BACKGROUND
4.1
79
Background
For a wide variety of applications such as equalization in wireless communication
when more than one receiver/transmitter are involved, or active control of sound
and vibration in cases where the acoustic environment or the dynamic system of
interest is complex and a number of primary sources excite the system, multi-channel
adaptive filtering (control) schemes are required [1]. A brief description of the multichannel implementation of the FxLMS algorithm in Section 4.1.1 will support the
observation in Reference [33] that “compared to single-channel algorithms, multichannel adaptive schemes are significantly more complex”. As Reference [33] points
out, successful application of the classical multi-channel adaptive algorithms has been
limited to cases involving repetitive noise with a few harmonics [43,49,13]. This
observation agrees with the results in Ref. [3] where a significant noise reduction is only
achieved for periodic noises∗ . In contrast to the classical approaches to multi-channel
adaptive filter (control) design, this chapter will show that, for the new estimationbased approach, the multi-channel design is virtually identical to the single channel
case. Furthermore, the analysis of the multi-channel adaptive system in the new
framework is a straightforward extension of the analysis used in the single channel
case.
4.1.1
Multi-Channel FxLMS Algorithm
Figure 4.1 shows the general block diagram of a multi-channel ANC system in which
reference signals can be affected by the output of the adaptive filters. Simulations in
this chapter, however, are based on the assumption that the effects of feedback are
negligible and that the reference signal is available to the adaptation scheme through
a noisy measurement. Measurement noise however is independent from the reference
signal itself. Note that, this section is only intended as a brief review of the multichannel FxLMS adaptive algorithm. For a detailed treatment of the subject see [33],
∗
Another interesting conclusion in Ref. [3] is the following: the performance of a Multi-Channel
implementation of the FxLMS algorithm is similar to the performance of an H∞ controller which
was directly designed for noise cancellation.
4.1. BACKGROUND
80
Chapter 5, and the references therein.
Referring to Figure 4.1, the ANC adaptive filter has J reference input signals
denoted by xT (k) = [x1 (k) x2 (k) · · · xJ (k)]. The controller generates K secondary
signals that are elements of the control vector uT (k) = [u1 (k) u2 (k) · · · uK (k)].
Therefore, a K ×J matrix of adaptive FIR filters can be used to describe the adaptive
control block in Figure 4.1,

···
T
W11
(k)

Ω(k) = 


T
W1J
(k)



..
.
T
Wkj
(k)
..
.
T
WK1
(k)
···
T
WKJ
(k)
(4.1)
where
def
T
Wkj
(k) = [wkj,1(k) · · · wkj,(L−1) (k)]
(4.2)
is the adaptive filter relating the j-th reference signal to the k-th control command.
Note that L is the length of the adaptive FIR filters. Defining the reference signal
vector as
XT (k) =
xT1 (k) xT2 (k) · · · xTJ (k)
(4.3)
where xTj (k) = [xj (k) xj (n − 1) · · · xj (n − L + 1)] is the last L samples of the j-th
reference signal, and the adaptive weight vector as
def
W(k) = [ Ω(1,:) (k) · · · Ω(K,:)(k) ]
(4.4)
with Ω(k,:) (k) referring to the k-th row of matrix Ω(k), the control vector u(k) can
be defined as
u(k) = X T (k)W(k)
where




X (k) = 


X(k)
0
..
.
0
(4.5)

···
0
X(k) · · ·
..
.
0
0
0
···
0
0
X(k)






(4.6)
4.2. ESTIMATION-BASED ADAPTIVE ALGORITHM FOR MULTI CHANNEL CASE
81
is a (JKL)×K matrix. The error signal vector can now be defined as
e(k) = d(k) − S(k) ⊕ u(k)
(4.7)
where S(k) is the impulse response of the K-input/M-output secondary path, and
⊕ indicates the convolution operation. Defining the instantaneous squared error,
eT (k)e(k), as an approximation for the sum of the mean-square errors, the gradient
based update equation for the weight vector W(k) is [33],
W(k + 1) = W(k) + µX 0 (k)e(k)
(4.8)
where X 0 (k) is the matrix of the filtered reference signals obtained by the Kronecker
convolution operation on the reference signal vector

ŝ (k) ⊕ X(k) ŝ21 (k) ⊕ X(k) · · · ŝM 1 (k) ⊕ X(k)
 11
 ŝ (k) ⊕ X(k) ŝ (k) ⊕ X(k) · · · ŝ (k) ⊕ X(k)
22
M2
 12
X 0 (k) = 
.
.
..
.
..
..
..

.

ŝ1K (k) ⊕ X(k) ŝ2K (k) ⊕ X(k) · · · ŝM K (k) ⊕ X(k)







(4.9)
Note that the adaptation algorithm has access to an estimate of the secondary path
only (e.g. through system identification) and therefore ŝij (an estimate of the impulse
response for the single-channel transfer function from input j to output i) are used
in calculation of the filtered reference signals.
4.2
Estimation-Based Adaptive Algorithm for Multi
Channel Case
The underlying concept for estimation based adaptive filtering algorithm is the same
for both single-channel and multi-channel systems. Figure 4.1 is the multi-channel
block diagram representation of Fig. 2.1 (in Chapter 2) where the main components of
active noise cancellation problem were described. Therefore, an estimation interpretation identical to the one in the single-channel case can be used to translate a given
4.2. ESTIMATION-BASED ADAPTIVE ALGORITHM FOR MULTI CHANNEL CASE
82
adaptive filtering (control) problem into an equivalent robust estimation problem.
See Chapter 2 for a detailed treatment of the steps involved in this translation.
As in Chapter 2, state space models for the adaptive filter and the secondary path
are used to construct an approximate model for the unknown primary path. This
approximate model replicates the structure of the adaptive path from the primary
source, ref (k), to the output of the secondary path, y(k) (see Fig. 4.2). Note that for a
given disturbance input, there is an “optimal” (but unknown) setting of adaptive filter
parameters for which the difference between the primary path and its approximate
model is minimized. Finding this optimal setting is the objective of the estimation
based approach which can be summarized as follows:
1. Devise an estimation strategy that recursively improves the estimate of the
optimal values of the adaptive filter parameters in the approximate model of
the primary path,
2. Set the actual value of the weight vector in the adaptive filter to the best
available estimate of the parameters obtained from the estimation strategy.
Note that in Figure 4.2,
e(k) = d(k) − y(k) + Vm (k)
(4.10)
where (a) e(k) ∈ RM ×1 is the measured error vector, (b) Vm (k) ∈ RM ×1 is the exogenous disturbance that captures measurement noise, modeling error and uncertainty
in the initial condition of the secondary path, and (c) y(k) = S(k) ⊕ u(k) (also in
RM ×1 ) is the output of the secondary path. S(k) is the impulse response of the secondary path and ⊕ denotes convolution. Here u(k) obeys Eq. (4.5) with the same
definitions for X (k) and W(k) as in (4.6) and (4.8), respectively. Equation (4.10)
can be rewritten as
e(k) + y(k) = d(k) + Vm (k)
(4.11)
where the left hand side is a noisy measurement of the output of the primary path
d(k). Since y(k) is not directly measurable (neither is d(k)), the adaptive algorithm should generate an internal copy of y(k) (referred to as ycopy (k)). The derived
4.2. ESTIMATION-BASED ADAPTIVE ALGORITHM FOR MULTI CHANNEL CASE
83
measured quantity can then be defined as
4
m(k) = e(k) + y(k) = d(k) + Vm (k)
(4.12)
which will be used in formulating the estimation problem. The only assumption
involved in constructing m(k) is the assumed knowledge of the initial condition of
the secondary path. In Chapter 2 it is shown that for a linear, stable secondary path (a
realistic assumption in practice), any error in y(k) due to an initial condition different
from what is assumed by the algorithm remains bounded (hence it can be treated as
a component of the measurement disturbance). Furthermore, for sufficiently large k,
this error decays to zero (i.e. ycopy (k) → y(k)). In Figure 4.3, the state space model
for the secondary path is
θ(k + 1) = As (k)θ(k) + Bs (k)u(k)
(4.13)
y(k) = Cs (k)θ(k) + Ds (k)u(k)
(4.14)
where θ(k) is the state variable capturing the dynamics of the secondary path. The
weight vector of the adaptive filter
T
T
T
T
T
T
T
(k) · · · W1J
(k) W21
(k) · · · W2J
(k) · · · WK1
(k) · · · WKJ
(k) (4.15)
W(k) = W11
is also treated as the state vector that captures the dynamics of the FIR filter.
Note that Wkj (k) is itself a vector of length L (length of each FIR filter). ξkT =
WT (k) θT (k) is then the state vector for the overall system. The state space
representation of the system is then
"
#
"
#"
#
W(k + 1)
I(JKL)×(JKL)
W(k)
0
=
θ(k + 1)
θ(k)
Bs (k)X ∗ (k) As (k)
4
ξk+1 = Fk ξk
(4.16)
where X (k), defined by Equation (4.6), captures the effect of the reference input
vector X(k). For this system, the derived measured output is
#
"
i W(k)
h
+ Vm (k)
m(k) =
Ds (k)X ∗ (k) Cs (k)
θ(k)
4
= Hk ξk + Vm (k)
(4.17)
4.2. ESTIMATION-BASED ADAPTIVE ALGORITHM FOR MULTI CHANNEL CASE
84
where m(k) is defined in Equation (4.12). Noting the objective of the adaptive
filtering problem in Fig. 4.1, s(k) = d(k) is the quantity to be estimated. Therefore,
"
#
h
i W(k)
s(k) =
Ds (k)X ∗ (k) Cs (k)
θ(k)
4
= Lk ξk
(4.18)
Here m(k) ∈ RM ×1 , s(k) ∈ RM ×1 , θ(k) ∈ RNs ×1 where Ns is the order of the
secondary path. All matrices are then of appropriate dimensions. Note that Equations (4.16) through (4.18) are identical to Equations (2.5) through (2.7). The only
difference is in the dimension of the variable involved, and the fact that h(k) is replaced by X ∗ (k).
Choosing the H∞ criterion to generate a filtering (prediction) estimate of s(k), the
equivalent estimation problem will also be identical to that in Chapter 2 (i.e. Eq. (2.11)
for the filtering estimate, and Eq. (2.12) for the prediction estimate).
Defining
4
ŝ(k|k) = F (m(0), · · · , m(k)) as the filtering estimate of s(k), the objective in the
filtering solution is to find ŝ(k|k) such that the worst case energy gain from the measurement disturbance and the initial condition uncertainty to the error in the filtering
estimate is properly bounded, i.e.
M
X
sup
Vm , ξ0
[s(k) − ŝ(k|k)]∗ [s(k) − ŝ(k|k)]
n=0
ˆ
(ξ0 − ξˆ0 )∗ Π−1
0 (ξ0 − ξ0 ) +
M
X
≤ γf2
(4.19)
∗
Vm
(k)Vm (k)
n=0
4
In a similar way, defining ŝ(k) = F (m(0), · · · , m(k − 1)) as the prediction estimate
of s(k), the objective in the prediction solution is to find ŝ(k) such that
M
X
sup
Vm , ξ0
[s(k) − ŝ(k)]∗ [s(k) − ŝ(k)]
n=0
(ξ0 − ξˆ0 )
∗
Π−1
0 (ξ0
− ξˆ0 ) +
M
X
≤ γp2
(4.20)
∗
Vm
(k)Vm (k)
n=0
Solutions to these two problems are discussed in detail in Chapter 2 and therefore
the next section only briefly presents the simplified solutions.
4.3. SIMULATION RESULTS
4.2.1
85
H∞ -Optimal Solution
Theorem 4.1: Invoking Theorem 2.3 in Chapter 2, for the state space representation
of the block diagram in Figure 4.3 (described by Equations (4.16)-(4.18)), and for
Lk = Hk , the central H∞ -optimal solution to the filtering problem in Eq. (4.19) is
obtained for γf = 1, and is described by
ξˆk+1 = Fk ξˆk + Kf,k m(k) − Hk ξˆk , ξˆ0 = 0
−1
ŝ(k|k) = Lk ξˆk + (Lk Pk Hk∗ ) RHe,k
m(k) − Hk ξˆk
(4.21)
(4.22)
with
−1
Kf,k = (Fk Pk Hk∗ ) RHe,k
and RHe,k = Ip + Hk Pk Hk∗
(4.23)
where Pk satisfies the Lyapunov recursion
Pk+1 = Fk Pk Fk∗ , P0 = Π0 .
(4.24)
Theorem 4.2: Invoking Theorem 2.4 in Chapter 2, for the state space representation
of the block diagram in Figure 4.3 (described by Equations (4.16)-(4.18)), and for
Lk = Hk , if (I − Pk L∗k Lk ) > 0, then the central H∞ -optimal prediction solution to the
linearized problem is obtained for γp = 1, and is given by
ξˆk+1 = Fk ξˆk + Kp,k m(k) − Hk ξˆk , ξˆ0 = 0
(4.25)
ŝ(k) = Lk ξˆk
(4.26)
with Kp,k = Fk Pk Hk∗ where Pk satisfies the Lyapunov recursion (4.24).
4.3
Simulation Results
The implementation scheme for the EBAF algorithm in multi-channel case is identical
to the implementation scheme in the single-channel case (see Chapter 2 for the FIR
case, and Chapter 3 for the IIR case), and therefore it is not repeated here.
4.3. SIMULATION RESULTS
4.3.1
86
Active Vibration Isolation
The Vibration Isolation Platform (VIP) (see Figure 4.4 and 4.5) is an experimental
set up which is designed to capture the main features of a real world payload isolation
and pointing problem. Payload isolation refers to the vibration isolation of payload
structures with instruments or equipments requiring a very quiet mounting [1]. VIP
is designed such that the base supporting the payload (middle mass in Figure 4.5)
can emulate spacecraft dynamics. Broadband as well as narrowband disturbances
can be introduced to the middle mass (emulating real world vibration sources such
as solar array drive assemblies, reaction wheels, control moment gyros, crycoolers,
and other disturbance sources that generate on-orbit jitter) via a set of three voice
coil actuators. The positioning of a second set of voice coil actuators allows for
the implementation of an adaptive/active isolation system. More specifically, the
Vibration Isolation Platform consists of the following main components:
1. Voice-Coil Actuators: 6 voice-coil actuators are mounted on the middle mass
casing. Three of these actuators (positioned 120 degrees apart on a circle of
radius 4.400 inches) are used to shake the middle mass and act as the source of
disturbance to the platform. They can also be used to introduce some desired
dynamics for the middle mass that supports the payload. As shown in Figure
4.5, these actuators act against the ground. The other three actuators (placed
120 degrees apart on a circle of radius 4.000 inches) act against the middle mass
and are used to isolate the payload (top mass) from the motions of the middle
mass. Note that the two circles on which control and disturbance actuators are
mounted are concentric, and one set of actuators is rotated in the horizontal
plane by 60 degrees with respect to the other.
2. Sensors: VIP is equipped with two sets of sensors,
(a) Position Sensors: Each actuator is equipped with a colocated position
(gap) sensor which is physically inside the casing of the actuator. Three
additional position sensors are used as truth sensors (Figure 4.5) to measure the displacement of the payload in the inertial frame.
4.3. SIMULATION RESULTS
87
(b) Load Cells: Three load cells are used to measure the interaction forces
between the middle mass and the payload. These sensors are colocated
with the point of contact of the control actuators and the payload. It is
important to note that any interaction force between the payload and the
rest of the VIP system is transfered via these load cells.
A state-space model for the VIP platform is identified using the FORSE system
identification software developed at MIT [30]. The detailed description of the identification process, the identified model, and the model validation process is discussed
in Appendix C. Figure 4.6 shows the singular value plots for the MIMO transfer
functions from control ([u])/disturbance ([d]) actuators to the load cells ([lc]) and
scoring sensors ([sc]).
In all simulations that follow, the length of the adaptive FIR filters, i.e. L in
Equation 4.2, is 4 (unless stated otherwise). This length is found to be sufficient for
an acceptable performance of the adaptive algorithms. The sampling frequency for
all the simulations in this section is 1000 Hz. Furthermore, all measurements are
subject to band limited white noise with power 0.008.
Figures 4.7 and 4.8 show the reading of the scoring sensors (i.e. the variations of
the payload from the equilibrium position in the inertial frame) for the multi-channel
implementation of the EBAF and the FxLMS adaptive algorithms, respectively. Disturbance actuators apply sinusoidal excitation of amplitude 0.1 Volts at 4 Hz to the
middle mass. The phase for the excitation of the disturbance actuator #1 is assumed
to be zero, while disturbance actuators #2 and #3 are 22.5 and 45 degrees out of
phase with the first actuator. Only a noisy measurement of the primary disturbance
is assumed to be available to the adaptive algorithms. The signal to noise ratio for
the available reference signal is 3.0. For simulations in Figures 4.7 and 4.8 the control
signal starts at t = 30 seconds. Figure 4.7 shows that the amplitude of the transient
vibrations of the payload under the EBAF adaptive algorithm (for 30 ≤ t ≤ 60)
does not exceed that of the open loop vibrations. In contrast, the amplitude of the
transient vibrations under the FxLMS, Figure 4.8, exceeds twice the amplitude of
the open loop vibrations in the system. For a smaller amplitude during transient
4.3. SIMULATION RESULTS
88
vibrations, the adaptation rate for the FxLMS algorithm should be reduced. This
will result in an even slower convergence of the adaptive algorithm. Note that, for the
results in Figure 4.8, the adaptation rate is 0.0001. Even with this adaptation rate,
FxLMS algorithm requires approximately 20 more seconds (compared to the EBAF
case) to converge to its steady state value. In the steady state, the EBAF algorithm
achieves a 20 times reduction in the amplitude of the payload vibrations. For the
FxLMS algorithm in this case, the measured reduction is approximately 16 times.
Figures 4.9 and 4.10 show the reading of the scoring sensors when the primary
disturbances are multi-tone sinusoids. The primary disturbance consists of sinusoidal
signal of amplitudes 0.1 and 0.2 volts at 4 and 15 Hz, respectively. As in the single tone
case, both components of the excitation for the disturbance actuator #1 are assumed
to have zero phase. Each sinusoidal component of the excitation for the disturbance
actuator #2 (#3) is assumed to have a phase lag of 22.5 (45) degrees with respect to
the corresponding component of the excitation in actuator #1. Figures 4.9 and 4.10
demonstrate a trend similar to that discussed for the single tone scenario. For the
FxLMS algorithm, a trade off between the amplitude of the transient vibrations of
the payload and the speed of the convergence exists. The adaptation rate here is the
same as the single tone case. Slower adaptation rates can reduce the amplitude of the
transient vibrations at the expense of the speed of the convergence. For the EBAF
algorithm, however, better transient behavior and faster convergence are observed. In
the steady state, the EBAF algorithm provides a 15 times reduction in the amplitude
of the vibrations of the payload. For the FxLMS algorithm a 9 times reduction is
recorded.
In Fig. 4.11 the effect of feedback contamination in the performance of the EBAF
algorithm is examined. As in the previous simulations, control actuators are switched
on at t = 30 seconds. The reference signal available to the adaptation algorithm is
the output of the load cells which measure the forces transfered to the payload. Obviously, load cell measurements contain the effect of both primary disturbances (single
tone sinusoids at 4 Hz for this example), and control actuators (and hence the classical feedback contamination problem). Here, no special measure to counter feedback
contamination is taken. Figure 4.11 shows that an average of 4 times reduction in the
4.3. SIMULATION RESULTS
89
magnitude of the vibrations transfered to the payload is achieved (i.e. a degraded performance when compared to the case without feedback contamination). The EBAF
algorithm, however, maintains its stability in the face of feedback contamination
without any additional measures, hence exhibiting robustness of the algorithm to the
contamination of the reference signal. Note that for the FxLMS adaptive algorithm,
with the adaptation rate similar to that in Figure 4.10, feedback contamination leads
to an unstable behavior. The adaptation rate should be reduced substantially (hence
extremely slow convergence of the FxLMS algorithm), in order to recover the stability
of the adaptation scheme.
The simulations in this section have shown that the multi-channel implementation
of the estimation-based adaptive filtering algorithm provides the same advantages
observed for the single channel case. More specifically, the multi-channel EBAF
algorithm achieves desirable transient behavior and fast convergence without compromising steady-state performance of the adaptive algorithm. It also demonstrates
robustness to feedback contamination. It is important to note that, the above mentioned advantages are achieved by an approach which is essentially identical to the
single-channel version of the algorithm.
4.3.2
Active Noise Cancellation
Consider the one dimensional acoustic duct shown in Figure 2.5. Here, disturbances
enter the duct via Speaker #1. The objective of the multi-channel noise cancellation
is to use both available speakers to simultaneously cancel the effect of the incoming
disturbance at Microphones #1 and #2. The control signal is supplied to each speaker
via an adaptive FIR filter (i.e. in the case of Speaker #1 added to the primary
disturbance). Figure 4.12 shows the output of the microphones when the primary
disturbance (applied to Speaker #1) is a multi-tone sinusoid with 150 Hz and 200
Hz frequencies. The length of each FIR filter in this simulation is 8. For the first
two seconds the controller is off (i.e both adaptive filters have zero outputs). At
t = 2.0 seconds the controller is switched on. The initial value for the Riccati matrix
is P0 = diag(0.0005I2(N +1) , 0.00005INs ) (where N + 1 is the length of each FIR filter,
4.3. SIMULATION RESULTS
90
and Ns is the order of the secondary path). It is clear that the error at Microphone
#2 is effectively canceled in 0.2 seconds. For Microphone #1 however the cancellation
time is approximately 5 seconds. A 30 times reduction in disturbance amplitude is
measured at Microphone #2 in approximately 10 seconds. For Microphone #1 this
reduction is approximately 15 times. Note that the distance between Speaker #2
and Microphone #1 (46 inches) is much greater than the distance between Speaker
#1 and Microphone #1 (6 inches). Due to this physical constraint, Speaker #2
alone, is not enough for an acceptable noise cancellation at both microphones. The
experimental data in Figures 2.9 and 2.10 in which the result of a single-channel
implementation of the EBAF algorithm, aimed at noise cancellation at Microphone
#2, was shown, confirm this observation. Using a multi-channel approach however,
allows for a substantial reduction in the amplitude of the measured noise at both
microphones. Nevertheless, noise cancellation at the position of Microphone #1 tends
to be slower than the noise cancellation at the position of Microphone #2.
A similar scenario with band limited white noise as the primary disturbance is
shown in Fig. 4.13. Here the length of each FIR filter is 32. The performance of the
adaptive multi-channel noise cancellation problem in the frequency domain is shown
in Fig. 4.14. Once again the cancellation at Microphone #2 is superior to that at
Microphone #1.
4.3. SIMULATION RESULTS
Primary Source
ref (k)
91
J
Vm (k)
M
Primary
Path
d(k)
+
+
+
−
Feedback
Path
e(k)
+
+
−
y(k)
Adaptive
Filter
K
Secondary
Path
u(k)
x(k)
Fig. 4.1: General block diagram for a multi-channel Active Noise Cancellation (ANC) problem
Primary Path
Modeling Error
x(k)
A Copy Of
Secondary Path
An FIR Filter
J
K
Secondary Path
Adaptive FIR Filter
u(k)
Vm (k)
d(k)
+
+ M
+
−
+
e(k)
y(k)
Adaptation Algorithm
Fig. 4.2: Pictorial representation of the estimation interpretation of the adaptive control problem:
Primary path is replaced by its approximate model
4.3. SIMULATION RESULTS
x(k)
A Copy Of Adaptive Filter
92
A Copy Of Secondary Path
Vm (k)
Ds (k)
A Matrix Of
Adaptive
Filters
Bs (k)
+
Z −1
Cs (k)
As (k)
Fig. 4.3: Approximate Model for Primary Path
Fig. 4.4: Vibration Isolation Platform (VIP)
+
+
d(k)
m(k)
4.3. SIMULATION RESULTS
93
Fig. 4.5: A detailed drawing of the main components in the Vibration Isolation Platform (VIP).
Of particular importance are: (a) the platform supporting the middle mass (labeled as component
#5), (b) the middle mass that houses all six actuators (of which only two, one control actuator and
one disturbance actuator) are shown (labeled as component #11), and (c) the suspension springs
to counter the gravity (labeled as component #12). Note that the actuation point for the control
actuator (located on the left of the middle mass) is colocated with the load cell (marked as LC1).
The disturbance actuator (located on the right of the middle mass) actuates against the inertial
frame.
4.3. SIMULATION RESULTS
94
SVD for [d]−>[lc]
SVD for [u]−>[lc]
Volts/volts
0
0
10
10
−5
−5
10
10
0
2
10
0
10
SVD for [d]−>[sc]
10
SVD for [u]−>[sc]
0
Volts/volts
2
10
0
10
10
−5
−5
10
10
0
2
10
10
Frequency (Hz)
0
2
10
10
Frequency (Hz)
Fig. 4.6: SVD of the MIMO transfer function
4.3. SIMULATION RESULTS
95
Scoring Sensor Readouts - EBAF
e(1)
10
0
−10
0
20
40
60
80
100
120
0
20
40
60
80
100
120
0
20
40
60
80
100
120
e(2)
10
0
−10
e(3)
10
0
−10
Time (sec.)
Fig. 4.7: Performance of a multi-channel implementation of EBAF algorithm when disturbance
actuators are driven by out of phase sinusoids at 4 Hz. The reference signal available to the adaptive
algorithm is contaminated with band limited white noise (SNR=3). The control signal is applied
for t ≥ 30 seconds.
4.3. SIMULATION RESULTS
96
Scoring Sensor Readouts - FxLMS
e(1)
10
0
−10
0
20
40
60
80
100
120
0
20
40
60
80
100
120
0
20
40
60
80
100
120
e(2)
10
0
−10
e(3)
10
0
−10
Time (sec.)
Fig. 4.8: Performance of a multi-channel implementation of FxLMS algorithm when simulation
scenario is identical to that in Figure 4.7.
4.3. SIMULATION RESULTS
97
Scoring Sensor Readouts - EBAF
e(1)
10
0
−10
0
20
40
60
80
100
120
0
20
40
60
80
100
120
0
20
40
60
80
100
120
e(2)
10
0
−10
e(3)
10
0
−10
Time (sec.)
Fig. 4.9: Performance of a multi-channel implementation of EBAF algorithm when disturbance
actuators are driven by out of phase multi-tone sinusoids at 4 and 15 Hz. The reference signal
available to the adaptive algorithm is contaminated with band limited white noise (SNR=4.5). The
control signal is applied for t ≥ 30 seconds.
4.3. SIMULATION RESULTS
98
Scoring Sensor Readouts - FxLMS
e(1)
10
0
−10
0
20
40
60
80
100
120
0
20
40
60
80
100
120
0
20
40
60
80
100
120
e(2)
10
0
−10
e(3)
10
0
−10
Time (sec.)
Fig. 4.10: Performance of a multi-channel implementation of FxLMS algorithm when simulation
scenario is identical to that in Figure 4.9.
4.3. SIMULATION RESULTS
99
Scoring Sensors Readouts for Single-Tone at 4 Hz With Feedback Contamination
e(1)
5
0
−5
0
20
40
60
80
100
120
0
20
40
60
80
100
120
0
20
40
60
80
100
120
e(2)
5
0
−5
e(3)
5
0
−5
Time (sec.)
Fig. 4.11: Performance of a Multi-Channel implementation of the EBAF for vibration isolation
when the reference signals are load cell outputs (i.e. feedback contamination exists). The control
signal is applied for t ≥ 30 seconds.
4.3. SIMULATION RESULTS
100
Multi-Channel Active Noise Cancellation in Acoustic Duct
Microphone #1 (Volts)
4
2
0
−2
−4
−6
0
1
2
3
4
0
1
2
3
4
5
6
7
8
9
10
5
6
7
8
9
10
Microphone #2 (Volts)
4
2
0
−2
−4
Time (sec.)
Fig. 4.12: Performance of the Multi-Channel noise cancellation in acoustic duct for a multi-tone
primary disturbance at 150 and 200 Hz. The control signal is applied for t ≥ 2 seconds.
4.3. SIMULATION RESULTS
101
Multi-Channel Active Noise Cancellation in Acoustic Duct
Microphone #1 (Volts)
15
10
5
0
−5
−10
0
1
2
3
4
0
1
2
3
4
5
6
7
8
9
10
5
6
7
8
9
10
Microphone #2 (Volts)
10
5
0
−5
−10
Time (sec.)
Fig. 4.13: Performance of the Multi-Channel noise cancellation in acoustic duct when the
primary disturbance is a band limited white noise. The control signal is applied for t ≥ 2 seconds.
4.3. SIMULATION RESULTS
102
Speaker #1 → Microphone #1
Oloop
Cloop
Magnitude
0
10
−1
10
−2
10
2
3
10
10
Speaker #1 → Microphone #2
Oloop
Cloop
Magnitude
0
10
−1
10
−2
10
2
10
3
10
Fig. 4.14: Closed-loop vs. open-loop transfer functions for the steady state performance of the
EBAF algorithm for the simulation scenario shown in Figure 4.13.
4.4. SUMMARY
4.4
103
Summary
The estimation-based synthesis and analysis of multi-channel adaptive (FIR) filters is
shown to be identical to the single-channel case . Simulations for a 3-input/3-output
Vibration Isolation Platform (VIP), and a multi-channel noise cancellation in the one
dimensional acoustic duct are used to demonstrate the feasibility of the estimationbased approach. The new estimation-based adaptive filtering algorithm is shown to
provide both faster convergence (with an acceptable transient behavior), and improved steady state performance when compared to a multi-channel implementation
of the FxLMS algorithm.
Chapter 5
Adaptive Filtering via Linear
Matrix Inequalities
In this chapter Linear Matrix Inequalities (LMIs) are used to synthesize adaptive
filters (controllers). The ability to cast the synthesis problem as LMIs is a direct
consequence of the systematic nature of the estimation-based approach to the design
of adaptive filters proposed in Chapters 2 and 3 of this thesis. The question of internal
stability of the overall system is directly addressed as a result of the Lyapunovbased nature of the LMIs formulation. LMIs also provide a convenient framework for
the synthesis of multi-objective (H2 /H∞ ) control problems. This chapter describes
the process of augmenting the H∞ criterion that serves as the center piece of the
estimation-based adaptive filtering algorithm with H2 performance constraints, and
investigates the characteristics of the resulting adaptive filter. As in Chapters 2 and 3,
an Active Noise Cancellation (ANC) scenario is used to study the main features of
the proposed formulation.
5.1
Background
A detailed discussion of the estimation-based approach to the design of adaptive
filters (controllers) is presented earlier in Chapters 2 and 3. The discussion here will
104
5.1. BACKGROUND
105
be kept brief and serves more as a notational introduction. Figure 5.1 is a block
diagram representation of the active noise cancellation problem, originally shown in
Figure 2.1. Clearly, the objective here is to generate a control signal u(k) such that the
output of the secondary path, y(k), is close to the output of the primary path, d(k).
In Chapter 2, it is shown that an estimation interpretation of the adaptive filtering
(control) problem can be used to formulate an equivalent estimation problem. It is
this equivalent estimation problem that admits LMIs formulation.
To mathematically describe the equivalent estimation problem, state space models for the adaptive filter and the secondary path are needed. As in Chapter 2,
[As (k), Bs (k), Cs (k), Ds (k)] is the state space model for the secondary path. The
state variable for the secondary path is θ(k). For the adaptive filter the weight vector, W (k) = [ w0 (k) w1 (k) · · · wN (k) ]T , will be treated as the state variable. The
state space description for the approximate model of the primary path can then be
described as:
"
W (k + 1)
#
"
=
θ(k + 1)
I(N +1)×(N +1)
Bs (k)h∗k
0
#"
As (k)
W (k)
#
θ(k)
ξk+1 = Fk ξk
(5.1)
where
h(k) = [x(k) x(k − 1) · · · x(k − N)]T
(5.2)
captures the effect of the reference input x(k). Note that in Figure 5.1
e(k) = d(k) − y(k) + Vm (k)
(5.3)
where e(k) is the available error measurement, Vm (k) is the exogenous disturbance
that captures measurement noise, modeling error and uncertainty in the initial condition of the secondary path, and y(k) is the output of the secondary path. To formulate
the estimation problem, a derived measurement for the output of the primary path is
needed. Rewriting Eq. (5.3) as
4
m(k) = e(k) + y(k) = d(k) + Vm (k)
(5.4)
the right hand side, i.e. d(k) + Vm (k), can be regarded as the “noisy measurement”
of the primary path output. Note that on the left hand side of Eq. (5.4) only e(k) is
5.1. BACKGROUND
106
directly measurable. In general, an internal copy of the output of the secondary path,
referred to as ycopy (k), should be generated by the adaptive algorithm. Section 2.7
discusses the ramifications of the introduction of this internal copy in detail, and
shows that for the stable linear secondary path the difference between ycopy (k) and
y(k) will decay to zero for sufficiently large k. Now, the derived measured output for
the equivalent estimation problem can be defined as
#
"
h
i W (k)
m(k) =
+ Vm (k)
Ds (k)h∗k Cs (k)
θ(k)
= Hk ξk + Vm (k)
(5.5)
where m(k) is defined in Equation (5.4). Noting the objective of the noise cancellation
problem, s(k) = d(k) is chosen as the quantity to be estimated:
#
"
h
i W (k)
s(k) =
Ds (k)h∗k Cs (k)
θ(k)
= Lk ξk
(5.6)
Note that m(·) ∈ R1×1 , s(·) ∈ R1×1 , θ(·) ∈ R1×1 , and W (·) ∈ R(N +1)×1 . All matrices
are then of appropriate dimensions. ξkT = W T (k) θT (k) is clearly the state vector
for the overall approximate model of the primary path.
4
The following H∞ criterion can be used to generate ŝ(k) = F (m(0), · · · , m(k − 1))
(the prediction estimate of the desired quantity s(k)) such that the worst case energy
gain from the measurement disturbance and the initial condition uncertainty to the
error in the causal estimate of s(k) is properly limited, i.e.
M
X
sup
Vm , ξ0
[s(k) − ŝ(k)]∗ [s(k) − ŝ(k)]
≤ γ2
k=0
ξ0∗ Π−1
0 ξ0
+
M
X
(5.7)
∗
Vm
(k)Vm (k)
k=0
The Riccati-based solution to this problem is discussed in Chapter 2 in detail. It is
sometimes desirable, however, to have the adaptive filter meet some H2 performance
criteria in addition to the H∞ constraint in Equation (5.7). Linear matrix inequalities
5.2. LMI FORMULATION
107
offer a convenient framework for formulating the mixed H2 /H∞ synthesis problem.
Furthermore, numerically sound algorithms exist that can solve these linear matrix
inequalities very efficiently. Therefore, next section pursues a first principle derivation
of the LMI formulation for the design of adaptive filters.
5.2
LMI Formulation
Assume the following specific structure for the estimator
ˆ
ˆ
ˆ
ξk+1 = Fk ξk + Γk m(k) − Hk ξk
ŝ(k) = Lk ξˆk
(5.8)
(5.9)
in which Γk is the design parameter to be chosen such that the H∞ criterion is met.
Now, the augmented system can be defined as follows
"
#
"
#"
# "
#
ξk+1
0
Fk
ξk
0
=
+
Vm (k) (5.10)
ξˆk+1
ξˆk
Fk Hk Fk − Γk Hk
−Γk
"
#
i ξ
h
4
k
Zk = s(k) − ŝ(k) =
(5.11)
Lk −Lk
ˆ
ξk
4
Introducing a new variable ξ˜k = ξk − ξˆk , the augmented system can be described as
#
"
#"
# "
#
"
0
Fk
ξk
0
ξk+1
=
+
Vm (k)
0 Fk − Γk Hk
Γk
ξ˜k+1
ξ˜k
ηk+1 = Φk ηk + Ψk Vm (k)
(5.12)
with
4
Zk = s(k) − ŝ(k) =
i
h
0 Lk
"
ξk
ξ˜k
#
= Ωk ηk
(5.13)
The LMI solution for the design of adaptive filters finds a Lyapunov function for the
augmented system in (5.12)-(5.13) at each step. In other words, at each time step, an
infinite horizon problem is solved, and the solution is implemented for the next step.
5.2. LMI FORMULATION
108
Introducing the quadratic function V (ηk ) = ηk T P ηk (where P > 0), it is straight
forward [8] to show that for the augmented system at time k, (5.7) holds if
V (ηk+1 ) − V (ηk ) < γ 2 Vm (k)T Vm (k) − ZkT Zk
(5.14)
Note that the inclusion of the energy of the initial condition error will only increase
the right hand side of the inequality in Eq. (5.14). Replacing for Zk and ηk+1 from
(5.12)-(5.13), and after some elementary algebraic manipulations the inequality in
(5.14) can be written as
"
#"
#
h
i ΦT P Φ − P + ΩT Ω
T
η
Φ
P
Ψ
k
k
k
k
k
k
k
T
<0
ηkT Vm
(k)
T
T
Ψk P Φk
Ψk P Ψk − γ 2 I
Vm (k)
(5.15)
Now, due to the block diagonal structure of the matrix Φk in (5.12), the Lyapunov
matrix P can be chosen with a block diagonal structure
"
#
R 0
P =
0 S
(5.16)
Replacing for P , (5.15) will reduce to

F T RFk − R
0
0
 k

0
(Fk − Γk Hk )T S (Fk − Γk Hk ) − S + LTk Lk (Fk − Γk Hk )T SΓk

0
ΓTk S (Fk − Γk Hk )
ΓTk SΓk − γ 2 I


<0

(5.17)
in which R is independent of Γk and S. To formulate the 2 × 2 block in (5.17) as an
LMI in S and Γk , note that
#
"
(Fk − Γk Hk )T S (Fk − Γk Hk ) − S + LTk Lk (Fk − Γk Hk )T SΓk
ΓTk S (Fk − Γk Hk )
z"
D
}|
T
Lk Lk − S
0
0
−γ 2 I
#{
z"
+
ΓTk SΓk − γ 2 I
C
=
}|
B
#{ A−1 z
}|
T
i{
h
(Fk − Γk Hk ) S z}|{
−1
S
(5.18)
S (Fk − Γk Hk ) SΓk
ΓTk S
5.2. LMI FORMULATION
109
and hence using Schur complement the inequality in (5.18) can be written as the
following set of linear matrix inequalities
 

−S
S (Fk − Γk Hk ) SΓk




  (F − Γ H )T S
LTk Lk − S
0
k k
 k

0
−γ 2 I
ΓTk S






<0

(5.19)
S>0
Note that the LMI corresponding to the (1, 1) block in Eq. (5.17), i.e.
FkT RFk − R ≤ 0
(5.20)
can never be strict (Fk has eigenvalues on unit circle). Most SDP-solvers, however,
only solve strictly feasible LMIs. Therefore, Fk should be first put in its modal form
"
#
I
0
Fk = Θ−1
Θk
(5.21)
k
0 ΛAs,k
for some diagonalizing transformation matrix Θk to isolate the poles on the unit circle. Straightforward matrix manipulations then lead to the following LMI in S, Rs ,
4
T = SΓk , and γ 2 which can be strictly feasible:
Minimize γ 2 > 0 subject to
 

−S
S (Fk − Γk Hk )





T


LTk Lk − S

  (Fk − Γk Hk ) S
ΓTk S








"
where R = ΘTk
0

SΓk
0
−γ 2 I

<0

(5.22)
ΛTAs,k Rs ΛAs,k − Rs < 0
S > 0, Rs > 0
I
0
#
Θk . The solution to (5.22) provides estimator gain, Γk , as
0 Rs
well as the Lyapunov matrix P which ensures that the quadratic cost V (ηk ) decreases
over time. It is shown in Chapter 2 that for the Riccati-based solution to Eq. (5.7)
the optimal value of γ is 1. In the absence of H2 constraints, γ in Eq. (5.22) can
be set to 1. This reduces the LMI solution to a feasibility problem in which S > 0,
Rs > 0, and T should be found.
5.2. LMI FORMULATION
5.2.1
110
Including H2 Constraints
Augmenting the above mentioned H∞ objective with appropriate H2 performance
constraints (such as the H2 norm of the transfer function from exogenous disturbance
Vm (k) to the cancellation error Zk in the augmented system described by (5.12)(5.13)) is also straight forward. Recall that [8]
k TZVm k22 = Tr Ωk Wc ΩTk
(5.23)
where Wc satisfies
ΦTk Wc Φk − W c + Ψk ΨTk = 0
(5.24)
To translate this into LMI constraints [10], note that bounding the H2 norm by ν 2 is
equivalent to
 "
#
T

Φ
W
Φ
−
W
Ψ

c
k
c
k
k

<0



ΨTk
−I


#
"
Q
Ωk Wc

>0



Wc ΩTk
Wc




Tr (Q) − ν 2 < 0
(5.25)
Allowing some conservatism, we pick Wc = P to augment (5.25) with the LMI constraints in (5.22). After some algebraic manipulation, we first express (5.25) as

 
T

F
RF
−
R
0
0
k

  k


T


<0

0
(F
−
Γ
H
)
S
(F
−
Γ
H
)
−
S
−Γ
k
k
k
k
k
k
k






T

0
−Γ
−I

k



(5.26)
Q
0 Lk S





 0

R 0 

>0





SLTk 0
S




Tr (Q) − ν 2 < 0
To formulate the 2 × 2 block in the first inequality in (5.26) as an LMI, note that
"
#
(Fk − Γk Hk )T S (Fk − Γk Hk ) − S −Γk
=
−ΓTk
−I
"
# "
#
h
i(5.27)
FkT SFk − S − HkT ΓTk SFk − FkT SΓk Hk −Γk
HkT ΓTk S
+
S −1 SΓk Hk 0
−ΓTk
−I
0
5.3. ADAPTATION ALGORITHM
111
and that the LMI constraint on R can be decoupled from the rest. Therefore the
H2 /H∞ problem can be formulated as the following LMI in S, T , Q, and Γk
Minimize























αγ 2 + βTr(Q) (α and β are known constants) subject to (5.22) and


−S
SΓk Hk
0


 H T ΓT S F T SFk − S − H T ΓT SFk − F T SΓk Hk −Γk  < 0
k
k k
k
 k k

T
0
−Γk
−I
"
#
(5.28)
Q Lk S
>0
SLTk
S
Tr (Q) − ν 2 < 0
Note that T = SΓk and Γk are not independent variables, and an appropriate linear
matrix inequality should be used to reflect this interdependence. For an alternative
derivation of the LMI formulation for the mixed H2 /H∞ synthesis problem see [10]
and references therein.
5.3
Adaptation Algorithm
The implementation scheme is similar to that of the adaptive FIR filter discussed in
Chapter 2. For easier reference, the main signals involved in the description of the
adaptive algorithm are briefly introduced here. For a more detailed description please
see Chapter 2. In what follows (a) Ŵ (k) is the estimate of the adaptive weight vector,
4
(b) θ̂(k) is the estimate of the state of the secondary path, (c) u(k) = h∗ (k)Ŵ (k)
is the actual control input to the secondary path, (d) y(k) and d(k) are the actual
outputs of the secondary and primary paths, respectively, (e) e(k) is the actual error
measurement, and (f) θcopy (k) and ycopy (k) are the adaptive algorithm’s internal copy
of the state and output of the secondary path which are used in constructing m(k)
according to Eq. (5.4). Now, the implementation algorithm can be outlined as follows:
1. Start with Ŵ (0) = Ŵ0 , θ̂(0) = θ̂0 as estimator’s best initial guess for the
state vector in the approximate model of the primary path. Also assume that
5.3. ADAPTATION ALGORITHM
112
θcopy (0) = θcopy,0 is adaptive algorithm’s initial condition for the internal copy
of the state of the secondary path.
2. For 0 ≤ k ≤ M(f inite horizon):
(a) Form h(k) according to Eq. (5.2),
(b) Form the control signal u(k) = h∗ (k)Ŵ (k), to be applied to the secondary
path. Note that applying u(k) to the secondary path produces
y(k) = Cs (k)θ(k) + Ds (k)u(k)
(5.29)
at the output of the secondary path. This in turn leads to the following
error signal measured at time k:
e(k) = d(k) − y(k) + Vm (k)
(5.30)
which is available to the adaptive algorithm to perform the state update
at time k.
(c) Propagate the internal copy of the state vector and the output of the
secondary path as
θcopy (k + 1) = As (k)θcopy (k) + Bs (k)u(k)
(5.31)
ycopy (k) = Cs (k)θcopy (k) + Ds (k)u(k)
(5.32)
(d) Form the derived measurement, m(k), using the direct measurement e(k)
and the controller’s copy of the output of the secondary path
m(k) = e(k) + ycopy (k)
(5.33)
Note that e(k) is the error measurement after the control signal u(k) is
applied.
(e) Use the LMI formulation in (5.22) (the LMI in (5.28) in case H2 constraints
exist) to find Γk (estimator’s gain).
(f) Update the state estimate according to Eq. (5.8), and extract Ŵ (k + 1)
from ξˆk+1 as the new value for the adaptive weight vector in the adaptive
filter.
(g) If k ≤ M, go to (a).
5.4. SIMULATION RESULTS
5.4
113
Simulation Results
In this section the feasibility of the design procedure is examined in the context of
Active Noise Cancellation in the one dimensional acoustic duct shown in Figure 2.5.
The identified transfer function for the acoustic duct is shown in Figures 2.6 and 2.7.
The single channel noise cancellation scenario is considered here (i.e. Speaker #2 is
used to cancel, at the position of Microphone #2, the effect of the disturbance that
enters the acoustic duct via Speaker #1). The length of the adaptive FIR filter in
this example is 8. The results presented in this section are intended to demonstrate
the feasibility of the LMI formulation in adaptive FIR filter design.
Figure 5.2 compares the output of the secondary path, y(k), to the output of the
primary path, d(k), when the primary disturbance input is a sinusoid at 30 Hz. The
adaptive algorithm has full access to the primary disturbance in this case. The error
in noise cancellation, i.e. the measurement of Microphone #2, is also plotted in this
figure. Note that, error measurements are subject to band limited white Gaussian
noise with power 0.008. Effectively, after 0.2 seconds the output of the adaptive filter
reaches its steady-state value. An approximately 10 times reduction in the amplitude
of the disturbances at Microphone #2 is recorded. A typical behavior of the elements
of the adaptive weight vector is shown in Figure 5.3. The variations in the elements
of the weight vector during steady-state operation of the algorithm are small, and
effectively after 0.2 seconds they assume their steady-state values.
Figure 5.4 shows the simulation results for a multi-tone primary disturbance that
consists of 30 and 45 Hz frequencies. In this case the output of the adaptive filter
effectively reaches its steady-state in about 0.35 seconds. The reduction in the magnitude of the measured disturbances at the position of Microphone #2 is approximately
5 times in this case. The simulation conditions are similar to the single-tone case. A
typical behavior of the elements of the adaptive FIR weight vector is shown in Figure 5.5. The elements of the weight vector display more variation during steady-state
operation of the adaptive FIR filter.
Even though the objective of the noise cancellation problem in this chapter is
the same as that in Chapter 2, the LMI-based adaptive algorithm is computationally
5.4. SIMULATION RESULTS
114
more expensive. Furthermore, simulation results presented in this section indicate
that, for the formulation of the problem presented in this chapter, the performance
of the Riccati-based EBAF is better than that of the LMI-based EBAF. This can
be associated with the conservatism introduced in the formulation, in particular the
diagonal structure assumed for the matrix P (Eq. (5.16)). The assumption of a
constant P matrix also results in additional conservatism. Nevertheless, the problem
formulation as LMIs provides a potent framework in which multiple design objectives
can be handled. Furthermore, the uncertainty in system model can be systematically
addressed in the LMI framework.
5.4. SIMULATION RESULTS
x(k)
115
Vm (k)
d(k)
Primary Path
(Unknown)
+
+
+
−
Adaptive
Filter
u(k)
Secondary Path
(Known)
y(k)
Adaptation
Algorithm
Fig. 5.1: General block diagram for an Active Noise Cancellation (ANC) problem
e(k)
5.4. SIMULATION RESULTS
116
LMI Single-Tone Noise Cancellation
d(k) vs. y(k) (Volts)
1
d(k)
y(k)
0.5
0
−0.5
−1
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
1
Microphone #2 (Volts)
e(k)
0.5
0
−0.5
−1
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
Time (sec.)
Fig. 5.2: Cancellation Error at Microphone #1 for a Single-Tone Primary Disturbance
0.5
5.4. SIMULATION RESULTS
117
0.2
0.2
Wk (1)
Filter Weight
0.15
0.1
0.1
0.05
0.05
0
0
−0.05
0
0.1
0.2
0.3
0.4
0.5
0.2
Filter Weight
−0.05
0
0.1
0.2
0.3
0.4
0.5
0.2
Wk (5)
0.15
0.1
0.05
0.05
0
0
0
0.1
0.2
0.3
Time(sec.)
0.4
Wk (8)
0.15
0.1
−0.05
Wk (3)
0.15
0.5
−0.05
0
0.1
0.2
0.3
Time(sec.)
0.4
0.5
Fig. 5.3: Typical Elements of Adaptive Filter Weight Vector for Noise Cancellation Problem in
Fig. 5.2
5.4. SIMULATION RESULTS
118
LMI Multi-Tone Noise Cancellation
d(k) vs. y(k) (Volts)
2
d(k)
y(k)
1
0
−1
−2
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
Microphone #2 (Volts)
2
e(k)
1
0
−1
−2
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
Time (sec.)
Fig. 5.4: Cancellation Error at Microphone #1 for a Multi-Tone Primary Disturbance
0.5
5.4. SIMULATION RESULTS
119
0.2
0.2
Wk (1)
Filter Weight
0.15
0.1
0.1
0.05
0.05
0
0
−0.05
0
0.1
0.2
0.3
0.4
0.5
0.2
Filter Weight
−0.05
0
0.1
0.2
0.3
0.4
0.5
0.2
Wk (6)
0.15
0.1
0.05
0.05
0
0
0
0.1
0.2
Time (sec.)
0.3
0.4
Wk (8)
0.15
0.1
−0.05
Wk (3)
0.15
0.5
−0.05
0
0.1
0.2
0.3
0.4
0.5
Time (sec.)
Fig. 5.5: Typical Elements of Adaptive Filter Weight Vector for Noise Cancellation Problem in
Fig. 5.4
5.5. SUMMARY
5.5
120
Summary
This chapter suggests that LMI-based synthesis tools can be used to design adaptive
filters. The feasibility of this approach is demonstrated in a typical adaptive ANC
scenario. One clear benefit is that the framework is suitable for designing adaptive
filters in which performance and robustness concerns are systematically addressed.
Chapter 6
Conclusion
6.1
Summary of the Results and Conclusions
In this dissertation, a new estimation-based procedure for the systematic synthesis
and analysis of adaptive filters (controllers) in “Filtered” LMS problems has been
presented. This is a well known nonlinear control problem for which “good” systematic synthesis and analysis techniques are not yet available. This dissertation
has proposed a two step solution to the problem. First, it developed an estimation
interpretation of the adaptive filtering (control) problem. Based on this interpretation, the original adaptive filtering (control) problem is replaced with an equivalent
estimation problem. The weight vector of the adaptive filter (controller) is treated
as the state variable in this equivalent estimation problem. In the original adaptive
control problem, a measured error signal (i.e. the difference between a desired signal,
d(k) in Fig. 2.1, and a controlled signal, y(k) in Fig. 2.1) reflects the success of the
adaptation scheme. The equivalent estimation problem has been constructed such
that this error signal remains a valid measure for successful estimation. The second
step, is then to solve the corresponding estimation problem. An observer structure
for the estimator has been chosen, so that “estimates” of the optimal weight vector
can be formed. The weight vector in the adaptive filter is then tuned according to
these state estimate.
121
6.1. SUMMARY OF THE RESULTS AND CONCLUSIONS
122
Both H2 and H∞ measures can be used as estimation criteria. The H∞ criterion
was chosen for the development of the adaptive algorithm in this Thesis. More specifically, the equivalent estimation problem seeks an estimate of the adaptive weight
vector such that the energy gain from the exogenous disturbances and the initial
condition uncertainty to the cancellation error (i.e. the error between d(k) and y(k)
in Fig. 2.1) is minimized. This objective function is justified by the nature of the
disturbances in the applications of interest (i.e. active noise cancellation and active
vibration isolation). The following is a summary of the results in this Thesis:
1. In the case of adaptive FIR filters:
(a) The equivalent estimation problem is shown to be linear. Given a bound on
energy gain γ > 0, the robust estimation literature provides exact filtering
and prediction solutions for this problem. The work in this Thesis proves
that γ = 1 is the optimal energy gain, and derives the conditions under
which this bound is achievable.
(b) The optimality arguments in this Thesis provide the conditions under
which the existence of an optimal filtering (prediction) solution for the
problem is guaranteed. This eliminates the possibility of the solution
breakdown which could have prevented real-time implementation of the
algorithm.
(c) It is shown that the filtering and prediction solutions only require one
Riccati recursion. The recursion propagates forward in time, and does not
rely on any information about the future of the system or the reference
signal (thus allowing the resulting adaptive algorithm to be implementable
in real-time). This has come at the expense of restricting the controller to
an FIR structure in advance.
(d) For the optimal value of γ = 1, the above mentioned Riccati recursion simplifies to a Lyapunov recursion. This leads to a computational complexity
that is comparable to that of a classical filtered LMS adaptive algorithm,
such as the FxLMS.
6.1. SUMMARY OF THE RESULTS AND CONCLUSIONS
123
(e) Experimental results, along with extensive simulations have been used to
demonstrate the improved transient and steady-state performance of the
EBAF algorithm over classical adaptive filtering algorithms such as the
FxLMS and the Normalized FxLMS.
(f) A clear connection between the limiting behavior of the EBAF algorithm
and the FxLMS (Normalized-FxLMS) adaptive algorithm has been established. In particular, it is shown that the gain vector in the predictionbased (filtering-based) EBAF algorithm converges to the gain vector in the
FxLMS (Normalized FxLMS) as k → ∞. The error terms however, are
shown to be different. Thus, the classical FxLMS (Normalized FxLMS)
adaptive algorithms can be thought of as an approximation to the EBAF
algorithm in this Thesis. This connection might explain the observed
improvement in both transient and steady-state performance of the new
EBAF algorithm.
2. For the EBAF algorithm in the IIR case, it has been shown that the equivalent
estimation problem is nonlinear. An exact solution for the nonlinear robust
estimation problem is not yet available. A linearizing approximation that makes
systematic synthesis of adaptive IIR filter tractable has been adopted in this
Thesis. The performance of the EBAF algorithm in this case has been compared
to the performance of the Filtered-U LMS (FuLMS) adaptive algorithm. The
proposed EBAF algorithm has been shown to provide improved steady-state
and transient performance.
3. The treatment of feedback contamination problem has been shown to be identical to the IIR adaptive filter design in the new estimation-based framework.
4. A multi-channel extension of the EBAF algorithm has been provided to demonstrate that the treatment of the single-channel and multi-channel adaptive filtering (control) problems in the new estimation based framework is virtually
identical. Simulation results for the problem of vibration isolation (in a 3input/3-output vibration isolation platform (VIP)), and noise cancellation in a
6.2. FUTURE WORK
124
one dimensional acoustic duct have been shown to prove the feasibility of the
EBAF algorithm in the multi-channel case.
5. The new estimation-based framework has been shown to be amenable to a
Linear Matrix Inequality (LMI) formulation. The LMI formulation is used to
explicitly address the stability of the overall system under adaptive algorithm by
producing a Lyapunov function. Augmentation of an H2 performance constraint
to the H∞ disturbance rejection criterion has also been discussed.
6.2
Future Work
There are several possible directions for future work. The first direction is to address
the question of stability for the EBAF-based adaptive IIR filters. During extensive
simulations, the estimation-based adaptive IIR filter was observed to be stable. Obtaining a formal proof for the stability of the system, however, is a difficult problem.
Exploring the role of a different linearizing approximation, in reducing the nonlinear
robust estimation problem encountered in the IIR filter design into a tractable linear
estimation problem, is another interesting avenue for further research. It is also interesting to examine the possibility of formulating other classes of adaptive filtering
problems from the estimation point of view.
It is possible to formulate a systematic approach to the synthesis of “optimal”
adaptive filters (e.g. adaptive filters of optimal length) by augmenting the existing
estimation criterion with an appropriate constraint or objective function. The systematic synthesis of a robust adaptive filter, in which an error in the modeling of
the secondary path is explicitly handled during the synthesis process, would be another interesting extension to the work in this Thesis. The LMI formulation of the
estimation-based approach, in particular, offers a rich machinery for addressing such
problems.
Application of this approach to adaptive equalization is the subject of ongoing
research. For adaptive equalization, adopting an H2 estimation criterion is well justified. With an H2 objective, the estimation based adaptive filtering algorithm in
6.2. FUTURE WORK
125
this Thesis is in fact an extension of the RLS algorithm to the more general class of
filtered LMS problems.
Appendix A
Algebraic Proof of Feasibility
A.1
Feasibility of γf = 1
As Chapter 2 points out, it is possible to directly establish the feasibility of γf = 1.
To do so, it should be shown that Rk and Re,k (in Theorem 2.1) have the same inertia,
for all k, if γf = 1. First, define
"
4
Z =
A B
C D
#


= 

−
Pk−1
Hk
Hk
!
Hk∗
I
Hk∗
0

! 


(A.1)
0 −γ 2 I
Then, apply UDL decomposition to A.1 to get
#"
#"
"
#
#
"
A B
I −A−1 B
I
0
A
0
4
Z1 =
=
C D
−CA−1 I
0
I
0 ∆A
(A.2)
(where ∆A = D − CA−1 B is in fact Re,k ). Note that Z and Z1 have the same inertia.
Now, perform LDU decomposition to get
#"
#"
#
"
#
"
A B
I
0
∆D 0
I −BD −1
4
Z2 =
=
0
I
C D
−D −1 C I
0
D
(A.3)
(where ∆D = A − BD −1 C will reduce to A for γ = 1). Clearly, Z and Z2 have the
same inertia as well. Therefore, Z1 and Z2 should have the same inertia. Since the
126
A.1. FEASIBILITY OF γF = 1
127
(1, 1) block matrices for Z1 and Z2 are the same (i.e. A = −Pk−1 ), the (2, 2) block
matrices, i.e. Re,k in Z1 and Ri in Z2 , should have the same inertia, which is the
condition 2.13 in Theorem 2.1.
Appendix B
Feedback Contamination Problem
Figure B.1 shows the block diagram for an approximate model of the primary path
when a feedback path from the output of the IIR filter to its input exists. The
presentation here follows the discussions in Sections 3.2 and 3.3, and therefore it is
kept brief. The state space description for the block diagram in Figure B.1 is as
follows:


W (k + 1)



I(2N +1)×(2N +1)



 θ(k + 1)  =  Bs (k)h∗
k



ϕ(k + 1)
Bf (k)h∗k
0
As (k)
0
0
W (k)


 θ(k) 
0 


ϕ(k)
Af (k)
ξ(k+1) = Fk ξ(k)
(B.1)
where θk is the state variable for the secondary path, ϕk is the state variable for the
feedback path, and hk and W (k) are defined in Section 3.2.1. Note that
x(k) = ref (k) + Df (k)h∗ (k)W (k) + Cf (k)ϕ(k)
(B.2)
r(k) = a0 x(k) + b1 r(k − 1) + · · · + bN r(k − N), r(−1) = · · · = r(−N) = 0
(B.3)
where the contamination of the reference signal with the feedback from the output of
the adaptive filter is evident. For this system the derived measured output, m(k), is
128
129
described as


W (k)
h
Ds (k)h∗k
m(k) =
i

 + Vm (k)
Cs (k) 0 
θ(k)


ϕ(k)
= Hk ξ(k)
(B.4)
while the quantity to be estimated, s(k), is


h
s(k) =
W (k)
Ds (k)h∗k Cs (k) 0
i

 θ(k) 


ϕ(k)
= Lk ξ(k)
(B.5)
Similar to the derivations in Chapters 2 and 3, Hk = Lk , i.e. s(k) = d(k). It is desired
to find an H∞ causal filter ŝ(k|k) = F (m(0), m(1), · · · , m(k)) such that
M
X
sup
Vm , ξ0
(s(k) − ŝ(k|k))∗ (s(k) − ŝ(k|k))
k=0
ξ0∗ Π−1
0 ξ0 +
M
X
≤ γf2
(B.6)
Vm (k)∗ Vm (k)
k=0
Equivalently, an strictly causal predictor ŝ(k) = F (m(0), m(1), · · · , m(k − 1)) can be
found such that
M
X
sup
Vm , ξ0
(s(k) − ŝ(k))∗ (s(k) − ŝ(k))
k=0
ξ0∗ Π−1
0 ξ0 +
M
X
≤ γp2
(B.7)
Vm (k)∗ Vm (k)
k=0
Here γf (γp ) are positive numbers. Note that Vm (k) is assumed to be an L2 signal.
Obviously, Equations (B.1), (B.4) and (B.5) are nonlinear in IIR filter parameters
and therefore the estimation problem in Eq. (B.6) (or Eq. (B.7)) is a nonlinear H∞
problem. As in Sections 3.2 and 3.3, the IIR filter parameters in (B.2) and (B.3)
are replaced with their best available estimate to obtain a linear time-variant system
130
dynamics. For this linearized system the solution in Section 3.3 will exactly apply,
and hence it is not repeated here. It is clear that the above mentioned discussion
holds true when the adaptive IIR filter is replaced with an adaptive FIR filter.
131
Approximate Model for Primary Path With Feedback Contamination
Df (k)
ref (k)
+
Cf (k)
Z −1
+
Bf (k)
Af (k)
Vm (k)
Ds (k)
a0
+
r(k)
+
x(k)
Bs (k)
+
Z −1
u(k)
Z −1
Cs (k)
+
+
d(k) m(k)
As (k)
a1
b1
Z −1
b2
r(k-2)
a2
Second Order IIR Filter
Fig. B.1: Block diagram of the approximate model for the primary path in the presence of the
feedback path
Appendix C
System Identification for Vibration
Isolation Platform
C.1
Introduction
Advanced control techniques are model based techniques. The achievable performance for these control techniques, therefore, depends on the accuracy of the available model. Thus, the objective of system modeling is to capture the dynamics of a
system in the form of a mathematical model as accurately as possible. In general,
two approaches to the mathematical modeling of a system are used; (a) Analytical
Approach: applies rules of physics (that govern system dynamics) to derive a physical
model, and (b) System Identification Approach: uses the experimental input/output
data to construct a mathematical model.
Analytical methods represent not only the input/output behavior of the system
but also capture the internal mechanics and physics of the system. Identification based
methods, on the other hand, are mainly concerned with input/output behavior of the
system. Reference [22] provides a detailed discussion on the role of each approach in
a control design problem. This appendix discusses the state space model derived for
the Vibration Isolation Platform (VIP) based on an advanced method of fitting the
measured transfer function measurements. The main components of the VIP, along
132
C.2. IDENTIFIED MODEL
133
with their operational role are described in Chapter 4.
C.2
Identified Model
This section discusses the system identification strategy used to extract a state space
model for the Vibration Isolation Platform (VIP) based on transfer function measurements obtained over 0.50 − 660.0 Hz frequency range. The process of data collection
is described first. The consistency of the collected data for the load cells and truth
(scoring) sensors with the physical behavior of the system is examined next. Load
cell measurements reflect the interacting forces between the middle mass and the payload, while truth (scoring) position sensors measure the displacement of the payload
from its equilibrium in inertial frame. It is therefore important to develop a model
in which the readings from these two sets of sensors agree with the true physics of
the system. For this modeling to be successful, the collected data should correctly
captured the true physics of the problem, and the discussion on the consistency of the
measurements demonstrates this important fact. Finally, the identified state space
model is presented and compared with the original measurements.
C.2.1
Data Collection Process
A program that controls both actuator excitation and sensor measurements was developed to measure the transfer functions. This program runs on a sun SPARC1e
real time processor. The measurements are conducted as follows:
1. At a given frequency, a sinusoidal excitation is applied to only one actuator at
a time. Other actuators are commanded to zero. A total of 209 data points
(logarithmically spaced) are used to cover the 0.5 − 660 Hz frequency range.
The amplitude of the sinusoidal excitation is set to 0.2 volts to ensure linear
behavior for the system throughout the measurements.
2. The program maintains this excitation for 30 seconds (to assure that the system
reaches its steady state behavior) before it starts recording sensor readings. The
C.2. IDENTIFIED MODEL
134
sampling period for analog to digital conversion (performed by Tustin 2100 data
conversion device) is 63 micro-seconds. The same period is used for the digitalto-analog converter that drives the actuators.
3. The time data for the measurements is temporarily stored. With the known
excitation frequency, a least squares algorithm is applied to the measurements
over 20 cycles of the recorded data to extract the amplitude and phase of the
measurements for each sensor.
With this data for each test frequency point, the transfer functions from each actuator to all existing sensors can be constructed. This data is then used by system
identification routine to extract a state space model for the system.
C.2.2
Consistency of the Measurements
To examine the consistency of the available measurements, the physical relationship
between the load cell and truth sensor measurements should be explored. Note that,
from measurements, transfer function matrices Hlc,u , Hlc,d , Hsc,u , and Hsc,d defined
by


LC1

 

=
LCu = 
LC
2
 

LC3
 

LC1
 



LCd =  LC2 
=
LC3
 

SC1
 

=
SCu = 
SC
2
 

SC3
 

SC1
 

 
SCd = 
SC
2  = 

SC3

H1,1 H1,2 H1,3

U1


 U2  = Hlc,u U
H2,1 H2,2 H2,3 


H3,1 H3,2 H3,3
U3


H1,4 H1,5 H1,6
D1




H2,4 H2,5 H2,6   D2 
 = Hlc,dD
H3,4 H3,5 H3,6
D3


H10,1 H10,2 H10,3
U1


 U2  = Hsc,uU
H11,1 H11,2 H11,3 


H12,1 H12,2 H12,3
U3


H10,4 H10,5 H10,6
D1


 D2  = Hsc,dD
H11,4 H11,5 H11,6 


H12,4 H12,5 H12,6
D3
(C.1)
(C.2)
(C.3)
(C.4)
C.2. IDENTIFIED MODEL
135
are available. Meanwhile, straightforward dynamics suggest that
LC
= SCALElc · Mx,y,z · X, and SC = SCALEsc · X
(j2πf )2
(C.5)
where LC is the vector of load cell force measurements, M is the mass/inertia matrix, X are the position measurements, and SC is truth (scoring) sensor’s position
measurement. The factor SCALEsc = diag(Ssc1 , Ssc2, Ssc3 ) accounts for the scaling
differences in the position measurement as it is seen by the truth sensor. With the
orthogonal transformation

4 
T =

0
√1
1.5
√1
3
−1
√
2
−0.5
√
1.5
√1
3
√1
2
−0.5
√
1.5
√1
3




the effects of the actuators and the measurements of the sensors can be decomposed
into two tilt motions (about the two perpendicular axes in x-y plane), and a piston
motion in z direction. Now, the load cell reading can be related to that of the scoring
sensor as follows:
z
SCALE−1
lc
−1
Slc1

T ·
 0
0
}|
0
−1
Slc2
0
z
{

0

1
0 
 · (j2πf )2
−1
Slc3
SCALE−1
sc
−1
Ssc1

T ·
 0
0
}|
0
−1
Ssc2
0

Ix
LC1



 =  0
·
LC2



0
LC3
{ 
0
Mθx ,θy ,z
z

SC1

}|
{ 
0 0
θx


Iy 0 
  θy
z
0 M


 (C.6)


θx




 SC2  =  θy 
0 




−1
SC3
Ssc3
z
(C.7)
and therefore
1
−1
· Hlc,u · Hsc,u
= SCALElc · T ∗ · Mθx ,θy ,z · T · SCALE−1
sc
(j2πf )2
(C.8)
Note that a similar relationship can be derived for measurements from the disturbance
actuators. The left hand side in Eq. (C.8) is the available measurement data, and
C.2. IDENTIFIED MODEL
136
hence an optimization problem can be set up to find the optimal scaling factors as
well as the inertia and mass parameters. Instead of solving this problem at each
frequency, the left hand side of Eq. (C.8) over the frequency range [0.5 − 30] Hz is
averaged to obtain


0.2174 0.0668 0.0635


Hlc,u
−1
−3 
= 10 ·  0.0697 0.1931 0.0586 
Hsc,u

2
(j2πf )
averaged
0.0705 0.0650 0.1751
(C.9)
Now a simple optimization algorithm yields the following optimal scaling factors and
inertia/mass parameters for the system
SCALElc = diag(1.1043, 1.0352, 0.9560) × 22.24(N/Volts)
SCALEsc = diag(1.0887, 1.0122, 0.9935) × 0.25(mm/Volts)
Ix = 0.6579 Kg.m2
Iy = 0.6184 Kg.m2
M = 1.7801 Kg
Applying the orthogonal transformation T and the optimal scaling factors for the
load cells and the scoring sensors,
Mθmeasured
x ,θy ,z
=T ·
SCALE−1
lc
Hlc,u
−1
· Hsc,u SCALEsc · T ∗
2
(j2πf )
the matrix in Equation (C.9) is transformed into


0.1223
0.0004
0.0065


Hlc,u
−1
−3 

=
10
·
H
0.0019
0.1203
−0.0060
sc,u


(j2πf )2
diagonalized
−0.0049 0.0024
0.3070
(C.10)
in which the diagonal dominance of the matrix is evident. This, in essence, indicates
that the relationship between load cell and scoring (truth) sensor measurements conforms with what the physics of the VIP system predicts.
Using the optimal scaling for the sensors and the orthogonal transformation T ,
−1
(Hlc,u /(2πf )2)·Hsc,u
can be diagonalized for each frequency. Figures C.1 and C.2 show
C.2. IDENTIFIED MODEL
137
−1
−1
and (Hlc,d/(2πf )2) · Hsc,d
,
the elements of the 3 × 3 matrices (Hlc,u /(2πf )2 ) · Hsc,u
respectively, for the raw data. All elements of these matrices are roughly of the same
order of magnitude. Figures C.3 and C.4 on the other hand reflect the elements of
similar matrices for the diagonalized version. The diagonal dominance in these matrices is maintained for all frequencies. Also note that the numerical values for the
diagonal elements in Figures C.3 and C.4 agree with those obtained by the optimization process, indicating a truthful reflection of system dynamics for frequencies up to
50 Hz in the collected data.
C.2.3
System Identification
This section discusses the identified model for a 6-input/12-output continuous time
state space representation of the VIP system. As mentioned earlier, the three control
actuators (denoted by U) along with the three disturbance actuators (denoted by
D) form the inputs to the VIP system. Three load cells (colocated with the control
actuators and denoted by LC), six displacement sensors (one colocated within each
actuator stator and denoted by CP for the ones colocated with control actuators
and DP for the ones colocated with disturbance actuators), and three scoring sensors
(measuring the inertial motion of the payload and denoted by SC) constitute the 12
outputs for this system.
For the results presented here, the Frequency domain Observability Range Space
Extraction (FORSE) algorithm (see Section C.3, and Ref. [30], Section 3 for a detailed
description of the identification algorithm) is used. FORSE is a model synthesis technique that operates directly on the transfer function data without an inverse Fourier
transform, and has numerical robustness properties similar to time domain synthesis
techniques (such as ERA and q-Markov methods). The discussion in Ref. [30] proves
that a specifically constructed matrix from the frequency domain transfer function
measurements has the same basis as the observability matrix for the system, and thus
the A and C matrices can be found directly. The rest of the realization (i.e. B and
D matrices) can then be found using the least squares method.
The available frequency domain data (0.50 − 660.0 Hz) is not enough to allow
C.2. IDENTIFIED MODEL
138
the identification algorithm to capture the true behavior of the system at low frequencies. Therefore, the transfer function measurements from the 6 actuators to the
load cells are scaled by K/(j2πf )2 (for an appropriate K) effectively translating force
measurements into position measurements. This data is then fed to the FORSE algorithm to extract a state space model (Ã, B̃, C̃LC , D̃LC ) for the transfer functions
[U, D] → K · LC/(j2πf )2 . Note that, this is indeed a MIMO (6-input/3-output)
identification process in which the maximum singular value of the modeling error is
minimized. Keeping à and B̃ fixed, the transfer function data for [U, D] → [CP ],
[U, D] → [DP ], and [U, D] → [SC] are then used to derive the state space model
for each of the above mentioned subsystems. Once again the available data does not
provide enough information to correctly capture the high frequency behavior of the
system. Therefore, in the course of modeling the feed-through terms for the transfer
functions (i.e. D̃LC , D̃CP , D̃DP , and D̃SC ) are forced to be zero.
To cancel the effect of the double integration and scaling of the load cell data, the
output of the transfer function [U, D] → K ·LC/(j2πf )2 should be scaled by 1/K and
differentiated twice. To approximate the required scaling and double differentiation,
a second order filter of the form
F (s) =
ω02 (j2πf )2
K(j2πf + ω0 )2
(with ω0 = 2πf0 and for some appropriate f0 ), can be used to filter the load cell
output (i.e. ỹ = C̃LC · x + D̃LC · u). This approach however introduces 6 extraneous
states, the elimination of which can effect the quality of the identified model. Instead,
the state space model [Ã, B̃, C̃LC , D̃LC ] is used to perform the scaling and double
differentiation as follows:
o
1 ¨
1 n
· ỹ =
· C̃ Ã2 · x + C̃ ÃB̃ · u + C̃ B̃ · u̇
K
K
where if (1/K) · C̃ B̃ ∼
= 0, then the state space model
1
1
2
· C̃ Ã , DLC =
· C̃ ÃB̃
A = Ã, B = B̃, CLC =
K
K
(C.11)
is an adequate representation for the transfer function [U, D] → [LC]. Note that no
extraneous states are introduced here, hence the state space model for the Vibration
C.2. IDENTIFIED MODEL
139
Isolation Platform is as follows:

A


CLC


 CCP = C̃CP

 C = C̃
DP
 DP
CSC = C̃SC
B
DLC
DCP = D̃CP
DDP = D̃DP









DSC = D̃SC
A good fit to the data was found by using 36 states for the system. Increasing the
number of the states did not appear to improve the fit. Figures C.5-C.6 compare
the singular value plots for the measured MIMO transfer functions and the transfer
functions of the model generated using FORSE before and after double differentiation.
As Figure C.5 clearly indicate, the original model provides a good match for the scaled
data over the entire 0.5 − 660 Hz frequency range. Figure C.6 then shows that the
differentiation process does not introduce significant error in the identified model, as
the singular value plots for the un-scaled load cell data closely match those of the
final load cell model. Note that all significant modes are properly captured in both
cases. Figures C.7-C.9 compare the singular value plots of the identified model and
the measured data for other input/output channels. As these figures indicate, the
identified model closely matches the measured data in all cases. Note that double
differentiation of the load cell output does not affect the identified model from the
actuators to the other sensors (i.e. CP, DP, and SC sensors).
It is important to point out that despite the presence of a pole around 15 Hz that
dominates the response, the identification algorithm has been reasonably successful
in capturing the MIMO zeros in this frequency range (Figures C.8-C.9). Figure C.10
shows the identified model over a wider frequency range. The low frequency behavior
of the load cells, as well as the high frequency behavior of the position sensors is
correctly reflected in this model. As mentioned earlier in this model DLC , DCP , DDP ,
and DSC are all set to zero. Figure C.11 shows the final identified model (i.e. after
double differentiation and the scaling) for the VIP system. Both low frequency and
high frequency behavior for the load cells are correctly reflected. Also the model
remains truthful at high frequencies for the position sensors. For future reference,
natural frequencies and damping factors for the poles of the identified model are
C.2. IDENTIFIED MODEL
140
listed here:
Pole Natural
Damping Ratio
Frequency (Hz)
C.2.4
0.9733
0.0195
1.0156
0.0291
1.8925
0.8298
2.2409
0.1549
4.0462
0.0074
4.3862
0.0201
4.4216
0.0109
15.2601
0.0159
16.8545
1.0000
40.0545
0.0100
44.3453
0.0121
48.1396
0.0062
50.3646
0.0090
224.5517
0.0021
323.1164
0.0029
393.0650
0.0032
439.2836
0.0054
468.8545
0.0044
Control design model analysis
As a final test for the quality of the identified model, a simple H2 output feedback
controller was designed based on the identified mathematical model. Note that no
attempt has been made to obtain the best possible H2 controller for the VIP system.
The only objective here is to examine the consistency of the identified model and the
measured data from a control design point of view. For the results presented here:
ẋ = A x + B1 u + B2 d
y = CLC x + DLC,1 u + DLC,2 d
C.3. FORSE ALGORITHM
141
"
z =
zx
zu
#
"
=
CLC DLC,1
0
w∗I
#"
x
#
u
where u is the 3 × 1 vector of control actuator command, d is the 3 × 1 vector of
disturbances. B1 = B(:, 4 : 6), B2 = B(:, 1 : 3), DLC,1 = DLC (:, 4 : 6), DLC,2 = DLC (:
, 1 : 3), and w is some weight (chosen to be 0.05 here) on the control command. Figure
C.12 compares the closed and open loop singular value plots for the MIMO transfer
function from three disturbance actuators to the load cell and scoring sensors. The
important feature of this plot is not the performance of the designed controller, but
the consistency of the closed loop behavior of the system as it is seen by the two
different sets of sensors (i.e. load cell and scoring sensors). This same controller is
then used to close the loop on the measured frequency response data. The comparison
of open and closed loop SVD plots in this case is shown in Figure C.13. Once again,
the performance of the controller on the closed loop transfer function is consistently
reflected in both load cell and scoring sensor measurements. Note the close similarities
between the closed-loop behavior of the system under this H2 controller when either
the model or the real data is used. This close correlation of the closed-loop behavior
indicates that this identified model is a good control design model.
C.3
FORSE algorithm
The identification technique used for this work was originally develop at MIT as part
of the MACE program [30,22]. The approach integrates the Frequency domain Observability Range Space Extraction (FORSE) identification algorithm, the Balanced
Realization (BR) model reduction algorithm, and the logarithmic and additive Least
Square (LS) modal parameter estimation algorithms for low order highly accurate
model identification. The algorithm is called Integrated Frequency domain Observability Range Space Extraction and Least Square parameter estimation algorithm
(IFORSELS).
The Balanced Realization (BR) model reduction algorithm transforms a state
space model to a balanced coordinate and reduces the model by deleting the states
C.3. FORSE ALGORITHM
142
associated with the smallest principle values. The LS estimation algorithms improve the fitting of reduced models to experimental data by updating state space
parameters in modal coordinates. Models derived by both the FORSE and BR algorithms are non-optimal models, but their computations are non-iterative. While
the LS algorithms compute optimal parameter estimates, its computation may deviate to non-optimal parameters if initial estimates are very inaccurate. Integrating
these algorithms in an iterative manner avoids the computational difficulties of the
LS algorithms and improves modeling accuracies of the FORSE and BR algorithms.
As a result, the IFORSELS identification algorithm is capable of generating highly
accurate, low order models.
Assume that the following frequency response samples were obtained from experiments, Ĝ(ωk ) , k = 1, 2, · · · , K. The objective of frequency domain state space model
identification is to minimize the cost function
J=
K
X
kĜ(ωk ) − (C(jωk I − A)−1 B + D)k22 ,
(C.12)
k=1
where A, B, C, D are system matrices of the following state space equation,
(
ẋ(t) = A x(t) + B u(t)
.
y(t) = C x(t) + D u(t)
(C.13)
It is apparent that this is a nonlinear problem. The FORSE algorithm computes a
suboptimal estimate of these matrices using a subspace-based approach. To make
the cost J very small, the state space model is usually over parameterized. If model
order is not constrained, this FORSE model can be used by the LS algorithms as an
initial estimate to derive a more accurate estimate of the A, B, C, D matrices.
The entire IFORSELS algorithm consists of the FORSE identification algorithm,
the BR model reduction algorithm and two LS parameter estimation algorithms. It is
important to point out that one can use other subspace identification, model reduction and parameter estimation algorithms to form such an integrated identification
algorithm. In fact, the Eigensystem Realization Algorithm was first used to form the
algorithm [1]. The FORSE algorithm is now used because it is a frequency domain
algorithm and does not suffer from time domain aliasing errors. Other advantages
C.3. FORSE ALGORITHM
143
of the FORSE algorithm include its frequency domain weighting feature and its capability of processing data which is not uniformly distributed along the frequency
axis [2]. The BR model reduction algorithm is used simply because its computational
software is readily available in MATLAB. Other model reduction algorithms, such as
the weighted BR algorithm, the Projection algorithm and the Component Cost algorithm can also be incorporated into this integrated algorithm. The two LS parameter
estimation algorithms used are the Additive LS (ALS) algorithm, which minimizes
the error cost J in Equation C.12, and the Logarithmic LS (LLS) algorithm, which
minimizes the following logarithmic error cost
Jlog =
ny nu
K X
X
X
k log(Ĝij (ωk )) − log((C(jωk I − A)−1 B + D)ij )k22 ,
(C.14)
k=1 i=1 j=1
where the subscript ij indicates the (i, j)-th element of the matrix, and nu and ny
are the numbers of inputs and outputs of the system, respectively.
In the IFORSELS algorithm, the LLS parameter estimation algorithm is used
in the early stages of model reduction and updating. The logarithmic cost of the
LLS algorithm weights data samples equally, regardless of their magnitude [2], while
the FORSE and ALS algorithms place more emphasis on data samples which have
high magnitudes. Hence, in the early stages of model reduction and updating, more
emphasis needs to be placed on fitting in the frequency ranges where the magnitude
of data samples is low (such as areas around system transmission zeros). For the
identification process presented in this report the ALS algorithm is not used.
In summary, the entire IFORSELS identification algorithm consists of the following steps. Initially, the FORSE algorithm is used to generate an over parameterized
model. This model is then forwarded to the model reduction and updating iterations
using the BR and LLS algorithms. The identified model is then judged (by human)
based on the accuracy of the fit and the order of the model. If necessary, modes
can be deleted or added to this model and the process of IFORSELS identification
algorithm can be repeated again until a satisfactory model is obtained.
C.3. FORSE ALGORITHM
Scaling Factors between Load Cell & Truth Sensor Measurements (Raw Data), 04/12/98
−3
−3
10
10
−3
10
abs([U−>(LC/W2)]/[U−>SC])
144
−4
−4
10
−5
20
30
40
50
−3
abs([U−>(LC/W2)]/[U−>SC])
10
−6
10
20
30
40
50
−3
−4
−5
−5
30
40
50
−3
10
−4
20
30
40
50
10
50
10 20 30 40
Frequency (Hz)
50
−5
10
−6
50
40
−4
−5
10 20 30 40
Frequency (Hz)
30
10
10
−6
20
−3
−4
−5
10
10
10
10
10
−6
10
10
10
50
−5
−3
10
40
10
−6
20
30
10
10
10
20
−4
10
10
10
10
−4
10
10
−3
10
−6
abs([U−>(LC/W2)]/[U−>SC])
10
−6
10
10
10
−5
10
−6
10
10
−5
10
10
−4
10
−6
10 20 30 40
Frequency (Hz)
50
10
Fig. C.1: Magnitude of the scaling factor relating load cell’s reading of the effect of control
actuators to that of the scoring sensor
C.3. FORSE ALGORITHM
Scaling Factors between Load Cell & Truth Sensor Measurements (Raw Data), 04/12/98
−3
−3
10
10
−3
10
abs([D−>(LC/W2)]/[D−>SC])
145
−4
−4
10
−5
20
30
40
50
−3
abs([D−>(LC/W2)]/[D−>SC])
10
−6
10
20
30
40
50
−3
−4
−5
−5
30
40
50
−3
10
−4
20
30
40
50
10
50
10 20 30 40
Frequency (Hz)
50
−5
10
−6
50
40
−4
−5
10 20 30 40
Frequency (Hz)
30
10
10
−6
20
−3
−4
−5
10
10
10
10
10
−6
10
10
10
50
−5
−3
10
40
10
−6
20
30
10
10
10
20
−4
10
10
10
10
−4
10
10
−3
10
−6
abs([D−>(LC/W2)]/[D−>SC])
10
−6
10
10
10
−5
10
−6
10
10
−5
10
10
−4
10
−6
10 20 30 40
Frequency (Hz)
50
10
Fig. C.2: Magnitude of the scaling factor relating load cell’s reading of the effect of disturbance
actuators to that of the scoring sensor
C.3. FORSE ALGORITHM
−3
abs([U−>(LC/W2)]/[U−>SC])
10
146
Scaling Factors between Load Cell & Truth Sensor Measurements (Diagonalized Data), 04/12/98
−3
−3
10
10
−4
−4
10
−5
20
30
40
50
−3
abs([U−>(LC/W2)]/[U−>SC])
10
−6
10
20
30
40
50
−3
−4
−5
−5
30
40
50
−3
10
−4
20
30
40
50
10
50
10 20 30 40
Frequency (Hz)
50
−5
10
−6
50
40
−4
−5
10 20 30 40
Frequency (Hz)
30
10
10
−6
20
−3
−4
−5
10
10
10
10
10
−6
10
10
10
50
−5
−3
10
40
10
−6
20
30
10
10
10
20
−4
10
10
10
10
−4
10
10
−3
10
−6
abs([U−>(LC/W2)]/[U−>SC])
10
−6
10
10
10
−5
10
−6
10
10
−5
10
10
−4
10
−6
10 20 30 40
Frequency (Hz)
50
10
Fig. C.3: Magnitude of the scaling factor relating load cell’s reading of the effect of control
actuators to that of the scoring sensor after diagonalization
C.3. FORSE ALGORITHM
−3
abs([D−>(LC/W2)]/[D−>SC])
10
147
Scaling Factors between Load Cell & Truth Sensor Measurements (Diagonalized Data), 04/12/98
−3
−3
10
10
−4
−4
10
−5
20
30
40
50
−3
abs([D−>(LC/W2)]/[D−>SC])
10
−6
10
20
30
40
50
−3
−4
−5
−5
30
40
50
−3
10
−4
20
30
40
50
10
50
10 20 30 40
Frequency (Hz)
50
−5
10
−6
50
40
−4
−5
10 20 30 40
Frequency (Hz)
30
10
10
−6
20
−3
−4
−5
10
10
10
10
10
−6
10
10
10
50
−5
−3
10
40
10
−6
20
30
10
10
10
20
−4
10
10
10
10
−4
10
10
−3
10
−6
abs([D−>(LC/W2)]/[D−>SC])
10
−6
10
10
10
−5
10
−6
10
10
−5
10
10
−4
10
−6
10 20 30 40
Frequency (Hz)
50
10
Fig. C.4: Magnitude of the scaling factor relating load cell’s reading of the effect of disturbance
actuators to that of the scoring sensor after diagonalization
C.3. FORSE ALGORITHM
148
SVD for Transfer Function [ U D ] −> [ LC/S2 ], 05/18/98
2
SVD (Volts/Volts), Model(−) vs. Measured (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
Fig. C.5: Comparison of SVD plots for the transfer function to the scaled/double-integrated
load cell data
SVD for Transfer Function [ U D ] −> [ LC ] after Double Differentiation, 05/18/98
2
SVD (Volts/Volts), Model(−) vs. Measured (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
Fig. C.6: Comparison of SVD plots for the transfer function to the actual load cell data
C.3. FORSE ALGORITHM
149
SVD for Transfer Function [ U D ] −> [ SC ], 05/18/98
2
SVD (Volts/Volts), Model(−) vs. Measured (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
Fig. C.7: Comparison of SVD plots for the transfer function to the scoring sensors
SVD for Transfer Function [ U D ] −> [ CP ], 05/18/98
2
SVD (Volts/Volts), Model(−) vs. Measured (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
Fig. C.8: Comparison of SVD plots for the transfer function to the position sensors colocated
with the control actuators
C.3. FORSE ALGORITHM
150
SVD for Transfer Function [ U D ] −> [ DP ], 05/18/98
2
SVD (Volts/Volts), Model(−) vs. Measured (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
Fig. C.9: Comparison of SVD plots for the transfer function to the position sensors colocated
with the disturbance actuators
C.3. FORSE ALGORITHM
151
SVD for [ U D ] −> [ LC/S2 ]
SVD for [ U D ] −> [ CP ], 05/18/98
2
2
SVD (Volts/Volts), Model Only
10
10
0
0
10
10
−2
−2
10
10
−4
−4
10
10
−6
10
−6
0
5
10
5
10
Frequency (Hz)
SVD for [ U D ] −> [ SC ]
SVD for [ U D ] −> [ DP ]
2
SVD (Volts/Volts), Model Only
0
10
Frequency (Hz)
2
10
10
0
0
10
10
−2
−2
10
10
−4
−4
10
10
−6
10
10
10
−6
0
5
10
10
Frequency (Hz)
10
0
5
10
10
Frequency (Hz)
Fig. C.10: The identified model for the system beyond the frequency range for which measurements are available
C.3. FORSE ALGORITHM
152
SVD for [ U D ] −> [ LC ]
SVD for [ U D ] −> [ CP ], 05/18/98
2
2
SVD (Volts/Volts), Model Only
10
10
0
0
10
10
−2
−2
10
10
−4
−4
10
10
−6
10
−6
0
5
10
10
5
10
Frequency (Hz)
SVD for [ U D ] −> [ SC ]
SVD for [ U D ] −> [ DP ]
2
SVD (Volts/Volts), Model Only
0
10
Frequency (Hz)
2
10
10
0
0
10
10
−2
−2
10
10
−4
−4
10
10
−6
10
10
−6
0
5
10
10
Frequency (Hz)
10
0
5
10
10
Frequency (Hz)
Fig. C.11: The final model for the system beyond the frequency range for which measurements
are available
C.3. FORSE ALGORITHM
153
H2 Controller Applied to the Load Cell Model
2
SVD Open−Loop (−) vs. Closed−Loop (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
H2 Controller Applied to the Scoring Sensor Model
2
SVD Open−Loop (−) vs. Closed−Loop (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
Fig. C.12: The comparison of the closed loop and open loop singular value plots when the
controller is used to close the loop on the identified model
C.3. FORSE ALGORITHM
154
H2 Controller Applied to Real Load Cell Data
2
SVD Open−Loop (−) vs. Closed−Loop (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
H2 Controller Applied to Real Scoring Sensor Data
2
SVD Open−Loop (−) vs. Closed−Loop (−−)
10
1
10
0
10
−1
10
−2
10
−3
10
−4
10
0
10
1
10
Frequency (Hz)
2
10
Fig. C.13: The comparison of the closed loop and open loop singular value plots when the
controller is used to close the loop on the real measured data
Bibliography
[1] E.H. Anderson and J.P. How. Active Vibration Isolation Using Adaptive Feedforward Control. ACC97, pages 1783–1788, July 1997.
[2] K.J. Astrom and B. Wittenmark. Adaptive Control. Addison-Wesley, 1989.
[3] M.R. Bai and Z. Lin. Active Noise Cancellation for a Three-Dimensional Enclosure by Using Multiple Channel Adaptive Control and H∞ Control. ASME
Journal of Vibration and Acoustics, 120:958–964, october 1998.
[4] D.S. Bernstein and W.M. Haddad. LQG Control with an H∞ Performance
Bound: A Riccati Equation Approach. IEEE Trans. on Auto. Control, 34:293–
305, 1989.
[5] N.J. Bershad, P.L. Feintuch, F.A. Reed, and B. Fisher. Tracking Characteristics
of the LMS Adaptive LineEnhancer-Response to a Linear Chirp Signal in Noise.
IEEE Trans. on Acoust., Speech, Signal Processing, 28:504–516, October 1980.
[6] N.J. Bershad and O.M. Macchi. Adaptive Recovery of a Chirped Sinusoid in
Noise, Part 2: Performance of the LMS Algorithm. IEEE Trans. on Signal
Processing, 39:595–602, March 1991.
[7] E. Bjarnason. Analysis of the Filtered-X LMS Algorithm. IEEE Trans. on Speech
and Audio Processing, 3:504–514, November 1995.
[8] Stephen Boyd, L. El Ghaoui, Eric Feron, and V. Balakrishan. Linear Matrix
Inequalities in System and Control Theory. SIAM, 1994.
155
BIBLIOGRAPHY
156
[9] A.E. Bryson and Y. Ho. Applied Optimal Control. Hemisphere Pub. Corp., New
York, NY, 1975.
[10] M. Chilali and P. Gahinet. H∞ Design With Pole Placement Constraints: An
LMI Approach. IEEE Trans. on Auto. Control, 41:358–367, March 1996.
[11] M. Dentino, J. McCool, and B. Widrow. Adaptive Filtering in Frequency Domain. Proceedings of the IEEE, 66:1658–1659, December 1978.
[12] J.C. Doyle, K. Glover, P. Khargonekar, and B. Francis. State-Space Solutions
to Standard H2 and H∞ Control Problems. IEEE Trans. Automat. Control,
34:831–847, August 1989.
[13] S.J. Elliott, I.M. Stothers, and P.A. Nelson. A Multiple Error LMS Algorithm
and Its Application to the Active Control of Sound and Vibration. IEEE Trans.
Acoust., Speech, Signal Processing, 35:1423–1434, October 1987.
[14] L.J. Eriksson. Development of the Filtered-U Algorithm for Active Noise Control.
J. Acoust. Soc. Am., 89:257–265, Jan. 1991.
[15] L.J. Eriksson, M.C. Allie, and R.A. Greiner. The Selection and Adaptation of
an IIR Adaptive Filter for Use in Active Sound Attenuation. IEEE Trans. on
Acoust., Speech, Signal Processing, 35:433–437, April 1987.
[16] P.L. Feintuch. An Adaptive Recursive LMS Filter. Proc. IEEE, 64:1622–1624,
Nov. 1976.
[17] P.L. Feintuch, N.J. Bershad, and A.K. Lo. A Frequency Domain Model for Filtered LMS Algorithm-Stability Analysis, Design, and Elemination of the Training Mode. IEEE Trans. on Signal Processing, 41:1518–1531, April 1993.
[18] P. Gahinet and P. Apkarian. A Linear Matrix Inequality Approach to H∞ Control. Int. J. of Robust and Nonlinear Control, 4:421–448, 1994.
[19] P. Gahinet, A. Nemirovski, A.J. Laub, and M. Chilali. LMI Control Toolbox.
The Math Works Inc., 1995.
BIBLIOGRAPHY
157
[20] A. Gelb. Applied Optimal Estimation. The M.I.T Press, Cambridge, MA, 1988.
[21] R.D. Gitlin, H.C. Meadors, and S.B. Weinstein. The Tap-Leakage Algorithm:
An Algorithm for the Stable Operation of a Digitally Implemented, Fractional
Adaptive Space Equalizer. Bell System Tech. Journal, 61:1817–1839, October
1982.
[22] R.M. Glaese and K. Liu. On-Orbit Modeling and System Identification of the
Middeck Active Control Experiment. IFAC 13th World Congress, July 1996.
[23] J.R. Glover. Adaptive Noise Cancellation Applied to Sinusoidal Interferences.
IEEE Trans. on Acoust., Speech, Signal Processing, 25:484–491, December 1977.
[24] M. Green and D.J.N. Limebeer. Linear Robust Control. Prentice-Hall, Englewood
Cliffs, NJ, 1995.
[25] R.W. Harris, D.M. Chabries, and F.A. Bishop. A Variable Step Adaptive Filter
Algorithm. IEEE Trans. on Acoust., Speech, Signal Processing, 34:309–316, April
1986.
[26] B. Hassibi, A.H. Sayed, and T. Kailath. H∞ Optimality of the LMS Algorithm.
IEEE Trans. on Signal Processing, 44:267–280, February 1996.
[27] B. Hassibi, A.H. Sayed, and T. Kailath. Indefinite Quadratic Estimation and
Control. SIAM Studies in Applied Mathematics, 1998.
[28] Babak Hassibi. Indefenite Metric Spaces In Estimation, Control and Adaptive
Filtering. PhD thesis, Stanford University, 1996.
[29] S. Haykin. Adaptive Filter Theory. Prentice-Hall, Englewood Cliffs, NJ, 1996.
[30] R.N. Jacques. On-Line System Identification and Control Design for Flexible
Structures. PhD thesis, MIT Department of Aeronautics and Astronautics, 1994.
[31] C.R. Johnson. Adaptive IIR Filtering: Current Results and Open Problems.
IEEE Trans. Inform. Theory, 30:237–250, March 1984.
BIBLIOGRAPHY
158
[32] P.P. Khargonekar and K.M. Nagpal. Filtering and Smoothing in an H∞ Setting.
IEEE Trans. on Automat. Control, 36:151–166, 1991.
[33] S.M. Kuo and D.R. Morgan. Adaptive Noise Control Systems. John Wiley &
Sons, New York, NY, 1996.
[34] R. W. Lucky. Automatic Equalization for Digital Communication. Bell Syst.
Tech. J., 44:547–588, 1965.
[35] D.R. Morgan. An Analysis of Multiple Correlation Cancellation Loops with a
Filter in the Auxiliary Path. IEEE Trans. on Acoust., Speech, Signal Processin,
28:454–467, August 1980.
[36] K.S. Narendra. Parameter Adaptive Control-The End ... or The Beginning. 33rd
Conf. on Decision and Control, pages 2117–2125, December 1994.
[37] K.S. Narendra and A.M. Annaswamy. Stable Adaptive Systems. Prentice-Hall,
Englewood Cliffs, NJ, 1989.
[38] D. Parikh, N. Ahmed, and S.D. Stearns. An Adaptive Lattice Algorithm for
Recursive Filters. IEEE Trans. on Acoust., Speech, Signal Processing, 28:110–
111, Feb. 1980.
[39] S.R. Popovich, D.E. Melton, and M.C. Allie. New Adaptive Multi-channel Control Systems for Sound and Vibration. Proc. Inter-Noise, pages 405–408, July
1992.
[40] J.G. Proakis. Adaptive Equalization for TDMA Digital Mobile radio. IEEE
Trans. on Vehicular Tech., 40:333–341, 1991.
[41] S. Qureshi. Adaptive Equalization. Proc. IEEE, 73:1349–1387, 1985.
[42] E.A. Robinson and S. Treitel. Geophysical Signal Analysis. Prentice-Hall, Englewood Cliffs, NJ, 1980.
BIBLIOGRAPHY
159
[43] S.P. Rubenstein, S.R. Popovich, D.E. Melton, and M.C. Allie. Active Cancellation of Higher Order Modes in a Duct Using Recursively-Coupled Multi-Channel
Adaptive Control System. Proc. Inter-Noise, pages 337–339, July 1992.
[44] A.H. Sayed and T. Kailath. A State-Space Approach to Adaptive RLS Filtering.
IEEE Signal Processing Magazine, pages 18–60, July 1994.
[45] C.W. Scherer. Multiobjective H2 /H∞ Control. IEEE Trans. on Auto. Control,
40:1054–1062, June 1995.
[46] Y. Shaked and Y. Theodor. H∞ -Optimal Estimation: A Tutorial. In Proc. IEEE
Conf. Decision Contr., volume 2, December 1992.
[47] T.J. Shan and T. Kailath. Adaptive Algorithms with an Automatic Gain Control
Feature. IEEE Trans. Circuits Syst., 35:122–127, January 1988.
[48] J.J. Shynk. Adaptive IIR Filtering Using Parallel-Form Realization. IEEE Trans.
on Acoust., Speech, Signal Processing, 37:519–533, April 1989.
[49] Scott D. Sommerfeldt. Multi-Channel Adaptive Control of Structural Vibration.
Noise Control Engineering Journal, 37:77–89, Sep.-Oct. 1991.
[50] B. Widrow. Adaptive filters i: Fundamentals. Technical report, Stanford Electron. Labs, Stanford University, 1966.
[51] B. Widrow and S.D. Stearns. Adaptive Signal Processing. Prentice-Hall, Englewood Cliffs, NJ, 1985.
[52] G. Zames. Feedbck Optimal Sensitivity: Model Preference Transformation, Multiplicative Seminorms and Approximate Inverses. IEEE Trans. on Automat.
Control, 26:301–320, 1981.
Download