Lecture Notes No.

advertisement
2.160 System Identification, Estimation, and Learning
Lecture Notes No. 11
March 15, 2006
6.5 Times-Series Data Compression
b2
u(t)
b1
y(t)
FIR
b3
t
Finite Impulse Response Model
Consider a FIR Model
y (t ) = b1u (t − 1) + b2 u (t − 2) + ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ + bm u (t − m)
The transfer functions of (11) are then: G (q ) = B ( q ), H (q ) = 1
One-Step-Ahead Predictor
[
]
yı(t θ ) = H −1 (q)G ( q)u(t ) + 1 − H −1 (q) y (t )
= G ( q )u ( t )
yı (t θ ) = ϕ T (t )θ : linear regression
Given input data {u(t ), 1 ≤ t ≤ N } , the least square estimate of the parameter vector
was obtained as
θıNLS = arg min
θ
1
N
2
N
1
( y(t ) − yı (t | θ )) = (R( N ))−1 f (N )
∑
t =1 2
(40)
where
1
R(N ) =
N
N
∑ϕ (t )ϕ (t )
T
= ΦΦ
T
t =1
1
and f ( N ) =
N
N
∑ ϕ (t ) y (t )
(41)
t =1
Pro’s and Con’s of FIR Modeling
Pros.
LSE gives a consistent estimate
limθı
N →∞
LS
N
= θ0
as long as the input sequence {u (t )} is
uncorrelated with the noise e(t), which may be colored.
Cons
Two failure scenarios
• The impulse response has a slow decaying mode
1
• The sampling rate is high
The number of parameters, m, is too large to estimate.
can hardly be satisfied.
The persistently exciting condition rankΦ = full rank
T
Check the eigenvalues of ΦΦ (or the singular values of Φ )
λ1 ≥ λ2 ≥ � ≥ λm
It is likely
λ1 ≥ λ2 ≥ � ≥ λn ≥ λn+1 ≅0 = � = λm
Often m becomes more than 50 and it is difficult to obtain such an input series having
50 non-zero singular values.
Time series data compression is an effective method for coping with this difficulty.
Before formulating the above least square estimate problem, data are processed so
that the information contained in the series of regressor may be represented in
compact, low-dimensional form.
ϕm
ϕ1
ϕn
.
. . .
. .
.
λ ≅0
ϕ1
ϕ2
Filter Design
Lk(q)
i
Rm
Rn
n<m
Coordinate Transformation (Data Compression)
L1(q)
u(t)
L2(q)
b1
b2
∑
y(t)
�
Ln(q)
bn
2
6.6 Continuous-Time Laguerre Series Expansion
Let us begin with continuous-time Laguerre expansion.
[Theorem 6.6]
Assume a transfer function G(s) is
•
Strictly proper
•
Analytic in Re(s) > 0
•
Continuous in Re(s) ≥ 0
lim G (
s ) = 0
G(∞) = 0
s →∞
Then, there exists a sequence
(45)
No pole on RHP
{g k }
such that
∞
G (s ) = ∑ g k
k =1
2a ⎛ s − a ⎞
⎜
⎟
s+a ⎝s+a⎠
k −1
(46)
where a > 0
Proof:
Consider the transformation given by
s+a
z =
s − a
sz − az = s + a
Im
s−a
s
s ( z − 1) = a ( z + 1)
∴ s = a
Bilinear transformation
z =
z-plane
(47)
s+a
s−a
z +1
z −1
(48)
Im
=1
for s = jω
(42)
1
∠z = ∠( s + a ) − ∠( s − a )
-1
Re
s−a
z =
Re
s
s+a
s−a
1
s + a The
imaginary axis is
mapped to the unit circle.
>1
(43)
The right half plane is
mapped to the outside of
the unit circle.
3
Im
Im
G(s) is analytic in Re( s ) > 0
1
-1
1
Re
Consider
Re
⎛
⎞
⎜ z +1⎟
G (z) = G⎜ a
⎟
z −�1 ⎟
⎜ ��
⎝ s ⎠
G (z ) is analytic outside
(44)
the unit circle.
Therefore, there exists a Laurent expansion for G (z ) :
∞
G ( z ) =
∑ g k z − k
z >1
(49)
k =1
s +
a
=
1 ∴
G ( z )
= 0 at z=1
s →∞ s − a
Considering that G (∞)
=
0
lim
G ( z ) =
∞
1
(1 −
z −1 )
∑ g k z −( k −1)
2a
k =1
(50)
Substituting (40) into (46)
s
− a ∞
1
⎛ s
+ a
⎞
⎛ s +
a
⎞
G⎜
(1
−
)∑
g k ⎜
⎟ = G ( s ) =
⎟
s
+ a k =1 ⎝
s −
a ⎠
2a
⎝ s − a ⎠
∞
∴
G ( s ) =
∑ g k
k =1
Lk (s ) =
2a ⎛ s − a ⎞
⎜
⎟
s +
a
⎝
s
+
a ⎠
2a ⎛ s − a ⎞
⎜
⎟
s +
a ⎝
s + a ⎠
Low Pass Filter
− (k −1)
(k −1)
(k −1)
(51)
s−a
s+a
jω − a
=1
jω + a
All-pass Filter
The Laplace transform of the Laguerre functions
The Main Point
The above Laguerre series expansion can be used for data compression if the
parameter a , called a Laguerre pole, is chosen such that slow poles, i.e. dominant
poles, of the original system are close to the Laguerre pole. Let pi be a slow real pole
of the original transfer function G(s). If the Laguerre pole a is chosen such that
4
a ≅pi , then a truncated Laguerre expansion:
n
Gn (s) = ∑ g k
k =1
2a ⎛ s − a ⎞
⎜
⎟
s+a ⎝s+a⎠
k −1
→ G( s )
(52)
converges to the original G(s) quickly.
- Recall -
For a continuous-time system, a pole close to the imaginary axis is slow to converge,
while a pole far from the imaginary axis converges quickly. Likewise, in discrete time, a pole close to the origin of a z-plane quickly converges, while the ones near the unit
circle are slow.
z-plane
Im
s-plane
Fast
Slow
Fast
Re
Slow
Choose a such that p1 ≅−a , where p1 is a slow, stable pole. Using the bilinear
transform, this slow pole in the s-plane can be transferred to a fast pole in the z-plane,
as shown below. Representing in the z-plane, the transfer function can be truncated;
just a few terms can approximate the impulse response since it converges quickly.
p1
Im
Re
a ≅p1
{g k }
Transformation
Im
z=
1
s+a
s−a
Truncated
-1
1
time
Re
Fast Convergence
-1
Fast pole z ≅0
When the original transfer function has many poles, the Laguerre pole is placed near
the dominant pole so that most of the slow poles may be approximated by the
Laguerre pole.
5
Im
Im
Dominant pole
1
-1
1
z
Re
Re
In the transformed plane, all
the poles are confined within
a circle of radius r<1.
−a
With the bilinear transform, these slow poles are transferred to the ones near the
origin of the z-plane. They are fast, hence the transfer function can be approximated
to a low order model.
6.7 Discrete-Time Laguerre Series Expansion
[Theorem 6.7]
Assume that a Z-transform G(z) is
•
Strictly proper
•
Analytic in z > 1
•
Continuous in z ≥ 1
G(∞) = 0
RHP
Then
∞
G (s ) = ∑ g k
k =1
where
K ⎛ 1 − az ⎞
⎜
⎟
z−a⎝ z−a ⎠
k −1
(53)
-1<a<1 and
K = (1 − a 2 )T
Proof
Consider the bilinear transformation:
z−a
w=
1 − az
w − azw = z − a
w + a = z (
aw + 1) therefore, z =
(54)
(55)
w+a
aw + 1
(56)
6
Transformation
z−a
w=
1 − az
z-plane
w-plane
⎛ w+a ⎞
G ( w) = G ⎜
⎟
⎝ aw + 1 ⎠
z >1
w >1
⎛ w
+
a ⎞
G (w) =
G
⎜
⎟ is analytic in w > 1, and is proper G(∞
) =
0
⎝
aw
+
1
⎠
lim
w
=−
z →∞
1
a
1
G
(
− ) = 0
a
(57)
Therefore
G ( w) =
∞
T
(a
+
w
−1 )∑ g k w −( k −1)
K
k =1
(58)
−(k −1)
1
−
az ∞
⎛ z −
a ⎞ T
⎛ z−a ⎞
G( z ) = G ⎜
)∑ g k ⎜
⎟ =
(a +
⎟
z −
a k =1 ⎝
1
−
az ⎠
⎝ 1
−
az ⎠
K
∞
∴ G ( s ) =
∑ g k
k =1
(59)
k −1
K ⎛ 1 −
az
⎞
⎜
⎟
z −
a
⎝
z −
a ⎠
(53)
Q.E.D.
Now we can write
y (t ) =
G ( q)u(t )
k −1
n
K ⎛
1
−
aq
⎞
=
∑
g k
⎜⎜
⎟⎟ u(t ) =
∑
g k Lk ( q)u(t )
q − a
⎝
q
−
a ⎠
k =1
k =1
n
where Lk ( q ), k =
1,�, n is a series of filters. Once the original input data are filtered
7
with Lk ( q ), k = 1,�, n ,
xk = Lk ( q )u(t ), k = 1,�, n
(60)
The output y (t ) is represented as a moving average of the transformed input xk ,
that is, a FIR model.
L1(q)
u(t)
L2(q)
Ln (q )
x1
x2
xn
g1
g2
y(t)
∑
gn
Furthermore, xk (t ) can be computed recursively.
x1 (t) = L1 ( q)u(t ) =
K
Kq −1
u( t ) =
u (t )
q−a
1 − aq −1
x1 (t) − ax1 (t − 1) = Ku(t − 1)
x1 (t) = ax1 (t − 1) + Ku(t − 1)
x2 (t) =
1 − aq
q −1 − a
x1 (t) =
x1 (t)
q−a
1 − aq −1
x2 (t) − ax 2 (t − 1) = x1 (t − 1) − ax1 (t )
x2 (t) = ax2 (t − 1) + x1 (t − 1) − ax1 (t )
x
k (t) = axk (t − 1) + xk −1 (t − 1) − axk −1 (t)
(61)
Recursive Filters
8
Download