Sec 4.7

advertisement
Chapter 4
Section 4.7
Similarity Transformations and
Diagonalization
Similar Matrices
If A,B are two ๐‘› × ๐‘› and there is a third ๐‘› × ๐‘› matrix S
that is non-singular (i.e. invertible) such that ๐ต = ๐‘† −1 ๐ด๐‘†
then A and B are called similar matrices or A is similar to B.
A is similar to B if there
is a matrix S such that:
๐ต = ๐‘† −1 ๐ด๐‘†
The reason we use the word similar is that the two matrices share many of the same features
as some of the next results will show.
Theorem
If A and B are similar matrices then the characteristic polynomial for A call it ๐‘ ๐‘ก is equal to
the characteristic polynomial for B call it ๐‘ž ๐‘ก and the eigenvalues of A are the same as the
eigenvalues of B with the same algebraic multiplicities.
Let S be the non-singular matrix so that ๐ต = ๐‘† −1 ๐ด๐‘†, that means that det ๐‘† ≠ 0.
๐‘ž ๐‘ก = det ๐ต − ๐‘ก๐ผ = det ๐‘† −1 ๐ด๐‘† − ๐‘ก๐ผ
= det ๐‘† −1 ๐ด๐‘† − ๐‘ก๐‘† −1 ๐‘† = det ๐‘† −1 ๐ด − ๐‘ก๐ผ ๐‘†
= det ๐‘† −1 det ๐ด − ๐‘ก๐ผ det ๐‘†
1
=
det ๐‘† det ๐ด − ๐‘ก๐ผ
det ๐‘†
= det ๐ด − ๐‘ก๐ผ
=๐‘ ๐‘ก
If the matrices have the same
characteristic polynomials they will both
factor the same way and have the same
roots which are the eigenvalues. The
number of time they occur in factored
form will also be equal, giving equal
algebraic multiplicities.
Careful similar matrices may not have the same eigenvectors but they are related.
Theorem
If A and B are similar with non-singular matrix S, such that ๐ด = ๐‘† −1 ๐ต๐‘† and ๐œ† is an eigenvalue
for A with corresponding eigenvector ๐ฏ ≠ ๐œฝ then the eigenvector that corresponds to the
eigenvalue ๐œ† for B is the vector ๐‘†๐ฏ.
Since S is non-singular and ๐ฏ ≠ ๐œฝ then ๐‘†๐ฏ ≠ 0.
๐ต ๐‘†๐ฏ = ๐ผ ๐ต๐‘† ๐ฏ = ๐‘†๐‘† −1 ๐ต๐‘† ๐ฏ = ๐‘† ๐‘† −1 ๐ต๐‘† ๐ฏ = ๐‘† ๐ด๐ฏ = ๐‘† ๐œ†๐ฏ = ๐œ†๐‘†๐ฏ
Therefore ๐œ† is an eigenvalue of B with corresponding eigenvector ๐‘†๐ฏ.
Corollary
If A and B are similar with non-singular matrix S, such that ๐ด = ๐‘† −1 ๐ต๐‘† and ๐œ†1 , โ‹ฏ , ๐œ†๐‘˜ are
eigenvalues of A with corresponding eigenvectors ๐ฏ1 , โ‹ฏ , ๐ฏ๐‘˜ then ๐œ†1 , โ‹ฏ , ๐œ†๐‘˜ are eigenvalues of
B with corresponding eigenvectors S๐ฏ1 , โ‹ฏ , S๐ฏ๐‘˜ .
Example
For the matrix A to the right find the eigenvalues (there are 3)
and corresponding eigenvectors ๐ฏ1 , ๐ฏ2 , ๐ฏ3 . Let ๐‘† = ๐ฏ1 , ๐ฏ2 , ๐ฏ3 ,
find ๐‘† −1 and use it to compute ๐‘† −1 ๐ด๐‘†.
4
3
๐ด = −6 −5
0
0
Use a 3rd row expansion to find the characteristic polynomial.
4−๐‘ก
3
2
4−๐‘ก
๐‘ ๐‘ก = −6 −5 − ๐‘ก
0 = 2−๐‘ก
−6
0
0
2−๐‘ก
= 2 − ๐‘ก ๐‘ก2 + ๐‘ก − 2 = 2 − ๐‘ก ๐‘ก + 2 ๐‘ก − 1
3
= 2−๐‘ก
−5 − ๐‘ก
4 − ๐‘ก −5 − ๐‘ก + 18
The eigenvalues are : -2, 1, 2
2
0
2
1 12
6
3 2
๐ด + 2๐ผ = −6 −3 0 , row reduces to 0 0
0
0 4
0 0
3
3
๐ด − ๐ผ = −6 −6
0
0
2
๐ด − 2๐ผ = −6
0
2
1
0 , row reduces to 0
1
0
1
0
0
1 0
3 2
−7 0 , row reduces to 0 1
0 0
0 0
0
−12
−1
1 , vector solution: ๐‘ฅ2 1 , let ๐ฏ1 = 2
0
0
0
0
−1
−1
1 , vector solution: ๐‘ฅ2 1 , let ๐ฏ2 = 1
0
0
0
7
−72
7
2
3 , vector solution: ๐‘ฅ3 −3 , let ๐ฏ3 = −6
2
0
1
−1 −1
Forming the matrix with eigenvector columns the matrix ๐‘† = ๐ฏ1 , ๐ฏ2 , ๐ฏ3 = 2
1
0
0
1
1 −12
Augmenting the matrix with I and row reducing it we get that ๐‘† −1 = −2 −1 4
1
0
0
2
1
1
๐‘† −1 ๐ด๐‘† = −2 −1
0
0
−12 4
4 −6
1
0
2
3
−5
0
2 −1 −1 7
−2 0
0 2
1 −6 = 0 1
2 0
0
2
0 0
0
0
2
Observe that the result of the matrix computation ๐‘† −1 ๐ด๐‘† is a diagonal matrix with the
eigenvalues of the matrix A going down the diagonal! This is not a coincidence.
7
−6 .
2
Theorem
Let A be a ๐‘› × ๐‘› matrix with eigenvalues ๐œ†1 , โ‹ฏ , ๐œ†๐‘˜ if the basis vectors for the eigenspaces
๐ธ๐œ†1 , โ‹ฏ , ๐ธ๐œ†๐‘˜ form a complete basis for โ„๐‘› (i.e. the are n vectors ๐ฏ1 , โ‹ฏ , ๐ฏ๐‘› in the combined
bases for all the eigenspaces) then if we set ๐‘† = ๐ฏ1 , โ‹ฏ , ๐ฏ๐‘› the result of ๐‘† −1 ๐ด๐‘† will be a
diagonal matrix with the eigenvalues of A going down the diagonal.
The reason for this comes from making 2 observations
1. If ๐ฏ1 , โ‹ฏ , ๐ฏ๐‘› are eigenvectors of A then ๐ด๐ฏ1 =
๐œ†1 ๐ฏ1 , โ‹ฏ , ๐ด๐ฏ๐‘› = ๐œ†๐‘› ๐ฏ๐‘› .
2. Since ๐‘† −1 ๐‘† = ๐ผ = ๐ž1 , โ‹ฏ , ๐ž๐‘› then ๐‘† −1 ๐ฏ1 , โ‹ฏ , ๐ฏ๐‘› =
๐‘† −1 ๐ฏ1 , โ‹ฏ , ๐‘† −1 ๐ฏ๐‘› = ๐ž1 , โ‹ฏ , ๐ž๐‘› or that in other words
๐‘† −1 ๐ฏ1 = ๐ž1 , โ‹ฏ , ๐‘† −1 ๐ฏ๐‘› = ๐ž๐‘› .
๐‘† −1 ๐ด๐‘† = ๐‘† −1 ๐ด ๐ฏ1 , โ‹ฏ , ๐ฏ๐‘›
= ๐‘† −1 ๐ด๐ฏ1 , โ‹ฏ , ๐ด๐ฏ๐‘›
= ๐‘† −1 ๐œ†1 ๐ฏ1 , โ‹ฏ , ๐œ†๐‘› ๐ฏ๐‘›
= ๐‘† −1 ๐œ†1 ๐ฏ1 , โ‹ฏ , ๐‘† −1 ๐œ†๐‘› ๐ฏ๐‘›
= ๐œ†1 ๐‘† −1 ๐ฏ1 , โ‹ฏ , ๐œ†๐‘› ๐‘† −1 ๐ฏ๐‘›
= ๐œ†1 ๐ž1 , โ‹ฏ , ๐œ†๐‘› ๐ž๐‘›
Now write out the matrix ๐œ†1 ๐ž1 , โ‹ฏ , ๐œ†๐‘› ๐ž๐‘› we se that ๐œ†1 ๐ž1 , โ‹ฏ , ๐œ†๐‘› ๐ž๐‘›
๐œ†1
= โ‹ฎ
0
โ‹ฏ 0
โ‹ฑ โ‹ฎ
โ‹ฏ ๐œ†๐‘›
Diagonalizable Matrices
A ๐‘› × ๐‘› matrix A is diagonalizable (or diagonalizes) if there is a non-singular matrix S such
that ๐‘† −1 ๐ด๐‘† = ๐ท where the matrix D is a diagonal matrix.
Theorem
If the ๐‘› × ๐‘› matrix A is not defective then the matrix A is diagonalizable.
Since the algebraic multiplicity is equal to the geometric multiplicity the total number of basis
vector in all eigenspaces will be n.
Example
Determine if the matrix A to the right is diagonalizable. If the matrix is
diagonalizable find a matrix S such that ๐‘† −1 ๐ด๐‘† = ๐ท where D is a
diagonal matrix.
๐ด=
8
−6
9
−7
Begin by computing the characteristic polynomial ๐‘ ๐‘ก and eigenvalues.
๐‘ ๐‘ก =
8−๐‘ก
−6
9
= 8 − ๐‘ก −7 − ๐‘ก + 54 = ๐‘ก 2 − ๐‘ก − 2 = ๐‘ก + 1 ๐‘ก − 2
−7 − ๐‘ก
We see from the above polynomial the eigenvalues are -1 and 2. The eigenvalues are distinct
so the matrix is not defective. Since A is not defective it is diagonalizable. To find S we need to
compute a basis for each eigenspace.
๐ด+๐ผ =
9
9
1 1
−1
−1
, row reduces to
, vector solution : ๐‘ฅ2
, let ๐ฏ1 =
−6 −6
0 0
1
1
1
6
9
๐ด − 2๐ผ =
, row reduces to
−6 −9
0
So ๐‘† =
3
2
0
, vector solution : ๐‘ฅ2
−32
−3
, let ๐ฏ2 =
2
1
1 2
−1 −3
, use the 2 × 2 matrix inverse formula for ๐‘† −1 = 1
1
2
−1
3
2
3
=
−1
−1 −1
We can check this with the following computation:
8
9 −1 −3
−1
2
3
2
3
1 −6
๐‘† −1 ๐ด๐‘† =
=
=
0
−1 −1 −6 −7 1
2
−1 −1 −1 4
0
2
Matrix Functions and Diagonal Matrices
Diagonal matrices behave just like numbers for matrix functions that are polynomials. Begin
by considering the product of two diagonal matrices which is another diagonal matrix.
๐‘‘11
โ‹ฎ
0
โ‹ฏ
0
โ‹ฑ
โ‹ฎ
โ‹ฏ ๐‘‘๐‘›๐‘›
๐‘’11
โ‹ฎ
0
โ‹ฏ 0
๐‘‘11 ๐‘’11
โ‹ฑ
โ‹ฎ =
โ‹ฎ
โ‹ฏ ๐‘’๐‘›๐‘›
0
You can just multiply the corresponding diagonal
entries. Taking a ๐‘š๐‘กโ„Ž power of a diagonal matrix is
a matter of taking a power of each diagonal entry.
๐‘‘11
โ‹ฎ
0
โ‹ฏ
โ‹ฑ
โ‹ฏ
0
โ‹ฎ
๐‘‘๐‘›๐‘› ๐‘’๐‘›๐‘›
โ‹ฏ
0
โ‹ฑ
โ‹ฎ
โ‹ฏ ๐‘‘๐‘›๐‘›
๐‘š
๐‘š
๐‘‘11
= โ‹ฎ
0
โ‹ฏ
0
โ‹ฑ
โ‹ฎ
๐‘š
โ‹ฏ ๐‘‘๐‘›๐‘›
Multiplying a diagonal matrix by a scalar multiplies each diagonal entry. Adding two diagonal
matrices adds the corresponding diagonal entries. Now if the function ๐‘“ ๐‘ก is a polynomial
function with ๐‘“ ๐‘ก = ๐‘๐‘š ๐‘ก ๐‘š + โ‹ฏ + ๐‘1 ๐‘ก + ๐‘0 we get the following for a diagonal matrix D:
๐‘“ ๐ท = ๐‘๐‘š
๐‘‘11
โ‹ฎ
0
โ‹ฏ
0
โ‹ฑ
โ‹ฎ
โ‹ฏ ๐‘‘๐‘›๐‘›
๐‘š
+ โ‹ฏ + ๐‘0
๐‘“ ๐‘‘11
1 โ‹ฏ 0
โ‹ฎ
โ‹ฎ โ‹ฑ โ‹ฎ =
0
0 โ‹ฏ 1
โ‹ฏ
0
โ‹ฑ
โ‹ฎ
โ‹ฏ ๐‘“ ๐‘‘๐‘›๐‘›
The difficulty is what if the matrix is not a diagonal matrix? Finding a value of a certain
matrix function could involve a great number of calculations considering how you do matrix
multiplication. If the matrix you want to evaluate in the function is diagonalizable then there
is a huge short cut.
Example
If matrix A is similar to B with non-singular matrix S such that ๐ต = ๐‘† −1 ๐ด๐‘† show that ๐ต2 =
๐‘† −1 ๐ด2 ๐‘† and ๐ต3 = ๐‘† −1 ๐ด3 ๐‘†.
๐ต2 = ๐‘† −1 ๐ด๐‘† 2 = ๐‘† −1 ๐ด๐‘† ๐‘† −1 ๐ด๐‘† = ๐‘† −1 ๐ด ๐‘† −1 ๐‘† ๐ด๐‘† = ๐‘† −1 ๐ด๐ผ๐ด๐‘† = ๐‘† −1 ๐ด2 ๐‘†
๐ต3 = ๐‘† −1 ๐ด๐‘† ๐‘† −1 ๐ด๐‘† ๐‘† −1 ๐ด๐‘† = ๐‘† −1 ๐ด ๐‘† −1 ๐‘† ๐ด ๐‘† −1 ๐‘† ๐ด๐‘† = ๐‘† −1 ๐ด๐ผ๐ด๐ผ๐ด๐‘† = ๐‘† −1 ๐ด3 ๐‘†
Computing Powers of Similar Matrices
If A and B are similar matrices with ๐ต = ๐‘† −1 ๐ด๐‘† for a non-singular matrix S
then for any positive integer m, ๐ต๐‘š = ๐‘† −1 ๐ด๐‘š ๐‘† and ๐ด๐‘š = ๐‘†๐ต๐‘š ๐‘† −1.
๐ต๐‘š = ๐‘† −1 ๐ด๐‘š ๐‘†
๐ด๐‘š = ๐‘†๐ต๐‘š ๐‘† −1
๐ต๐‘š = ๐‘† −1 ๐ด๐‘† โ‹ฏ ๐‘† −1 ๐ด๐‘† = ๐‘† −1 ๐ด ๐‘† −1 ๐‘† โ‹ฏ ๐‘† −1 ๐‘† ๐ด ๐‘† = ๐‘† −1 ๐ด๐ผ โ‹ฏ ๐ผ๐ด ๐‘† = ๐‘† −1 ๐ด๐‘š ๐‘†
๐‘š ๐‘ก๐‘–๐‘š๐‘’๐‘ 
๐‘š ๐‘ก๐‘–๐‘š๐‘’๐‘ 
๐‘š ๐‘ก๐‘–๐‘š๐‘’๐‘ 
Example
Compute ๐ด5 where A is the matrix to the right from the previous example.
We can take advantage of the fact that ๐‘† −1 ๐ด๐‘† = ๐ท a diagonal
matrix to compute this where S and D are given to the right.
−1 −3 −1 0 5 2
๐ด = ๐‘†๐ท ๐‘† =
0 2 −1
1
2
−3
98
99
−1 −3 −2
=
=
−66 −67
1
2 −32 −32
5
5 −1
3
−1
=
−1
1
๐‘†=
−1
1
๐ด=
8
−6
9
−7
−1 0
−3
,๐ท =
0 2
2
−3 −1 0
2
3
0 32 −1 −1
2
Instead of four matrix multiplications we only needed to do two or 50% faster!
Computing Matrix Functions of Similar Matrices
If A and B are similar matrices with ๐ต = ๐‘† −1 ๐ด๐‘† for a non-singular
matrix S then and if ๐‘“ ๐‘ก = ๐‘๐‘š ๐‘ก ๐‘š + โ‹ฏ + ๐‘1 ๐‘ก + ๐‘0 then the
matrices ๐‘“ ๐ด and ๐‘“ ๐ต can be computed as given to the right.
๐‘“ ๐ต = ๐‘† −1 ๐‘“ ๐ด ๐‘†
๐‘“ ๐ด = ๐‘†๐‘“ ๐ต ๐‘† −1
If we can express a function as a polynomial (that is defined for all values) then we can
evaluate what the matrix is using this method. For example the function sin ๐‘ฅ is expressed as
an infinite polynomial in the following way:
๐‘ฅ 3 ๐‘ฅ 5 ๐‘ฅ 7 ๐‘ฅ 9 ๐‘ฅ 11
sin ๐‘ฅ = ๐‘ฅ −
+ − + −
±โ‹ฏ
3! 5! 7! 9! 11!
Example
Compute sin ๐ด , for the previous matrix A given to the right.
We can take advantage of the fact that ๐‘† −1 ๐ด๐‘† = ๐ท a diagonal
matrix to compute this where S and D are given to the right.
8
9
−1
= ๐‘† sin
−6 −7
0
−1 −3 2 sin −1 3 sin
=
− sin 2
sin
1
2
sin
๐ด=
๐‘†=
0
−1 −3 sin −1
๐‘† −1 =
0
2
1
2
−1
−2 sin −1 + 3 sin 2
=
2
2 sin −1 − 2 sin 2
−1
1
8
−6
9
−7
−1 0
−3
,๐ท =
0 2
2
0
2 3
sin 2 −1 1
−3 sin −1 − 3 sin 2
3 sin −1 + 2 sin 2
Orthogonal Matrices
If the columns of a ๐‘› × ๐‘› matrix ๐‘„ = ๐‘„1 , โ‹ฏ , ๐‘„๐‘› form an orthonormal basis for โ„๐‘› . That
is the set of vectors ๐‘„1 , โ‹ฏ , ๐‘„๐‘› is a basis such that ๐‘„๐‘–๐‘‡ ๐‘„๐‘— = 0 ๐‘“๐‘œ๐‘Ÿ ๐‘– ≠ ๐‘— and ๐‘„๐‘–๐‘‡ ๐‘„๐‘– = 1.
The matrix Q is called an orthogonal matrix and has the property that ๐‘„−1 = ๐‘„๐‘‡ .
Example
Apply Gram-Schmidt to the basis for โ„3 given to the right to find an
orthogonal matrix Q.
1
0
0
0
−1
๐ฐ1 = 0 , ๐ฐ2 = 1 − ๐‘๐‘Ÿ๐‘œ๐‘—๐ฐ1 1 = 1 − 2
−1
1
1
1
2
2
2
2
0
๐ฐ3 = 1 − ๐‘๐‘Ÿ๐‘œ๐‘—๐ฐ1 1 − ๐‘๐‘Ÿ๐‘œ๐‘—๐ฐ2 1 = 1 − 2
2
2
2
2
๐ฐ1
๐ฐ1
1
2
= 0 ,
−1
1
2
๐‘„= 0
−1
2
1
6
2
6
1
6
๐ฐ1
๐ฐ1
1
3
−1
3
1
3
=
1
6
2
6
1
6
, ๐‘„๐‘‡ =
,
๐ฐ1
๐ฐ1
1
2
1
6
1
3
=
0
2
6
−1
3
1
3
−1
3
1
3
−1
2
1
6
1
3
1
1
2
0 = 1 , ๐ฐ2
1
−1
2
1
1
6
0 −6 2 =
−1
1
1
0 2
0 , 1 , 1
−1 1 2
1
= 2๐ฐ2 = 2
1
1
−1
1
This gives
us an
orthogonal
basis for
โ„3 .
Normalize each vector by dividing by its
length to get a vector of length 1. This is
now an orthonormal basis for โ„3 .
1
2
๐‘‡
๐‘„๐‘„ = 0
−1
2
1
6
2
6
1
6
1
3
−1
3
1
3
1
2
1
6
1
3
0
2
6
−1
3
−1
2
1
6
1
3
1 0
= 0 1
0 0
0
0
1
Download