Continuing from the last lecture... Example. An example of repeated

advertisement
Continuing from the last lecture...
Example. An example of repeated eigenvalue having only two eigenvectors.


0 1 1
A =  1 0 1 .
1 1 0
Solution: Recall,
Steps to find eigenvalues and eigenvectors:
1. Form the characteristic equation
det(λI − A) = 0.
2. To find all the eigenvalues of A, solve the characteristic equation.
3. For each eigenvalue λ, to find the corresponding set of eigenvectors,
solve the linear system of equations
(λI − A)~x = 0
Step 1. Form the Characteristic Equation.
The characteristic equation is:
λ −1 −1 −1 λ −1 = 0
−1 −1 λ λ(λ2 − (−1)(−1)) − (−1)((−1)λ − (−1)(−1)) − 1((−1)λ − (−1)(−1)) = 0
λ3 − 3λ2 − 2 = 0
(λ + 1)2 (λ − 2) = 0
Step 2. Find the eigenvalues.
We need to solve the characteristic equation. i.e. we need to factorize
the characteristic polynomial. We can factorize it by either using long
division or by directly trying to spot a common factor.
Method 1: Long Division.
We want to factorize this cubic polynomial. In general it is quite difficult to guess what the factors may be. We try λ = ±1, ±2, ±3, etc.
and hope to quickly find one factor. Let us try λ = −1. We divide the
polynomial λ3 + 2λ2 − 29λ − 30 by λ + 1, to get,
λ2
3
−λ −2
−3λ −2;
λ + 1 |λ
−λ3 +λ2
−λ2 −3λ
− −λ2 −λ
−2λ
− −2λ
−2
−2
−2
0
The quotient is λ2 − λ − 2.
The remainder is 0.
Therefore λ3 − 3λ − 2 = (λ + 1)(λ2 − λ − 2) + 0.
λ3 − 3λ − 2 = 0
(λ + 1)(λ2 − λ − 2) = 0
(λ + 1)(λ + 1)(λ − 2) = 0
Therefore the eigenvalues are: {−1, 2}.
Method 2: Direct factorization by spotting common factor.
λ −1 −1 −1 λ −1 = 0
−1 −1 λ λ(λ2 − (−1)(−1)) − (−1)((−1)λ − (−1)(−1)) − 1((−1)λ − (−1)(−1))
λ(λ2 − 1) + 1(−λ − 1) − 1(1 + λ)
λ(λ − 1)(λ + 1) − 1(λ + 1) − 1(1 + λ)
(λ + 1)(λ(λ − 1) − 1 − 1)
(λ + 1)(λ2 − λ − 2)
(λ + 1)2 (λ − 2)
Therefore the eigenvalues of A are: {−1, 2}.
=
=
=
=
=
=
0
0
0
0
0
0
Step 3. Find Eigenvectors corresponding to each Eigenvalue:
We now need to find eigenvectors corresponding to each eigenvalue.
case(i) λ1 = −1.
The eigenvectors are the solution space of




x1
−1 −1 −1
 −1 −1 −1   x2  = 
−1 −1 −1
x3
−x1 − x2 − x3 = 0;
the following system:

0
0 
0
x1 = −x2 − x3
The set of eigenvectors corresponding to λ1 = −1 is,


x1
{  x2  |x1 = −x2 − x3 }
x3


−x2 − x3
 | at least one of x2 and x3 is a non-zero real number}
x2
{ 
x3




−1
−1
{ x2  1  + x3  0  | at least one of x2 and x3 is a non-zero real number}
0
1
We therefore can get two linearly independent eigenvectors corresponding
λ1 = −1:
 to 

−1
−1
~v1 =  1 , ~v2 =  0 .
0
1
case(ii) λ2 = 2.
The eigenvectors are the solution space of the following system:



 
2 −1 −1
x1
0
 −1 2 −1   x2  =  0 
−1 −1 2
x3
0
2x1 − x2 − x3 = 0
−x1 + 2x2 − x3 = 0
−x1 − x2 + 2x3 = 0
Since this system of equations looks fairly complicated, it may be
a good idea to use row reduction to simplify the system.




2 −1 −1 0
−1 −1 2 0
↔R1
 −1 2 −1 0  R3→
 −1 2 −1 0 
−1 −1 2 0
2 −1 −1 0


1
1 −2 0
R1 =(−1)R1
 −1 2 −1 0 
→
2 −1 −1 0


1 1 −2 0
R =R2 −2R1
−−3−−−
−−→  0 3 −3 0 
R2 =R1 +R2
0 −3 3 0


1 1 −2 0
R2 =1/3R2
 0 1 −1 0 
→
0 −3 3 0


1 1 −2 0
R3 =R3 +3R2
 0 1 −1 0 
→
0 0 0 0
The resulting set of equations is:
x2 − x3 = 0 =⇒ x2 = x3
x1 + x2 − 2x3 = 0 =⇒ x1 = x3
The set of eigenvectors corresponding to λ2 = 2 is,


x1
{  x2  |x1 = x2 = x3 }
x3


x3
{  x3  |x3 is a non-zero real number}
x3
 
1

1  |x3 is a non-zero real number}
{ x3
1
 
1
An eigenvector corresponding to λ2 = 2 is  1 .
1
Example. An example of a repeated eigenvalue having only one linearly
independent eigenvector.


6 3 −8
A =  0 −2 0  .
1 0 −3
Solution: The characteristic polynomial for A:
λ − 6 −3
8
0
λ+2
0 −1
0
λ+3 (λ − 6)((λ + 2)(λ + 3) − 0) − 3(0 − 0) + 8(0 − (−1)(λ + 2))
λ3 − λ2 − 16λ − 20
(λ + 2)(λ2 − 3λ − 10)
(λ + 2)2 (λ − 5)
= 0
=
=
=
=
Therefore the eigenvalues of A are: λ1 = −2, λ2 = 5.
case(i) λ1 = −2.
The eigenvectors are the solution space of



−8 −3 8
x1
 0
0 0   x2  =
x3
−1 0 1
−x1 + x3 =
−8x1 − 3x2 + 8x3 =
The set of eigenvectors

x1
{  x2
x3

x3

0
{
x3

the following system:
 
0
 0 
0
0 =⇒ x1 = x3
0 =⇒ x2 = 0
corresponding to λ1 = −2 is,

 |x1 = x3 , x2 = 0}

 |x3 is a non-zero real number}

1
{ x3  0  |x3 is a non-zero real number}
1
0
0
0
0

1
An eigenvector corresponding to λ1 = −2 is  0 .
1

case(ii) λ2 = 5.
The eigenvectors are the solution space of



x1
−1 −3 8
 0
7 0   x2  =
−1 0 8
x3
−x1 + 8x3 =
7x2 =
the following system:
 
0
 0 
0
0 =⇒ x1 = 8x3
0 =⇒ x2 = 0
The set of eigenvectors corresponding to λ2 = 5 is,


x1
{  x2  |x1 = 8x3 , x2 = 0}
x3


8x3
{  0  |x3 is a non-zero real number}
x3
 
8

0  |x3 is a non-zero real number}
{ x3
1


8
An eigenvector corresponding to λ2 = 5 is  0 .
1
Example. An example of an eigenvalue repeated three times having only
two linearly independent eigenvectors.


4 0 −1
A =  0 3 0 .
1 0 2
Solution: The characteristic equation for A:
det (λI − A) = 0
λ−4
0
1 0
λ−3
0 = 0
−1
0
λ−2 λ3 − 9λ2 + 27λ − 27 = 0
(λ − 3)3 = 0
Therefore A has only one eigenvalue: λ = 3.
The eigenvectors corresponding to λ = 3 are the solution space of the
following system:

 


0
−1 0 1
x1
 0 0 0   x2  =  0 
x3
0
−1 0 1
−x1 + x3 = 0 =⇒ x1 = x3
The set of eigenvectors corresponding to λ = 3 is,


x1
{  x2  |x1 = x3 }
x3


x3
{  x2  | at least one of x2 and x3 is a non-zero real number}
x3
 
 
0
1
{ x2  1  + x3  0  | at least one on x2 and x3 is a non-zero real number}
0
1
Therefore we can find two linearly independent eigenvectors corresponding
to the eigenvalue λ = 3:
 
 
1
0
 1  and  0  .
1
0
8.3
Dimension of Eigenspace
Definition. The rank of a matrix is defined to be the number of linearly
independent rows or columns.
Remark. We use the following results to determine the rank of a matrix.
• Let A be a square matrix and let Aechelon be the row echelon form of
the matrix. Then both A and Aechelon have the same number of linearly
independent rows.
• Consequently, Rank(A) = Rank(Aechelon ).
• The number of linearly independent rows of a matrix in row echelon
form is just the number of non-zero rows.
• We may conclude that
Rank(A) = Rank(Aechelon ) = number of non-zero rows in Aechelon .

Example.

2. Rank 

3. Rank 



−1 −1 −1
1 1 1
1. Rank  −1 −1 −1  = Rank  0 0 0  = 1.
−1 −1 −1
0 0 0



−8 −3 8
1 0 −1
0
0 0  = Rank  0 −3 0  = 2.
−1 0 1
0 0
0



−1 0 1
−1 0 1
0 0 0  = Rank  0 0 0  = 1.
−1 0 1
0 0 0
Definition. Let A be a square matrix of size n × n having an eigenvalue λ.
The set of eigenvectors corresponding to the eigenvalue λ, alongwith the zero
vector, is called the eigenspace corresponding to eigenvalue λ.
The dimension of the eigenspace is defined to be the maximum number of
linearly independent vectors in the eigenspace.
Theorem Let A be a matrix of size n × n and let λ be an eigenvalue of A
repeating k times. Then the dimension of the eigenspace of λ or equivalently
the maximum number of linearly independent eigenvectors corresponding to
λ is:
n − Rank(λI − A).
Example. Let us use
examples.

0 1
1. Recall A =  1 0
1 1
this result to confirm our calculations in the previous

1
1 
0
• has eigenvalue λ = −1 occurring with multiplicity 2.
• We found two linearly independent eigenvectors corresponding to this
eigenvalue.




−1
−1
~v1 =  1 , ~v2 =  0 .
0
1
We will use the above theorem to confirm that we should indeed expect
to find two linearly independent eigenvectors. By the above theorem, the
maximum number of linearly independent eigenvectors corresponding to
λ = −1 is:
= 3 − Rank(λI − A)


−1 −1 −1
= 3 − Rank  −1 −1 −1 
−1 −1 −1
=3−1
=2


6
3 −8
2. Recall A =  0 −2 0 
−1 0
1
• has eigenvalue λ = −2 occurring with multiplicity 2.
• We found only one linearly independent eigenvector corresponding
to this

 eigenvalue.
1
~v1 =  0 .
1
We will use the above theorem to confirm that we should indeed expect
to find only one linearly independent eigenvector. By the above theorem,
the maximum number of linearly independent eigenvectors corresponding
to λ = −2 is:
= 3 − Rank(λI − A)


−8 −3 8
0 0 
= 3 − Rank  0
−1 0 1
=3−2
=1


4 0 −1
3. Recall A =  0 3 0 
1 0 2
• has eigenvalue λ = 3 occurring with multiplicity 3.
• We found only two linearly independent eigenvectors corresponding
to this
 eigenvalue.

 
0
1



1 , ~v2 =
0 .
~v1 =
0
1
We will use the above theorem to confirm that we should indeed expect to
find only two linearly independent eigenvectors. By the above theorem,
the maximum number of linearly independent eigenvectors corresponding
to λ = 3 is:
= 3 − Rank(λI − A)


−1 0 1
= 3 − Rank  0 0 0 
−1 0 1
=3−1
=2
8.4
Diagonalization
Definition. A square matrix of
the form

d1 0
 0 d2

 . .

 . .

 0 .
0 .
size n is called a diagonal matrix if it is of
. .
0 .
.
.
.
dn−1
. .
0
0
0
.
.
0
dn








Definition. Two square matrices A and B of size n are said to be similar,
if there exists an invertible matrix P such that
B = P −1 AP
Definition. A matrix A is diagonalizable if A is similar to a diagonal matrix. i.e. there exists an invertible matrix P such that P −1 AP = D where D
is a diagonal matrix. In this case, P is called a diagonalizing matrix.
Example. Recall the example we did in the last chapter.
4 2
A=
.
1 3
Solution: In previous lectures we have found,
the eigenvalues of A are: λ1 = 2, λ2 = 5. −1
An eigenvector corresponding to λ1 = 2 is
.
1
2
An eigenvector corresponding to λ2 = 5 is
.
1
Define a new matrix P such that its columns are the eigenvectors of A.
−1 2
P =
.
1 1
Note that since eigenvectors corresponding to distinct eigenvalues are linearly
independent, the columns of P are linearly independent, hence P is an invertible matrix. (We can also confirm this by verifying that the determinant
of P is non-zero).
−1
1 −2
−1
P =
−1 −1
3
P
−1
AP =
=
=
=
−1
1
−1
3
−1
2
−5
3
−1 −6
0
3
2 0
,
0 5
−2
−1
4 2
1 3
−4
−5
−1 2
1 1
0
−15
−1 2
1 1
which is a diagonal matrix. Therefore A is diagonalizable and a diagonalizing
matrix for A is
−1 2
P =
,
1 1
and the corresponding diagonal matrix D is
2 0
D=
.
0 5
Theorem Let A be a square matrix of size n × n. Then A is diagonalizable
if and only if there exists n linearly independent eigenvectors. In this case a
diagonalizing matrix P is formed by taking as its columns the eigenvectors of
A,
P −1 AP = D.
where D is a diagonal matrix whose diagonal entries are eigenvalues corresponding to the eigenvectors of A. The ith diagonal entry is the eigenvalue
corresponding to the eigenvector in the ith column of P .
Example. An example where the matrix

6
3

0 −2
A=
−1 0
is not diagonalizable. Recall

−8
0 .
1
(i) Find the eigenvalues and eigenvectors of A.
(ii) Is A diagonalizable?
(iii) If it is diagonalizable, find a diagonalizing matrix P such that P −1 AP =
D, where D is a diagonal matrix.
Solution: In previous lectures we have found,
A had two eigenvalues:
λ1 = −1 occurring with multiplicity 2, and
λ2 = 5 occurring with multiplicity 1.
We could find only one linearly independent eigenvector corresponding to
λ1 = −1,
and one linearly independent eigenvector corresponding to λ2 = 5.
Therefore we could find only two linearly independent eigenvectors. By the
above theorem, A is not diagonalizable.
Proposition Let A be a square matrix of size n × n. If λ1 and λ2 are two
distinct eigenvalues of A, and ~v1 and ~v2 are eigenvectors corresponding to λ1
and λ2 respectively, then ~v1 and ~v2 are linearly independent.
Corollary Let A be a square matrix of size n. If A has n distinct eigenvalues then A is diagonalizable.
Example. Recall


4
0
1
A =  −1 −6 −2  .
5
0
0
(i) Find the eigenvalues and eigenvectors of A.
(ii) Is A diagonalizable?
(iii) Find a diagonalizing matrix P such that P −1 AP = D, where D is a
diagonal matrix.
Solution: (i) In previous lectures we have found,
The eigenvalues of A are: λ1 = −1, λ2 = −6 
and λ3= 5.
−1
An eigenvector corresponding to λ1 = −1 is 
5
−9
25
1
.

0
An eigenvector corresponding to λ2 = −6 is  1 .
 0
1
.
An eigenvector corresponding to λ3 = 5 is  −3
11
1

(ii) The matrix A is diagonalizable, since A is a square matrix of size 3 × 3
and we have found three linearly independent eigenvectors for A.
(iii) A diagonalizing matrix P has the eigenvectors as its columns. Let
 −1

0
1
5
.
1 −3
P =  −9
25
11
1 0 1
The corresponding diagonal matrix D will have the eigenvalues as diagonal entries corresponding to the columns of P . i.e. the ith diagonal
entry will be the eigenvalue corresponding to the eigenvector in the ith
column.


−1 0 0
D =  0 −6 0  .
0
0 5
We will now verify that P −1 AP = D.

  −1

4
0
1
0
1
5

1 −3
AP =  −1 −6 −2   −9
25
11
5
0
0
1 0 1


1/5 0
5
=  9/25 −6 −15/11 
−1
0
5
 −1


0 1
−1 0 0
5
  0 −6 0 
1 −3
P D =  −9
25
11
1 0 1
0
0 5


1/5 0
5

9/25 −6 −15/11 
=
−1
0
5
AP = P D
(1)
Since the columns of P are linearly independent, P has non-zero determinant and is therefore invertible. We multiply (1) by P −1 on both
sides.
P −1 AP = P −1 P D,
= ID
=D
Download