Basic Topics

advertisement
Basic Topics
Objectives
Get you comfortable and familiar with tools of linear algebra and other applications.
Will be given 12-14 exercise sets, will choose the best 10 (must serve at least 10)
80% of the grade is based on homework!
20% is based on the exam
Topics Planned to be Covered
-
Vector Spaces over โ„ and โ„‚.
Basic definitions connected with vector spaces
Matrices, Block of matrices
Gaussian elimination
Conservation of dimensions
Eigen values and Eigen vectors
Determinants
Jordan forms
Difference and differential equations
Normed linear spaces
Inner product spaces (orthogonality)
Symmetric, Hermitian, Normal matrices
Singular Value Decompositions (SVD)
Vecor Valued Functions of many variables
Fixed point theorems
Implicit function theorem
Extremal problems with constraints
General Idea of Linear Algebra
๐‘Ž11 ๐‘ฅ1 + ๐‘Ž12 ๐‘ฅ2 + โ‹ฏ + ๐‘Ž1๐‘ž ๐‘ฅ๐‘ž = ๐‘1
๐‘Ž21 ๐‘ฅ1 + ๐‘Ž22 ๐‘ฅ2 + โ‹ฏ + ๐‘Ž2๐‘ž ๐‘ฅ๐‘ž = ๐‘2
โ‹ฎ
๐‘Ž๐‘1 ๐‘ฅ1 + ๐‘Ž๐‘2 ๐‘ฅ2 + โ‹ฏ + ๐‘Ž๐‘๐‘ž ๐‘ฅ๐‘ž = ๐‘๐‘
Given ๐‘Ž๐‘–๐‘— , ๐‘– = 1, … , ๐‘, ๐‘— = 1, … , ๐‘ž; ๐‘1 , … , ๐‘๐‘
Find ๐‘ฅ1 , … , ๐‘ฅ๐‘ž
A choice of ๐‘ฅ1 , … , ๐‘ฅ๐‘ž that satisfies these equations is called a solution.
1) When can you guarantee at least one solution?
2) When can you guarantee at most one solution?
3) How can you find solutions?
4) Are there approximate solutions?
Short notation:
๐ด๐‘ฅ = ๐‘
๐‘Ž11 … ๐‘Ž1๐‘ž
โ‹ฎ ]
๐ด=[ โ‹ฎ
๐‘Ž๐‘1 … ๐‘Ž๐‘๐‘ž
๐‘ฅ1
๐‘ฅ=[โ‹ฎ ]
๐‘ฅ๐‘ž
A vector space ๐‘‰ over โ„ is a collection of objects called vectors on which two operations are
defined:
1) Vector addition – ๐‘ข ∈ ๐‘‰, ๐‘ฃ ∈ ๐‘‰ ๐‘ข + ๐‘ฃ ∈ ๐‘‰
2) Scalar multiplication – ๐›ผ ∈ โ„, ๐‘ข ∈ ๐‘‰ ๐›ผ๐‘ข ∈ ๐‘‰
Subjects to following constraints:
1) ๐‘ข + ๐‘ฃ = ๐‘ฃ + ๐‘ข
2) (๐‘ข + ๐‘ฃ) + ๐‘ค = ๐‘ข + (๐‘ฃ + ๐‘ค)
3) There is a zero vector, 0 s.t. ๐‘ข + 0 = ๐‘ข, 0 + ๐‘ข = ๐‘ข
4) ๐‘ข ∈ ๐‘ˆ, then there is a ๐‘ค s.t. ๐‘ข + ๐‘ค = 0
5) 1 ∈ โ„, 1๐‘ข = ๐‘ข
6) ๐›ผ, ๐›ฝ ∈ โ„, then ๐›ผ(๐›ฝ๐‘ข) = ๐›ผ๐›ฝ๐‘ข
7) ๐›ผ, ๐›ฝ ∈ โ„, then (๐›ผ + ๐›ฝ)๐‘ข = ๐›ผ๐‘ข + ๐›ฝ๐‘ข
8) ๐›ผ(๐‘ข + ๐‘ฃ) = ๐›ผ๐‘ข + ๐›ผ๐‘ฃ
Vector Space Example
โ„๐‘˜
๐‘ข1
๐‘ข2
[โ‹ฎ]
๐‘ข๐‘˜
๐‘ข๐‘– ∈ โ„
๐‘ข1
๐‘ฃ1
๐‘ข1 + ๐‘ฃ1
โ‹ฎ
โ‹ฎ
[ ]+[ ]= [ โ‹ฎ
]
๐‘ข๐‘˜
๐‘ฃ๐‘˜
๐‘ข๐‘˜ + ๐‘ฃ๐‘˜
โ„2๐‘‹3
๐‘Ž11 ๐‘Ž12
๐ด = [๐‘Ž
21 ๐‘Ž22
๐‘Ž13
๐‘Ž23 ]
๐ต=[
๐‘11
๐‘21
๐‘12
๐‘22
๐‘13
]
๐‘23
Claim: if ๐‘‰ is a vector space over โ„, then there is exactly one zero element.
Proof: Suppose that 0 and 0ฬƒ are both zero elements.→
0+๐‘ข =๐‘ข+0=๐‘ข
0ฬƒ + ๐‘ข = ๐‘ข + 0ฬƒ = ๐‘ข
Lets set ๐‘ข = 0ฬƒ
0 + 0ฬƒ = 0ฬƒ
Lets set ๐‘ฃ = 0
0ฬƒ + 0 = 0
Therefore 0 = 0ฬƒ
Claim: If ๐‘ข ∈ ๐‘‰, there is exactly one additive inverse
Suppose ๐‘ค and ๐‘ค
ฬƒ are both additive inverses of ๐‘ข.
๐‘ข+๐‘ค =0
๐‘ข+๐‘ค
ฬƒ =0
๐‘ค =๐‘ค+0
Lets mix things up:
๐‘ค = ๐‘ค + 0 = ๐‘ค + (๐‘ข + ๐‘ค
ฬƒ) = (๐‘ค + ๐‘ข) + ๐‘ค
ฬƒ =0+๐‘ค
ฬƒ →๐‘ค=๐‘ค
ฬƒ
We denote the additive inverse of ๐‘ข by – ๐‘ข.
๐‘ข + (- − -๐‘ข) = 0
Since this is clumsy to write we abbreviate this as ๐‘ข − ๐‘ข = 0
Note: We can replace โ„ by โ„‚!
Let ๐‘‰ be a vector space over ๐”ฝ Means we can replace ๐”ฝ by either โ„ or โ„‚ in all that follows this
statement.
Definitions
Let ๐‘‰ be a vector space over ๐”ฝ.
A subset ๐‘ˆ of ๐‘‰ is called a subspace of ๐‘‰ if:
1) ๐‘ข, ๐‘ค ∈ ๐‘ˆ → ๐‘ข + ๐‘ค ∈ ๐‘ˆ
2) ๐›ผ ∈ ๐”ฝ, ๐‘ข ∈ ๐‘ˆ → ๐›ผ๐‘ข ∈ ๐‘ˆ
If (1) and (2) hold, then ๐‘ˆ is a vector space of ๐‘‰ over ๐”ฝ.
Example
๐‘Ž
๐‘‰ = โ„2 = {[ ] |๐‘ค๐‘–๐‘กโ„Ž ๐‘Ž, ๐‘ ∈ โ„}
๐‘
๐‘Ž
๐‘ˆ = {[ ] |๐‘Ž ∈ โ„}
0
7
1
8
[ ]+[ ]= [ ]
0
0
0
๐‘Ž
๐›ผ๐‘Ž
๐›ผ๐‘Ž
๐›ผ[ ] = [ ] = [ ]
0
๐›ผ0
0
As we can see, this is a subspace…
However…
๐‘Ž
Let ๐‘ˆ = {[ ] |๐‘Ž ∈ โ„}
1
๐‘Ž
๐‘
๐‘Ž+๐‘
[ ]+[ ] = [
]
1
1
2
So this is not a subspace!
Span
A span {๐‘ฃ1 , … , ๐‘ฃ๐‘˜ } (๐‘ฃ1 , … , ๐‘ฃ๐‘˜ ∈ ๐‘‰)
๐‘˜
= {∑ ๐›ผ๐‘– : ๐›ผ1 , … , ๐›ผ๐‘˜ ∈ ๐”ฝ}
๐‘–=1
2๐‘ฃ1 + 7๐‘ฃ2 + 3๐‘ฃ3
2๐‘ฃ1 + 0๐‘ฃ2 + 0๐‘ฃ3
Check: span {๐‘ฃ1 , … , ๐‘ฃ๐‘˜ } is a subspace of ๐‘‰,
Clear
Span ๐‘ ๐‘๐‘Ž๐‘›{๐‘ฃ1 , ๐‘ฃ2 } ⊆ ๐‘ ๐‘๐‘Ž๐‘›{๐‘ฃ1 , ๐‘ฃ2 , ๐‘ฃ3 } ⊆ ๐‘ ๐‘๐‘Ž๐‘›{๐‘ฃ1 , ๐‘ฃ2 , ๐‘ฃ3 , ๐‘ฃ4 }
For example:
1
1 2
1 2 3
๐‘ ๐‘๐‘Ž๐‘› {[ ]} ⊆ ๐‘ ๐‘๐‘Ž๐‘› {[ ] , [ ]} ⊆ ๐‘ ๐‘๐‘Ž๐‘› {[ ] , [ ] , [ ]}
1
1 2
1 2 3
But they are all equal(!!):
1
2
3
1
๐›ผ [ ] + ๐›ฝ [ ] + ๐›พ [ ] = (๐›ผ + ๐›ฝ + ๐›พ) [ ]
1
2
3
1
Linear Dependency
๐‘ฃ1 , … , ๐‘ฃ๐‘˜ are said to be linearly dependent if there is a choice ๐›ผ1 , … , ๐›ผ๐‘˜ ∈ ๐”ฝ s.t.
๐›ผ1 ๐‘ฃ1 + โ‹ฏ + ๐›ผ๐‘˜ ๐‘ฃ๐‘˜ = 0
Not all of which are zero!
๐›ผ
๐›ผ
If say ๐›ผ1 ≠ 0, ๐‘ฃ1 = (-๐›ผ2 ) ๐‘ฃ2 + โ‹ฏ + (-๐›ผ๐‘˜ ) ๐‘ฃ๐‘˜
1
1
๐‘ฃ1 , … , ๐‘ฃ๐‘˜ are said to be linearly independent if the following statement is true:
๐›ผ1 ๐‘ฃ1 +, … , ๐›ผ๐‘˜ ๐‘ฃ๐‘˜ = 0 → ๐›ผ1 = ๐›ผ2 = โ‹ฏ = ๐›ผ๐‘˜ = 0
If ๐‘ฃ1 , … , ๐‘ฃ๐‘˜ are linearly independent, then
๐‘ข = ๐›ผ1 ๐‘ฃ1 + โ‹ฏ + ๐›ผ๐‘˜ ๐‘ฃ๐‘˜ = ๐›ฝ1 ๐‘ฃ1 + โ‹ฏ + ๐›ฝ๐‘˜ ๐‘ฃ๐‘˜ → ๐›ผ๐‘– = ๐›ฝ๐‘– for all ๐‘–.
A set of vectors ๐‘ข1 , … , ๐‘ข๐‘™ is said to be a basis for a vector space ๐‘ˆ over ๐”ฝ, if ๐‘ข1 , … , ๐‘ข๐‘™ ∈ ๐‘ˆ and:
1) ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข1 , … , ๐‘ข๐‘™ } = ๐‘ˆ (there’s enough vectors to span the entire space)
2) ๐‘ข1 , … , ๐‘ข๐‘˜ are linearly independent. (there are’t too many vectors)
Matrix Mutiplication
If ๐ด๐‘๐‘‹๐‘ž , ๐ด๐‘ž๐‘‹๐‘Ÿ are matrices,
๐‘ž
๐ถ = ๐ด๐ต is a ๐‘๐‘‹๐‘Ÿ matrix with ๐‘๐‘–,๐‘— = ∑๐‘ =1 ๐‘Ž๐‘–,๐‘  ๐‘๐‘ ,๐‘—
๐‘1,๐‘—
[๐‘Ž๐‘–,1 , … , ๐‘Ž๐‘–,๐‘ ] โˆ™ [ โ‹ฎ ]
๐‘๐‘ž,๐‘—
If ๐ด๐‘๐‘‹๐‘ž , ๐ต๐‘ž๐‘‹๐‘Ÿ , ๐ท๐‘Ÿ๐‘‹๐‘  are matrices, then (๐ด๐ต)๐ท = ๐ด(๐ต๐ท)
In general – ๐ด๐ต ≠ ๐ต๐ด!
0 1 1 1
0 0
1
][
]=[
]≠[
0 1 0 0
0 0
0
๐›ผ, ๐›ฝ ๐›ผ๐›ฝ = 0 → ๐›ผ = 0 or ๐›ฝ = 0
[
1 0 1
0
][
]=[
0 0 1
0
2
]
0
1
๐ด ∈ ๐”ฝ๐‘๐‘‹๐‘Ÿ Is said to be left invertible if there is ๐ต ∈ ๐”ฝ๐‘ž๐‘‹๐‘ s.t. ๐ต๐ด = ๐ผ ๐‘ž = [0
0
๐‘๐‘‹๐‘ž
๐‘ž๐‘‹๐‘
๐ด∈๐”ฝ
is right invertible if there’s a ๐ถ ∈ ๐”ฝ
s.t. ๐ด๐ถ = ๐ผ๐‘
If ๐ต๐ด = ๐ผ๐‘ž and ๐ด๐ถ = ๐ผ๐‘ then ๐ต = ๐ถ.
๐ต = ๐ต๐ผ๐‘ = ๐ต(๐ด๐ถ) = (๐ต๐ด)๐ถ = ๐ผ๐‘ž ๐ถ = ๐ถ
Later we shall show that this also forces ๐‘ = ๐‘ž.
Triangular matrices
๐ด๐‘๐‘‹๐‘ ∈ ๐”ฝ is said to be upper triangular if ๐‘Ž๐‘–,๐‘— = 0 for ๐‘– > 0
๐‘Ž11 …
๐‘Ž1๐‘
0
โ‹ฑ
โ‹ฎ ]
[
0
0 ๐›ผ_๐‘๐‘ ๐‘๐‘‹๐‘
๐ด๐‘๐‘‹๐‘ ∈ ๐”ฝ is said to be lower triangular if ๐‘Ž๐‘–,๐‘— = 0 for ๐‘– > 0
0 0
โ‹ฑ 0]
0 1 ๐‘ž๐‘‹๐‘ž
๐‘Ž11
[ โ‹ฎ
๐‘Ž๐‘1
0
โ‹ฑ
…
0
0 ]
๐›ผ_๐‘๐‘
๐‘๐‘‹๐‘
If both upper and lower than we denote it as simply triangular.
๐‘Ž11 0
0
โ‹ฑ
0 ]
[ 0
0 0 ๐›ผ_๐‘๐‘ ๐‘๐‘‹๐‘
Theorm: Let ๐ด ∈ ๐”ฝ๐‘๐‘‹๐‘ be a triangular matrix, then A is invertible if and only if all the diagonal
entries ๐‘Ž๐‘–๐‘– are nonzero.
Moreover, in this case, ๐ด is upper triangular ↔ its inverse is upper triangular
And ๐ด is lower triangular ↔ its inverse is lower triangular.
๐ด=[
๐‘Ž
0
๐‘
]
๐‘‘
Investigate ๐ด ๐ต = ๐ผ2
want to:
๐‘Ž
[
0
[
๐‘ ๐›ผ
][
๐‘‘ ๐›พ
๐‘Ž๐›ผ + ๐‘๐›พ
๐‘‘๐›พ
๐›ฝ
1
]=[
๐›ฟ
0
๐‘Ž๐›ฝ + ๐‘๐›ฟ
]
๐‘‘๐›ฟ
Want to choose ๐›ผ, ๐›ฝ, ๐›พ, ๐›ฟ so that:
๐‘Ž๐›ผ + ๐‘๐›พ = 1
๐‘‘๐›พ = 0
๐‘Ž๐›ฝ + ๐‘๐›ฟ = 0
๐‘‘๐›ฟ = 1
1
(4)→ ๐‘‘ ≠ 0, ๐›ฟ ≠ 0, ๐›ฟ = ๐‘‘
(2)+(๐‘‘ ≠ 0) → ๐›พ = 0
(1)+๐›พ = 0 → ๐‘Ž๐›ผ = 1 → ๐‘Ž ≠ 0, ๐›ผ ≠ 0,
1
1
๐›ผ
1
(3)๐›ฝ = − (๐‘Ž) ๐‘ (๐‘‘)
If ๐ด๐ต = ๐ผ2 it is necessary that ๐‘Ž ≠ 0 and ๐‘‘ ≠ 0.
Also we have a formula for B
Can show ๐ต๐ด = ๐ผ2
0
]
1
Theorm: Let ๐ด ∈ ๐”ฝ๐‘๐‘‹๐‘ , ๐ด is triangular, then A is invertible if and only if the diagonal entries of ๐ด
are all non-zero.
Moreover, if this condition is met, then A is upper triangular ↔ the inverse to ๐ด is upper
triangular. (lower↔lower)
๐ด invertible, there exists a matrix ๐ต ∈ ๐”ฝ๐‘๐‘‹๐‘
Such that, ๐ด๐ต = ๐ต๐ด = ๐ผ๐‘
Let ๐ด(๐‘˜+1)๐‘‹(๐‘˜+1) upper triangular matrix.
๐‘Ž=[
๐ด11
0
๐ด12
]
๐ด22
Plan: Suppose we know that if ๐ด๐‘˜๐‘‹๐‘˜ upper triangular, then ๐ด is invertible if and only if it’s
diagonal entries are nonzero.
Objective: Extend this to (๐‘˜ + 1)๐‘‹(๐‘˜ + 1) upper triangular matrices.
Suppose first that ๐ด is invertible.
๐ด11 ๐ด12 ๐ต11 (๐‘˜๐‘‹๐‘˜) ๐ต12 (๐‘˜๐‘‹1)
][
]=
๐ด21 ๐ด22 ๐ต21 (1๐‘‹๐‘˜) ๐ต22 (1๐‘‹1)
๐ด ๐ต + ๐ด12 ๐ต21 ๐ด11 ๐ต12 + ๐ด12 ๐ต22
[ 11 11
]
๐ด22 ๐ต21
๐ด22 ๐ต22
๐ด๐ต = [
TODO: Draw single vector multiplication
[๐‘Ž๐‘–1 ๐‘Ž๐‘–2 ๐‘Ž๐‘–3 ]
[ ]
(1)
(2)
(3)
(4)
๐ด11 ๐ต11 + ๐ด12 ๐ต21 = ๐ผ๐‘˜
๐ด22 ๐ต21 = 01๐‘‹๐‘˜
๐ด11 ๐ต12 + ๐ด12 ๐ต22 = 0
๐ด22 ๐ต22 = 1
(4)→ ๐ด22 ≠ 0, ๐ต22 ≠ 0 ๐‘Ž๐‘›๐‘‘ ๐‘22 =
1
๐ด22
(2)+๐ด22 ≠ 0 → ๐ต21 = 0
(1)+๐ต21 = 0 → ๐ด11 ๐ต11 = ๐ผ๐‘˜
๐ต11 ๐ด11
๐ต๐ด = ๐ผ๐‘˜+1
๐ต
๐ต12 ๐ด11 ๐ด12
๐ผ 0
[ 11
][
]=[๐‘˜
]
0 ๐ต22 0 ๐ด22
0 1
= ๐ผ๐‘˜ → ๐ด11 ๐‘–๐‘  ๐‘Ž ๐‘˜๐‘‹๐‘˜ ๐‘ข๐‘๐‘๐‘’๐‘Ÿ ๐‘ก๐‘Ÿ๐‘–๐‘Ž๐‘›๐‘”๐‘ข๐‘™๐‘Ž๐‘Ÿ ๐‘š๐‘Ž๐‘ก๐‘Ÿ๐‘–๐‘ฅ ๐‘กโ„Ž๐‘Ž๐‘ก ๐‘–๐‘  ๐‘–๐‘›๐‘ฃ๐‘’๐‘Ÿ๐‘ก๐‘–๐‘๐‘™๐‘’!
If the theorem is true for ๐‘˜๐‘‹๐‘˜ marices, then if ๐ด ∈ ๐”ฝ(๐‘˜+1)๐‘‹(๐‘˜+1) triangular upper matrix,
diagonal of ๐ด11 are nonzero and ๐ด22 ≠ 0!
Showed: ๐ด(๐‘˜+1)๐‘‹(๐‘˜+1) is upper triangular invertible.
Now take ๐ด(๐‘˜+1)๐‘‹(๐‘˜+1) upper triangular with non zero entries on its diagonal.
๐ด (๐‘˜๐‘‹๐‘˜)
๐ด12
๐ด = [ 11
]
0
๐ด22 (1๐‘‹1)
๐ด11 is ๐‘˜๐‘‹๐‘˜ upper triangular matrix with non-zero entries on its diagonal.
Theorm true for ๐‘˜๐‘‹๐‘˜ matrices, then this = there exists a ๐‘˜๐‘‹๐‘˜ ๐ถ ๐‘ . ๐‘ก.
๐ด11 ๐ถ = ๐ถ๐ด11 = ๐ผ๐‘˜
−1
๐ถ −๐ถ๐ด12 ๐ด−1
๐ด−1 −๐ด11
๐ด12 ๐ด−1
22
22
Lets define B=[[
] = [ 11
]
−1
0
๐ด22
0
๐ด−1
22
Can show – ๐ด๐ต = ๐ต๐ด = ๐ผ๐‘˜+1
[
๐ด11
0
−1
๐ด12 ๐ด11
][
๐ด22 0
−1
−1
−๐ด11
๐ด12 ๐ด−1
22
] = [๐ด11 ๐ด11
−1
๐ด22
0
๐ผ๐‘˜ - − ๐ผ๐‘˜ + ๐ผ๐‘˜
=[
]
0
∗
−1
−1
๐ด11 (−๐ด11
๐ด12 ๐ด−1
22 ) + ๐ด12 ๐ด22 ]
∗
๐ด๐‘‡ =transpose of ๐ด.
1 4
2 3
] = [2 5 ]
5 6
3 6
๐‘–๐‘— entry of ๐ด๐‘‡ is ๐‘—๐‘– entry of ๐ด
๐ด๐ป =Hermitian transpose - ฬ…ฬ…ฬ…ฬ…
๐ด๐‘‡ (take complex conjugate)
1
Example: ๐ด = [
4
1−๐‘–
3
๐ป
] → ๐ด = [ −2๐‘–
6 − 7๐‘–
3
๐ด ∈ ๐”ฝ๐‘๐‘‹๐‘ž , ๐’ฉ๐ด = {๐‘ฅ ∈ ๐”ฝ๐‘ž : ๐ด๐‘ฅ = 0} - subspace of ๐”ฝ๐‘ž
Nullspace of ๐ด
โ„›๐ด = {๐ด๐‘ฅ: ๐‘ฅ ∈ ๐”ฝ๐‘ž } - subspace of ๐”ฝ๐‘ž
1+๐‘–
๐ด=[
4
2๐‘–
5
๐ด๐‘ฅ = 0
๐ด๐‘ฆ = 0
→
๐ด(๐‘ฅ + ๐‘ฆ) = ๐ด๐‘ฅ + ๐ด๐‘ฆ = 0 + 0 = 0
๐ด๐›ผ๐‘ฅ = ๐›ผ(๐ด๐‘ฅ) = ๐›ผ0 = 0
Need to show all properties. But they exist…
4
5 ]
6 + 8๐‘–
Matrix Multiplication
๐ด๐‘๐‘‹๐‘ž , ๐ต๐‘ž๐‘‹๐‘Ÿ
๐‘Ž1
๐‘Ž2
๐ด = [๐‘Ž1 ๐‘Ž2 … ๐‘Ž๐‘ž ] = [ โ‹ฎ ]
๐‘Ž๐‘
๐‘1
๐‘2
๐ต = [๐‘1 ๐‘2 … ๐‘๐‘ž ] = [ ]
โ‹ฎ
๐‘๐‘
[
1 2
4 5
3
] = [๐‘ข1 ๐‘ข2 ๐‘ข3 ]
6
1
๐‘ข1 = [ ]
4
๐‘Ž1 ๐ต
๐‘Ž1
๐‘Ž
๐ต
๐ด๐ต = ๐ด[๐‘1 … ๐‘๐‘Ÿ ] = [๐ด๐‘1 ๐ด๐‘2 … ๐ด๐‘๐‘Ÿ ] = [ โ‹ฎ ] ๐ต = [ 2โ‹ฎ ]
๐‘Ž๐‘
๐‘Ž๐‘ ๐ต
๐‘ž
๐ด๐ต = ∑ ๐‘Ž๐‘– ๐‘๐‘–
๐‘–=1
๐ด๐ต = [๐‘Ž1 ๐‘Ž2 … ๐‘Ž1 ]๐ต = [๐‘Ž1 0 … 0]๐ต + [0 ๐‘Ž2 0 … 0]๐ต + โ‹ฏ + [0 … 0 ๐‘Ž๐‘ž ]๐ต
There are ๐‘๐‘‹๐‘ž matrices that are right invertible but not left invertible.
Similarly - there are ๐‘๐‘‹๐‘ž matrices that are left invertible and not right invertible.
This does not happen if ๐‘ = ๐‘ž.
1
1 0 0
[
] [0
0 1 0
๐‘Ž
0
1] = ๐ผ2
๐‘
Gaussian Elimination
๐ด
โŸ
๐‘”๐‘–๐‘ฃ๐‘’๐‘› ๐‘๐‘‹๐‘ž
โˆ™๐‘ฅ =
โŸ
๐‘
๐‘”๐‘–๐‘ฃ๐‘’๐‘› ๐‘ž๐‘‹1
Looking for solutions for ๐‘ฅ…
For instance:
3๐‘ฅ1 + 2๐‘ฅ2 = 6
2๐‘ฅ1 + 3๐‘ฅ2 = 4
A Gaussian Elimination is a method of passing to a new system of equations that is easier to
work with.
This new system must have the same solutions as the original one.
๐‘ˆ ๐‘ž๐‘‹๐‘ž is called upper echelon if the first nonzero entry in row ๐‘–, sits to the left of the first
nonzero entry in row ๐‘– + 1.
Example:
1
0
๐‘ˆ=[
0
0
3
6
0
0
4
0
0
0
2
0
4
0
1
2
]
1
0
The first nonzero entry in each row is called a pivot. In the example we have 3 pivots marked in
red.
Consider the equation:
๐ด๐‘ฅ = ๐‘
0 2 3 1
1
๐ด = [ 1 5 3 4] ๐‘ = [ 2]
2 6 3 2
3
Two operations:
1) Interchange rows
2) Substract a multiple of one row from another row
Each operation corresponds to multiplying on the left by an invertible matrix.
We can revert the process later:
→ ๐ถ๐ด๐‘ฅ = ๐ถ๐‘ → ๐ถ −1 (๐ถ๐ด๐‘ฅ) = ๐ถ −1 ๐ถ๐‘ → ๐ด๐‘ฅ = ๐‘
Steps:
1)
2)
3)
4)
0
ฬƒ
[๐ด๐‘]
Construct the augmented matrix ๐ด =
= [1
2
1 5 3 4 |
Interchange rows 1 and 2. ๐ดฬƒ1 = [0 2 3 1 |
2 6 3 2 |
1 5
ฬƒ
Subtract 2 times row 1 from row 3 ๐ด2 = [0 2
0 −4
1 5 3 4
Add 2 times row 2 to row 3 ๐ดฬƒ3 = [0 2 3 1
0 0 3 −4
2 3 1
5 3 4
6 3 2
2
1]
1
3
4
3
1
−3 −6
| 2
| 1]
| −1
| 1
| 2]
| 1
| 2
| 1]
| −3
2
5) Solve the system – ๐‘ˆ๐‘ฅ = [ 1 ]
−1
๐‘ฅ1
1 5 3 4 ๐‘ฅ
2
[0 2 3 1 ] [๐‘ฅ ]
3
0 0 3 −4 ๐‘ฅ
4
6) Work from bottom up
3๐‘ฅ3 − 4๐‘ฅ4 = −1
Solve the pivot variable: ๐‘ฅ3 = (−1 + 4๐‘ฅ4 )3
Note:
0 1 0
๐ดฬƒ1 = ๐‘ƒ1 ๐ดฬƒ, ๐‘ƒ1 = [1 0 0]
0 0 1
0 1 0 ๐‘Ž
๐‘
[ 1 0 0] [ ๐‘ ] = [ ๐‘Ž ]
0 0 1 ๐‘
๐‘
๐ดฬƒ2 = ๐ธ1๐ดฬƒ1
1 0 0
๐ธ1 = [ 0 1 0]
−2 0 1
๐‘Ž
๐‘Ž
๐ธ1 [๐‘] = [ ๐‘
]
−2๐‘Ž + ๐‘
๐‘
๐ดฬƒ3 = ๐ธ2 ๐ดฬƒ2
1 0 0
๐ธ2 = [0 1 0]
0 2 1
๐‘Ž
๐‘Ž
๐‘
๐ธ2 [ ] = [ ๐‘ ]
2๐‘ + ๐‘
๐‘
------- End of lesson 2
Gauss Elimination
0 2 3
๐ด = [1 5 3
2 6 3
1
๐‘ = [2]
1
Try to solve ๐ด๐‘ฅ
1
4]
2
=๐‘
๐ดฬƒ = [๐ด ๐‘]
Two operations:
1) To permute (interchange) rows
2) Subtract a multiple of one row from the other
(1)&(2) are implemented by multiplying ๐ด๐‘ฅ = ๐‘ on the left by the appropriate invertible matrix.
๐ด๐‘ฅ = ๐‘
๐‘ƒ1 ๐ด๐‘ฅ = ๐‘ƒ1 ๐‘ where ๐‘ƒ1 is an appropriate permutation matrix
๐ธ1 ๐‘ƒ1 ๐ด๐‘ฅ = ๐ธ1 ๐‘ƒ1 ๐‘ where ๐ธ2 is a lower triangular matrix
โ‹ฎ
Eventally you will have ๐‘ˆ๐‘ฅ = ๐‘
Such that ๐‘ˆ is an upper echelon matrix.
1 5 3 4
2
๐ดฬƒ → [0 2 3 1
1 ] = [๐‘ˆ ๐‘ ′ ]
0 0 3 −4 −1
Now we need to solve ๐‘ˆ๐‘ฅ = ๐‘.
๐‘ฅ1 + 5๐‘ฅ2 + 3๐‘ฅ3 + 4๐‘ฅ4 = 2
2๐‘ฅ2 + 3๐‘ฅ3 + ๐‘ฅ4
3๐‘ฅ3 − 4๐‘ฅ4 = −1
Work from the bottom up, Solve the pivot variables in terms of the others.
๐‘ฅ3 =
−1+4๐‘ฅ4
3
2๐‘ฅ2 = −3๐‘ฅ3 − ๐‘ฅ4 + 1 = 1 − 4๐‘ฅ4 − ๐‘ฅ4 + 1 = −5๐‘ฅ4 + 2
๐‘ฅ1 = 2 − 5๐‘ฅ2 − 3๐‘ฅ3 − 4๐‘ฅ4 = 2 −
−1 +
25
๐‘ฅ4 − 5๐‘ฅ4
2
=-−1+
5(2−5๐‘ฅ4 )
—1 +
2
15
๐‘ฅ
2 4
4๐‘ฅ4 − ๐‘ฅ4 + 1 =
9
−2
๐‘ฅ1
2
5
1
๐‘ฅ2
−2
1 + ๐‘ฅ4
๐‘ฅ = [๐‘ฅ ] =
−3
3
4
๐‘ฅ4
[0]
3
[1]
9
2
0
5
๐ด − 2 = [ 0]
4
0
3
[1]
Another example:
0
0
๐ด=[
0
0
0
1
2
0
3
0
3
6
๐‘1
7
๐‘
0
] , ๐‘ = [ 2]
๐‘3
8
๐‘4
14
4
0
6
8
Try to solve ๐ด๐‘ฅ = ๐‘
0
0
๐ดฬƒ = [๐ด ๐‘] → [
0
0
1
0
0
0
0
3
0
0
0
4
2
0
0
7
1
0
๐‘2
๐‘1
]
๐‘3 + 2๐‘2 − ๐‘1
๐‘4 − 2๐‘1
๐ด๐‘ฅ = ๐‘ ⇔
๐‘ฅ2 = ๐‘2
3๐‘ฅ3 + 4๐‘ฅ3 + 7๐‘ฅ4 = ๐‘1
2๐‘ฅ4 + ๐‘ฅ5 = ๐‘3 − 2๐‘2 − ๐‘1
0 = ๐‘4 − 2๐‘1
Solve for ๐‘ฅ2 , ๐‘ฅ3 , ๐‘ฅ4 and find that:
0
0
1
๐‘2
0
0
5
3๐‘1 +4๐‘2 −2๐‘3
๐‘ฅ=
+ ๐‘ฅ1 0 + ๐‘ฅ5 − 3
3
1
๐‘3 −2๐‘2 −๐‘1
0
−2
2
[0]
[1]
[
]
0
This is a solution of ๐ด๐‘ฅ = ๐‘ for any choice of ๐‘ฅ1 and ๐‘ฅ5 provided ๐‘4 = 2๐‘1.
Let’s track the Gaussian elmnimation from a mathematical point of view:
๐‘ƒ๐ธ๐‘ƒ๐ด … = ๐‘ˆ
๐ด ∈ ๐”ฝ๐‘×๐‘ž ๐ด ≠ 0๐‘×๐‘ž , then there exists an invertible matrix ๐บ๐‘×๐‘ such that ๐บ๐ด = ๐‘ˆ upper
echelon.
Theorem: ๐‘ˆ ∈ ๐”ฝ๐‘×๐‘ž be upper echelon with ๐‘˜ pivots, ๐‘˜ ≥ 1,
Then:
(1) ๐‘˜ ≤ min{๐‘, ๐‘ž}
(2) ๐‘˜ = ๐‘ž ⇔ ๐‘ˆ is left invertible⇔ ๐’ฉ๐‘ˆ = {0}
(3) ๐‘˜ = ๐‘ ⇔ ๐‘ˆ is right invertible⇔ โ„›๐‘ˆ = ๐”ฝ๐‘
Proof of (2):
Suppose ๐‘˜ = ๐‘ž.
โ–ญ
If ๐‘ = ๐‘ž, ๐‘ˆ = [
โ–ญ
] →no zero diagonal entries
โ‹ฑ
0
Otherwise ๐‘ > ๐‘ž,
โ–ญ
โ‹ฑ
โ–ญ
๐‘ž×๐‘ž
๐‘ˆ11
= [ ๐‘−๐‘ž×๐‘ž
]
− − − −
๐‘ˆ21
0 0 0 0
[0 0 0 0]
๐‘ž×๐‘ž
๐‘×๐‘−๐‘ž
The left invertible matrix ๐‘‰ = [๐‘‰11 , ๐‘‰12
]
๐‘ž×๐‘ž
๐‘‰11 ๐‘ˆ11 = ๐ผ
by definition
๐‘ˆ=
๐‘‰๐‘ˆ = [๐‘‰11
๐‘‰12 ] [
๐‘ˆ11
] = ๐‘‰11 ๐‘ˆ11 + ๐‘‰12 ๐‘ˆ21 = ๐ผ๐‘ž + 0 = ๐ผ๐‘ž
๐‘ˆ21
(2a) We shown that ๐‘˜ = ๐‘ž ⇒ ๐‘ˆ is left invertible.
(2b) ๐‘ˆ left invertible ⇒ ๐’ฉ๐‘ˆ = {0}.
If ๐ด๐‘×๐‘ž , ๐’ฉ๐ด = {๐‘ฅ|๐ด๐‘ฅ = 0}
Let ๐‘ฅ ∈ ๐’ฉ๐ด By definition, it means that ๐‘ˆ๐‘ฅ = 0
By assumption, ๐‘ˆ is left invertible⇒ there is a ๐‘‰ ∈ ๐”ฝ๐‘×๐‘ž ๐‘ . ๐‘ก. ๐‘‰๐‘ˆ = ๐ผ๐‘ž
0 = ๐‘‰0 = ๐‘‰(๐‘ˆ๐‘ฅ) = (๐‘‰๐‘ˆ)๐‘ฅ = ๐ผ๐‘ž ๐‘ฅ = ๐‘ฅ
(2c) ๐’ฉ๐‘ˆ = {0} ⇒ ๐‘ˆ has ๐‘ž pivots.
๐’ฉ๐‘ˆ = {0} ⇒ ๐‘ˆ๐‘ฅ = 0 ⇔ ๐‘ฅ = 0
๐‘ฅ1
๐‘ฅ2
[๐‘ข1 , ๐‘ข2 , … , ๐‘ข๐‘ž ] [ โ‹ฎ ] = 0 ⇔ ๐‘ฅ1 = ๐‘ฅ2 = โ‹ฏ = ๐‘ฅ๐‘ž = 0
๐‘ฅ๐‘ž
๐’ฉ๐‘ˆ = {0} ⇒Columns of ๐‘ˆ are independent.
That forces k=q
(3๐‘Ž)๐‘˜ = ๐‘ ⇒ ๐‘ˆ is right invertible
๐‘ž = ๐‘ ⇒ ๐‘ˆ upper triangular with nonzero diagonal entries ⇒ U is invertible
๐‘ž>๐‘
โ–ญ
| โˆ™ โˆ™
โ–ญ โˆ™ โˆ™ โˆ™ โˆ™
๐‘ ๐‘ค๐‘Ž๐‘ ๐‘๐‘œ๐‘™๐‘ข๐‘š๐‘›๐‘ 
[0 0 โˆ™ โˆ™ โˆ™ ]๐‘ƒ
=
โ–ญ
| โˆ™ โˆ™]
[
0 0 0 0 โ–ญ
โ–ญ | โˆ™ โˆ™
You can find such a ๐‘ƒ such that ๐‘ˆ๐‘ƒ = [๐‘ˆ11 ๐‘ˆ12 ] here ๐‘ˆ11 is ๐‘ × ๐‘ upper triangular with
nonzero diagonal elements, and ๐‘ˆ12 = 0.
๐‘‰
๐‘‰
๐‘‰ = [ 11 ] so that [๐‘ˆ11 ๐‘ˆ12 ] [ 11 ] = ๐ผ๐‘ ⇒ Therefore ๐‘ˆ11 is invertible. By VP!
๐‘‰21
๐‘‰21
And it doesn’t matter what ๐‘‰21 we choose.
(3b) ๐‘ฅ is the input, ๐ด๐‘ฅ is the output.
The range is the set of outputs.
โ„›๐ด = {๐ด๐‘ฅ|๐‘ฅ ∈ ๐”ฝ๐‘ }
Claim: Given any ๐‘ ∈ ๐”ฝ๐‘ , can find an ๐‘ฅ ∈ ๐”ฝ๐‘ž ๐‘ . ๐‘ก. ๐‘ˆ๐‘ฅ = ๐‘
Let W be a right inverse to ๐‘ˆ, let ๐‘ฅ = ๐‘Š๐‘.
๐‘ˆ๐‘ฅ = ๐‘ˆ(๐‘Š๐‘) = (๐‘ˆ๐‘Š)๐‘ = ๐ผ๐‘ ๐‘ = ๐‘
(3c)โ„›๐‘ˆ = ๐”ฝ๐‘ ⇒ ๐‘ ๐‘๐‘–๐‘ฃ๐‘œ๐‘ก๐‘ 
If ๐‘˜ < ๐‘ then it must look something like:
โ–ญ
โ‹ฑ
๐‘ˆ=
−
0
[0
−
0
0
−
0
0
โ–ญ
−
0
0
−
0
0]
We have ๐‘ − ๐‘˜ zero rows.
So all of our answers would always have ๐‘ − ๐‘˜ zeroes at the end! Therefore, we don’t cover all
๐”ฝ๐‘
โ‹ฎ
โ‹ฎ
๐‘ˆ๐‘ฅ = โ‹ฎ
โ‹ฎ
[0]
Theorem: If ๐ด ∈ ๐”ฝ๐‘×๐‘ž , ๐ต ∈ ๐”ฝ๐‘×๐‘ž and ๐ต๐ด = ๐ผ๐‘ž then ๐‘ ≥ ๐‘ž
Proof: Clear that ๐ด ≠ 0๐‘×๐‘ž .
If we apply Gaussian elimination, can find ๐บ๐‘×๐‘ invertible matrix such that ๐บ๐ด = ๐‘ˆ which is
upper echelon.
๐ผ๐‘ž = ๐ต๐ด = ๐ต(๐บ −1 ๐‘ˆ) = (๐ต๐บ −1 )๐‘ˆ ⇒ ๐‘ˆ is left invertible⇒ ๐‘ˆ has ๐‘ž pivots⇒ ๐‘ ≥ ๐‘ž.
Theorem: Let ๐’ฑ be a vector space over ๐”ฝ and let {๐‘ข1 , … , ๐‘ข๐‘˜ } be a basis for ๐’ฑ.
And let {๐‘ฃ1 , … , ๐‘ฃ๐‘™ } also be a basis for ๐’ฑ. Then ๐‘˜ = ๐‘™.
๐‘™
∀๐‘– = 1, … , ๐‘˜: ๐‘ข๐‘– = ∑ ๐‘๐‘ ๐‘– ๐‘ฃ๐‘ 
๐‘ =1
Define ๐ต as ๐‘™ × ๐‘˜
๐‘™
∀๐‘— = 1, … , ๐‘˜: ๐‘ฃ๐‘— = ∑ ๐‘Ž๐‘ก๐‘— ๐‘ข๐‘ก
๐‘ =1
Define ๐ด as ๐‘˜ × ๐‘™
๐‘™
๐‘˜
๐‘ =1
๐‘ก=1
๐‘˜
๐‘™
๐‘™
0 ๐‘–๐‘“ ๐‘ก ≠ ๐‘–
๐‘ข๐‘– = ∑ ๐‘๐‘ ๐‘– ∑ ๐‘Ž๐‘ก๐‘  ๐‘ข๐‘ก = ∑ ∑ ๐‘Ž๐‘ก๐‘  ๐‘ข๐‘ก ⇒ ∑ ๐‘Ž๐‘ก๐‘  ๐‘๐‘ ๐‘– = {
⇒ ๐ด๐ต = ๐ผ๐‘˜ => ๐‘™ ≥ ๐‘˜
1 ๐‘–๐‘“ ๐‘ก = ๐‘–
๐‘ก=1 ๐‘ =1
๐‘ =1
But I can do this again symmetrically with u’s replaced by v’s, I would get that ๐‘˜ ≥ ๐‘™
So ๐‘˜ = ๐‘™.
-------End of lesson 3
Theorem:
Let ๐’ฑ be a vector space over ๐”ฝ. Let {๐‘ข1 , … , ๐‘ข๐‘˜ } be a basis for ๐’ฑ and {๐‘ฃ1 , … , ๐‘ฃ๐‘™ } also a basis for ๐’ฑ,
then ๐‘™ = ๐‘˜.
That number is called the dimension of ๐’ฑ
The vector space 0 = {0}. Define its dimension to be 0.
A transformation (mapping, operator,…) from a vector space ๐’ฐ over ๐”ฝ into a vector space ๐’ฑ
over ๐”ฝ is a rule (algorithm, formula) that assigns a a vector ๐‘‡๐‘ข in ๐’ฑ for each ๐‘ข ∈ ๐’ฐ.
๐‘Ž
Example 1: Let ๐’ฐ = โ„2 = {[ ] |๐‘Ž, ๐‘ ∈ โ„} ๐’ฑ = โ„3
๐‘
2
๐‘Ž
๐‘Ž
๐‘‡1 [ ] = [ 2๐‘ 2 ]
๐‘
๐‘Ž2 + ๐‘ 2
๐‘Ž
Example 2: Let ๐’ฐ = โ„2 = {[ ] |๐‘Ž, ๐‘ ∈ โ„} ๐’ฑ = โ„3
๐‘
1
3
๐‘Ž
๐‘Ž
๐‘‡2 [ ] = [ 2 0] [ ] (by matrix multiplication)
๐‘
๐‘
−4 6
Definition: A transformation ๐‘‡ from ๐’ฐ into ๐’ฑ is said to be linear if:
1) ๐‘‡(๐‘ข1 + ๐‘ข2 ) = ๐‘‡๐‘ข1 + ๐‘‡๐‘ข2
2) ๐‘‡(๐›ผ๐‘ข) = ๐›ผ๐‘‡๐‘ข
Is ๐‘‡1 linear?
No!
4
1
1
1
1
๐‘‡1 ([ ] + [ ]) = [8] ≠ 2 โˆ™ ๐‘‡1 [ ] = 2 โˆ™ [2]
1
1
1
8
2
Every linear transformation can be expressed in terms of matrix multiplication.
If ๐‘‡ is a linear transformation from a vector space ๐’ฐ over ๐”ฝ into the vector space ๐’ฑ over ๐”ฝ. Then
define:
๐’ฉ๐‘‡ = {๐‘ข ∈ ๐’ฐ|๐‘‡๐‘ข = 0๐’ฑ } –null space of ๐‘‡.
โ„› ๐‘‡ = {๐‘‡๐‘ข|๐‘ข ∈ ๐’ฐ} - range space of ๐‘‡.
๐’ฉ๐‘‡ is a subspace of ๐’ฐ
โ„› ๐‘‡ is also a subspace of ๐’ฐ
๐‘ข1 ∈ ๐’ฉ๐‘‡ , ๐‘ข2 ∈ ๐’ฉ๐‘‡ ⇒ ๐‘‡๐‘ข1 = 0๐’ฑ , ๐‘‡๐‘ข2 = 0_๐’ฑ
๐‘‡(๐‘ข1 + ๐‘ข2 ) = ๐‘‡๐‘ข1 + ๐‘‡๐‘ข2 = 0๐’ฑ + 0๐’ฑ = 0๐’ฑ
๐‘ข1 + ๐‘ข2 ∈ ๐’ฉ๐‘‡
Conservation of dimension
Theorem: Let ๐‘‡ be a linear transformation from a finite dimension vector space ๐’ฐ over ๐”ฝ into a
vector space ๐’ฑ over ๐”ฝ.
Then dim ๐’ฐ = dim ๐’ฉ๐‘‡ + dim โ„› ๐‘‡
Proof:
Suppose ๐’ฉ๐‘‡ ≠ {0} and โ„› ๐‘‡ ≠ {0}
Let ๐‘ข1 , … , ๐‘ข๐‘˜ be a basis for ๐’ฉ๐‘‡ .
Claim โ„› ๐‘‡ is also a finite dimensional vector space. Let ๐‘ฃ1 , … , ๐‘ฃ๐‘™ be a basis for โ„› ๐‘‡ .
Since ๐‘ฃ๐‘– ∈ โ„› ๐‘‡ , so we can find ๐‘ฆ๐‘– ∈ ๐’ฐ ๐‘ . ๐‘ก. ๐‘‡๐‘ฆ๐‘– = ๐‘ฃ๐‘– .
Claim: {๐‘ข1 , … , ๐‘ข๐‘˜ ; ๐‘ฆ1 , … , ๐‘ฆ๐‘™ } are linearly independent.
To check: Suppose there are coefficients: ๐‘Ž1 , … , ๐‘Ž๐‘˜ , ๐‘1 , … , ๐‘๐‘™ ∈ ๐”ฝ ๐‘ . ๐‘ก.
๐‘Ž1 ๐‘ข1 + โ‹ฏ + ๐‘Ž๐‘˜ ๐‘ข๐‘˜ + ๐‘1 ๐‘ฆ1 + โ‹ฏ + ๐‘๐‘™ ๐‘ฆ๐‘™ = 0๐’ฐ
Let’s activate the transformation on both sides
๐‘‡(๐‘Ž1 ๐‘ข1 + โ‹ฏ + ๐‘Ž๐‘˜ ๐‘ข๐‘˜ + ๐‘1 ๐‘ฆ1 + โ‹ฏ + ๐‘๐‘™ ๐‘ฆ๐‘™ ) = ๐‘‡(0๐’ฐ )
๐‘Ž1 ๐‘‡๐‘ข1 + โ‹ฏ + ๐‘Ž๐‘˜ ๐‘‡๐‘ข๐‘˜ + ๐‘1 ๐‘‡๐‘ฆ1 + โ‹ฏ + ๐‘๐‘™ ๐‘‡๐‘ฆ๐‘™ = ๐‘‡๐‘œ๐’ฐ = 0๐’ฑ
Why ๐‘‡0๐’ฐ = 0๐’ฑ
๐‘‡0๐’ฐ = ๐‘‡(0๐’ฐ + 0๐’ฐ ) = ๐‘‡0๐’ฐ + 0๐’ฑ ⇒ ๐‘‡0๐’ฐ = 0๐’ฑ
Since ๐‘ข1 , … , ๐‘ข๐‘˜ ∈ ๐’ฉ๐‘‡ , their transformation is zero!
Linearly independent items must never be zero, otherwise you can multiply the zero one with
some value other than zero and get zeros.
{๐‘ฃ1 , … , ๐‘ฃ๐‘™ } is a basis for โ„› ๐‘‡ ⇒ ๐‘ฃ1 , … , ๐‘ฃ๐‘™ are linearly independent⇒ ๐‘1 = ๐‘2 = โ‹ฏ = ๐‘๐‘™ = 0.
⇒ ๐‘Ž1 ๐‘ข1 + โ‹ฏ + ๐‘Ž๐‘˜ ๐‘ข๐‘˜ = 0๐’ฐ
{๐‘ข1 , … , ๐‘ข๐‘˜ } is a basis (for ๐’ฉ๐‘‡ ) ⇒ ๐‘ข1 , … , ๐‘ข๐‘˜ are linearly independent⇒ ๐‘Ž1 = โ‹ฏ = ๐‘Ž๐‘˜ = 0.
Claim next: ๐‘ ๐‘๐‘Ž๐‘› {๐‘ข1 , … , ๐‘ข๐‘˜ , ๐‘ฃ1 , … , ๐‘ฃ๐‘™ } = ๐’ฐ.
Let ๐‘ข ∈ ๐’ฐ and consider ๐‘‡๐‘ข ∈ โ„› ๐‘‡ .
๐‘™
๐‘™
๐‘™
⇒ ๐‘‡๐‘ข = ∑ ๐‘‘๐‘— ๐‘ฃ๐‘— = ∑ ๐‘‘๐‘— ๐‘‡๐‘ฆ๐‘— = ๐‘‡ (∑ ๐‘‘๐‘— ๐‘ฆ๐‘— )
๐‘—=1
๐‘ฆ−1
๐‘™
๐‘‡ (๐‘ข − ∑ ๐‘‘๐‘— ๐‘ฆ๐‘— ) = 0๐’ฑ
๐‘ฆ−1
๐‘™
๐‘ข − ∑ ๐‘‘๐‘— ๐‘ฆ๐‘— ∈ ๐’ฉ๐‘‡ ⇒
๐‘ฆ−1
๐‘ฆ−1
๐‘™
๐‘˜
๐‘ข − ∑ ๐‘‘๐‘— ๐‘ฆ๐‘— ∈ ๐’ฉ๐‘‡ = ∑ ๐‘๐‘– ๐‘ข๐‘– ⇒ ๐‘ข ∈ ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข1 , … , ๐‘ข๐‘˜ , ๐‘ฆ1 , … , ๐‘ฆ๐‘™ } ⇒
๐‘ฆ−1
๐‘–=1
{๐‘ข1 , … , ๐‘ข๐‘˜ , ๐‘ฆ1 , … , ๐‘ฆ๐‘™ } is a basis for ๐’ฐ
⇒ dim ๐’ฐ = โŸ
๐‘˜ + โŸ๐‘™
dim ๐’ฉ๐‘‡
dim ๐’ฉ๐‘‡
โˆŽ
If โ„› ๐‘‡ = {0} it forces ๐‘‡ to be the zero transformation
Last time we showed that:
If ๐‘ˆ is an upper echelon ๐‘ × ๐‘ž with ๐‘˜ pivots.
(1) ๐‘˜ ≤ min{๐‘, ๐‘ž}
(2) ๐‘˜ = ๐‘ž ⇔ ๐‘ˆ left invertible⇔ ๐’ฉ๐‘ˆ = {0}
(3) ๐‘˜ = ๐‘ ⇔ ๐‘ˆ right invertible⇔ โ„›๐‘ˆ = ๐”ฝ๐‘
Objective: Develop analogous set of facts for ๐ด ∈ ๐”ฝ๐‘×๐‘ž not necessarily upper echelon.
(1) ๐ด ∈ ๐”ฝ๐‘×๐‘ž , ๐ด ≠ 0๐‘×๐‘ž , then there exists an invertible matrix ๐บ ๐‘ . ๐‘ก. ๐บ๐ด = ๐‘ˆ is upper
echelon.
๐ต1 , ๐ต2 , ๐ต3 are invertible ๐‘ × ๐‘ matrices, then ๐ต1 ๐ต2 ๐ต3 is also invertible.
(2) Let ๐ด ∈ ๐”ฝ๐‘×๐‘ž , ๐ต ∈ ๐”ฝ๐‘ž×๐‘ž , ๐ถ ∈ ๐”ฝ๐‘×๐‘ ๐‘ . ๐‘ก. ๐ต and ๐ถ are invertible. Then:
a. โ„›๐ด๐ต = โ„›๐ด dim ๐’ฉ๐ด๐ต = dim ๐’ฉ๐ด
b. dim โ„›๐ถ๐ด = dim โ„›๐ด ๐’ฉ๐ถ๐ด = ๐’ฉ๐ด
Suppose ๐‘ฅ ∈ โ„›๐ด๐ต ⇒ there is ๐‘ข ๐‘ . ๐‘ก. ๐‘ฅ = (๐ด๐ต)๐‘ข
But this also means ๐‘ฅ = ๐ด(๐ต๐‘ข) ⇒ ๐‘ฅ ∈ โ„›๐ด .
i.e. โ„›๐ด๐ต ⊆ โ„›๐ด .
Suppose ๐‘ฅ ∈ โ„›๐ด ⇒ ๐‘ฅ = ๐ด๐‘ฃ = ๐ด(๐ต๐ต−1 )๐‘ฅ = (๐ด๐ต)(๐ต−1 ๐‘ฃ) ⇒ ๐‘ฅ ∈ โ„›๐ด๐ต i.e. โ„›๐ด ⊆ โ„›๐ด๐ต
โ„›๐ด ⊆ โ„›๐ด๐ต ⊆ โ„›๐ด ⇒ โ„›๐ด = โ„›๐ด๐ต
Can we also show ๐’ฉ๐ด๐ต = ๐’ฉ๐ด ?
No. Or at least not always…
0 1
Let ๐ด = [
]
0 0
๐‘ฅ1
๐‘ฅ
0
0
Let ๐‘ฅ ∈ ๐’ฉ๐ด , ๐ด [๐‘ฅ ] = [ 2 ] = [ ] ⇒ ๐’ฉ๐ด = {[ ] |๐›ฝ ∈ ๐”ฝ}
๐›ฝ
0
0
2
0 1
1 0
2
Let ๐ต = [
], ๐ต = [
]. B is invertible.
1 0
0 1
๐›ผ
๐’ฉ๐ด๐ต = {[ ] |๐›ผ ∈ ๐”ฝ}
0
What we can show, is the dimensions of these spaces are equal.
Let {๐‘ข1 , … , ๐‘ข๐‘˜ } be a basis for ๐’ฉ๐ด
{๐ต−1 ๐‘ข1 , … , ๐ต−1 _๐‘ข๐‘˜ } ∈ ๐’ฉ๐ด๐ต
Easy to see: ๐ด๐ต(๐ต−1 ๐‘ข๐‘– ) = ๐ด๐‘ข๐‘– = 0.
Claim: they are linearly independent.
Proof:
๐›ผ1 ๐ต−1 ๐‘ข1 + โ‹ฏ + ๐›ผ๐‘˜ ๐ต−1 ๐‘ข๐‘˜ = 0
๐ต−1 (๐›ผ1 ๐‘ข1 + โ‹ฏ + ๐›ผ๐‘˜ ๐‘ข๐‘˜ ) = 0
๐›ผ1 ๐‘ข1 , … , ๐›ผ๐‘˜ ๐‘ข๐‘˜ = ๐ต0 = 0
But ๐‘ข1 , … , ๐‘ข๐‘˜ is a basis (therefore independent) so ๐›ผ1 = ๐›ผ2 = โ‹ฏ = ๐›ผ๐‘˜ = 0.
dim ๐’ฉ๐ด = ๐‘˜
dim ๐’ฉ๐ด๐ต ≥ ๐‘˜
So we’ve shown that: dim ๐’ฉ๐ด๐ต ≥ dim ๐’ฉ๐ด
Now take basis – {๐‘ฃ1 , … , ๐‘ฃ๐‘™ } of ๐’ฉ๐ด๐ต ⇒ ๐ด๐ต๐‘ฃ๐‘– = 0 ⇒ ๐ต๐‘ฃ๐‘– ∈ ๐’ฉ๐ด
Check {๐ต๐‘ฃ1 , … , ๐ต๐‘ฃ๐‘™ } are linearly independent. Same as before…
⇒ dim ๐’ฉ๐ด ≥ ๐‘™ = dim ๐’ฉ๐ด๐ต
So now we have two inequalities resulting I dim ๐’ฉ๐ด = dim ๐’ฉ๐ด๐ต
Definition: If ๐ด ∈ โ„๐‘×๐‘ž ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด = dim โ„›๐ด
โ„›๐ด = ๐‘ ๐‘๐‘Ž๐‘› ๐‘œ๐‘“ ๐‘กโ„Ž๐‘’ ๐‘๐‘œ๐‘™๐‘ข๐‘š๐‘›๐‘  ๐‘œ๐‘“ ๐ด
๐ด = [๐‘Ž1 ๐‘Ž2 ๐‘Ž3 ๐‘Ž4 ]
๐‘ฅ1
๐‘ฅ2
๐ด [๐‘ฅ ] = ๐‘ฅ1 ๐‘Ž1 + ๐‘ฅ2 ๐‘Ž2 + ๐‘ฅ3 ๐‘Ž3 + ๐‘ฅ4 ๐‘Ž4
3
๐‘ฅ4
โ–ญ 0 0 0
Say ๐‘ˆ = [ 0 0 โ–ญ 0] = [๐‘ข1 ๐‘ข2 ๐‘ข3 ๐‘ข4 ]
0 0 0 0
โ„›๐‘ˆ = {๐‘ข1 , ๐‘ข3 }
๐‘Ÿ๐‘Ž๐‘›๐‘˜๐‘ˆ = number of pivots in ๐‘ˆ.
Theorem: Let ๐ด ∈ ๐”ฝ๐‘×๐‘ž then
(1) ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด ≤ min(๐‘, ๐‘ž)
(2) ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด = ๐‘ž ⇔ ๐ด is left invertible ⇔ ๐’ฉ๐ด = {0}
(3) ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด = ๐‘ ⇔ ๐ด is right ivertible ⇔ โ„›๐ด = ๐”ฝ๐‘
Proof:
If ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด = 0 (1) is clearly true.
If ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด ≠ 0 ⇒ ∃๐บ ๐‘×๐‘ invertible such that ๐บ๐ด = ๐‘ˆ upper echelon.
This implies that ๐‘…๐‘Ž๐‘›๐‘˜๐บ๐ด = ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐‘ˆ = number of pivots in ๐‘ˆ.
Suppose ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด = ๐‘ž ⇒ ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐บ๐ด = ๐‘ž ⇒ ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐‘ˆ = ๐‘ž ⇒ ๐‘ˆ is left invertible ⇒ there is a
๐‘‰ ๐‘ . ๐‘ก. ๐‘‰๐‘ˆ = ๐ผ๐‘ž
Same as to say ๐‘‰(๐บ๐ด) = ๐ผ๐‘ž ⇒ (๐‘‰๐บ)๐ด = ๐ผ๐‘ž ⇒ ๐ด is left invertible.
(3)
------End of lesson 4
๐‘‡ linear transformation from a finite dimension vector spce ๐’ฐ over ๐”ฝ into a vector space ๐’ฑ over
๐”ฝ, then
dim ๐’ฐ = dim ๐’ฉ๐‘‡ + dim โ„› ๐‘‡
๐‘×๐‘ž
๐‘ž |๐ด๐‘ฅ
{๐‘ฅ
If ๐ด ∈ ๐”ฝ , ๐’ฉ๐ด =
∈๐”ฝ
= 0} (also a subspace of ๐”ฝ๐‘ž )
๐‘ž = dim ๐’ฉ๐ด + dim โ„›๐ด
Theorem: If ๐ด ∈ ๐”ฝ๐‘×๐‘ž , then:
(1) rank๐ด ≤ min{๐‘, ๐‘ž}
(2) rank๐ด = ๐‘ž ⇔ ๐ด is left invertible⇔ ๐’ฉ๐ด = {0}
(3) rank๐ด = ๐‘ ⇔ ๐ด is right invertible⇔ โ„›๐ด = ๐”ฝ๐‘ =”everything”
Exploited fact: if ๐‘ˆ upper echelon with ๐‘˜ pivots, then:
(1) ๐‘˜ ≤ min{๐‘š, ๐‘›}
(2) ๐‘˜ = ๐‘ž ⇔ ๐‘ˆ is left invertible⇔ ๐’ฉ๐‘ˆ = {0}
(3) ๐‘˜ = ๐‘ ⇔ ๐‘ˆ is right invertible⇔ โ„›๐‘ˆ = ๐”ฝ๐‘
Gaussian elimination corresponds to finding and invertible ๐บ ∈ ๐”ฝ๐‘×๐‘ such that
๐บ๐ด =
๐‘ˆ
โŸ
๐‘ข๐‘๐‘๐‘’๐‘Ÿ ๐‘’๐‘โ„Ž๐‘’๐‘™๐‘œ๐‘›
๐ต1 , ๐ต2 , ๐ต3 invertible ⇒ ๐ต1 ๐ต2 ๐ต3 is invertible.
Rank๐ด=dim โ„›๐ด
If ๐‘ˆ upper echelon with ๐‘˜ pivots ⇔rank๐‘ˆ = ๐‘˜
Implications
1) System of equaeions:
๐‘Ž11 …
๐ด๐‘ฅ = ๐‘ ๐ด = [ โ‹ฎ
๐‘Ž๐‘1 …
๐‘Ž1๐‘ž
๐‘1
โ‹ฎ ] ๐‘=[โ‹ฎ ]
๐‘๐‘
๐‘Ž๐‘๐‘ž
๐‘ฅ1
Looking or vectors ๐‘ฅ = [ โ‹ฎ ] ๐‘ . ๐‘ก. ๐ด๐‘ฅ = ๐‘ (if any).
๐‘ฅ๐‘ž
(a) As ๐ด left invertible, guarantees at most one solution.
To chek: Suppose ๐ด๐‘ฅ = ๐‘, ๐ด๐‘ฆ = ๐‘ ⇒
๐ด๐‘ฅ − ๐ด๐‘ฆ = ๐‘ − ๐‘ = 0
๐ด(๐‘ฅ − ๐‘ฆ) = 0
If ๐ต is a left inverse of ๐ด
๐‘ฅ − ๐‘ฆ = ๐ผ(๐‘ฅ − ๐‘ฆ) = ๐ต๐ด(๐‘ฅ − ๐‘ฆ) = ๐ต(๐ด(๐‘ฅ − ๐‘ฆ)) = ๐ต0 = 0
(b) ๐ด right invertible guarantees at least one solution.
Let ๐ถ be a right inverse of ๐ด and choose ๐‘ฅ = ๐ถ๐‘
Then ๐ด(๐ถ๐‘) = (๐ด๐ถ)๐‘ = ๐ผ๐‘ ๐‘ = ๐‘
2) If ๐ด ∈ ๐”ฝ๐‘×๐‘ž and ๐ด is both right invertible and left invertible then ๐‘ = ๐‘ž.
Earlier we showed that ๐ต, ๐ถ ∈ ๐”ฝ๐‘×๐‘ž ๐‘ . ๐‘ก.
๐ต๐ด = ๐ผ๐‘ž and ๐ด๐ถ = ๐ผ๐‘ then ๐ต = ๐ถ.
Rank๐ด = ๐‘˜.
๐ด left invertible ⇒ ๐‘˜ = ๐‘ž
๐ด right invertible ⇒ ๐‘˜ = ๐‘
So ๐‘ = ๐‘ž.
3) ๐ด ∈ ๐”ฝ๐‘×๐‘ž , ๐บ ∈ ๐”ฝ๐‘×๐‘ invertible, ๐บ๐ด = ๐‘ˆ is upper echelon.
And lwr ๐‘˜ = number of pivots in ๐‘ˆ.
⇒rank๐ด =rank๐‘ˆ = ๐‘˜.
Claim: The pivot columns of ๐‘ˆ are linearly independent and form a basis to โ„›๐‘ˆ
The corresponding columns in ๐ด are linearly independent and form a basis for โ„›๐ด
1 ๐‘ข12 ๐‘ข13 ๐‘ข14
๐‘ˆ = [0 0
2 ๐‘ข24 ] = [๐‘ข1 ๐‘ข2 ๐‘ข3 ๐‘ข4 ]
0 0
0
0
๐‘ข1 , ๐‘ข3 are lin independent, and span{๐‘ข1 , ๐‘ข3 } = โ„›๐‘ˆ
If ๐บ๐ด = ๐‘ˆ, ๐ด = [๐‘Ž1 ๐‘Ž2 ๐‘Ž3 ๐‘Ž4 ]
claim: ๐‘Ž1 , ๐‘Ž3 are linearly independent and span{๐‘Ž1 , ๐‘Ž3 } = โ„›๐ด
Suppose we can find coefficients such that ๐›ผ๐‘Ž1 + ๐›ฝ๐‘Ž3 = 0
๐ด = ๐บ −1 ๐‘ˆ
[๐‘Ž1 ๐‘Ž2 ๐‘Ž3 ๐‘Ž4 ] = [๐บ −1 ๐‘ข1 ๐บ −1 ๐‘ข2 ๐บ −1 ๐‘ข3 ๐บ −1 ๐‘ข4 ]
๐›ผ๐‘Ž1 + ๐›ฝ๐‘Ž3 = ๐›ผ๐บ −1 ๐‘ข1 + ๐›ฝ๐บ −1 ๐‘ข3 = 0 = ๐บ −1 (๐›ผ๐‘ข1 + ๐›ฝ๐‘ข3 ) = 0
๐›ผ๐‘ข1 + ๐›ฝ๐‘ข3 = ๐บ0 = 0 ⇒ ๐›ผ = ๐›ฝ = 0 since ๐‘ข1 and ๐‘ข3 are linearly independent.
4) Related application
Given ๐‘ฅ1 , … , ๐‘ฅ๐‘› ∈ ๐”ฝ๐‘
Find a basis for span{๐‘ฅ1 , … , ๐‘ฅ๐‘› }
Define ๐ด = [๐‘ฅ1 ๐‘ฅ2 … ๐‘ฅ๐‘› ] ๐‘ × ๐‘›
Bring to upper echelon form via Gaussian elimination.
The number of pivots in ๐‘ˆ = dim ๐‘ ๐‘๐‘Ž๐‘›{๐‘ฅ1 , … , ๐‘ฅ๐‘› }
and the corresponding columns will be a basis
5) Calculating inverses
Let ๐ด be 3 × 3 and it is known to be invertible.
How to calculate its inverse?
1
0
0
๐ด๐‘ฅ1 = [0] , ๐ด๐‘ฅ2 = [1] , ๐ด๐‘ฅ3 = [0]
0
0
1
1 0 0
๐ด[๐‘ฅ1 ๐‘ฅ2 ๐‘ฅ3 ] = [๐ด๐‘ฅ1 ๐ด๐‘ฅ2 ๐ด๐‘ฅ3 ] = [0 1 0]
0 0 1
Gauss-Seidel
Do all these calculations in one shot
๐ดฬƒ = [๐ด
๐ผ3 ]
๐บ๐ดฬƒ = [๐บ๐ด
๐บ ] ๐บ๐ด = ๐‘ˆ upper echelon
2 4
Suppose ๐‘ˆ = [0 3
0 0
1
2
6
6]
4
0 0
1 2 3
0 , ๐ท๐‘ˆ = [0 1 2]
0 0 1
1
[0 0 4 ]
๐‘Ž
1 0 −3 ๐‘Ž
๐‘Ž − 3๐‘
๐น1 [๐‘] = [0 1 −2] [๐‘ ] = [๐‘ − 2๐‘ ]
๐‘
0 0 1 ๐‘
๐‘
3
3−3
0
[
] [2] = [2 − 2] = [0]
1
1
1
D= 0
1
3
1 0 −3 1 2
[0 1 −2] [0 1
โŸ0 0 1 0 0
3
1
2] = [0
1
0
2 0
1 0]
0 1
๐น1 ๐ท๐‘ˆ
๐ดฬƒ = [๐ด ๐ผ3 ]
(1) Manipulate by permutations and subtracting multiplies of rows from lower rows.
(2) Multiplying through by a diagonal matrix [ ๐ท๐บ๐ด
โŸ
๐ท๐บ]
๐‘ข๐‘๐‘๐‘’๐‘Ÿ ๐‘ก๐‘Ÿ
๐น๐ท๐บ๐ด
โŸ
๐น๐ท๐บ
(3) Subtract multiplies of rows from higher rows [ ๐ผ
] inverse of ๐ด is ๐น๐ท๐บ.
๐‘
๐ด ∈ ๐”ฝ๐‘×๐‘ž , ๐ด๐‘‡
๐ด=[
๐‘–, ๐‘— entry of ๐ด is ๐‘—๐‘– entry of ๐ด๐‘‡
1 2
4 5
3
],
6
1
๐ด๐‘‡ = [2
3
Claim: rank๐ด =rank๐ด๐‘‡
๐บ๐ด = ๐‘ˆ upper echelon
rank๐ด =rank๐‘ˆ
(๐บ๐ด)๐‘‡ = ๐ด๐‘‡ ๐บ ๐‘‡ = ๐‘ˆ ๐‘‡
Rank ๐ด๐‘‡ =rank๐‘ˆ ๐‘‡
∗ โˆ™ โˆ™ โˆ™
๐‘ˆ = [0 0 ∗ โˆ™ ]
0 0 0 0
dim โ„›๐‘ˆ = 2
4
5]
6
∗ 0
โˆ™ 0
๐‘ˆ๐‘‡ = [
โˆ™ ∗
โˆ™ โˆ™
0
0
] , ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐‘ˆ ๐‘‡ =number of pivots in ๐‘ˆ=rank๐‘ˆ.
0
0
If ๐‘‡ is a linear transformation from a vector space ๐’ฐ over ๐”ฝ into itself (๐‘ฅ ∈ ๐’ฐ ⇒ ๐‘‡๐‘ฅ ∈ ๐’ฐ)
A subspace ๐’ณ of ๐’ฐ is said to be invariant under ๐‘‡ if for every ๐‘ฅ ∈ ๐’ณ, ๐‘‡๐‘ฅ ∈ ๐’ณ.
The simplest non-zero invariant subspace (if they exist) would be one dimensional spaces.
Suppose ๐’ณ is one dimensional and is invariant under ๐‘‡, then if you take any vector ๐‘ฅ ∈ ๐’ณ, ๐‘‡๐‘ฅ ∈
๐’ณ
๐’ณ = {๐›ผ๐‘ฆ|๐›ผ ∈ ๐ดฬƒ}
๐‘‡๐‘ฅ = โŸ
๐œ†๐‘ฅ
∈๐”ฝ
In this case, ๐œ† is said to be thean Eigen value of ๐‘‡ and ๐‘ฅ is said to be an Eigen vector.
Important – ๐‘ฅ ≠ 0!!!
In other words, a vector ๐‘ฅ is called an Eigenvector of ๐‘‡ if:
(1) ๐‘ฅ ≠ 0
(2) ๐‘‡๐‘ฅ = ๐œ†๐‘ฅ some ๐œ† ∈ โ„‚
That ๐œ† is called an Eigenvalue of ๐‘‡.
If ๐‘‡๐‘ฅ = ๐œ†๐‘ฅ, then ๐‘‡2๐‘ฅ = ๐œ†2๐‘ฅ.
So there’s a “flexibility” of stretching.
There isn’t the Eigenvector.
Theorem: Let ๐’ฑ be a vector space over โ„‚, ๐‘‡ a linear transformation from ๐’ฑ into ๐’ฑ and let ๐’ฐ be a
non-zero finite dimension subspace of ๐’ฑ that is invariant under ๐‘‡. Then, there exists a vector
๐‘ข ∈ ๐’ฐ and a ๐œ† ∈ โ„‚ ๐‘ . ๐‘ก. ๐‘‡๐‘ข = ๐œ†๐‘ข.
Proof:
Take any ๐‘ค ∈ ๐’ฐ, ๐‘ค ≠ 0. Suppose dim ๐’ฐ = ๐‘›
Consider ๐‘ค, ๐‘‡๐‘ค, … , ๐‘‡ ๐‘› ๐‘ค. That’s a list of ๐‘› + 1 vectors. In an ๐‘› dimensional space. Therefore,
they are linearly dependent.
Can find ๐‘0 , … , ๐‘๐‘› not all of which are zero such that ๐‘0 ๐‘ค + ๐‘1 ๐‘‡๐‘ค + โ‹ฏ + ๐‘๐‘› ๐‘‡ ๐‘› ๐‘ค = 0
Suppose that ๐‘˜ is the largest index that is non-zero.
So this expression reduces to ๐‘0 ๐‘ค + ๐‘1 ๐‘‡๐‘ค + โ‹ฏ + ๐‘๐‘˜ ๐‘‡ ๐‘˜ ๐‘ค = 0
๐‘0 ๐ผ + ๐‘1 ๐‘‡ + โ‹ฏ + ๐‘๐‘˜ ๐‘‡ ๐‘˜ = 0
Consider the following polynomial:
๐‘0 + ๐‘1 ๐‘ฅ + โ‹ฏ + ๐‘๐‘˜ ๐‘ฅ ๐‘˜
๐‘๐‘Ž๐‘› ๐‘๐‘’ ๐‘“๐‘Ž๐‘๐‘ก๐‘œ๐‘Ÿ๐‘’๐‘‘
=
(๐‘ฅ − ๐œ‡1 )(๐‘ฅ − ๐œ‡2 ) … (๐‘ฅ − ๐œ‡๐‘˜ ) =
๐‘๐‘˜ (๐‘‡ − ๐œ‡๐‘˜ ๐ผ)(๐‘‡ − ๐œ‡2 ๐ผ) … (๐‘‡ − ๐œ‡1 ๐ผ)๐‘ค = 0
Possibilities Either:
(1) (๐‘‡ − ๐œ‡1 ๐ผ)๐‘ค = 0
(2) (๐‘‡ − ๐œ‡1 )๐‘ค ≠ 0 but (๐‘‡ − ๐œ‡2 ๐ผ){(๐‘‡ − ๐œ‡1 ๐ผ)๐‘ค} = 0
(3) (๐‘‡ − ๐œ‡2 ๐ผ)(๐ผ − ๐œ‡3 ๐ผ) ≠ 0 but (๐‘‡ − ๐œ‡3 ๐ผ){(๐‘‡ − ๐œ‡2 ๐ผ)(๐‘‡ − ๐œ‡1 ๐ผ)๐‘ค} = 0
(4) โ‹ฎ
(5) (๐‘‡ − ๐œ‡๐‘˜ ๐ผ) … (๐‘‡ − ๐œ‡1 ๐ผ)๐‘ค ≠ 0 but (๐‘‡ − ๐œ‡๐‘˜ ๐ผ){(๐‘‡ − ๐œ‡๐‘˜ ๐ผ) … (๐‘‡ − ๐œ‡1 ๐ผ)๐‘ค} = 0
----- End of lesson 5
Previously – on linear algebra
Thorem: Let ๐’ฑ be a vector space over โ„‚, T a linear transformation from ๐’ฑ into ๐’ฑ, and let ๐’ฐ be a
finite dimensional subspace in ๐’ฑ that is invariant of ๐‘‡. (i.e. ๐‘ข ∈ ๐‘ˆ, then ๐‘‡๐‘ข ∈ ๐‘ˆ). Then there
exists a non-zero vector ๐‘ค ∈ ๐’ฐ and a ๐œ† ∈ โ„‚ ๐‘ . ๐‘ก. ๐‘‡๐‘ค = ๐œ†๐‘ค.
We could have just worked just with ๐’ฐ.
Note: ๐‘‡๐‘ค = ๐œ†๐‘ค ⇔ (๐‘‡ − ๐œ†๐ผ)๐‘ค = 0 ⇔ ๐’ฉ๐‘‡−๐œ†๐ผ ≠ {0}
Could have rephrased the conclusion as:
There is a finite dimension subspace ๐’ฐ of ๐’ฑ that is invariant under ๐‘‡ ⇔There is a one dimension
subspace ๐’ณ of ๐’ฐ that is invariant under ๐‘‡.
Implication
Let ๐ด ∈ โ„‚๐‘›×๐‘› , Then there exists a point ๐œ† ∈ โ„‚ ๐‘ . ๐‘ก. ๐’ฉ๐ด−๐œ†๐ผ๐‘› ≠ {0}
Because the transformation ๐‘‡ from โ„‚๐‘› into โ„‚๐‘› that is defined by the formula ๐‘‡๐‘ข = ๐ด๐‘ข
Proof: Take any vecto ๐‘ค ∈ โ„‚๐‘› , ๐‘ค ≠ 0.
Consider the set of vectors {๐‘ค, ๐ด๐‘ค, … , ๐ด๐‘› ๐‘ค}. These are ๐‘› + 1 vectors in โ„‚ - an ๐‘› dimensional
space.
Since they are independent, there are coefficients:
๐‘0 ๐‘ค + ๐‘1 ๐ด๐‘ค + โ‹ฏ + ๐‘๐‘› ๐ด๐‘› ๐‘ค = 0 such that not all coefficients are zero.
Same as to say (๐‘0 ๐ผ๐‘› + ๐ถ1 ๐ด + โ‹ฏ + ๐‘๐‘› ๐ด๐‘› )๐‘ค = 0
Let ๐‘˜ = max{๐‘—|๐‘๐‘— ≠ 0} Claim ๐‘˜ ≥ 1 (trivial)
(๐‘0 ๐ผ๐‘› + ๐‘1 ๐ด + โ‹ฏ + ๐‘๐‘˜ ๐ด๐‘˜ )๐‘ค = 0, ๐‘๐‘˜ ≠ 0
Claim that if we look at the ordinary polynomial:
๐‘0 + ๐‘1 ๐‘ฅ + ๐‘2 ๐‘ฅ 2 + โ‹ฏ + ๐‘๐‘˜ ๐‘ฅ ๐‘˜ = ๐‘๐‘˜ (๐‘ฅ − ๐œ‡1 )(๐‘ฅ − ๐œ‡2 ) … (๐‘ฅ − ๐œ‡๐‘˜ ), ๐‘๐‘˜ ≠ 0
⇒ ๐‘0 ๐ผ๐‘› + ๐‘1 ๐ด + โ‹ฏ ๐‘๐‘˜ ๐ด๐‘˜ = ๐‘๐‘˜ (๐ด − ๐œ‡1 ๐ผ๐‘› )(๐ด − ๐œ‡2 ๐ผ๐‘› ) … (๐ด − ๐œ‡๐‘˜ ๐ผ๐‘› )
๐‘๐‘˜ (๐ด − ๐œ‡๐‘˜ ๐ผ๐‘› ) … (๐ด − ๐œ‡1 ๐ผ๐‘› )๐‘ค = 0
๐‘Ÿ๐‘’๐‘ฃ๐‘’๐‘Ÿ๐‘ ๐‘’ ๐‘กโ„Ž๐‘’ ๐‘œ๐‘‘๐‘’๐‘Ÿ
=
Suppose ๐‘˜ = 3
Either:
(1) (๐ด − ๐œ‡1 ๐ผ๐‘› )๐‘ค = 0
(2) (๐ด − ๐œ‡1 ๐ผ๐‘› )๐‘ค ≠ 0 and (๐ด − ๐œ‡2 ๐ผ๐‘› ){(๐ด − ๐œ‡1 ๐ผ๐‘› )๐‘ค} = 0
(3) (๐ด − ๐œ‡2 ๐ผ๐‘› )(๐ด − ๐œ‡1 ๐ผ๐‘› )๐‘ค ≠ 0 and (๐ด − ๐œ‡3 ๐ผ๐‘› )(๐ด − ๐œ‡2 ๐ผ๐‘› )(๐ด − ๐œ‡1 ๐ผ๐‘› )๐‘ค = 0
We are looking fo ๐œ† ๐‘ . ๐‘ก. (๐ด − ๐œ†๐ผ๐‘› )๐‘ฅ = 0, ๐‘ฅ ≠ 0
Questions: Suppose ๐ด ∈ โ„๐‘›×๐‘›
Can one guarantee that ๐ด has at least one real Eigen value?
No!
0 1
]
−1 0
๐‘Ž
0 1 ๐‘Ž
Looking for [
] [ ] = ๐œ† [ ] ⇔ ๐‘ = ๐œ†๐‘Ž, −๐‘Ž = ๐œ†๐‘ ⇔ ๐‘ = −๐œ†2 ๐‘ ⇔ (1 + ๐œ†2 )๐‘ = 0
๐‘
−1 0 ๐‘
If ๐‘ = 0 ⇒ ๐‘Ž = 0 ⇒The entire vector is zero! Not acceptable…
So ๐œ†2 + 1 = 0 i.e. ๐œ† = ±๐’พ
๐ด=[
Result for ๐ด ∈ โ„‚๐‘›×๐‘› says that there is always a one dimensional subspace of โ„‚๐‘› that is invariant
under A.
๐ด ∈ โ„๐‘›×๐‘› - there always exists at least one two dimensional subspace of โ„๐‘› that is invariant
under ๐ด.
Implication:
Let ๐ด ∈ โ„‚๐‘›×๐‘› - then there exists a point ๐œ† ∈ โ„‚ such that ๐’ฉ๐ด−๐œ†๐ผ๐‘› ≠ {0}
Suppose ๐’ฉ๐ด−๐œ†๐ผ๐‘› ≠ {0} for some ๐‘˜ distinct points in โ„‚ ๐œ†1 , … , ๐œ†๐‘˜ .
That is to say, there are non-zero vectors ๐‘ข1 , … , ๐‘ข๐‘˜ of ๐ด๐‘ข๐‘— = ๐œ†๐‘— ๐‘ข๐‘— 1 ≤ ๐‘— ≤ ๐‘˜
Claim ๐‘ข1 , … , ๐‘ข๐‘˜ are linearly independent.
Let’s check for ๐‘˜ = 3
Suppose ๐‘1 ๐‘ข1 + ๐‘2 ๐‘ข2 + ๐‘3 ๐‘ข3 = 0
(๐ด − ๐œ†1 ๐ผ๐‘› ) (๐‘1 ๐‘ข1 + ๐‘2 ๐‘ข2 + ๐‘3 ๐‘ข3 ) = (๐ด − ๐œ†๐ผ๐‘› )0 = 0
Let’s break it up:
(๐ด − ๐›ผ๐ผ๐‘› )๐‘ข๐‘— = ๐ด๐‘ข๐‘— − ๐›ผ๐‘ข๐‘— = ๐œ†๐‘— ๐‘ข๐‘— − ๐›ผ๐‘ข๐‘— = (๐œ†๐‘— − ๐›ผ)๐‘ข๐‘—
Continue with our calculation:
๐‘1 (๐œ†1 − ๐œ†1 )๐‘ข1 + ๐‘2 (๐œ†2 − ๐œ†1 )๐‘ข2 + ๐‘3 (๐œ†3 − ๐œ†1 )๐‘ข3 = ๐‘2 (๐œ†2 − ๐œ†1 )๐‘ข2 + ๐‘3 (๐œ†3 − ๐œ†1 )๐‘ข3
We’ve reduced the problem!
(we know the lambda’s are different so the rest of the coefficients are not zero)
We can do it again to get
(๐ด − ๐œ†2 ๐ผ๐‘› )(๐‘2 (๐œ†2 − ๐œ†1 )๐‘ข2 + ๐‘3 (๐œ†3 − ๐œ†1 )๐‘ข3 ) =
๐‘2 (๐œ†2 − ๐œ†1 )(๐œ†2 − ๐œ†2 )๐‘ข2 + ๐‘3 (๐œ†3 − ๐œ†1 )(๐œ†3 − ๐œ†2 )๐‘ข3 = ๐‘3 (๐œ†3 − ๐œ†1 )(๐œ†3 − ๐œ†2 )๐‘ข3
Now ๐‘3 must be zero since the other scalars and vectors in the expression are non-zero.
We can in fact repeat the process to knock out ๐‘2 and also ๐‘1 .
This can also be generalized for ๐‘› other than 3.
๐ด๐‘ข๐‘— = ๐œ†๐‘— ๐‘ข๐‘— ๐‘— = 1, … , ๐‘˜ ๐œ†๐‘– ≠ ๐œ†๐‘— if ๐‘– ≠ ๐‘—
๐œ†1
๐ด[๐‘ข1 … ๐‘ข๐‘˜ ] = [๐ด๐‘ข1 … ๐ด๐‘ข๐‘˜ ] = [๐œ†1 ๐‘ข1 … ๐œ†๐‘˜ ๐‘ข๐‘˜ ] = [๐‘ข1 … ๐‘ข๐‘˜ ] [
๐œ†2
]
โ‹ฑ
๐œ†๐‘˜
๐ด ∈ โ„๐‘›×๐‘› , then there is at least one dimensional subspace of โ„‚๐‘› that is invariant under ๐ด.
Translate this to: There is ๐‘ข ∈ โ„‚๐‘› , ๐œ† ∈ โ„‚ ๐‘ . ๐‘ก. ๐ด๐‘ข = ๐œ†๐‘ข
๐‘ข = ๐‘ฅ + ๐’พ๐‘ฆ, ๐œ† = ๐›ผ + ๐’พ๐›ฝ
๐ด(๐‘ฅ + ๐’พ๐‘ฆ) = (๐›ผ + ๐’พ๐›ฝ)(๐‘ฅ + ๐’พ๐‘ฆ) = (๐›ผ๐‘ฅ − ๐›ฝ๐‘ฆ) + ๐’พ(๐›ฝ๐‘ฅ + ๐›ผ๐‘ฆ)
๐ด ∈ ๐‘… ๐‘›×๐‘›
๐ด๐‘ฅ = ๐›ผ๐‘ฅ − ๐›ฝ๐‘ฆ ๐ด๐‘ฆ = ๐›ฝ๐‘ฅ + ๐›ผ๐‘ฆ
๐’ฒ = ๐‘ ๐‘๐‘Ž๐‘›{๐‘ฅ, ๐‘ฆ} with real coefficients
๐‘ค ∈ ๐’ฒ ⇒ ๐ด๐‘ค ∈ ๐’ฒ
๐‘ค = ๐‘Ž๐‘ฅ + ๐‘๐‘ฆ
๐ด๐‘ค = ๐‘Ž๐ด๐‘ฅ + ๐‘๐ด๐‘ฆ = ๐‘Ž(๐›ผ๐‘ฅ + ๐›ฝ๐‘ฆ) + ๐‘(๐›ฝ๐‘ฅ + ๐›ผ๐‘ฆ) = (๐‘Ž๐›ผ + ๐‘๐›ฝ)๐‘ฅ + (−๐‘Ž๐›ฝ + ๐‘๐›ผ)๐‘ฆ
Example:
If ๐ด ∈ โ„‚๐‘›×๐‘› with ๐‘˜ distinct Eigenvalues, then ๐‘˜ ≤ ๐‘›.
Because, ๐‘ข1 , … , ๐‘ข๐‘˜ Eigenvectors are linearly independent. And they sit inside an ๐‘› dimensional
space โ„‚๐‘› .
The matrix of the Eigenvectors:
[๐‘ข1 … ๐‘ข๐‘˜ ] is an ๐‘› × ๐‘˜ matrix with rank ๐‘˜.
๐œ†1
If ๐‘˜ = ๐‘›, then ๐ด[๐‘ข1 … ๐‘ข๐‘› ] = [๐‘ข1 … ๐‘ข๐‘› ] [
๐œ†2
]
โ‹ฑ
๐œ†๐‘˜
๐ด๐‘ˆ = ๐‘ˆ๐ท
๐‘ˆ is invertible. (Rank k)
So we can rewrite this as: ๐ด = ๐‘ˆ๐ท๐‘ˆ −1
If you need to raise some matrix ๐ด to the power of 100
Suppose: ๐ด = ๐‘ˆ๐ท๐‘ˆ −1 , ๐ด2 = ๐‘ˆ๐ท๐‘ˆ −1 ๐‘ˆ๐ท๐‘ˆ −1 = ๐‘ˆ๐ท 2 ๐‘ˆ −1
So ๐ด100 = ๐‘ˆ๐ท100 ๐‘ˆ −1
๐›ผ 100
๐›ฝ100
๐ท100 = [
โ‹ฑ
]
๐›พ100
Sums of Subspaces
Let ๐’ฐ, ๐’ฑ be subspaces of a vector space ๐’ฒ over ๐”ฝ.
๐’ฐ + ๐’ฑ = {๐‘ข + ๐‘ฃ|๐‘ข ∈ ๐’ฐ, ๐‘ฃ ∈ ๐’ฑ}
Claim: ๐’ฐ + ๐’ฑ is a vector space.
(๐‘ข1 + ๐‘ฃ1 ) + (๐‘ข2 + ๐‘ฃ2 ) = โŸ
(๐‘ข1 + ๐‘ข2 ) + โŸ
(๐‘ฃ1 + ๐‘ฃ2 )
∈๐’ฐ
∈๐’ฑ
Lemma:
Let ๐’ฐ, ๐’ฑ subspaces of ๐’ฒ a finite dimensional vector space
dim(๐’ฐ + ๐’ฑ) = dim ๐’ฐ + dim ๐’ฑ − dim(๐’ฐ ∩ ๐’ฑ)
Another claim: \๐’ฐ ∩ ๐’ฑ is a vector space.
Proof: Suppose {๐‘ค1 , … , ๐‘ค๐‘˜ } is a basis for ๐’ฐ ∩ ๐’ฑ
๐’ฐ∩๐’ฑ ⊆๐’ฐ
if really ๐’ฐ ∩ ๐’ฑ ≠ ๐’ฐ
{๐‘ค1 , … , ๐‘ค๐‘˜ , ๐‘ข1 , … , ๐‘ข๐‘  } basis for ๐’ฐ
๐’ฐ ∩ ๐’ฑ ⊆ ๐’ฑ. Suppose ๐’ฐ ∩ ๐’ฑ ≠ ๐’ฑ
{๐‘ค1 , … , ๐‘ค๐‘˜ , ๐‘ฃ1 , … , ๐‘ฃ๐‘ก } is a basis for ๐’ฑ
Claim: {๐‘ค1 , … , ๐‘ค๐‘˜ , ๐‘ข1 , … , ๐‘ข๐‘  , ๐‘ฃ1 , … , ๐‘ฃ๐‘ก } basis for ๐’ฐ + ๐’ฑ
Need to show:
(a) ๐‘ข + ๐‘ฃ is a linear combination of these
๐‘ข = ๐‘1 ๐‘ค1 + โ‹ฏ + ๐‘๐‘˜ ๐‘ค๐‘˜ + ๐‘‘1 ๐‘ข1 + โ‹ฏ + ๐‘‘๐‘  ๐‘ข๐‘ 
๐‘ฃ = ๐‘Ž1 ๐‘ค1 + โ‹ฏ + ๐‘Ž๐‘˜ ๐‘ค๐‘˜ + ๐‘1 ๐‘ฃ1 + โ‹ฏ + ๐‘๐‘ก ๐‘ฃ๐‘ก
๐‘ข+๐‘ฃ =
(๐‘Ž1 + ๐‘1 )๐‘ค1 + โ‹ฏ + (๐‘Ž๐‘˜ + ๐‘๐‘˜ )๐‘ค๐‘˜ + โ‹ฏ ๐‘‘1 ๐‘ข1 + โ‹ฏ + ๐‘‘๐‘  ๐‘ข๐‘  + โ‹ฏ + ๐‘1 ๐‘ฃ1 + โ‹ฏ + ๐‘๐‘ก ๐‘ฃ๐‘ก
(b) Show {๐‘ค1 … ๐‘ค๐‘˜ , ๐‘ข1 , … , ๐‘ข๐‘  , ๐‘ฃ1 , … , ๐‘ฃ๐‘ก } are linearly independent.
Claim: b is correct.
Denote ๐‘ฃ = ๐‘1 ๐‘ฃ1 + โ‹ฏ + ๐‘๐‘ก ๐‘ฃ๐‘ก
Suppose:
๐‘Ž1 ๐‘ค1 + โ‹ฏ + ๐‘Ž๐‘˜ ๐‘ค๐‘˜ + ๐‘1 ๐‘ข1 + โ‹ฏ + ๐‘๐‘  ๐‘ข๐‘  + ๐‘1 ๐‘ฃ + โ‹ฏ + ๐‘๐‘ก ๐‘ฃ๐‘ก = 0
๐‘Ž1 ๐‘ค1 + โ‹ฏ + ๐‘Ž2 ๐‘ค๐‘˜ + ๐‘1 ๐‘ข1 + โ‹ฏ + ๐‘๐‘  ๐‘ข๐‘  = −๐‘ฃ
โŸ
โŸ
∈๐’ฐ
∈๐’ฑ
We don’t know who −๐‘ฃ is, but it is definitely in ๐’ฑ ∩ ๐’ฐ!
It has to be expressible this way:
(−๐‘1 )๐‘ฃ1 + โ‹ฏ + (−๐‘๐‘ก )๐‘ฃ๐‘ก = ๐›ผ1 ๐‘ค1 + โ‹ฏ + ๐›ผ๐‘˜ ๐‘ค๐‘˜ ⇒ ๐‘๐‘– = 0
dim(๐’ฐ + ๐’ฑ) = ๐‘˜ + ๐‘  + ๐‘ก = (๐‘˜ + ๐‘ ) + (๐‘˜ + ๐‘ก) − ๐‘˜ = dim ๐’ฐ + dim ๐’ฑ − dim ๐’ฐ ∩ ๐’ฑ
----- End of lesson 6
๐’ฐ, ๐’ฑ subspaces of a vector space ๐’ฒ over ๐”ฝ.
๐’ฐ + ๐’ฑ = {๐‘ข + ๐‘ฃ|๐‘ข ∈ ๐’ฐ, ๐‘ฃ ∈ ๐’ฑ}
๐’ฐ ∩ ๐’ฑ = {๐‘ฅ|๐‘ฅ ∈ ๐’ฐ, ๐‘ฅ ∈ ๐’ฑ}
Both of these are vector spaces.
We established that a dimension dim ๐’ฐ + ๐’ฑ = dim ๐’ฐ + dim ๐’ฑ − dim ๐’ฐ ∩ ๐’ฑ
The sum ๐’ฐ + ๐’ฑ is said to be direct (called a direct sum) if dim ๐’ฐ + ๐’ฑ = dim ๐’ฑ + dim ๐’ฐ
(consequence, the sum ๐’ฐ + ๐’ฑ is a direct sum ⇔ ๐’ฐ ∩ ๐’ฑ = {0})
In a direct sum, the decomposition is unique.
If ๐‘ฅ = ๐‘ข1 + ๐‘ฃ1 , such that ๐‘ข1 ∈ ๐’ฐ, ๐‘ฃ1 ∈ ๐’ฑ and also ๐‘ฅ = ๐‘ข2 + ๐‘ฃ2 such that ๐‘ข2 ∈ ๐’ฐ, ๐‘ฃ2 ∈ ๐’ฑ then
๐‘ข1 = ๐‘ข2 and ๐‘ฃ1 = ๐‘ฃ2 .
Why?
๐‘ข1 + ๐‘ฃ1 = ๐‘ข2 + ๐‘ฃ2
๐‘ข1 − ๐‘ข2 = โŸ
๐‘ฃ1 − ๐‘ฃ2
โŸ
∈๐’ฐ
∈๐’ฑ
So ๐‘ข1 − ๐‘ข2 ∈ ๐’ฐ ∩ ๐’ฑ. From our assumption, ๐‘ข1 − ๐‘ข2 = 0.
Similarly, ๐‘ฃ2 − ๐‘ฃ1 = 0.
This means ๐‘ข1 = ๐‘ข2 and ๐‘ฃ1 = ๐‘ฃ2 .
๐’ฑ1 , … , ๐’ฑ๐‘˜ subspaces of a vector space ๐’ฒ then
๐’ฐ1 + โ‹ฏ + ๐’ฐ๐‘˜ = {๐‘ข1 + โ‹ฏ + ๐‘ข๐‘˜ |๐‘ข๐‘— = ๐’ฐ๐‘— , ๐‘— = 1 … ๐‘˜}
๐’ฐ1 + โ‹ฏ + ๐’ฐ๐‘˜ is said to be direct if dim ๐’ฐ1 + โ‹ฏ + ๐’ฐ๐‘˜ = dim ๐’ฐ1 + โ‹ฏ + dim ๐’ฐ๐‘˜
It’s very tempting to jump to the conclusion (for example) that
๐’ฐ1 + ๐’ฐ2 + ๐’ฐ3 is a direct sum ⇔ ๐’ฐ1 ∩ ๐’ฐ2 = {0}, ๐’ฐ1 ∩ ๐’ฐ3 = {0}, ๐’ฐ2 ∩ ๐’ฐ3 = {0}
However, this is not enough.
1
0
1
Consider ๐’ฐ1 = ๐‘ ๐‘๐‘Ž๐‘› {[0]} , ๐’ฐ2 = ๐‘ ๐‘๐‘Ž๐‘› {[1]} , ๐’ฐ3 = ๐‘ ๐‘๐‘Ž๐‘› {[1]}
0
0
0
๐’ฐ1 ∩ ๐’ฐ2 = {0}, ๐’ฐ1 ∩ ๐’ฐ3 = {0}, ๐’ฐ2 ∩ ๐’ฐ3 = {0}
dim ๐’ฐ1 + dim ๐’ฐ2 + dim ๐’ฐ3 = 3
But dim ๐’ฐ1 + ๐’ฐ2 + ๐’ฐ3 = 2
So forget the false conclusions, and just remember dim ๐’ฐ1 + ๐’ฐ2 = dim ๐’ฐ1 + dim ๐’ฐ2.
Generally we say ๐’ฐ1 + โ‹ฏ + ๐’ฐ๐‘˜ is direct ⇔ every collection of non-zero vectors, at most one
from each subspace is linearly independent.
Suppose ๐‘˜ = 4. ๐’ฐ, ๐’ฑ, ๐’ณ, ๐’ด. If the sum ๐’ฐ + ๐’ฑ + ๐’ณ + ๐’ด is direct, then all non-zero ๐‘ข ∈ ๐’ฐ, ๐‘ฃ ∈
๐’ฑ, ๐‘ฅ ∈ ๐’ณ, ๐‘ฆ ∈ ๐’ด ⇒ {๐‘ข, ๐‘ฃ, ๐‘ฅ, ๐‘ฆ} are linearly independent.
Every ๐ด ∈ โ„‚๐‘›×๐‘› has at least one Eigenvalue. i.e. there is a point ๐œ† ∈ โ„‚ and a vector ๐‘ฅ ≠ 0 such
that ๐ด๐‘ฅ = ๐œ†๐‘ฅ (same as saying (๐ด − ๐œ†๐ผ)๐‘ฅ = 0
That is equivalent to saying ๐’ฉ(๐ด−๐œ†๐ผ) ≠ {0}
Suppose ๐ด has ๐‘˜ distinct Eigenvalues ๐œ†1 , … , ๐œ†๐‘˜ ∈ โ„‚. We shoed that if ๐ด๐‘ข๐‘— = ๐œ†๐‘— ๐‘ข๐‘— ๐‘— = 1 … ๐‘˜ then
๐‘ข1 , … , ๐‘ข๐‘˜ are linearly independent,.
Same as to say ๐’ฉ(๐ด−๐œ†1 ๐ผ) + โ‹ฏ + ๐’ฉ(๐ด−๐œ†๐‘˜ ๐ผ) is a direct sum.
Same as to say dim ๐’ฉ(๐ด−๐œ†1 ๐ผ) + โ‹ฏ + ๐’ฉ(๐ด−๐œ†๐‘˜ ๐ผ) = dim ๐’ฉ(๐ด−๐œ†1 ๐ผ) + โ‹ฏ + dim ๐’ฉ(๐ด−๐œ†๐‘˜ ๐ผ)
Criteria: A matrix ๐ด ∈ โ„‚๐‘›×๐‘› is diagonizable ⇔ can find ๐‘› linearly independent Eigenvectors.
Same as to say dim ๐’ฉ(๐ด−๐œ†1 ๐ผ) + โ‹ฏ + dim ๐’ฉ(๐ด−๐œ†๐‘˜๐ผ) = ๐‘›.
๐ด5×5 , ๐œ†1 , … , ๐œ†5 distinct Eigenvalues. ๐ด๐‘ข๐‘— = ๐œ†๐‘— ๐‘ข๐‘— , ๐‘— = 1, … ,5 ๐‘ข๐‘—
๐‘ˆ
๐œ†1
โž
๐ด [๐‘ข1 , … , ๐‘ข_5 ] = [๐ด๐‘ข1 … ๐ด๐‘ข5 ] = [๐œ†1 ๐‘ข1 … ๐œ†5 ๐‘ข5 ] = [๐‘ข1 … ๐‘ข5 ] [ 0
0
≠0
0 0
โ‹ฑ 0]
0 ๐œ†5
๐ด๐‘ˆ = ๐‘ˆ๐ท
Additional fact: ๐‘ข1 , … , ๐‘ข5 are linearly independent. Therefore, ๐’ฐ is invertible. dim โ„›๐‘ˆ = 5. ๐‘ˆ is
a 5 × 5 matrix.
So ๐ด = ๐‘ˆ๐ท๐‘ˆ −1
dim
โŸ ๐’ฉ(๐ด−๐œ†๐‘—๐ผ5 ) ≥ 1
๐›พ๐‘—
๐›พ1 + ๐›พ2 + โ‹ฏ + ๐›พ5 ≤ 5 ⇒ ๐›พ1 = ๐›พ2 = โ‹ฏ = ๐›พ5 = 1
๐ด is a 5 × 5 matrix.
2 distinct Eigenvalues ๐œ†1 , ๐œ†2 . ๐œ†1 ≠ ๐œ†2
Crucial issue is the dimension of the null spaces: dim ๐’ฉ(๐ด−๐œ†1 ๐ผ) = 3 and dim ๐’ฉ(๐ด−๐œ†2 ๐ผ) = 2
⇒ ๐ด๐‘ข1 = ๐œ†1 ๐‘ข1 , ๐ด๐‘ข2 = ๐œ†1 ๐‘ข2 , ๐ด๐‘ข3 = ๐œ†1 ๐‘ข3 , ๐ด๐‘ข4 = ๐œ†2 ๐‘ข4 , ๐ด๐‘ข5 = ๐œ†2 ๐‘ข5
{๐‘ข1 , ๐‘ข2 , ๐‘ข3 } a basis for ๐’ฉ(๐ด−๐œ†1 ๐ผ)
{๐‘ข4 , ๐‘ข5 } a basis for ๐’ฉ(๐ด−๐œ†2 ๐ผ)
๐œ†1 0 0 0 0
0 ๐œ†1 0 0 0
๐‘ข
๐‘ข
๐‘ข
๐‘ข
๐‘ข
๐‘ข
๐‘ข
๐‘ข
๐‘ข
๐‘ข
0 ๐œ†1 0 0
๐ด[ 1
2
3
4
5] = [ 1
2
3
4
5] 0
0 0 0 ๐œ†2 0
[ 0 0 0 0 ๐œ†2 ]
(1) Find Engenvalues ๐œ†1 , … , ๐œ†๐‘˜ of ๐ด (distinct Eigenvalues)
(2) Find basis for each space ๐’ฉ(๐ด−๐œ†๐‘–๐ผ) , ๐‘– = 1 … ๐‘˜
(3) Stack resulting vectors ๐ด๐‘ˆ = ๐‘ˆ๐ท
(columns of ๐‘ˆ are taken from basis ๐’ฉ(๐ด−๐œ†๐‘–๐ผ) , ๐‘– = 1 … ๐‘˜ and always linearly
independent)
Question: Do you always have enough columns? No. ๐‘ˆ could be non invertible, and then ๐ด is not
diagonizable.
Example:
2 1 0
๐ด = [0 2 1]
0 0 2
2−๐œ†
1
0
(๐ด − ๐œ†๐ผ3 ) = [ 0
2−๐œ†
0 ]
0
0
2−๐œ†
(๐ด − ๐œ†๐ผ3 ) is invertible (nullspace is {0}) if ๐œ† ≠ 2, not invertible (nullspace is not {0}) if ๐œ† = 2.
0 1 0
๐ด − 2๐ผ3 = [0 0 1]
0 0 0
Let’s look for vectors in the null space:
0
[0
0
1 0 ๐›ผ
๐›ฝ
0 1] [๐›ฝ ] = [ ๐›พ ] = 0
0 0 ๐›พ
0
1
So ๐›ฝ = ๐›พ = 0 ⇒the nullspace is ๐‘ ๐‘๐‘Ž๐‘› {[0]}
0
Only one Eigenvector!
1 0 0
๐‘ˆ = [ 0 0 0]
0 0 0
2
0
Suppose ๐ด = 0
0
[0
0
0
๐ด − 2๐ผ = 0
[
1
2
0
0
0
1
0
0
0
1
2
0
0
0
0
0
3
0
0
2−๐œ†
0
0
,
๐ด
−
๐œ†๐ผ
=
0
0
5
1
0
[ 0
3]
0
1
0
,
1
0
๐’ฉ๐ด−2๐ผ
1
1]
Remember: ๐ต๐‘×๐‘ž : ๐‘ž = dim ๐’ฉ๐ต + dim โ„›๐ต
1
2−๐œ†
0
0
0
1
0
= ๐‘ ๐‘๐‘Ž๐‘› 0
0
{[0]}
0
1
2−๐œ†
0
0
0
0
0
3−๐œ†
0
0
0
0
1
3 − ๐œ†]
−1 1
0 0 0
0 −1 1 0 0
๐ด − 3๐ผ = 0
0 −1 0 0 ,
0
0
0 0 1
[0
0
0 0 0]
0
0
(๐ด − 2๐ผ)2 = 0
0
[0
0
0
0
0
0
1
0
0
0
0
0
0
0
1
0
0
0
0 ,
2
1]
๐’ฉ๐ด−3๐ผ
[
๐ด11
0
0
0
= ๐‘ ๐‘๐‘Ž๐‘› 0
1
[
{ 0]}
๐ด๐‘˜
0 ๐‘˜
] = [ 11
๐ด22
0
0
]
๐ด๐‘˜22
๐’ฉ(๐ด−2๐ผ)2 = ๐‘ ๐‘๐‘Ž๐‘›{๐‘’1 , ๐‘’2 } – added eigenvector
0
0
(๐ด − 2๐ผ)3 = 0
0
[0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
∗
1]
๐’ฉ(๐ด−2๐ผ)2 = ๐‘ ๐‘๐‘Ž๐‘›{๐‘’1 , ๐‘’2 , ๐‘’3 } - added eigenvector
0
0
(๐ด − 2๐ผ)๐‘˜ , ๐‘˜ ≥ 4 will still be something of the form 0
0
[0
eigenvectors in the following powers.
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0 so we won’t get any new
∗
1]
๐’ฉ๐ด−3๐ผ = ๐‘ ๐‘๐‘Ž๐‘›{๐‘’4 }
๐’ฉ(๐ด−3๐ผ)2 = ๐‘ ๐‘๐‘Ž๐‘›{๐‘’4 , ๐‘’5 }
Summary: dim ๐’ฉ(๐ด−2๐ผ)3 + dim ๐’ฉ(๐ด−3๐ผ)2 = dim ๐’ฉ(๐ด−2๐ผ)5 + dim ๐’ฉ(๐ด−3๐ผ)5 = 5
1) ๐ต ∈ โ„‚๐‘›×๐‘› , ๐’ฉ๐ต ⊆ ๐’ฉ๐ต2 ⊆ โ‹ฏ
But there is a saturation! If ๐’ฉ๐ต๐‘˜ = ๐’ฉ๐‘๐‘˜+1 then ๐’ฉ๐ต๐‘˜+1 = ๐’ฉ๐‘๐‘˜+2 and so on.
Note: ๐’ฉ๐ต๐‘— = ๐’ฉ๐ต๐‘˜ ∀๐‘— ≥ ๐‘˜
If ๐‘ฅ ∈ ๐’ฉ๐ต๐‘— ⇒ ๐ต ๐‘—+1 ๐‘ฅ = ๐ต(๐ต ๐‘— ๐‘ฅ) = ๐ต0 = 0
Translation: ๐‘ฅ ∈ ๐’ฉ๐ต๐‘—+1 i.e. ๐’ฉ๐ต๐‘— ⊆ ๐’ฉ๐ต๐‘—+1
Suppose ๐’ฉ๐ต๐‘˜ = ๐’ฉ๐ต๐‘˜+1 . Take ๐‘ฅ ∈ ๐’ฉ๐ต๐‘˜+2 ⇒ ๐ต๐‘˜+2 ๐‘ฅ = 0 ⇒ ๐ต๐‘˜+1 (๐ต๐‘ฅ) = 0 ⇒
๐ต๐‘ฅ ∈ ๐’ฉ๐ต๐‘˜+1
But we assumed ๐’ฉ๐ต๐‘˜+1 = ๐’ฉ๐ต๐‘˜ ⇒ ๐ต๐‘ฅ ∈ ๐’ฉ๐ต๐‘˜ ⇒ ๐ต๐‘˜+1 ๐‘ฅ = 0 ⇒ ๐‘ฅ ∈ ๐’ฉ๐ต๐‘˜+1
2) ๐ต๐‘˜ ๐‘ฅ = 0 ⇒ ๐ต๐‘› ๐‘ฅ = 0 always.
suppose ๐‘› = 5, ๐ต 5 × 5, ๐ต3 ๐‘ฅ = 0 ⇒ ๐ต5 ๐‘ฅ = 0 (easy)
Interesting ๐ต6 ๐‘ฅ = 0 ⇒ ๐ต5 ๐‘ฅ = 0
Because if it is not true ⇒ ๐’ฉ๐ต5 โŠŠ ๐’ฉ๐ต6 ⇒ Saturation after 5. ⇒ dim ๐’ฉ๐ต6 ≥ 6. But 5 ×
5 matrices can’t have such degree!
Let ๐ด ∈ โ„‚๐‘›×๐‘› with ๐‘˜ distinct eigenvalues.
i.e. ๐’ฉ๐ด−๐œ†๐ผ๐‘› ≠ 0 ⇔ ๐œ† = ๐œ†1 or ๐œ† = ๐œ†2 or … ๐œ† = ๐œ†๐‘˜
๐›พ๐‘— = dim ๐’ฉ(๐ด−๐œ†๐‘—๐ผ๐‘› ) called the geometric multiplicity
๐›ผ๐‘— = dim ๐’ฉ(๐ด−๐œ†
๐‘›
๐‘— ๐ผ๐‘› )
called the algebraic multiplicity
It’s clear that ๐›พ๐‘— ≤ ๐›ผ๐‘— ⇒ ๐›พ1 + โ‹ฏ + ๐›พ๐‘˜ ≤ ๐›ผ1 + โ‹ฏ + ๐›ผ๐‘˜
3) ๐’ฉ๐ต๐‘› ∩ โ„›๐ต๐‘› = {0}
Let ๐‘ฅ ∈ ๐’ฉ๐ต๐‘› ∩ โ„›๐ต๐‘›
๐‘›๐‘œ๐‘ก ๐‘œ๐‘๐‘ฃ๐‘–๐‘œ๐‘ข๐‘  ๐‘Ž๐‘›๐‘‘ ๐‘ค๐‘–๐‘™๐‘™
๐‘๐‘’ ๐‘๐‘Ÿ๐‘œ๐‘ฃ๐‘’๐‘‘ ๐‘™๐‘Ž๐‘ก๐‘’๐‘Ÿ ๐‘œ๐‘›
=
๐ต๐‘ฆ 2
๐ต๐‘› ๐‘ฅ = 0, ๐‘ฅ = ๐ต๐‘› ๐‘ฆ ⇒ ๐ต๐‘› ๐ต๐‘› ๐‘ฆ = 0 ⇒ ๐‘ฆ ∈ ๐’ฉ๐ต2๐‘› ⇒
4) โ„‚๐‘› = ๐’ฉ๐ต๐‘› + โ„›๐ต๐‘› and this sum is direct
3 implies that the sum is direct. That means:
dim(๐’ฉ๐ต๐‘› + โ„›๐ต๐‘› ) = dim ๐’ฉ๐ต๐‘› + dim โ„›๐ต๐‘›
๐‘›
๐‘ฆ ∈ ๐’ฉ๐ต ๐‘› ⇒ ๐ต ๐‘› ๐‘ฆ = 0 ⇒ ๐‘ฅ = 0
๐‘๐‘œ๐‘›๐‘ ๐‘’๐‘Ÿ๐‘ฃ๐‘Ž๐‘ก๐‘–๐‘œ๐‘› ๐‘œ๐‘“ ๐‘‘๐‘–๐‘š๐‘’๐‘›๐‘ ๐‘–๐‘œ๐‘›๐‘ 
=
๐‘›
Remark: Is it true that the dim(๐’ฉ๐ถ + โ„›๐ถ ) = dim ๐’ฉ๐ถ + dim โ„›๐ถ for any square matrix ๐ถ?
Answer: NO!
0 1
1
1
Consider: ๐ถ = [
], ๐’ฉ๐ถ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]} , โ„›๐ถ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]}
0 0
0
0
dim ๐’ฉ๐ถ + dim โ„›๐ถ = 2 ≠ dim ๐’ฉ๐ถ + โ„›๐ถ = 1
---- end of lesson 7
๐ด ∈ โ„‚๐‘›×๐‘›
๐œ† Eigenvalue means ๐’ฉ๐ด−๐œ†๐ผ๐‘› ≠ {0}
๐ด has at least one Eigenvalue in โ„‚.
If ๐ด has ๐‘˜ distinct Eigenvalues, then
๐›พ๐‘— = dim ๐’ฉ๐ด−๐œ†๐ผ๐‘› - Geometric multiplicity
๐›ผ๐‘— = dim ๐’ฉ(๐ด−๐œ†๐ผ๐‘› )๐‘› – Algebric multiplicity
๐‘‚๐ต๐ฝ๐ธ๐ถ๐‘‡๐ผ๐‘‰๐ธ
๐›พ๐‘— ≤ ๐›ผ๐‘— always: ๐›พ1 + โ‹ฏ + ๐›พ๐‘˜ ≤ ๐›ผ1 + โ‹ฏ + ๐›ผ๐‘˜
=
๐‘›
EX. ๐ด 5 × 5 2 distinct Eigenvalues ๐œ†1 , ๐œ†2
If ๐›พ1 + ๐›พ2 = ๐‘› (equivalent to statement ๐’ฉ๐ด−๐œ†1 ๐ผ5 + ๐’ฉ๐ด−๐œ†2 ๐ผ5 = โ„‚5
Then ๐ด[๐‘ข1 … ๐‘ข5 ] = [๐‘ข1 … ๐‘ข5 ]๐ท
(5 linearly independent Eigenvectors and ๐ท is diagonal and ๐‘ˆ is invertible)
If ๐›พ1 = 3 and ๐›พ2 = 2
Then: ๐ด[๐‘ข1 … ๐‘ข5 ] = [๐œ†1 ๐‘ข1 ๐œ†1 ๐‘ข2 ๐œ†1 ๐‘ข3 ๐œ†2 ๐‘ข4 ๐œ†2 ๐‘ข5 ]
๐œ†1
๐œ†1
[๐‘ข1 ๐‘ข2 ๐‘ข3 ๐‘ข4 ๐‘ข5 ]
๐œ†1
๐œ†2
[
๐œ†2 ]
๐‘ฅ ∈ โ„‚5
5
๐‘ฅ = ∑ ๐‘’๐‘— ๐‘ข๐‘—
1
๐ด๐‘ฅ = ๐ด(๐‘1 ๐‘ข1 + โ‹ฏ + ๐‘5 ๐‘ข5 ) = ๐‘1 ๐œ†1 ๐‘ข1 + ๐‘2 ๐œ†1 ๐‘ข2 + ๐‘3 ๐œ†1 ๐‘ข3 + ๐‘4 ๐œ†2 ๐‘ข4 + ๐‘5 ๐œ†2 ๐‘ข5
Will show that ๐‘Ž ∈ ๐ถ ๐‘×๐‘ with ๐‘˜ distinct Eigenvalues ๐œ†1 , … , ๐œ†๐‘˜ can find invertible ๐‘ˆ๐‘›×๐‘› and
upper triangular ๐‘› × ๐‘› ๐ฝ of special form such that ๐ด๐‘ˆ = ๐‘ˆ๐ฝ
๐ต๐œ†1 |
−
+
−
+
| ๐ต๐œ†2 |
๐ฝ=
+
−
+
โ‹ฑ
[
๐ต๐œ†๐‘˜ ]
๐›พ๐‘— cells in ๐ต๐œ†๐‘—
Theorem: ๐ด ∈ ๐ถ ๐‘›×๐‘› , ๐œ†1 , … , ๐œ†๐‘˜ distinct Eigenvalues, then:
๐ถ ๐‘› = ๐’ฉ(๐ด−๐œ†1 ๐ผ๐‘›)๐‘› + โ‹ฏ + ๐’ฉ(๐ด−๐œ†๐‘˜ ๐ผ๐‘›)๐‘› and the sum is direct.
Suppose ๐‘˜ = 3
If e.g. ๐‘˜ = 3
{๐‘ข1 , … , ๐‘ข๐‘Ÿ } basis for ๐’ฉ(๐ด−๐œ†1 ๐ผ๐‘›)๐‘›
{๐‘ฃ1 , … , ๐‘ฃ๐‘Ÿ } basis for ๐’ฉ(๐ด−๐œ†2 ๐ผ๐‘›)๐‘›
{๐‘ค1 , … , ๐‘ค๐‘Ÿ } basis for ๐’ฉ(๐ด−๐œ†3 ๐ผ๐‘›)๐‘›
Then {๐‘ข1 , … , ๐‘ข๐‘Ÿ , ๐‘ฃ1 , … , ๐‘ฃ๐‘  , … , ๐‘ค1 , … , ๐‘ค๐‘ก } is a basis for โ„‚๐‘› .
1) ๐ต ∈ ๐ถ ๐‘›×๐‘› , โ„›๐ต๐‘› ∩ ๐’ฉ๐ต๐‘› = {0}
2) ๐ถ ๐‘› = ๐’ฉ๐ต๐‘› + โ„›๐ต๐‘› and this sum is direct
3) If ๐›ผ ≠ ๐›ฝ, then ๐’ฉ(๐ด−๐›ผ๐ผ๐‘› )๐‘› ⊆ โ„›(๐ด−๐›ฝ๐ผ๐‘› )๐‘›
Binomial theorem:
๐‘›
๐‘›
๐‘Ž, ๐‘ ∈ โ„‚ ⇒ (๐‘Ž + ๐‘) = ∑ ( ๐‘— ) ๐‘Ž ๐‘— ๐‘ ๐‘›−๐‘—
๐‘›
๐‘—=0
2
2
2
(๐‘Ž + ๐‘)2 = ( ) ๐‘Ž0 ๐‘ 2 + ( ) ๐‘Ž๐‘ + ( ) ๐‘Ž2 ๐‘ 0 = ๐‘ 2 + 2๐‘Ž๐‘ + ๐‘Ž2
0
1
2
Can we do something similar for matrices?
2
2
2
(๐ด + ๐ต)๐‘› = ( ) ๐ด0 ๐ต2 + ( ) ๐ด๐ต + ( ) ๐ด2 ๐ต0 = ๐ต2 + 2๐ด๐ต + ๐ด2
0
1
2
(๐ด + ๐ต)2 = (๐ด + ๐ต)๐ด + (๐ด + ๐ต)๐ต = ๐ด2 + ๐ต๐ด + ๐ด๐ต + ๐ต2
AB=BA
Correct only if ๐ด๐ต = ๐ต๐ด
We accept it as correct for ๐‘› though we can verify it ourselves.
๐ด − ๐›ผ๐ผ๐‘› = ๐ด − ๐›ฝ๐ผ๐‘› + (๐›ฝ − ๐›ผ)๐ผ๐‘›
๐ด − ๐›ฝ๐ผ๐‘› commutative with (๐›ฝ − ๐›ผ)๐ผ๐‘›
๐‘›
(๐ด − ๐›ผ๐ผ๐‘›
)๐‘›
๐‘›
= ((๐ด − ๐›ฝ๐ผ๐‘› ) + (๐›ฝ − ๐›ผ)๐ผ๐‘› ) = ∑ ( ๐‘— ) (๐ด − ๐›ฝ๐ผ๐‘› )๐‘˜ ((๐›ฝ − ๐›ผ )๐ผ๐‘› )๐‘›−๐‘—
๐‘›
๐‘—=0
๐‘›
๐‘›
๐‘›−๐‘—
๐‘›
= ( ) (๐ด − ๐›ฝ๐ผ๐‘› )0 (๐›ฝ − ๐›ผ)๐‘› ๐ผ๐‘› + ∑ ( ๐‘— ) (๐ด − ๐›ฝ๐ผ๐‘› )๐‘— ((๐›ฝ − ๐›ผ)๐ผ๐‘› )
0
๐‘—=1
๐‘ฅ ∈ ๐’ฉ(๐ด−๐›ผ๐ผ๐‘› )๐‘› ⇒ (๐ด − ๐›ผ๐ผ๐‘› )๐‘› ๐‘ฅ = 0
๐‘›
๐‘›
0 = (๐›ฝ − ๐›ผ) ๐‘ฅ + ∑ ( ๐‘— ) (๐›ฝ − ๐›ผ)๐‘›−๐‘— (๐ด − ๐›ฝ๐ผ๐‘› )๐‘— ๐‘ฅ
๐‘›
๐‘—=1
๐‘›
๐‘ฅ=
−1
๐‘›
∑ ( ๐‘— ) (๐›ฝ − ๐›ผ)๐‘›−๐‘— (๐ด − ๐›ฝ๐ผ๐‘› )๐‘— ๐‘ฅ =
๐‘›
(๐›ฝ − ๐›ผ)
๐‘—=1
๐‘›
−1
๐‘›
(๐ด − ๐ต๐ผ๐‘› ) {
∑ ( ๐‘— ) (๐›ฝ − ๐›ผ)๐‘›−๐‘— (๐ด − ๐›ฝ๐ผ๐‘› )๐‘— ๐‘ฅ } ๐‘ฅ
(๐›ฝ − ๐›ผ)๐‘›
๐‘—=1
๐‘ฅ = (๐ด − ๐›ฝ๐ผ๐‘› )๐‘ƒ(๐ด)๐‘ฅ
So we can replace ๐‘ฅ with it’s polynomial:
(๐ด − ๐›ฝ๐ผ๐‘› )๐‘ƒ(๐ด)2 ๐‘ฅ = (๐ด − ๐›ฝ๐ผ๐‘› )๐‘› ๐‘ƒ(๐ด)๐‘› ๐‘ฅ = (๐ด − ๐›ฝ๐ผ๐‘› )๐‘› ⇒ ๐‘ฅ ∈ โ„›(๐ด−๐›ฝ๐ผ๐‘› )๐‘›
4) If ๐’ฒ = ๐’ฐ + ๐’ฑ and the sum is direct
Let ๐’ณ be a subspace of ๐’ฒ, then:
๐’ณ ∩ ๐’ฒ = ๐’ณ ∩ ๐’ฐ + ๐’ณ ∩ ๐’ฑ this is not always true!
โ„2 = ๐’ฒ
1
๐’ฐ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]} ,
0
0
๐’ฑ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]}
1
1
๐’ณ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]}
1
๐’ณ ∩ ๐‘ˆ = {0}
๐’ณ ∩ ๐’ฑ = {0}
๐’ณ ≠ ๐’ณ∩๐’ฐ+๐’ณ∩๐’ฑ
Not always true!
It is true if ๐’ฐ is a subset of ๐’ณ or ๐’ฑ is a subset of ๐’ณ
Let ๐’ฐ ⊆ ๐’ณ, ๐‘ฅ ∈ ๐’ณ, ๐’ณ ⊆ ๐’ฒ ⇒
๐‘ฅ ∈ ๐’ฒ ⇒ ๐‘ฅ = ๐‘ข + ๐‘ฃ some ๐‘ข ∈ ๐’ฐ, ๐‘ฃ ∈ ๐’ฑ
If also ๐’ฐ ⊆ ๐’ณ ⇒ ๐‘ข ∈ ๐’ณ ⇒ ๐‘ฃ = ๐‘ฅ − ๐‘ข ∈ ๐’ณ
๐’ณ = ๐’ฐ∩๐’ณ+๐’ฑ∩๐’ณ
5) โ„‚๐‘› = ๐’ฉ(๐ด−๐œ†1 ๐ผ๐‘› )๐‘› + โ‹ฏ + ๐’ฉ(๐ด−๐œ†๐‘˜ ๐ผ๐‘› )๐‘› sum is direct.
For simplicity fix ๐‘˜ = 3
To simplify the writing ๐’ฉ๐‘— = ๐’ฉ(๐ด−๐œ†
๐‘›
๐‘— ๐ผ๐‘› )
โ„›๐‘— = โ„›(๐ด−๐œ†
๐‘›
๐‘— ๐ผ๐‘› )
Wish to show โ„‚๐‘› = ๐’ฉ1 + ๐’ฉ2 + ๐’ฉ3 (sum is direct)
(2) ⇒
1- ๐ถ ๐‘› = ๐’ฉ1 โˆ” โ„›1
โ„‚๐‘› = ๐’ฉ2 โˆ” โ„›2
โ„‚๐‘› = ๐’ฉ3 โˆ” โ„›3
โˆ” means the sum is direct
(3) ⇒ ๐’ฉ2 ⊆ โ„›1
โ„‚๐‘› = ๐’ฉ2 โˆ” โ„›2
2- [โ„›1 = โ„›1 ∩ โ„‚๐‘› = โ„›1 ∩ ๐’ฉ2 โˆ” โ„›1 ∩ โ„›2 = ๐’ฉ2 โˆ” โ„›1 ∩ โ„›2 ]
3- [โ„›1 ∩ โ„›2 = โ„›1 ∩ โ„›2 ∩ ๐’ฉ3 โˆ” โ„›1 ∩ โ„›2 ∩ โ„›3 = ๐’ฉ3 โˆ” โ„›1 + โ„›2 + โ„›3 ]
โ„‚๐‘› = ๐’ฉ3 โˆ” โ„›3
๐’ฉ3 ⊆ โ„›2 , ๐’ฉ3 ⊆ โ„›1 ⇒ ๐’ฉ3 ⊆ โ„›1 ∩ โ„›2
โ„‚๐‘› = ๐’ฉ1 โˆ” (๐’ฉ2 โˆ” โ„›1 ∩ โ„›2 ) = ๐’ฉ1 โˆ” (๐’ฉ2 โˆ” {๐’ฉ3 โˆ” โ„›1 ∩ โ„›2 ∩ โ„›3 })
โ„‚๐‘› = ๐’ฉ1 โˆ” ๐’ฉ2 โˆ” ๐’ฉ3 โˆ” โ„›1 ∩ โ„›2 ∩ โ„›3
โ„›1 ∩ โ„› 2 ∩ โ„› 3
Claim: โ„›1 ∩ โ„›2 ∩ โ„›3 is invariant under A.
i.e. ๐‘ฅ ∈ โ„›1 ∩ โ„›2 ∩ โ„›3 ⇒ ๐ด๐‘ฅ ∈ โ„›1 ∩ โ„›2 ∩ โ„›3
โ„›๐‘— = โ„›(๐ด−๐œ†
๐‘›
๐‘— ๐ผ๐‘› )
๐‘›
๐‘ฅ ∈ โ„›๐‘— ⇒ ๐‘ฅ = (๐ด − ๐œ†๐‘— ๐ผ๐‘› ) ๐‘ฆ
๐‘›
๐ด๐‘ฅ = ๐ด(๐ด − ๐œ†๐‘— ๐ผ๐‘› ) ๐‘ฆ
๐‘‡โ„Ž๐‘’๐‘ ๐‘’ ๐‘š๐‘Ž๐‘ก๐‘Ÿ๐‘–๐‘๐‘’๐‘  ๐‘Ž๐‘Ÿ๐‘’ ๐‘–๐‘›๐‘ก๐‘’๐‘Ÿ๐‘โ„Ž๐‘Ž๐‘›๐‘”๐‘’๐‘๐‘™๐‘’
=
๐‘›
(๐ด − ๐œ†๐‘— ๐ผ๐‘› ) (๐ด๐‘ฆ)
โ„›1 ∩ โ„›2 ∩ โ„›3 is a vector space, subspace of โ„‚๐‘› . It’s invariant under ๐ด.
⇒Either โ„›1 ∩ โ„›2 ∩ โ„›3 = {0} , or ∃๐‘ข ∈ โ„›1 ∩ โ„›2 ∩ โ„›3 and a ๐œ† ∈ โ„‚ such that ๐ด๐‘ข = ๐œ†๐‘ข.
๐œ† = ๐œ†1 or ๐œ†2 or ๐œ†3 .
If ๐œ† = ๐œ†1 , ๐‘ข ∈ โ„›1 and (๐ด − ๐œ†1 ๐ผ)๐‘ข = 0 ∈ ๐’ฉ1
So the second possibility cannot happen.
So we conclude that โ„›1 ∩ โ„›2 ∩ โ„›3 = {0}
Therefore: โ„‚๐‘› = ๐’ฉ1 โˆ” ๐’ฉ2 โˆ” ๐’ฉ3
โ„‚๐‘› = ๐’ฉ(๐ด−๐œ†1 ๐ผ๐‘› )๐‘› โˆ” ๐’ฉ(๐ด−๐œ†2 ๐ผ๐‘› )๐‘› โˆ” ๐’ฉ(๐ด−๐œ†3 ๐ผ๐‘› )๐‘›
Let
{๐‘ข1 , … , ๐‘ข๐‘Ÿ } be any basis for ๐’ฉ1
{๐‘ฃ1 , … , ๐‘ฃ๐‘  } be any basis for ๐’ฉ2
{๐‘ค1 , … , ๐‘ค๐‘ก } be any basis for ๐’ฉ3
Then
{๐‘ข1 , … , ๐‘ข๐‘Ÿ , ๐‘ฃ1 , … , ๐‘ฃ๐‘  , … , ๐‘ค1 , … , ๐‘ค๐‘ก } is a basis for โ„‚๐‘›
๐ด[๐‘ข1 … ๐‘ข๐‘Ÿ ๐‘ฃ1 … ๐‘ฃ๐‘  ๐‘ค1 … ๐‘ค๐‘ก ]
๐ด๐’ฉ๐‘— ⊆ ๐’ฉ๐‘—
๐‘Ž๐‘๐‘๐‘Ÿ๐‘’๐‘ฃ๐‘–๐‘Ž๐‘ก๐‘’
=
๐ด[๐‘ˆ ๐‘‰ ๐‘Š]
๐‘›
If ๐‘ฅ ∈ ๐’ฉ๐‘— then (๐ด − ๐œ†๐‘— ๐ผ๐‘› ) ๐‘ฅ = 0
Is ๐ด๐‘ฅ ∈ ๐’ฉ๐‘— ?
Yes. We can interchange:
๐‘›
๐‘›
(๐ด − ๐œ†๐‘— ๐ผ๐‘› ) ๐ด๐‘ฅ = ๐ด(๐ด − ๐œ†๐‘— ๐ผ๐‘› ) ๐‘ฅ = ๐ด0 = 0
๐ด๐‘ข๐‘— ∈ ๐‘ ๐‘๐‘Ž๐‘› {๐‘ข1 , … , ๐‘ข๐‘Ÿ } ๐‘— = 1, … , ๐‘Ÿ
๐ด๐‘ฃ๐‘— ∈ ๐‘ ๐‘๐‘Ž๐‘› {๐‘ฃ1 , … , ๐‘ฃ๐‘  } ๐‘— = 1, … , ๐‘ 
๐ด๐‘ค๐‘— ∈ ๐‘ ๐‘๐‘Ž๐‘› {๐‘ค1 , … , ๐‘ค๐‘ก } ๐‘— = 1, … , ๐‘ก
๐‘ฅ ∈ ๐‘ ๐‘๐‘Ž๐‘› {๐‘ข1 , … , ๐‘ข๐‘Ÿ } ⇒ ๐‘ฅ = ๐‘ˆ๐‘Ž ⇒ ๐ด๐‘ฅ = ๐ด๐‘ˆ๐‘Ž
๐บ1
−
๐ด[๐‘ˆ ๐‘‰ ๐‘Š] = [๐‘ˆ ๐‘‰ ๐‘Š]
−
[
๐บ1
−
๐ด=๐‘‡
−
[
|
+
|
+
|
−
๐บ2
−
|
+ −
| ๐บ2
+ −
|
|
+ −
|
+ −
| ๐บ3 ]
|
+ −
|
๐‘‡ −1
+ −
| ๐บ3 ]
Next we will choose basis for each of the spaces of ๐’ฉ๐‘— which is useful!
----- End of lesson 8
Theorem: If ๐ด ∈ โ„‚๐‘›×๐‘› with ๐‘˜ distinct Eigenvalues ๐œ†1 , … , ๐œ†๐‘˜ , then
โ„‚๐‘› = ๐’ฉ๐ด−๐œ†1 ๐ผ + โ‹ฏ + ๐’ฉ๐ด−๐œ†๐‘˜ ๐ผ and this sum is direct.
Implication: (write for ๐‘˜ = 3)
If {๐‘ข1 , … , ๐‘ข๐‘Ÿ } is a basis for ๐’ฉ๐ด−๐œ†1 ๐ผ
And {๐‘ฃ1 , … , ๐‘ฃ๐‘  } is a basis for ๐’ฉ๐ด−๐œ†2 ๐ผ
And {๐‘ค1 , … , ๐‘ค๐‘ก } is a basis for ๐’ฉ๐ด−๐œ†3 ๐ผ
Then {๐‘ข1 , … , ๐‘ข๐‘Ÿ , ๐‘ฃ1 , … , ๐‘ฃ๐‘  , ๐‘ค1 , … , ๐‘ค๐‘ก } is a basis for โ„‚๐‘›
๐›พ๐‘— = dim ๐’ฉ๐ด−๐œ†๐‘—๐ผ – Geometric multiplicity.
๐›ผ๐‘— = dim ๐’ฉ(๐ด−๐œ†1 ๐ผ)๐‘› - Algebraic multiplicity.
๐‘Ÿ = ๐›ผ1 , ๐‘  = ๐›ผ2 , ๐‘ก = ๐›ผ3
Recall also:
๐’ฉ(๐ด−๐œ† ๐ผ)๐‘› is invariant under ๐ด.
๐‘—
i.e. ๐‘ฅ ∈ ๐’ฉ(๐ด−๐œ†
๐‘›
๐‘— ๐ผ)
⇒ ๐ด๐‘ฅ ∈ ๐’ฉ(๐ด−๐œ†
๐‘›
๐‘— ๐ผ)
๐ด๐‘ข1 ∈ ๐’ฉ(๐ด−๐œ†1 ๐ผ)๐‘› ⇒ ๐ด๐‘ข1 ∈ ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข1 , … , ๐‘ข๐‘Ÿ }
So
๐‘ฅ11
๐‘ฅ21
๐ด๐‘ข1 = [๐‘ข1 … ๐‘ข๐‘Ÿ ] [ โ‹ฎ ]
๐‘ฅ๐‘Ÿ1
Same as ๐‘ฅ11 ๐‘ข1 + ๐‘ฅ21 ๐‘ข2 + โ‹ฏ + ๐‘ฅ๐‘Ÿ1 ๐‘ข๐‘Ÿ
๐ด๐‘ข2 = [๐‘ข1
๐‘ฅ12
๐‘ฅ22
… ๐‘ข๐‘Ÿ ] [ ]
โ‹ฎ
๐‘ฅ๐‘Ÿ2
[๐ด๐‘ข1
๐ด๐‘ข2
… ๐ด๐‘ข๐‘Ÿ ] = [๐‘ข1
So
๐‘ข1
๐ด [โŸ
… ๐‘ข๐‘Ÿ ] = [๐‘ข1
… ๐‘ข๐‘Ÿ ]๐‘‹ ๐‘Ÿ×๐‘Ÿ
๐‘ˆ
๐‘ฃ1
๐ด [โŸ
… ๐‘ฃ๐‘  ] = [๐‘ฃ1
… ๐‘ฃ๐‘  ]๐‘Œ ๐‘ ×๐‘ 
๐‘‰
๐‘ค1
๐ด [โŸ
… ๐‘ค๐‘ก ] = [๐‘ค1
๐‘Š
… ๐‘ค๐‘ก ]๐‘ ๐‘ก×๐‘ก
๐‘ฅ11
๐‘ฅ21
… ๐‘ข๐‘Ÿ ] [
โ‹ฎ
๐‘ฅ๐‘Ÿ1
๐‘ฅ12
๐‘ฅ22
โ‹ฎ
๐‘ฅ๐‘Ÿ2
… ๐‘ฅ1๐‘Ÿ
…
โ‹ฎ
โ‹ฎ ]
… ๐‘ฅ๐‘Ÿ๐‘Ÿ
๐ด[๐‘ˆ
๐‘‰
๐‘Š ] = [๐‘ˆ๐‘‹
๐‘‰๐‘Œ
๐‘Š๐‘] = [๐‘ˆ
๐‘‹
๐‘Š] [ 0
0
๐‘‰
0
๐‘Œ
0
0
0]
๐‘
This much holds for any choice of basis for each of the spaces ๐’ฉ(๐ด−๐œ†
Next objective is to choose the basis in ๐’ฉ(๐ด−๐œ†
๐‘›
๐‘— ๐ผ)
๐‘›
๐‘— ๐ผ)
.
to give nice results.
If ๐‘˜ = 3, we would like ๐‘‹, ๐‘Œ, ๐‘ to be matrices which are easy to work with.
Lemma: Let ๐ต ∈ โ„‚๐‘›×๐‘› , then:
(1) dim ๐’ฉ๐ต๐‘˜+1 dim ๐’ฉ๐ต๐‘˜ ≤ dim ๐’ฉ๐ต๐‘˜ − dim ๐’ฉ๐ต๐‘˜−1 ๐‘˜ = 2,3, …
(2) dim ๐’ฉ๐ต2 − dim ๐’ฉ๐ต ≤ dim ๐’ฉ๐ต
Proof (1): We know that always dim ๐’ฉ๐ต ≤ dim ๐’ฉ๐ต2 ≤ dim ๐’ฉ๐ต3 ≤ โ‹ฏ and somewhere it
saturates.
If the left side is zero, it’s not an interesting statement, because the right side is always bigger.
Assume dim ๐’ฉ๐ต๐‘˜+1 > dim ๐’ฉ๐ต๐‘˜
But then dim ๐’ฉ๐ต๐‘˜ > dim ๐’ฉ๐ต๐‘˜−1
Let {๐‘Ž1 , … , ๐‘Ž๐‘Ÿ } be a basis for ๐’ฉ๐ต๐‘˜−1
Let {๐‘Ž1 , … , ๐‘Ž๐‘Ÿ , ๐‘1 , … , ๐‘๐‘  } be a basis for ๐’ฉ๐ต๐‘˜
Let {๐‘Ž1 , … , ๐‘Ž๐‘Ÿ , ๐‘1 , … , ๐‘๐‘  , ๐‘1 , … , ๐‘๐‘ก } be any basis for ๐’ฉ๐ต๐‘˜+1
Claim: (๐‘Ÿ + ๐‘  + ๐‘ก) − (๐‘Ÿ + ๐‘ ) ≤ (๐‘Ÿ + ๐‘ ) − ๐‘Ÿ
i.e. wish to show that ๐‘ก ≤ ๐‘ 
1) ๐ต๐‘˜ ๐‘1 + โ‹ฏ + ๐ต๐‘˜ ๐‘๐‘ก are linearly independent
Suppose ๐›พ1 ๐ต๐‘˜ ๐‘1 + โ‹ฏ + ๐›พ๐‘ก ๐ต๐‘˜ ๐‘๐‘ก = 0
๐ต๐‘˜ (๐›พ1 ๐‘1 + โ‹ฏ + ๐›พ๐‘ก ๐‘๐‘ก ) = 0
This means that ๐›พ1 ๐‘1 + โ‹ฏ + ๐›พ๐‘ก ๐‘๐‘ก ∈ ๐’ฉ๐ต๐‘˜
However, we know that
๐’ฉ๐ต๐‘˜ = ๐’ฉ๐ต๐‘˜−1 โˆ” ๐‘ ๐‘๐‘Ž๐‘›{๐‘1 , … , ๐‘๐‘  }
๐’ฉ๐ต๐‘˜+1 = ๐’ฉ๐ต๐‘˜ โˆ” ๐‘ ๐‘๐‘Ž๐‘›{๐‘1 , … , ๐‘๐‘ก }
And we also know that ๐›พ1 ๐‘1 + โ‹ฏ + ๐›พ๐‘ก ๐‘๐‘ก ∈ ๐’ฉ๐ต๐‘˜+1
But we know that
๐‘๐‘– ๐‘  ๐‘Ž๐‘Ÿ๐‘’ ๐‘–๐‘›๐‘‘๐‘’๐‘๐‘’๐‘›๐‘‘๐‘’๐‘›๐‘ก!
๐’ฉ๐ต๐‘˜ ∩ ๐‘ ๐‘๐‘Ž๐‘›{๐‘! , … , ๐‘๐‘ก } = {0} ⇒ ๐›พ1 ๐‘1 + โ‹ฏ + ๐›พ๐‘ก ๐‘๐‘ก = 0 ⇒
๐›พ1 = ๐›พ๐‘ก = 0
2) ๐ต๐‘˜−1 ๐‘1 , … , ๐ต๐‘˜−1 ๐‘๐‘† are linearly independent. The proof goes similarly to the proof of 1.
3) ๐ต๐‘˜ ๐‘๐‘– ∈ ๐‘ ๐‘๐‘Ž๐‘›{๐ต๐‘˜−1 ๐‘1 , … , ๐ต๐‘˜−1 ๐‘๐‘  }
๐‘๐‘– ∈ ๐’ฉ๐ต๐‘˜+1
๐ต๐‘˜+1 ๐‘๐‘– = 0 ⇒ ๐ต๐‘˜ (๐ต๐‘๐‘– ) = 0 ⇒ ๐ต๐‘๐‘– ∈ ๐’ฉ๐ต๐‘˜
Due to this observation, we can write:
๐ต๐‘๐‘– = ๐‘ข + ๐‘ฃ such that ๐‘ข ∈ ๐‘ ๐‘๐‘Ž๐‘›{๐‘Ž1 , … , ๐‘Ž๐‘Ÿ } = ๐’ฉ๐ต๐‘˜−1 , ๐‘ฃ = ๐‘ ๐‘๐‘Ž๐‘›{๐‘1 , … , ๐‘๐‘  } = ๐’ฉ๐ต๐‘˜
๐‘˜
๐ต๐‘๐‘– = ๐‘ข + ∑ ๐›ฝ๐‘— ๐‘๐‘—
๐‘—=1
๐‘˜
๐ต๐‘˜−1 (๐ต๐‘๐‘– )
=๐ต
๐‘˜−1
๐‘ข + ∑ ๐›ฝ๐‘— ๐ต
๐‘—=1
๐‘ 
๐‘˜−1
๐‘๐‘— = ∑ ๐›ฝ๐‘— ๐ต๐‘˜−1 ๐‘๐‘—
1
4) ๐‘ ๐‘๐‘Ž๐‘›{๐ต๐‘˜ ๐‘1 , … , ๐ต๐‘˜ ๐‘๐‘ก } ⊆ ๐‘ ๐‘๐‘Ž๐‘›{๐ต๐‘˜+1 ๐‘1 , … , ๐ต๐‘˜+1 ๐‘๐‘  }
Since all ๐ต๐‘˜ ๐‘1 are linearly independent, so ๐‘  ≥ ๐‘ก!
Lemma: Suppose ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข1 , … , ๐‘ข๐‘˜ } ⊆ ๐‘ ๐‘๐‘Ž๐‘›{๐‘ฃ1 , … , ๐‘ฃ๐‘™ } and ๐‘ข1 , … , ๐‘ข๐‘˜ are linearly independent
and ๐‘ฃ1 , … , ๐‘ฃ๐‘™ are linearly independent. Then can choose find ๐‘™ − ๐‘˜ of the vectors ๐‘ฃ1 , … , ๐‘ฃ๐‘™ call
them ๐‘ค1 , … , ๐‘ค๐‘™−๐‘˜ , then ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข1 , … , ๐‘ข๐‘˜ , ๐‘ค1 , … , ๐‘ค๐‘™−๐‘˜ } = ๐‘ ๐‘๐‘Ž๐‘›{๐‘ฃ1 , … , ๐‘ฃ๐‘™ ๐‘’}
Let ๐‘˜ = 3, ๐‘™ − 5
๐‘ ๐‘
TODO: Fill in!!!!
๐ด ∈ โ„‚๐‘›×๐‘› with ๐œ†1 , … , ๐œ†๐‘˜ distinct Eigenvalues
Then โ„‚๐‘› = ๐’ฉ(๐ด−๐œ†1 ๐ผ๐‘› ) โˆ” … โˆ” ๐’ฉ(๐ด−๐œ†๐‘˜ ๐ผ๐‘› )
In other words:
Take any basis of ๐’ฉ(๐ด−๐œ†1 ๐ผ๐‘› ) with ๐›ผ1 vectors
Take any basis of ๐’ฉ(๐ด−๐œ†2 ๐ผ๐‘› ) with ๐›ผ2 vectors
โ‹ฎ
Take any basis of ๐’ฉ(๐ด−๐œ†๐‘˜ ๐ผ๐‘› ) with ๐›ผ๐‘˜ vectors
Take all those ๐›ผ1 + โ‹ฏ + ๐›ผ๐‘˜ vectors together. It will form a basis for โ„‚๐‘›
Next step, is to choose the basis for each space in a good way.
Let ๐ต = ๐ด − ๐œ†๐‘— ๐ผ๐‘› ๐’ฉ๐ต ≠ {0}
๐’ฉ๐ต โŠŠ ๐’ฉ๐ต2 โŠŠ ๐’ฉ๐ต3 = ๐’ฉ๐ต4
{๐‘Ž1 , … , ๐‘Ž๐‘Ÿ } basis for ๐’ฉ๐ต
{๐‘Ž1 , … , ๐‘Ž๐‘Ÿ , ๐‘1 , … , ๐‘๐‘  } basis for ๐’ฉ๐ต2
{๐‘Ž1 , … , ๐‘Ž๐‘Ÿ , ๐‘1 , … , ๐‘๐‘  , ๐‘1 , … , ๐‘๐‘ก } basis for ๐’ฉ๐ต3
๐ต2 ๐‘1 , … , ๐ต2 ๐‘๐‘ก โŸ
๐ต๐‘1 , … , ๐ต๐‘๐‘ 
โŸ
๐‘–๐‘›๐‘‘๐‘’๐‘๐‘’๐‘›๐‘‘๐‘’๐‘›๐‘ก
๐‘–๐‘›๐‘‘๐‘’๐‘๐‘’๐‘›๐‘‘๐‘’๐‘›๐‘ก
๐‘Ž1 , … , ๐‘Ž๐‘Ÿ
โŸ
๐‘–๐‘›๐‘‘๐‘’๐‘๐‘’๐‘›๐‘‘๐‘’๐‘›๐‘ก
All in the nullspace of ๐ต
๐‘ก + ๐‘Ÿ + ๐‘  vecors in ๐’ฉ๐ต !
But we only need ๐‘Ÿ vectors (since the dimension of the nullspace).
And also
๐‘ ๐‘๐‘Ž๐‘›{๐ต2 ๐‘1 , … , ๐ต2 ๐‘๐‘ก } ⊆ ๐‘ ๐‘๐‘Ž๐‘›{๐ต๐‘1 , … , ๐ต๐‘๐‘  }
๐‘ ๐‘๐‘Ž๐‘›{๐ต๐‘1 , … , ๐ต๐‘๐‘  } ⊆ ๐‘ ๐‘๐‘Ž๐‘›{๐‘Ž1 , … , ๐‘Ž๐‘Ÿ }
So we will keep:
all ๐ต2 ๐‘1 , … , ๐ต2 ๐‘๐‘ก
๐‘  − ๐‘ก of ๐ต๐‘1 , … , ๐ต๐‘๐‘ 
๐‘Ÿ − ๐‘  of ๐‘Ž1 , … , ๐‘Ž๐‘Ÿ
Leaves us with ๐‘ก + (๐‘  − ๐‘ก) + (๐‘Ÿ − ๐‘ ) = ๐‘Ÿ vectors.
Can find ๐‘  − ๐‘ก of ๐ต๐‘1 , … , ๐ต๐‘๐‘  such that ๐‘ ๐‘๐‘Ž๐‘›{๐ต2 ๐‘1 , … , ๐ต2 ๐‘๐‘ก , ๐‘กโ„Ž๐‘œ๐‘ ๐‘’ ๐‘  − ๐‘ก} = ๐‘ ๐‘๐‘Ž๐‘›{๐ต๐‘1 , … , ๐ต๐‘๐‘  }
Add ๐‘Ÿ − ๐‘  of the {๐‘Ž1 , … , ๐‘Ž๐‘Ÿ } vectors, so the whole collection will equal to ๐‘ ๐‘๐‘Ž๐‘›{๐‘Ž1 , … , ๐‘Ž๐‘Ÿ } = ๐’ฉ๐ต
Let’s take an example with numbers:
๐‘ก = 2, ๐‘  = 3, ๐‘Ÿ = 5
{๐‘Ž1 , … , ๐‘Ž5 } basis for ๐’ฉ๐ต
{๐‘Ž1 , … , ๐‘Ž5 , ๐‘1 , ๐‘2 , ๐‘3 } basis for ๐‘ ๐‘๐‘Ÿ๐‘–๐‘๐‘ก๐ต2
{๐‘Ž1 , … , ๐‘Ž5 , ๐‘1 , ๐‘2 , ๐‘3 , ๐‘1 , ๐‘2 } basis for ๐‘ ๐‘๐‘Ÿ๐‘–๐‘๐‘ก๐ต3
๐ต2 ๐‘1 ๐ต2 ๐‘2 ๐ต๐‘1 ๐ต๐‘2 ๐ต๐‘3 ๐‘Ž1 ๐‘Ž2 ๐‘Ž3 ๐‘Ž4 ๐‘Ž5
Suppose ๐ต๐‘2 is linearly independent of ๐ต2 ๐‘1 , ๐ต2 ๐‘2
And ๐‘Ž2 , ๐‘Ž4 are linearly independent of ๐ต2 ๐‘1 , ๐ต2 ๐‘2 , ๐ต๐‘2
So keep
๐ต2 ๐‘1 ๐ต2 ๐‘2 ๐ต๐‘2 ๐‘Ž2 ๐‘Ž4
๐ต๐‘1 ๐ต๐‘2
๐‘2
๐‘1
๐‘2
Claim: Those 10 vectors are Linearly independent.
To this point, we only know ๐ต2 ๐‘1 , ๐ต2 ๐‘2 , ๐ต๐‘2 , ๐‘Ž2 , ๐‘Ž4 are linearly independent.
๐›พ1 ๐ต2 ๐‘1 + ๐›พ2 ๐ต๐‘1 + ๐›พ3 ๐‘1 + ๐›พ4 ๐ต2 ๐‘2 + ๐›พ5 ๐ต๐‘2 + ๐›พ6 ๐‘2 + ๐›พ7 ๐ต๐‘2 + ๐›พ8 ๐‘2 + ๐›พ9 ๐‘Ž2 + ๐›พ10 ๐‘Ž4 = 0
Let’s apply ๐ต2 on both sides.
What’s remains of these two terms is:
๐›พ3 ๐ต2 ๐‘1 + ๐›พ6 ๐ต2 ๐‘2 = 0 ⇒ ๐›พ3 = ๐›พ6 = 0 Since they are linearly independent!
Let’s apply ๐ต!
๐›พ2 ๐ต2 ๐‘1 + ๐›พ5 ๐ต2 ๐‘2 + ๐›พ8 ๐ต๐‘2 = 0 ⇒ ๐›พ2 = ๐›พ5 = ๐›พ8 because they are linearly independent!
๐›พ1 ๐ต2 ๐‘2 + ๐›พ4 ๐ต2 ๐‘2 + ๐›พ7 ๐ต๐‘2 + ๐›พ9 ๐‘Ž2 + ๐›พ10 ๐‘Ž4 = 0 ⇒ ๐›พ1 = ๐›พ4 = ๐›พ7 = ๐›พ9 = ๐›พ10 = 0 because they
are linearly independent!
So all coefficients must be zero, therefore all coefficients are linearly independent.
Since dim ๐’ฉ๐ต3 = 10, the vectors in this array form a basis f=r ๐’ฉ๐ต3 .
Claim: ๐ด [๐ต
โŸ2 ๐‘1
๐ต๐‘1
๐‘1
๐ต2 ๐‘2
๐ต๐‘2
๐‘2
๐ต๐‘2
๐‘2
๐‘Ž2
๐‘Ž4 ]
๐‘ˆ๐‘—
(3)
๐ถ๐œ†๐‘—
๐ด๐‘ˆ๐‘— = ๐‘ˆ๐‘—
0
0
0
0
(3)
๐ถ๐œ†๐‘—
0
0
0
0
๐ถ๐œ†๐‘—
(2)
0
0
0
0
๐ถ๐œ†๐‘—
[ 0
0
0
0
(1)
๐œ†๐‘—
0
0
0
0
0
0
0 = ๐‘ˆ๐‘—
0
0
0
0
(1)
๐ถ๐œ†๐‘— ]
0
[0
1
๐œ†๐‘—
0
0
0
0
0
0
0
0
0
1
๐œ†๐‘—
0
0
0
0
0
0
0
0
0
0
๐œ†๐‘—
0
0
0
0
0
0
0
0
0
1
๐œ†๐‘—
0
0
0
0
0
0
0
0
0
1
๐œ†๐‘—
0
0
0
0
0
0
0
0
0
0
๐œ†๐‘—
0
0
0
0
0
0
0
0
0
1
๐œ†๐‘—
0
0
0
0
0
0
0
0
0
0
๐œ†๐‘—
0
0
0
0
0
0
0
0
0
0
๐œ†๐‘— ]
If ๐ด has 3 Eigenvalues ๐œ†1 , ๐œ†2 , ๐œ†3
๐ท๐œ†1
0
0
0 ]
๐‘ˆ3 ] [ 0 ๐ท๐œ†2
0
0 ๐ท๐œ†3
๐ท๐œ†๐‘— is an ๐›ผ๐‘— × ๐›ผ๐‘— with ๐œ†๐‘— on the diagonal and ๐›พ๐‘— Jordan cells in its “decomposition”.
๐ด[๐‘ˆ1
๐‘ˆ2
๐‘ˆ3 ] = [๐‘ˆ1
๐‘ˆ2
dim ๐’ฉ(๐ด−๐œ†๐‘—๐ผ) = number of Jordan cells.
Analyzing ๐’ฉ
๐‘˜
(๐ด−๐œ†๐‘— ๐ผ๐‘› )
๐‘˜ = 1,2, …
For simplicity, denote ๐ด − ๐œ†๐‘— ๐ผ๐‘› = ๐ต
dim ๐’ฉ๐ต = 5 = ๐‘Ÿ
dim ๐’ฉ๐ต2 = 8 = ๐‘Ÿ + ๐‘ 
dim ๐’ฉ๐ต3 = 10 = ๐‘Ÿ + ๐‘  + ๐‘ก
Just by these numbers you can find out how ๐‘ˆ should look like:
๐‘‹
๐‘‹
๐‘‹
๐‘‹
๐‘‹
๐‘‹
๐‘‹
๐‘‹
๐‘‹
๐‘‹
5
8−5=3
1−8=2
The Jordan form:
2 3 × 3 Jordan cells
1 2 × 2 Jordan cells
2 1 × 1 Jordan cells
1 2 0 0 1
0 1 2 0 0
๐ด= 0 0 1 0 0
0 0 0 2 0
[0 0 0 2 2]
The objective is to find ๐‘ˆ 5 × 5 such that ๐ด๐‘ˆ = ๐‘ˆ๐ฝ ←Jordan form
(1) Find the Eigenvalues – ๐ด has 2 distinct Eigenvalues – ๐œ†1 = 1, ๐œ†2 = 2, ๐›ผ1 = 3, ๐›ผ2 = 2
(2) First set ๐ต = ๐ด − ๐œ†1 ๐ผ5 = ๐ด − ๐ผ5 , calculate ๐’ฉ๐ต , ๐’ฉ๐ต2 , …
0 2 0 0 1
0 0 2 0 0
๐ต= 0 0 0 0 0
0 0 0 1 0
[0 0 0 2 1]
๐’ฉ๐ต = ๐‘ ๐‘๐‘Ž๐‘›{๐‘’1 } ๐‘’๐‘— = ๐‘—th column of ๐ผ5
1
0
๐‘’1 = 0 , …
0
[ 0]
๐’ฉ๐ต2 = ๐‘ ๐‘๐‘Ž๐‘›{๐‘’1 , ๐‘’2 }
๐’ฉ๐ต3 = ๐‘ ๐‘๐‘Ž๐‘›{๐‘’1 , ๐‘’2 , ๐‘’3 }
0 2 0
0 0 2
0 0 0
๐ต2 =
− − −
0 0 0
[0 0 0
| 0 1 0
| 0 0 0
| 0 0 0
+ − − −
| 1 0 0
| 2 1] [ 0
(๐ต )2
[ 11
0
2 0 | 0 1
0 2 | 0 0
๐ต
0 0 | 0 0
= [ 11
0
− − + − −
0 0 | 1 0
0 0 | 2 1]
๐ต11 ๐ต12 + ๐ต12 ๐ต22
]=
(๐ต22 )2
0 0 0 |
0 0 0 |
0 0 0 |
๐ต3 =
− − − +
0 0 0 |
[0 0 0 |
0 0 0 |
0 0 0 |
0 0 0 |
๐ต4 =
− − − +
0 0 0 |
[0 0 0 |
∗
∗
∗
−
@
@
∗
∗
∗
−
@
@
∗
∗
∗
−
@
@]
∗
∗
∗
−
@
@]
So we only keep {๐‘’1 = ๐‘1 }
In this case dim ๐’ฉ − ๐œ†1 ๐ผ5 = 1 ⇒ 1 Jordan cell for this Eigenvalue
Calculate ๐’ฉ๐ต , ๐’ฉ๐ต2
๐‘’3 , ๐ต๐‘’3 , ๐ต2 ๐‘’3
๐ต๐‘’3 = 2๐‘’2
2
๐ต ๐‘’3 = ๐ต(๐ต๐‘’3 ) = ๐ต2 ๐‘’2 = 4๐‘’1
A detour to old history:
0
๐ต[๐ต2 ๐‘1 ๐ต๐‘1 ๐‘1 ] = [๐ต3 ๐‘1 ๐ต2 ๐‘1 ๐ต๐‘1 ] [0
0
So
๐œ†1 1 0
๐ด[๐‘ข1 ๐‘ข2 ๐‘ข3 ] = [๐‘ข1 ๐‘ข2 ๐‘ข3 ] [ 0 ๐œ†1 1 ]
0 0 ๐œ†1
1 0
0 1]
0 0
๐ต12 ๐ต11
][
๐ต22 0
๐ต12
]=
๐ต22
−1 2
0 0 1
0 −1 2 0 0
Now let ๐ต = ๐ด − ๐œ†2 ๐ผ5 = ๐ด − 2๐ผ5 = 0
0 −1 0 0
0
0
0 0 0
[0
0
0 2 0]
1
0
๐ต 0 =0
0
[−1]
๐’ฉ๐ต = ๐‘ ๐‘๐‘Ž๐‘›{๐‘’1 + ๐‘’5 }
๐‘ฅ1 − 4๐‘ฅ2 − 4๐‘ฅ3 + 2๐‘ฅ4 − ๐‘ฅ5
−1 −4 −4 2 −1 ๐‘ฅ1
0 −1 −4 0 0 ๐‘ฅ2
−๐‘ฅ2 − 4๐‘ฅ3
๐ต2 ๐‘ฅ = 0
0
1 0 0 ๐‘ฅ3 =
๐‘ฅ3
0
0
0 0 0 ๐‘ฅ4
0
[0
]
๐‘ฅ
[
]
[
]
0
0 0 0
5
0
๐‘ฅ1
−2๐‘ฅ4 + ๐‘ฅ5
−2
1
๐‘ฅ2
0
0
0
๐‘ฅ3 =
0
= ๐‘ฅ4 0 + ๐‘ฅ5 0
๐‘ฅ4
๐‘ฅ4
1
0
[0]
[1]
[๐‘ฅ5 ] [
๐‘ฅ5
]
1
0
๐’ฉ๐ต = ๐‘ ๐‘๐‘Ž๐‘› 0
0
{[1]}
1 −2
0
0
2
๐’ฉ๐ต = ๐‘ ๐‘๐‘Ž๐‘› 0 , 0
0
1
{[1] [ 0 ]}
0
๐ต[๐ต๐‘1 ๐‘1 ] = [๐ต2 ๐‘1 ๐ต๐‘1 ] = [0 ๐ต๐‘1 ] = [๐ต๐‘1 ๐‘1 ] [
0
2
1
So ๐ด[๐‘ข4 ๐‘ข5 ] = [๐‘ข4 ๐‘ข5 ] [
]
0 2
๐ด[๐‘ข1
---- end of lesson 9
๐‘ข2
๐‘ข3
๐‘ข4
๐‘ข5 ] = [๐‘ข1
๐‘ข2
๐‘ข3
1
]
0
๐‘ข4
1
0
๐‘ข5 ] 0
0
[0
1
1
0
0
0
0
1
1
0
0
0
0
0
2
0
0
0
0
1
2]
Determinants
๐ด ∈ ๐ถ ๐‘›×๐‘› Special number called its determinant.
Show that there exists exactly one function from ๐ถ ๐‘›×๐‘› to โ„‚
(1) ๐‘“(๐ผ๐‘› ) = 1
(2) If ๐‘ƒ a simple permutation, then ๐‘“(๐‘ƒ๐ด) = −๐‘“(๐ด)
(3) ๐‘“(๐ด) is linear in each row of ๐ด separately.
Let ๐ด ∈ โ„‚3×2
๐‘Ž1
๐ด = [๐‘Ž2 ], if say ๐‘Ž2 = ๐›ผ๐‘ข + ๐›ฝ๐‘ฃthen
๐‘Ž3
๐‘Ž1
๐‘Ž1
๐‘Ž1
๐‘Ž1
๐‘Ž1
๐›ผ๐‘ข
๐‘ข
๐›ผ๐‘ข
+
๐›ฝ๐‘ฃ
๐›ฝ๐‘ฃ
๐‘“ ([
]) = ๐‘“ ([ ]) + ๐‘“ ([ ]) = ๐›ผ๐‘“ ([ ]) + ๐›ฝ๐‘“ ([ ๐‘ฃ ])
๐‘Ž3
๐‘Ž3
๐‘Ž3
๐‘Ž3
๐‘Ž3
Properties 1,2,3 imply automatically:
(4) If 2 rows of ๐ด coincide, then ๐‘“(๐ด) = 0
๐‘Ž1
๐‘›×๐‘›
๐ด∈โ„‚
๐‘Ž๐‘– = ๐‘Ž๐‘— ≠ ๐‘— ๐ด = [ โ‹ฎ ]
๐‘Ž๐‘›
Let ๐‘ƒ be the simple permutation that interchanges row ๐‘– with ๐‘—.
But −๐‘“(๐ด) = ๐‘“(๐‘ƒ๐ด) = ๐‘“(๐ด)
2๐‘“(๐ด) = 0 ⇒ ๐‘“(๐ด) = 0
Example:
1 2 3
0 0 1
๐ด = [ 4 5 6] ,
๐‘ƒ = [ 0 1 0]
1 2 3
1 0 0
1 2 3
๐‘ƒ๐ด = [4 5 6] = ๐ด
1 2 3
๐‘Ž11
๐ด = [๐‘Ž
๐‘Ž12
[๐‘Ž11 ๐‘Ž12 ] = ๐‘Ž11 [1 0] + ๐‘Ž12 [0 1]
๐‘Ž22 ] ,
1
0
0
1
๐‘“(๐ด) = ๐‘Ž11 ๐‘“ ([
]) + ๐‘Ž12 ๐‘“ ([
])
๐‘Ž21 ๐‘Ž22
๐‘Ž21 ๐‘Ž22
[๐‘Ž21 ๐‘Ž22 ] = ๐‘Ž21 [1 0] + ๐‘Ž22 [0 1]
21
So
[1 0]
[1 0]
[0 1]
[0 1]
]) + ๐‘Ž11 ๐‘“ ([
]) + ๐‘Ž12 ๐‘“ ([
]) + ๐‘Ž12 ๐‘“ ([
]) =
๐‘Ž21 [1 0]
๐‘Ž22 [0 1]
๐‘Ž21 [1 0]
๐‘Ž22 [0 1]
1 0
1 0
0 1
0 1
๐‘Ž11 ๐‘Ž21 ๐‘“ ([
]) + ๐‘Ž11 ๐‘Ž22 ๐‘“ ([
]) + ๐‘Ž12 ๐‘Ž21 ๐‘“ ([
]) + ๐‘Ž12 ๐‘Ž22 ๐‘“ ([
]) =
1 0
0 1
1 0
0 1
0 + ๐‘Ž11 ๐‘Ž22 − ๐‘Ž12 ๐‘Ž21 + 0 = ๐‘Ž11 ๐‘Ž22 − ๐‘Ž12 ๐‘Ž21
๐‘“(๐ด) = ๐‘Ž11 ๐‘“ ([
๐ด∈โ„‚
3×3
๐‘Ž1
๐‘Ž11
๐‘Ž
๐‘Ž
, ๐ด = [ 2 ] = [ 21
๐‘Ž3
๐‘Ž31
๐‘Ž12
๐‘Ž22
๐‘Ž32
๐‘Ž13
๐‘Ž23 ] , ๐‘’1 , ๐‘’2 , ๐‘’3 columns of ๐ผ3 .
๐‘Ž33
๐‘Ž1 = ๐‘Ž11 ๐‘’1 + ๐‘Ž12 ๐‘’2 + ๐‘Ž13 ๐‘’3
๐‘Ž1
๐‘’1๐‘‡
๐‘’2๐‘‡
๐‘’3๐‘‡
๐‘“(๐ด) = ๐‘“ ([๐‘Ž2 ]) = ๐‘Ž11 ๐‘“ ([๐‘Ž2 ]) + ๐‘Ž12 ๐‘“ ([๐‘Ž2 ]) + ๐‘Ž13 ๐‘“ ([๐‘Ž2 ])
๐‘Ž3
๐‘Ž3
๐‘Ž3
๐‘Ž3
Let’s call them 1,2,3 accordingly.
So
1 = ๐‘Ž2 = ๐‘Ž21 ๐‘’1 + ๐‘Ž22 ๐‘’2 + ๐‘Ž23 + ๐‘’3
๐‘’1๐‘‡
๐‘’1๐‘‡
๐‘’1๐‘‡
๐‘‡
๐‘‡
1 = ๐‘Ž11 ๐‘Ž21 ๐‘“ ([๐‘’1 ]) + ๐‘Ž11 ๐‘Ž22 ๐‘“ ([๐‘’2 ]) + ๐‘Ž11 ๐‘Ž23 ๐‘“ ([๐‘’3๐‘‡ ])
๐‘Ž3
๐‘Ž3
๐‘Ž3
The first value is zero! We only have to deal with the last two terms.
๐‘Ž3 = ๐‘Ž31 ๐‘’1๐‘‡ + ๐‘Ž32 ๐‘’2๐‘‡ + ๐‘Ž33 ๐‘’3๐‘‡
๐‘’1๐‘‡
๐‘’1๐‘‡
So 1 will be: ๐‘Ž11 ๐ด22 ๐‘Ž33 ๐‘“ ([๐‘’2๐‘‡ ]) + ๐‘Ž11 ๐ด23 ๐‘Ž32 ๐‘“ ([๐‘’3๐‘‡ ])
๐‘’3๐‘‡
๐‘’2๐‘‡
So
1 = ๐‘Ž11 ๐‘Ž22 ๐‘Ž33 − ๐‘Ž11 ๐‘Ž23 ๐‘Ž32
2 = Two more terms
3 = Two more terms
TODO: Draw diagonals
๐‘Ž1
(5) If ๐ด = [ โ‹ฎ ] , ๐ต = ๐ธ๐ด, ๐ธ lower triangular, 1’s on the diagonal and exactly one non-zero
๐‘Ž๐‘›
entry below diagonal.
Then det(๐ธ๐ด) = det ๐ด
1 0
๐ธ = [0 1
0 ๐›ผ
๐‘Ž1
๐ธ๐ด = [ ๐‘Ž2 ]
๐›ผ๐‘Ž2 + ๐‘Ž3
๐ด ∈ โ„‚3×3 ,
0
0]
1
๐‘Ž1
๐‘Ž1
det ๐ธ๐ด = αdet [๐‘Ž2 ] + det [๐‘Ž2 ] = det ๐ด
๐‘Ž3
๐‘Ž2
Same as to say: adding a multiple of one row to another row does not change the determinant.
(6) If ๐ด has a row which is all zeroes, then det(๐ด) = 0
(7) If two rows are linearly dependent, then det ๐ด = 0
(8) If ๐ด ∈ โ„‚๐‘›×๐‘› is triangular, then det ๐ด = ๐‘Ž11 ๐‘Ž22 … ๐‘Ž๐‘›๐‘›
๐‘Ž11
๐ด=[ 0
0
๐‘Ž12
๐‘Ž22
0
๐‘Ž13
๐‘Ž23 ]
๐‘Ž33
By the linearity in row 3⇒ det ๐ด =
๐‘Ž11 ๐‘Ž12 ๐‘Ž13 ๐‘๐‘Ÿ๐‘œ๐‘๐‘’๐‘Ÿ๐‘ก๐‘ฆ 5
๐‘Ž11 ๐‘Ž12 0
๐‘Ž33 det [ 0 ๐‘Ž22 ๐‘Ž23 ]
=
๐‘Ž33 det [ 0 ๐‘Ž22 0] =
0
0
1
0
0 1
๐‘Ž11 ๐‘Ž12 0
๐‘Ž11 0 0
1 0
๐‘Ž33 ๐‘Ž22 det [ 0
1 0] = ๐‘Ž33 ๐‘Ž22 det [ 0 1 0] = ๐‘Ž33 ๐‘Ž22 ๐‘Ž11 det [0 1
0 0
0
0 1
0 0 1
(9) det ๐ด ≠ 0 ⇔ ๐ด is invertible
๐ด nonzero … ๐ธ3 ๐‘ƒ3 ๐ธ2 ๐‘ƒ2 ๐ธ1 ๐‘ƒ1 ๐ด = ๐‘ˆ upper echelon
๐‘ƒ1 - Permutation of the first row with one of others or just ๐ผ
๐‘ƒ2 - Permutation of the second row with either 3rd, 4th … or ๐ผ
So you can generalize it to:
๐ธ๐‘ƒ๐ด = ๐‘ˆ
๐ธ-lower triangular, ๐‘ƒ-permutation.
1
๐‘ƒ2 ๐ธ1 = ๐‘ƒ2 [๐›ผ
๐›ฝ
0
1
0
0
0]
1
๐‘ƒ2 interchanges 2nd with 3rd row
1 0 0
๐‘ƒ2 = [0 0 1]
0 1 0
๐‘ƒ2 ๐ธ1 ≠ ๐ธ1 ๐‘ƒ2 (not always!)
0 0
๐‘ƒ2 (๐ผ3 + [๐›ผ 0
๐›ฝ 0
0 0
This is the same as ๐‘ƒ2 + [๐›ฝ 0
๐›ผ 0
However, it is equal to
๐ธ1′ ๐‘ƒ2 ๐‘ƒ1
0
0 0
0]) = ๐‘ƒ2 + ๐‘ƒ2 [๐›ผ 0
0
๐›ฝ 0
0
0 0
0] ๐‘ƒ2 = (๐ผ3 + [๐›ฝ 0
0
๐›ผ 0
0
0
0] = ๐‘ƒ2 + [๐›ฝ
0
๐›ผ
0
0]) ๐‘ƒ2 = ๐ธ1′ ๐‘ƒ2
0
๐‘ƒ2 ๐ธ1 ๐‘ƒ1 ≠ ๐ธ1 ๐‘ƒ2 ๐‘ƒ1
where has the same form as ๐ธ1
๐ธ1′
0 0
0 0]
0 0
0
0]
1
If ๐ด ∈ โ„‚๐‘›×๐‘› , ๐ด ≠ 0, then there exists an ๐‘› × ๐‘› permutation matrix ๐‘ƒ, and an ๐‘› × ๐‘› lower
triangular ๐ธ with 1’s on the diagonal such that:
๐ธ๐‘ƒ๐ด = ๐‘ˆ ๐‘› × ๐‘› upper echelon.
In fact ๐‘ˆ is upper triangular!
det ๐ธ๐‘ƒ๐ด = det ๐‘ˆ = ๐‘ข11 ๐‘ข22 … ๐‘ข๐‘›๐‘›
By iterating property 5, det ๐ธ๐‘ƒ๐ด = det ๐‘ƒ๐ด = ± det ๐ด
So we know that |det ๐ด| = |๐‘ข11 ๐‘ข22 … ๐‘ข๐‘›๐‘› |
Claim: ๐ธ๐‘ƒ๐ด = ๐‘ˆ
Since ๐ธ and ๐‘ƒ are invertible, ๐ด is invertible ⇔ ๐‘ˆ is invertible ⇔ all values on the diagonal are
non zero ⇔ |det ๐ด| ≠ 0 ⇔ det ๐ด ≠ 0
(10)๐ด, ๐ต ∈ โ„‚๐‘›×๐‘› , then det ๐ด๐ต = det ๐ด โˆ™ det ๐ต
Case (1) – ๐’ฉ๐ต ≠ {0} ⇒ ๐’ฉ๐ด๐ต ≠ {0}
So ๐ต not invertible means ๐ด๐ต not invertible. So by 9, det ๐ต = 0 and det ๐ด๐ต = 0. Which also
means that det ๐ด โˆ™ det ๐ต = 0 โˆ™ ๐‘ ๐‘œ๐‘š๐‘’๐‘กโ„Ž๐‘–๐‘›๐‘” = 0 = det ๐ด๐ต
Case (2) – ๐ต is invertible.
๐œ‘(๐ด) =
det(๐ด๐ต)
det ๐ต
What are the properties of this function?
๐œ‘(๐ผ) = 1
if ๐‘ƒ is a simple permutation
๐œ‘(๐‘ƒ๐ด) =
det ๐‘ƒ(๐ด๐ต)
det ๐ด๐ต
=−
= −๐œ‘(๐ด)
det ๐ต
det ๐ต
Claim: ๐œ‘ is linear in each row of ๐ด
In the 3 × 3 case
๐‘Ž1
๐‘Ž
๐ด = [ 2]
๐‘Ž3
๐‘Ž1 ๐ต
det [ ๐‘Ž2 ๐ต ]
๐ด3 ๐ต
๐œ‘(๐ด) =
det ๐ต
Say that ๐‘Ž1 = ๐›ผ๐‘ข + ๐›ฝ๐‘ฃ
So ๐‘Ž1 ๐ต = ๐›ผ(๐‘ข๐ต) + ๐›ฝ(๐‘ฃ๐ต)
๐‘ข๐ต
๐‘ฃ๐ต
(๐›ผ det [๐‘Ž2 ๐ต] + ๐›ฝ det [๐‘Ž2 ๐ต])
๐›ผ๐‘ข๐ต + ๐›ฝ๐‘ฃ๐ต
๐›ผ๐‘ข + ๐›ฝ๐‘ฃ
๐‘Ž3 ๐ต
๐‘Ž3 ๐ต
๐‘Ž2 ๐ต
๐œ‘ ([ ๐‘Ž2 ]) = det [
]=
det
๐ต
๐‘Ž
๐‘Ž ๐ต
3
3
1
(11) If ๐ด is invertible, then det ๐ด−1 = det ๐ด. Easily obtained by calculating det ๐ด โˆ™ ๐ด−1 which
must be 1 and according to property 3 is also the multiplication of determinants.
(12)det ๐ด๐‘‡ = det ๐ด
๐ธ๐‘ƒ๐ด = ๐‘ˆ upper echelon
So according to previous properties we know that det ๐‘ƒ โˆ™ det ๐ด = ๐‘ข11 ๐‘ข22 … ๐‘ข๐‘›๐‘›
But we also know that ๐ด๐‘‡ ๐‘ƒ๐‘‡ ๐ธ ๐‘‡ = ๐‘ˆ ๐‘‡
So:
det ๐ด๐‘‡ โˆ™ det ๐‘ƒ๐‘‡ โˆ™ det ๐ธ ๐‘‡ = det ๐‘ˆ ๐‘‡
We know that det ๐ธ = det ๐ธ ๐‘‡ = 1
Also when we flip ๐‘ˆ, we still have a triangular matrix, just instead of a upper triangular we have
a lower triangular. So its determinant stays the same. So, so far this is true:
det ๐‘ƒ โˆ™ det ๐ด = det ๐‘ƒ๐‘‡ โˆ™ det ๐ด๐‘‡
Let ๐‘ƒ be some permutation.
Claim: ๐‘ƒ๐‘‡ ๐‘ƒ = ๐ผ
Let’s multiply both sides by det ๐‘ƒ !
(det ๐‘ƒ)2 โˆ™ det ๐ด = det ๐‘ƒ โˆ™ det ๐‘ƒ๐‘‡ โˆ™ det ๐ด๐‘‡ = det ๐‘ƒ๐‘ƒ๐‘‡ โˆ™ det ๐ด๐‘‡
(±1)2 โˆ™ det ๐ด = 1 โˆ™ det ๐ด๐‘‡ ⇒ det ๐ด = det ๐ด๐‘‡
๐ด ∈ โ„‚๐‘›×๐‘› , ๐œ†1 , … , ๐œ†๐‘˜ distinct Eigenvalues.
Then ๐ด๐‘ˆ = ๐‘ˆ๐ฝ where ๐ฝ is in Jordan form.
๐ด with ๐œ†1 , ๐œ†2 algebraic multiplicities ๐›ผ1 , ๐›ผ2
๐›ผ๐‘— = dim ๐’ฉ(๐ด−๐œ†
๐‘›
๐‘— ๐ผ๐‘› )
det ๐œ†๐ผ๐‘› − ๐ด = det(๐œ†๐ผ๐‘› − ๐‘ˆ๐ฝ๐‘ˆ −1 ) = det ๐‘ˆ(๐œ†๐ผ๐‘› − ๐ฝ)๐‘ˆ −1 = det ๐‘ˆ โˆ™ det ๐œ†๐ผ๐‘› − ๐ฝ โˆ™ det ๐‘ˆ −1
= det(๐œ†๐ผ๐‘› − ๐ฝ) = (๐œ† − ๐œ†1 )๐›ผ1 (๐œ† − ๐œ†2 )๐›ผ2
---- end of lesson 10
Main Properties of Determinants - Again
Let ๐ด ∈ โ„‚๐‘›×๐‘›
det ๐ผ๐‘› = 1
det ๐‘ƒ๐ด = − det ๐ด if ๐‘ƒ is a simple permutation
det ๐ด is linear in each row of ๐ด
๐ด in invertible ⇔ det ๐ด ≠ 0
Let ๐ต ∈ โ„‚๐‘›×๐‘›
det ๐ด๐ต = det ๐ด โˆ™ det ๐ต = det ๐ต๐ด
If ๐ด triangular then det ๐ด = ๐‘Ž11 ๐‘Ž22 … ๐‘Ž๐‘›๐‘›
Determinants in the aid of Eigenvalues
Recall ๐œ† is an eigenvalue of ๐ด if ๐’ฉ๐ด−๐œ†๐ผ๐‘› ≠ {0}
๐ต matrix. It’s Left invertible ⇔ ๐’ฉ๐ต = {0}
If ๐ต ∈ โ„‚๐‘›×๐‘› it’s invertible ⇔ ๐’ฉ๐ต = {0}
det(๐œ†๐ผ − ๐ด) ≠ 0 ⇔ ๐œ†๐ผ − ๐ด invertible ⇔ ๐’ฉ๐œ†๐ผ−๐ด = {0}
So the opposite is:
det(๐œ†๐ผ − ๐ด) = 0 ⇔ ๐’ฉ๐œ†๐ผ−๐ด ≠ {0}
๐ด ∈ โ„‚๐‘›×๐‘› has ๐‘˜ distinct eigenvalues, then there exists an invertible matrix ๐‘ˆ and an upper
triangular matrix ๐ฝ of special form (cough cough, the Jordan form cough cough)
Such that:
๐ด๐‘ˆ = ๐‘ˆ๐ฝ equivalently ๐ด = ๐‘ˆ๐ฝ๐‘ˆ −1
If ๐‘˜ = 3:
๐œ†1
0
0
−
0
0
0
0
0
0
0
โ‹ฑ
โ‹ฑ
0
−
0
0
0
0
0
0
0
0
โ‹ฑ
๐œ†1
−
0
0
0
0
0
0
0
| 0
| 0
| 0
+ −
| ๐œ†2
| 0
| 0
+ −
0 0
0 0
0 0
0 0
0 0
0 0
− −
โ‹ฑ 0
โ‹ฑ โ‹ฑ
0 ๐œ†2
− −
0 0
0 0
0 0
0 0
0 0
0 0
+ 0
| 0
| 0
| 0
+ −
| ๐œ†3
| 0
| 0
0 0
0 0
0 0
0 0
0 0
0 0
0 0
− −
โ‹ฑ 0
โ‹ฑ โ‹ฑ
0 ๐œ†3
det(๐œ†๐ผ − ๐ด) = det(๐œ†๐ผ๐‘› − ๐‘ˆ๐ฝ๐‘ˆ −1 ) = det(๐‘ˆ(๐œ†๐ผ๐‘› − ๐ฝ)๐‘ˆ −1 ) = det ๐‘ˆ โˆ™ det(๐œ†๐ผ๐‘› − ๐ฝ) โˆ™ det ๐‘ˆ −1 =
det(๐œ†๐ผ๐‘› − ๐ฝ)
So if ๐‘˜ = 3 this would leave us with (๐œ† − ๐œ†1 )๐›ผ1 (๐œ† − ๐œ†2 )๐›ผ2 (๐œ† − ๐œ†3 )๐›ผ3
If we set ๐œ† = 0 we get (−๐œ†1 )๐›ผ1 โˆ™ (−๐œ†2 )๐›ผ2 โˆ™ (−๐œ†3 )๐›ผ3
So det −๐ด = (−๐œ†1 )๐›ผ1 โˆ™ (−๐œ†2 )๐›ผ2 โˆ™ (−๐œ†3 )๐›ผ3 ⇒ (−1)๐‘› det ๐ด = (−1)๐‘› (๐œ†1 )๐›ผ1 (๐œ†2 )๐›ผ2 (๐œ†3 )๐›ผ3
So det ๐ด = (๐œ†1 )๐›ผ1 (๐œ†2 )๐›ผ2 (๐œ†3 )๐›ผ3
๐ด ∈ โ„‚๐‘›×๐‘›
๐‘›
๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐ด = ∑ ๐‘Ž๐‘–๐‘–
๐‘–=1
๐‘›×๐‘›
๐ด = ๐‘ˆ๐ฝ๐‘ˆ −1 ,
๐ด, ๐ต ∈ โ„‚
,
๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐ด๐ต = ๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐ต๐ด
๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐ด = ๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐‘ˆ(๐ฝ๐‘ˆ −1 ) = ๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’(๐ฝ๐‘ˆ −1 )๐‘ˆ = ๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐ฝ = ๐›ผ1 ๐œ†1 + โ‹ฏ + ๐›ผ๐‘˜ ๐œ†๐‘˜
๐ด11
0
Claim: det ๐ด = det ๐ด11 โˆ™ det ๐ด22
0
],
๐ด22
๐ด=[
๐ด11 ∈ โ„‚๐‘×๐‘ ,
๐ด22 ∈ โ„‚๐‘ž×๐‘ž
๐ด11 = ๐‘ˆ๐ฝ๐‘ˆ −1 , ๐ฝ upper triangular
๐ด22 = ๐‘‰๐ฝฬƒ๐‘‰ −1 , ๐ฝฬƒ upper triangular
det ๐ด11 = det ๐ฝ , det ๐ด22 = det ๐ฝฬƒ
๐‘ˆ๐ฝ๐‘ˆ −1
0
๐‘ˆ
๐ด=[
]=[
0
0
๐‘‰๐ฝฬƒ๐‘‰ −1
๐ฝ 0
So det ๐ด = det [
] = det ๐ฝ det ๐ฝฬƒ
0 ๐ฝฬƒ
0 ๐ฝ 0 ๐‘ˆ −1
][
][
๐‘‰ 0 ๐ฝฬƒ
0
๐‘ฅ11
๐‘ฅ22
๐ฝ 0
[
]=
0 ๐ฝฬƒ
๐‘ฅ33
๐‘ฆ11
๐‘ฆ22 ]
[
It is still upper triangular!
A more sophisticated claim:
๐ด=[
๐ด11 (๐‘ × ๐‘) ๐ด12 (๐‘ × ๐‘ž)
]
0(๐‘ž × ๐‘) ๐ด22 (๐‘ž × ๐‘ž)
Claim: det ๐ด = det ๐ด11 det ๐ด22
๐ด=[
๐ผ๐‘
0
0 ๐ด11
][
๐ด22 0
๐ด22
๐ผ๐‘ž ]
0 ]
๐‘‰ −1
So we now know that det ๐ด = det [
๐ผ๐‘
0
๐‘Ž11
๐‘Ž21
๐‘Ž31
0
[ 0
0
๐ด11
] det [ 0
๐ด22
๐‘Ž12
๐‘Ž22
๐‘Ž32
0
0
๐‘Ž13
๐‘Ž23
๐‘Ž33
0
0
|
|
|
|
|
๐ด22
๐ผ๐‘ž ]
๐‘Ž14
๐‘Ž24
๐‘Ž34
1
0
๐‘Ž15
๐‘Ž25
๐‘Ž35
0
1 ]
I can subtract rows from other rows and the determinant still stays the same. So I can zero out
all the rows in the right (since the rest of the values are zeroes)
So we get:
๐‘Ž11 ๐‘Ž12 ๐‘Ž13 | 0 0
๐‘Ž21 ๐‘Ž22 ๐‘Ž23 | 0 0
๐‘Ž31 ๐‘Ž32 ๐‘Ž33 | 0 0
0
0
0 | 1 0
0
0
0 | 0 1]
[
But now it’s of the form as in the previous claim.
๐ผ๐‘
0
๐ด11 ๐ด22
๐ด11 0
So det ๐ด = det [
] det [ 0
๐ผ๐‘ž ] = det ๐ด22 det [ 0 ๐ผ๐‘ž ] = det ๐ด22 det ๐ด11
0 ๐ด22
An even more complicated claim - Schur Complemnts:
๐ด
๐ด12
๐ด = [ 11
]
๐ด21 ๐ด22
(all blocks are squares)
If ๐ด22 is invertible, then:
−1
๐ผ๐‘
๐ผ ๐ด12 ๐ด−1
0
22 ๐ด11 − ๐ด12 ๐ด22 ๐ด21
๐ด=[๐‘
][
] [ −1
0
๐ผ๐‘ž
0
๐ด22 ๐ด22 ๐ด21
−1
From this it follows that det ๐ด = det(๐ด11 − ๐ด12 ๐ด22 ๐ด21 ) โˆ™ det ๐ด22
Suppose ๐ด11 is invertible
๐ผ๐‘
0 ๐ด11
0
๐ผ๐‘
๐ด=[
][
][
−1
−1
๐ด12 0
๐ด21 ๐ด11 ๐ผ๐‘ž 0 ๐ด22 − ๐ด21 ๐ด11
−1
det ๐ด = det ๐ด11 det(๐ด22 − ๐ด21 ๐ด11
๐ด12 )
1
0
๐ด= 0
0
[0
−1
๐ด11
๐ด22
]
๐ผ๐‘ž
2 0 | 0
1 2 | 0
0 1 | 0
0 0 | 2
0 0 | 2
๐ด๐‘ˆ = ๐‘ˆ −1 ๐ฝ
๐œ†1 = 1,
๐œ†2 = 2
๐›ผ1 = 3,
๐›ผ2 = 2
1
0
0
0
2]
0
]
๐ผ๐‘ž
๐œ†−1
2
0
|
0
1
0
๐œ†−1
2
|
0
0
0
0
๐œ†−1 |
0
0
det(๐œ†๐ผ − ๐ด) = det
=
−
−
−
+
−
−
0
0
0
| ๐œ†−2
0
0
0
|
2
๐œ† − 2]
[ 0
๐œ† − 1 −2
0
๐œ†−2
0
det [ 0
] = (๐œ† − 1)3 โˆ™ (๐œ† − 2)2
๐œ† − 1 −2 ] โˆ™ det [
−2 ๐œ† − 2
0
0
๐œ†−1
Expansions of Minors
If ๐ด ∈ โ„‚๐‘›×๐‘› let ๐ด{๐‘–,๐‘—} = determinant of (๐‘› − 1) × (๐‘› − 1) matrix which is obtained by erasing
row ๐‘– and column ๐‘—
1 2 3
๐ด = [4 5 6]
7 8 9
1 2
๐ด{2,3} = det [
] = 8 − 14 = −6
7 8
det ๐ด = ∑๐‘›๐‘–=1(−1)๐‘–+๐‘— โˆ™ ๐‘Ž๐‘–๐‘— ๐ด{๐‘–๐‘—} for each ๐‘—, ๐‘— = 1, … , ๐‘› (expansion along the j’th column)
det ๐ด = ∑๐‘›๐‘—=1(−1)๐‘–+๐‘— โˆ™ ๐‘Ž๐‘–๐‘— ๐ด{๐‘–๐‘—} for each ๐‘–, ๐‘– = 1, … , ๐‘› (expansion along the I’th row)
So if you have a row with a lot of zeroes for instance:
0 0 0 4
โˆ™ โˆ™ โˆ™ โˆ™
[
]
โˆ™ โˆ™ โˆ™ โˆ™
โˆ™ โˆ™ โˆ™ โˆ™
Then
๐‘›
det ๐ด = ∑(−1)๐‘–+๐‘— โˆ™ ๐‘Ž๐‘–๐‘— ๐ด{๐‘–๐‘—}
๐‘—=1
So if we choose ๐‘– = 1 we can get all zeroes but the 4th column.
Example:
๐‘Ž11 ๐‘Ž12 ๐‘Ž13
๐‘Ž1
Suppose that ๐ด = [๐‘Ž21 ๐‘Ž22 ๐‘Ž23 ] = [๐‘Ž2 ]
๐‘Ž31 ๐‘Ž32 ๐‘Ž33
๐‘Ž3
๐‘Ž1 = ๐‘Ž11 [1 0 0] + ๐‘Ž12 [0 1
So we can use linearity of the determinant:
1
0
0
0
det ๐ด = ๐‘Ž11 det [๐‘Ž21 ๐‘Ž22 ๐‘Ž23 ] + ๐‘Ž12 det [๐‘Ž21
๐‘Ž31 ๐‘Ž32 ๐‘Ž33
๐‘Ž31
0] + ๐‘Ž13 [0
1
๐‘Ž22
๐‘Ž32
0 1]
0
0
๐‘Ž23 ] + ๐‘Ž13 det [๐‘Ž21
๐‘Ž33
๐‘Ž31
0
๐‘Ž22
๐‘Ž32
1
๐‘Ž23 ]
๐‘Ž33
1
0
0 ๐‘ ๐‘ข๐‘๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘ก ๐‘กโ„Ž๐‘’ ๐‘“๐‘–๐‘Ÿ๐‘ ๐‘ก ๐‘Ÿ๐‘œ๐‘ค ๐‘š๐‘ข๐‘™๐‘ก๐‘–๐‘๐‘™๐‘–๐‘’๐‘‘ ๐‘“๐‘Ÿ๐‘œ๐‘š ๐‘กโ„Ž๐‘’ 2๐‘›๐‘‘ ๐‘Ž๐‘›๐‘‘ 3๐‘Ÿ๐‘‘
1) ๐‘Ž11 det [๐‘Ž21 ๐‘Ž22 ๐‘Ž23 ]
=
๐‘Ž31 ๐‘Ž32 ๐‘Ž33
1 0
0
๐‘Ž22 ๐‘Ž23
๐‘Ž11 det [0 ๐‘Ž22 ๐‘Ž23 ] = ๐‘Ž11 det [๐‘Ž
] = ๐‘Ž11 ๐ด{11}
32 ๐‘Ž33
0 ๐‘Ž32 ๐‘Ž33
0
1
0 ๐‘ ๐‘ข๐‘๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘ก ๐‘กโ„Ž๐‘’ ๐‘“๐‘–๐‘Ÿ๐‘ ๐‘ก ๐‘Ÿ๐‘œ๐‘ค ๐‘š๐‘ข๐‘™๐‘ก๐‘–๐‘๐‘™๐‘–๐‘’๐‘‘ ๐‘“๐‘Ÿ๐‘œ๐‘š ๐‘กโ„Ž๐‘’ 2๐‘›๐‘‘ ๐‘Ž๐‘›๐‘‘ 3๐‘Ÿ๐‘‘
2) ๐‘Ž12 det [๐‘Ž21 ๐‘Ž22 ๐‘Ž23 ]
=
๐‘Ž31 ๐‘Ž32 ๐‘Ž33
0 1 0
0
0 1
๐‘Ž21 ๐‘Ž23
๐‘Ž
0
๐‘Ž
๐‘Ž
๐‘Ž
๐‘Ž12 det [ 21
]=
23 ] = −๐‘Ž12 det [ 21
23 0] = −๐‘Ž12 det [๐‘Ž
31 ๐‘Ž33
๐‘Ž31 0 ๐‘Ž33
๐‘Ž31 ๐‘Ž33 0
−๐‘Ž12 ๐ด{12}
3) ๐‘Ž13 ๐ด{13}
Theorem: ๐ด ∈ โ„‚๐‘›×๐‘› , ๐ถ is the ๐‘› × ๐‘› matrix with ๐‘–๐‘— entries ๐‘๐‘–๐‘— = (−1)๐‘–+๐‘— ๐ด{๐‘—๐‘–} , then
๐ด๐ถ = (det ๐ด)๐ผ
๐ด 3×3
๐‘ฅ
๐‘ฆ
๐‘ง
๐‘Ž
๐‘Ž
๐‘Ž
det [ 21
22
23 ] = ๐‘ฅ๐ด{11} − ๐‘ฆ๐ด{12} + ๐‘ง๐ด{13}
๐‘Ž31 ๐‘Ž32 ๐‘Ž33
Suppose ๐‘ฅ = ๐‘Ž11 , ๐‘ฆ = ๐‘Ž12 , ๐‘ง = ๐‘Ž13 ⇒ det ๐ด = ๐‘Ž11 ๐ด{11} − ๐‘Ž12 ๐ด{12} + ๐‘Ž13 ๐ด{13}
Suppose ๐‘ฅ = ๐‘Ž21 , ๐‘ฆ = ๐‘Ž22 , ๐‘ง = ๐‘Ž23 ⇒ 0 = ๐‘Ž21 ๐ด{11} − ๐‘Ž22 ๐ด{12} + ๐‘Ž23 ๐ด{13}
Suppose ๐‘ฅ = ๐‘Ž31 , ๐‘ฆ = ๐‘Ž32 , ๐‘ง = ๐‘Ž33 ⇒ 0 = ๐‘Ž31 ๐ด{11} − ๐‘Ž32 ๐ด{12} + ๐‘Ž33 ๐ด{13}
๐‘Ž11
๐‘Ž
[ 21
๐‘Ž31
๐‘Ž12
๐‘Ž22
๐‘Ž32
๐‘Ž13 ๐ด{11}
det ๐ด
๐‘Ž23 ] [−๐ด{12} ] = [ 0 ]
๐‘Ž33 ๐ด{13}
0
(to be continued!)
We can do the same for the second row!
๐‘Ž11 ๐‘Ž12 ๐‘Ž13
๐‘ฆ
๐‘ง ] = ๐‘ฅ๐ด{21} − ๐‘ฆ๐ด{22} + ๐‘ง๐ด{23}
det [ ๐‘ฅ
๐‘Ž31 ๐‘Ž32 ๐‘Ž33
[๐‘ฅ ๐‘ฆ ๐‘ง] = ๐‘Ž1 ⇒ 0 = ๐‘Ž11 ๐ด{21} − ๐‘Ž12 ๐ด{22} + ๐‘Ž13 ๐ด{23}
[๐‘ฅ ๐‘ฆ ๐‘ง] = ๐‘Ž2 ⇒ det ๐ด = ๐‘Ž21 ๐ด{21} − ๐‘Ž22 ๐ด{22} + ๐‘Ž23 ๐ด{23}
[๐‘ฅ ๐‘ฆ ๐‘ง] = ๐‘Ž2 ⇒ 0 = ๐‘Ž31 ๐ด{21} − ๐‘Ž32 ๐ด{22} + ๐‘Ž33 ๐ด{23}
So we can fill in the matrix a bit more:
๐ด{11} −๐ด{21}
๐‘Ž11 ๐‘Ž12 ๐‘Ž13
det ๐ด
[๐‘Ž21 ๐‘Ž22 ๐‘Ž23 ] [−๐ด{12} ๐ด{22} ] = [ 0
๐‘Ž31 ๐‘Ž32 ๐‘Ž33
๐ด{13} −๐ด{23}
0
0
det ๐ด]
0
And if we do the same thing again for the third row we get:
๐‘Ž11
๐‘Ž
[ 21
๐‘Ž31
๐‘Ž12
๐‘Ž22
๐‘Ž32
๐ด{11}
๐‘Ž13
๐‘Ž23 ] [−๐ด{12}
๐‘Ž33
๐ด{13}
−๐ด{21}
๐ด{22}
−๐ด{23}
๐ด{31}
det ๐ด
−๐ด{32} ] = [ 0
๐ด{33}
0
0
0
det ๐ด
0 ]
0
det ๐ด
๐‘“ (๐‘ฅ) ๐‘“12 (๐‘ฅ)
๐น(๐‘ฅ) = [ 11
]
๐‘“21 (๐‘ฅ) ๐‘“22 (๐‘ฅ)
๐‘“(๐‘ฅ) = det ๐น(๐‘ฅ)
๐‘“(๐‘ฅ + ๐œ–) − ๐‘“(๐‘ฅ)
๐‘“ ′ (๐‘ฅ) = lim
ฯต→0
๐œ–
Shorthand notations: โƒ—โƒ—โƒ—
๐‘“1 = [๐‘“11
โƒ—โƒ—โƒ—2 = [๐‘“21
๐‘“12 ], ๐‘“
๐‘“22 ]
๐‘“11 (๐‘ฅ + ๐œ–) ๐‘“12 (๐‘ฅ + ๐œ–)
๐‘“ (๐‘ฅ) ๐‘“12 (๐‘ฅ)
] − det [ 11
]
๐‘“21 (๐‘ฅ + ๐œ–) ๐‘“22 (๐‘ฅ + ๐œ–)
๐‘“21 (๐‘ฅ) ๐‘“22 (๐‘ฅ)
=
๐œ–
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—โƒ— (๐‘ฅ + ๐œ–) − ๐‘“
โƒ—โƒ—โƒ—โƒ—1 (๐‘ฅ) + โƒ—โƒ—โƒ—
๐‘“ (๐‘ฅ)
๐‘“ (๐‘ฅ)
๐‘“
๐‘“1 (๐‘ฅ)
`
det [ 1 `
] + det [ 1
] − det [ 1 ]
โƒ—โƒ—โƒ—2 (๐‘ฅ + ๐œ–)
โƒ—โƒ—โƒ—
๐‘“2 (๐‘ฅ + ๐œ–)
๐‘“
๐‘“2 (๐‘ฅ)
det [
But we know that:
โƒ—โƒ—โƒ— (๐‘ฅ)
โƒ—โƒ—โƒ—1 (๐‘ฅ)
๐‘“
๐‘“
det [ 1
] = det det [
]
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—2 (๐‘ฅ + ๐œ–) − โƒ—โƒ—โƒ—
๐‘“2 (๐‘ฅ + ๐œ–)
๐‘“
๐‘“2 (๐‘ฅ) + โƒ—โƒ—โƒ—
๐‘“2 (๐‘ฅ)
So the entire formula is:
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—
๐‘“ (๐‘ฅ + ๐œ–) − โƒ—โƒ—โƒ—
๐‘“1 (๐‘ฅ)
๐‘“1 (๐‘ฅ)
det [ 1
] + det [
]
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—2 (๐‘ฅ + ๐œ–) − โƒ—โƒ—โƒ—
๐‘“2 (๐‘ฅ + ๐œ–)
๐‘“
๐‘“2 (๐‘ฅ)
Or
1
= det [ ๐œ–
1
โƒ—โƒ—โƒ—
๐‘“1 (๐‘ฅ + ๐œ–) − โƒ—โƒ—โƒ—
๐‘“1 (๐‘ฅ)
] + det [
] det [
0
โƒ—โƒ—โƒ—
๐‘“2 (๐‘ฅ)
0
โƒ—โƒ—โƒ—
๐‘“1 (๐‘ฅ)
1] det [
]=
โƒ—โƒ—โƒ—
๐‘“2 (๐‘ฅ + ๐œ–) − โƒ—โƒ—โƒ—
๐‘“2 (๐‘ฅ)
๐œ–
1
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—
๐‘“1 (๐‘ฅ + ๐œ–) − โƒ—โƒ—โƒ—
๐‘“1 (๐‘ฅ)
๐‘“1 (๐‘ฅ)
det [
+
det
]
[
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—2 (๐‘ฅ)]
๐œ–
๐‘“2 (๐‘ฅ + ๐œ–) − ๐‘“
โƒ—โƒ—โƒ—
๐‘“2 (๐‘ฅ)
๐œ–
′ (๐‘ฅ)
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—
๐‘“ (๐‘ฅ)
๐‘“
So as ๐œ– goes to zero it equals det [ 1
] + det [ 1
]
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—
๐‘“ ′ (๐‘ฅ)
๐‘“ (๐‘ฅ)
1
---- end of lesson 11
2
๐ด ∈ โ„‚๐‘›×๐‘›
๐ด{๐‘–,๐‘—} determinant of the (๐‘› − 1) × (๐‘› − 1) matrix that is obtained from ๐ด by erasing the ๐‘–’th
row and ๐‘—’th column
Let ๐ถ ∈ โ„‚๐‘›×๐‘›
๐‘๐‘–๐‘— = (−1)๐‘–+๐‘— ๐ด{๐‘—,๐‘–}
๐ด๐ถ = det ๐ด ๐ผ๐‘›
๐ถ
If det ๐ด ≠ 0 then ๐ด is invertible and ๐ด−1 = det ๐ด
Cramer’s rule
Let ๐ด be an invertible ๐‘› × ๐‘› matrix, consider ๐ด๐‘ฅ = ๐‘
๐‘›
๐‘›
๐‘—=1
๐‘—=1
๐ถ๐‘
1
1
⇒๐‘ฅ=๐ด ๐‘=
⇒ ๐‘ฅ๐‘– =
∑ ๐‘๐‘–๐‘— ๐‘๐‘— =
∑(−1)๐‘–+๐‘— ๐ด{๐‘—,๐‘–} ๐‘๐‘— =
det ๐ด
det ๐ด
det ๐ด
−1
det[๐‘Ž1 … ๐‘Ž๐‘–−1 ๐‘ ๐‘Ž๐‘–+1 … ๐‘Ž๐‘› ]
det ๐ด
det ๐ดฬƒ๐‘– where ๐‘Ž๐‘– is replaced by ๐‘
๐‘“
′ (๐‘ก)
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—1 (๐‘ก)
๐‘“1 (๐‘ก)
๐‘“
๐‘“1′ (๐‘ก)
โƒ—โƒ—โƒ—2 (๐‘ก)] =
= det [ โƒ—โƒ—โƒ—
๐‘“2′ (๐‘ก)] + det [๐‘“
๐‘“2 (๐‘ก) ] + det [โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—
โƒ—โƒ—โƒ—3 (๐‘ก)
โƒ—โƒ—โƒ—3 (๐‘ก)
๐‘“3′ (๐‘ก)
๐‘“
๐‘“
∑ ๐‘“1′ (๐‘ก)(−1)1+๐‘— ๐ด{1,๐‘—} + ∑ ๐‘“2′ (๐‘ก)(−1)2+๐‘— ๐ด{2,๐‘—} + ∑ ๐‘“3′ (๐‘ก)(−1)3+๐‘— ๐ด{3,๐‘—} =
∑ ๐‘“1,๐‘— (๐‘ก)๐‘๐‘—1 (๐‘ก) + ∑ ๐‘“2,๐‘— (๐‘ก)๐‘๐‘—2 (๐‘ก) + ∑ ๐‘“3,๐‘— (๐‘ก)๐‘๐‘—3 (๐‘ก) =
3
3 3
๐‘๐‘—๐‘– (๐‘ก)
′
′
(๐‘ก)๐‘”๐‘—๐‘– (๐‘ก)
๐‘“(๐‘ก) ∑ ∑ ๐‘“๐‘–,๐‘— (๐‘ก)
= ๐‘“(๐‘ก) ∑ ∑ ๐‘“๐‘–,๐‘—
๐‘“(๐‘ก)
๐‘–=1 ๐‘—=1
๐‘–=1 ๐‘—=1
๐‘“ ′ (๐‘ก)
′ (๐‘ก)
−1 }
๐‘“(๐‘ก)
3
= ๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’{๐น
๐น(๐‘ก)
A reminder: ๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐ด = ∑ ๐‘Ž๐‘–๐‘–
๐‘›
๐‘›
๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐ด๐ต = ∑ (∑ ๐‘Ž๐‘–๐‘— ๐‘๐‘—๐‘– ) = ๐‘ก๐‘Ÿ๐‘Ž๐‘๐‘’๐ต๐ด
๐‘–=1
๐‘—=1
0 0
]
0 1
To calculate eigenvalues find the root of the polynomial:
๐œ†
0
det(๐œ†๐ผ2 − ๐ด) = det [
] = ๐œ†(๐œ† − 1) ๐œ† = 0,1
0 ๐œ†−1
๐ด=[
0 0
]
0 1
๐‘ฃ2 ] = [๐‘ฃ1
๐ด๐‘‰ = ๐‘‰ [
๐ด[๐‘ฃ1
0
0
๐ด= 0
0
[−๐‘Ž0
1
0
0
0
−๐‘Ž1
๐‘ฃ2 ] [0 0] = [0๐‘ฃ1
0 1
0
1
0
0
−๐‘Ž2
1
0
In general ๐ด = [
0
−๐‘Ž0
0
0
1
0
−๐‘Ž3
1๐‘ฃ2 ]
0
0
0
1
−๐‘Ž4 ]
0
0
โ‹ฑ
0
]
0
1
… −๐‘Ž๐‘›−1
To calculate Eigenvalues, need to find roots of det(๐ด๐ผ๐‘› − ๐ด)
0
1
๐‘› = 2,
๐ด=[
]
−๐‘Ž0 −๐‘Ž1
๐œ†
−1
๐œ†๐ผ1 − ๐ด = [
]
๐‘Ž0 ๐œ† + ๐‘Ž1
det(๐œ†๐ผ3 − ๐ด) = ๐œ†(๐œ† + ๐‘Ž1 ) + ๐‘Ž0 = ๐‘Ž0 + ๐‘Ž1 ๐œ† + ๐œ†2
det(๐œ†๐ผ3 − ๐ด) = ๐‘Ž0 + ๐‘Ž1 ๐œ† + ๐‘Ž2 ๐œ†2 + ๐œ†3
๐‘›=5
๐œ†
0
๐œ†๐ผ5 − ๐ด = 0
0
[−๐‘Ž0
−1
๐œ†
0
0
−๐‘Ž1
0
−1
๐œ†
0
−๐‘Ž2
0
0
−1
๐œ†
−๐‘Ž3
0
0
0
−1
๐‘Ž4 + ๐œ†]
๐‘ข๐‘ ๐‘–๐‘›๐‘” ๐‘กโ„Ž๐‘’ ๐‘Ÿ๐‘–๐‘”โ„Ž๐‘š๐‘œ๐‘ ๐‘ก ๐‘๐‘œ๐‘™๐‘ข๐‘š๐‘›
(−1)(−1)4+5 ๐ด{45} + (๐‘Ž4 + ๐œ†)(−1)5×5 ๐ด{55} =
det(๐œ†๐ผ5 − ๐ด)
=
๐œ† −1
๐œ† −1
det [
] + (๐œ† + ๐‘Ž4 )๐œ†4
๐œ†
−1
๐‘Ž0 ๐‘Ž1 ๐‘Ž2 (๐‘Ž3 − ๐œ†) + ๐œ†
๐‘Ž0 + ๐‘Ž1 ๐œ† + ๐‘Ž2 ๐œ†2 + (๐‘Ž3 − ๐œ†)๐œ†3 + ๐œ†4 + ๐‘Ž4 ๐œ†4 + ๐œ†5
IF ๐‘Ž is an ๐‘› × ๐‘› companion matrix then:
1) det(๐œ†๐ผ๐‘› − ๐ด) = ๐‘Ž0 + ๐‘Ž1 ๐œ† + โ‹ฏ + ๐‘Ž๐‘›−1 ๐œ†๐‘›−1 + ๐œ†๐‘›
1
1
0
๐œ†
๐ด[
]=[
0
โ‹ฎ
−๐‘Ž0
๐œ†๐‘›−1
0
0
1
โ‹ฑ
0
๐œ†
][
]=
0
1
โ‹ฎ
… −๐‘Ž๐‘›−1 ๐œ†๐‘›−1
๐œ†
๐œ†2
โ‹ฎ
๐œ†๐‘›−1
[−(๐‘Ž0 + ๐‘Ž1 ๐œ†+. . +๐‘Ž๐‘›−1 ๐œ†๐‘›−1 )]
1
๐œ†
Denote ๐‘ฃ(๐œ†) = [
]
โ‹ฎ
๐œ†๐‘›−1
๐‘(๐œ†) = det(๐œ†๐ผ๐‘› − ๐ด)
0
โ‹ฎ
๐ด๐‘ฃ(๐œ†) = ๐œ†๐‘ฃ(๐œ†) − ๐‘(๐œ†) [ ]
0
1
Suppose ๐‘(๐œ†1 ) = 0
๐ด๐‘ฃ(๐œ†1 ) = ๐œ†1 ๐‘‰(๐œ†1 )
1
๐œ†
So [
] is an eigenvector
โ‹ฎ
๐œ†๐‘›−1
0
1
′ (๐œ†)
2๐œ†
Denote ๐‘ฃ
=
โ‹ฎ
[(๐‘› − 1)๐œ†๐‘›−2 ]
0
โ‹ฎ
๐ด๐‘ฃ ′ (๐œ†) = ๐‘ฃ(๐œ†) + ๐œ†๐‘ฃ ′ (๐œ†) − ๐‘′ (๐œ†) [ ]
0
1
๐‘(๐œ†) = (๐œ† − ๐œ†1 )2 ๐‘ž(๐œ†)
๐‘′ (๐œ†) = 2(๐œ† − ๐œ†1 )๐‘ž(๐œ†) + (๐œ† − ๐œ†1 )2 ๐‘ž′ (๐œ†)
So ๐‘′ (๐œ†1 ) = 0
1. ๐ด๐‘ฃ(๐œ†) = ๐œ†๐‘ฃ(๐œ†) − ๐‘(๐œ†)๐‘’๐‘›
2. ๐ด๐‘ฃ ′ (๐œ†) = ๐œ†๐‘ฃ ′ (๐œ†) + ๐‘ฃ(๐œ†) − ๐‘′ (๐œ†)๐‘’๐‘›
3. ๐‘Ž๐‘ฃ ′′ (๐œ†) = 2๐‘ฃ ′ (๐œ†) + ๐œ†๐‘ฃ ′′ (๐œ†) − ๐‘′′ (๐œ†)๐‘’๐‘›
4. ๐‘Ž๐‘ฃ ′′′ (๐œ†) = 3๐‘ฃ ′′ (๐œ†) + ๐œ†๐‘ฃ ′′ (๐œ†) − ๐‘′′′ (๐œ†)๐‘’๐‘›
Suppose ๐‘(๐œ†1 ) = ๐‘′ (๐œ†1 ) = ๐‘′′ (๐œ†1 ) = ๐‘′′′ (๐œ†1 ) = 0
Denote:
๐‘ฃ1 = ๐‘‰(๐œ†1 )
๐‘ฃ2 = ๐‘ฃ ′ (๐œ†1 )
๐‘ฃ ′′ (๐œ†1 )
2!
๐‘ฃ ′′′ (๐œ†1 )
๐‘ฃ4 =
3!
๐‘ฃ3 =
1⇒ ๐ด๐‘ฃ1 = ๐œ†1 ๐‘ฃ1
2⇒ ๐ด๐‘ฃ2 = ๐œ†1 ๐‘ฃ2 + ๐‘ฃ1
3⇒ ๐ด๐‘ฃ3 = ๐œ†1 ๐‘ฃ3 + ๐‘ฃ2
4⇒ ๐ด๐‘ฃ4 = ๐œ†1 ๐‘ฃ4 + ๐‘ฃ3
๐ด๐‘ฃ3 = 2๐‘ฃ2 + ๐œ†1 ๐‘ฃ ′′ (๐œ†1 )
๐ด[๐‘ฃ1
๐‘ฃ2
๐‘ฃ3
๐‘ฃ4 ] = [๐‘ฃ1
๐‘ฃ2
๐‘ฃ3
๐œ†1
๐‘ฃ4 ] [ 0
0
0
1
๐œ†1
0
0
0
1
๐œ†1
0
0
0
]
1
๐œ†1
2) ๐ด is invertible⇔ ๐‘Ž0 ≠ 0
3) The geometric multiplicity of each eigenvalue of ๐ด is equal to 1
๐œ†
−1
0
0
0
0
๐œ†
−1
0
0
0
๐œ†
−1
0
๐œ†๐ผ5 − ๐ด = 0
0
0
0
๐œ†
−1
[−๐‘Ž0 −๐‘Ž1 −๐‘Ž2 −๐‘Ž3 ๐‘Ž4 + ๐œ†]
Claim that for every ๐œ† ∈ โ„‚ − dim โ„›๐œ†๐ผ5 −๐ด ≥ 4
Or generally: dim โ„›๐œ†๐ผ๐‘› −๐ด ≥ ๐‘› − 1
Since we have 4 independent columns!! Maybe the first column is dependent, and
maybe not.
๐œ† is an eigenvalue, then
๐›ฝ = dim ๐’ฉ๐ด−๐œ†๐ผ ≥ 1
A 5 × 5 companion matrix
det(๐œ†๐ผ5 − ๐ด) = (๐œ† − ๐œ†1 )5 (๐œ† − ๐œ†2 )2 ,
Find an invertible matrix ๐‘ˆ, ๐ฝ in Jordan form such that:
๐ด๐‘ˆ = ๐‘ˆ๐ฝ
Since the geometrical multiplicity is one, then:
๐œ†1 1 0 | 0 0
0 ๐œ†1 1 | 0 0
0 0 ๐œ†1 | 0 0
๐ฝ=
− − − + − −
0 0 0 | ๐œ†2 0
[ 0 0 0 | 0 ๐œ†2 ]
For sure!
๐œ†1 ≠ ๐œ†2
1
๐œ†1
๐‘ˆ = (๐œ†1 )2
(๐œ†1 )3
[(๐œ†1 )4
0
1
2๐œ†1
3(๐œ†1 )2
4(๐œ†1 )3
0
0
1
3๐œ†1
6(๐œ†1 )2
|
1
|
๐œ†2
| (๐œ†2 )2
| (๐œ†2 )3
| (๐œ†2 )4
0
1
2๐œ†2
3(๐œ†2 )2
4(๐œ†2 )3 ]
๐‘ฅ2 = ๐‘Ž๐‘ฅ1 + ๐‘๐‘ฅ0
๐‘ฅ3 = ๐‘Ž๐‘ฅ2 + ๐‘๐‘ฅ1
๐‘ฅ4 = ๐‘Ž๐‘ฅ3 + ๐‘๐‘ฅ2
โ‹ฎ
๐‘ฅ๐‘› = ๐‘Ž๐‘ฅ๐‘›−1 + ๐‘๐‘ฅ๐‘›−2
๐‘ฅ๐‘›−2
๐‘ฅ๐‘› = [๐‘ ๐‘Ž] [๐‘ฅ ]
๐‘›−1
๐‘ฅ๐‘›−2
๐‘ฅ๐‘›−1 = [0 1] [๐‘ฅ ]
๐‘›−1
๐‘ฅ๐‘›−1
0 1 ๐‘ฅ๐‘›−2
[ ๐‘ฅ ]=[
][
]
๐‘ ๐‘Ž ๐‘ฅ๐‘›−1
๐‘›
Given ๐‘ฅ0 , ๐‘ฅ1 want a recipe for ๐‘ฅ๐‘›
0 1
Denote ๐ด = [
]
๐‘ ๐‘Ž
๐‘ฅ๐‘—
๐‘ข๐‘— = [๐‘ฅ + 1]
๐‘—
๐‘ฅ0
๐‘ข0 = [ ๐‘ฅ ]
1
๐‘ฅ1
๐‘ข1 = [๐‘ฅ ]
2
๐‘ข2 = ๐ด๐‘ข1 = ๐ด2 ๐‘ข0
๐‘ข3 = ๐ด๐‘ข2 = ๐ด3 ๐‘ข0
๐‘ข๐‘› = ๐ด๐‘› ๐‘ข0
So we take ๐ด, and we want to write it in a form such that ๐ด = ๐‘ˆ๐ฝ๐‘ˆ −1
๐‘ข๐‘› = ๐‘ˆ๐ฝ๐‘ˆ −1 ๐‘ข0
det(๐œ†๐ผ2 − ๐ด) =?
This is a companion matrix!
0 1
So let’s change the notation such that ๐ด = [
]
๐‘Ž0 ๐‘Ž1
So det(๐œ†๐ผ2 − ๐ด) = ๐‘Ž0 + ๐‘Ž1 ๐œ† + ๐œ†2 = −๐‘ − ๐‘Ž๐œ† + ๐œ†2
๐œ†
0
๐œ†
1
(๐œ† − ๐œ†1 )2 ⇒ [ 1
] ๐‘œ๐‘Ÿ [ 1
]
0 ๐œ†1
0 ๐œ†1
det(๐œ†๐ผ2 − ๐ด) = {
๐œ†
0
(๐œ† − ๐œ†1 )(๐œ† − ๐œ†2 ) ⇒ [ 1
]
0 ๐œ†2
๐œ†
0
But the case [ 1
] cannot happen since the geometric multiplicity is 1!
0 ๐œ†1
In the first case:
๐‘ข๐‘› = [
1
๐œ†1
0 ๐œ†1
][
1 0
1 1
][
๐œ†1 ๐œ†1
0 −1
] ๐‘ข0
1
1 ๐œ†1
][
๐œ†2 0
0 1
][
๐œ†2 ๐œ†1
1 −1
] ๐‘ข0
๐œ†2
Or in the second case:
๐‘ข๐‘› = [
--- end of lesson
1
๐œ†1
0
0
๐ด= 0
0
[−๐‘Ž0
1
0
0
0
−๐‘Ž1
0
1
0
0
−๐‘Ž2
0
0
1
0
−๐‘Ž3
0
0
0 5 × 5 companion matrix
1
−๐‘Ž4 ]
(1) det(๐œ†๐ผ5 − ๐ด) = ๐‘Ž0 + ๐‘Ž1 ๐œ† + โ‹ฏ + ๐‘Ž4 ๐œ†4 + ๐œ†5
(2) If ๐œ†๐‘— is an eigenvalue of ๐ด, then ๐›พ๐‘— =Geometrical multiplication= 1 ⇒
only one Jordan cell connected with ๐œ†๐‘—
det(๐œ†๐ผ5 − ๐ด) = (๐œ† − ๐œ†1 )3 (๐œ† − ๐œ†1 )2 ,
๐œ†1 ≠ ๐œ†2
Suppose we know that:
๐‘ฅ3 = ๐‘Ž๐‘ฅ0 + ๐‘๐‘ฅ1 + ๐‘๐‘ฅ2
๐‘ฅ4 = ๐‘Ž๐‘ฅ1 + ๐‘๐‘ฅ2 + ๐‘๐‘ฅ3
โ‹ฎ
๐‘ฅ๐‘›+3 = ๐‘Ž๐‘ฅ๐‘› + ๐‘๐‘ฅ๐‘›+1 + ๐‘๐‘ฅ๐‘›+2
๐‘ฅ0
๐‘ฅ3 = [๐‘Ž ๐‘ ๐‘] [๐‘ฅ1 ]
๐‘ฅ2
๐‘ฅ1
0 1 0 ๐‘ฅ0
[๐‘ฅ2 ] = [0 0 1] [๐‘ฅ1 ]
๐‘ฅ3
๐‘Ž ๐‘ ๐‘ ๐‘ฅ2
๐‘ฅ๐‘—
๐‘ฃ๐‘— = [๐‘ฅ๐‘—+1 ]
๐‘ฅ๐‘—+2
๐‘ฃ1 = ๐ด๐‘ฃ0
๐‘ฃ2 = ๐ด๐‘ฃ1 = ๐ด2 ๐‘ฃ0
๐‘ฃ๐‘› = ๐ด๐‘› ๐‘ฃ0
๐‘ฅ๐‘›
๐‘ฅ0
๐‘› ๐‘ฅ
๐‘ฅ
[ ๐‘›+1 ] = ๐ด [ 1 ]
๐‘ฅ๐‘›+2
๐‘ฅ2
๐ด is a companion matrix!
det(๐œ†๐ผ5 − ๐ด) = ๐‘Ž0 + ๐‘Ž1 ๐œ† + ๐‘Ž2 ๐œ†2 + ๐œ†3 = −๐‘Ž − ๐‘๐œ† − ๐‘๐œ†2 + ๐œ†3
3 cases:
(๐œ† − ๐œ†1 )(๐œ† − ๐œ†2 )(๐œ† − ๐œ†3 ) 3 ๐‘‘๐‘–๐‘ ๐‘ก๐‘–๐‘›๐‘๐‘ก ๐‘’๐‘–๐‘”๐‘’๐‘› ๐‘ฃ๐‘Ž๐‘™๐‘ข๐‘’๐‘ 
(๐œ† − ๐œ†1 )2 (๐œ† − ๐œ†2 ) 2 ๐‘‘๐‘–๐‘ ๐‘ก๐‘–๐‘›๐‘๐‘ก ๐‘’๐‘–๐‘”๐‘’๐‘› ๐‘ฃ๐‘Ž๐‘™๐‘ข๐‘’๐‘ 
det(๐œ†๐ผ5 − ๐ด) = {
(๐œ† − ๐œ†1 ) 1 ๐‘‘๐‘–๐‘ ๐‘ก๐‘–๐‘›๐‘๐‘ก ๐‘’๐‘–๐‘”๐‘’๐‘› ๐‘ฃ๐‘Ž๐‘™๐‘ข๐‘’
(1) ๐ฝ๐‘› = [
๐œ†1๐‘›
๐œ†๐‘›2
] , ๐ด = ๐‘ˆ๐ฝ๐‘ˆ −1
๐œ†๐‘›3
๐‘ฅ๐‘›
๐‘ฅ
[ ๐‘›+1 ] = ๐‘ฃ๐‘› = ๐‘ˆ๐ฝ๐‘› ๐‘ˆ −1 ๐‘ฃ0
๐‘‹๐‘›+2
๐‘ฅ๐‘› = [1 0 0]๐‘ฃ๐‘› = [1 0 0
]๐‘ˆ๐ฝ๐‘›
๐‘ˆ
−1
๐‘ฃ0 = [๐›ผ
๐›ฝ
๐œ†1๐‘›
๐›พ] [ 0
0
0
๐œ†๐‘›2
0
0 ๐‘‘
0 ] [๐‘’ ] =
๐œ†๐‘›3 ๐‘“
๐›ผ๐‘‘๐œ†1๐‘› + ๐›ฝ๐‘’๐œ†๐‘›2 + ๐›พ๐‘“๐œ†๐‘›3
๐‘ฅ๐‘› = ๐›ผฬƒ๐œ†๐‘›1 + ๐›ฝฬƒ ๐œ†๐‘›2 + ๐›พฬƒ๐›พ3๐‘›
We must be able to assume ๐‘Ž ≠ 0 otherwise, we are talking about second order and not
third order!
So assume ๐‘Ž ≠ 0
๐‘ฅ0 = ๐›ผฬƒ + ๐›ฝฬƒ + ๐›พฬƒ
๐‘ฅ1 = ๐›ผฬƒ๐œ†1 + ๐›ฝฬƒ ๐œ†2 + ๐›พฬƒ๐œ†3
๐‘ฅ1 = ๐›ผฬƒ๐œ†21 + ๐›ฝฬƒ ๐œ†22 + ๐›พฬƒ๐œ†23
1 1 1 ๐›ผฬƒ
๐‘ฅ0
๐œ†
๐‘ฅ
[ 1 ] = [ 1 ๐œ†2 ๐œ†3 ] [๐›ฝฬƒ ]
๐‘ฅ2
๐œ†12 ๐œ†22 ๐œ†23 ๐›พฬƒ
๐œ†1
(2) ๐ฝ = [ 0
0
๐œ†1
(3) ๐ฝ = [ 0
0
1
๐œ†1
0
1
๐œ†1
0
0
0]
๐œ†2
0
1]
๐œ†1
Detour: Analyze ๐ฝ๐‘› for case (3)
0 1 0
๐ฝ = ๐œ†1 ๐ผ3 + ๐‘,
๐‘ = [ 0 0 1]
0 0 0
๐‘˜
๐‘˜
๐‘˜
๐‘˜
๐‘˜
๐ฝ = (๐œ†1 ๐ผ3 + ๐‘) = ∑ ( ) (๐œ†๐ผ3 )๐‘˜−๐‘— ๐‘๐‘— = ( ) ๐œ†๐‘˜ ๐ผ + ( ) ๐œ†๐‘˜−1 ๐‘ + ( ) ๐œ†๐‘˜−2 ๐‘ 2 =
๐‘—
0
1
2
๐‘˜
๐‘˜
๐‘—=0
๐œ†1๐‘˜
๐‘˜
( ) ๐œ†1๐‘˜−1
1
0
๐œ†1๐‘˜
[0
0
And for case 2:
๐‘˜
( ) ๐œ†1๐‘˜−2
2
๐‘˜ ๐‘˜−1
( ) ๐œ†1
1
๐œ†1๐‘˜ ]
๐œ†1๐‘› ๐‘›๐œ†1๐‘›−1 0
๐‘ฅ๐‘› = [1 0 0]๐‘ˆ [ 0
๐œ†1๐‘›
0 ] = [๐›ผ
0
0
๐œ†๐‘›2
๐›ผ๐‘‘๐œ†1๐‘› + ๐›ผ๐‘’๐‘›๐œ†1๐‘›−1 + ๐›ฝ๐‘’๐œ†1๐‘› + ๐‘“๐œ†๐‘›2
๐‘ฅ๐‘› = ๐›ผฬƒ๐œ†๐‘›1 + ๐›ฝฬƒ ๐‘š๐œ†1๐‘› + ๐›พฬƒ๐œ†๐‘›2
๐‘ฅ๐‘› = ๐‘”๐œ†1๐‘› + โ„Ž๐‘›๐œ†1๐‘› + ๐‘™๐œ†๐‘›2
๐›ฝ
๐‘‘
๐›พ][ ] [ ๐‘’ ] = [๐›ผ
๐‘“
๐›ฝ
๐‘‘๐œ†1๐‘› + ๐‘’๐œ†1๐‘› ๐‘‘
๐›พ] [ ๐‘’๐œ†1๐‘› ] [ ๐‘’ ] =
๐‘“
๐‘“๐œ†1๐‘›
๐‘ฅ0 = ๐‘”
๐‘ฅ1 = ๐‘”๐œ†1 + โ„Ž๐œ†1 + ๐‘™๐œ†2
๐‘ฅ2 = ๐‘”๐œ†12 + โ„Ž๐œ†12 + ๐‘™๐œ†22
1
๐‘ฅ0
[๐‘ฅ1 ] = [๐œ†1
๐‘ฅ2
๐œ†12
0
๐œ†1
2๐œ†12
1 ๐‘”
๐œ†2 ] [ โ„Ž ]
๐œ†22 ๐‘™
Go back to case 3:
๐‘ฅ๐‘› = [1 0 0]๐‘ˆ
๐œ†1๐‘›
๐‘›๐œ†1๐‘›−1
0
๐œ†1๐‘›
0
[0
๐‘›
๐‘›
2 4
๐‘ฅ๐‘› = ๐‘”๐œ†1 + โ„Ž๐‘›๐œ†1 + ln ๐œ†1
๐‘›(๐‘› − 1) ๐‘›−2
๐œ†1
2
๐‘ˆ −1 ๐‘ฃ0
๐‘›๐œ†1๐‘›−1
๐œ†1๐‘›
]
๐‘ฅ0 = ๐‘”
๐‘ฅ1 = ๐‘”๐œ†1 + โ„Ž๐œ†1 + ๐‘™๐œ†1
๐‘ฅ2 = ๐‘”๐œ†12 + โ„Ž2๐œ†12 + ๐‘™4๐œ†12
1
๐‘ฅ0
๐‘ฅ
๐œ†
[ 1] = [ 1
๐‘ฅ2
๐œ†12
0
๐œ†1
2๐œ†12
0 ๐‘”
๐œ†1 ] [โ„Ž ]
4๐œ†12 ๐‘™
General Algorithm
๐‘ฅ5 = ๐‘Ž๐‘ฅ0 + ๐‘๐‘ฅ1 + ๐‘๐‘ฅ2 + ๐‘‘๐‘ฅ3 + ๐‘’๐‘ฅ4 ๐‘Ž ≠ 0
๐‘ฅ๐‘›+5 = ๐‘Ž๐‘ฅ๐‘› + ๐‘๐‘ฅ๐‘›+1 + ๐‘๐‘ฅ๐‘›+2 + ๐‘‘๐‘ฅ๐‘›+3 + ๐‘’๐‘ฅ๐‘›+4
๐‘—
(1) Replace ๐‘ฅ๐‘— by ๐œ† - ๐œ†๐‘›+5 = ๐‘Ž๐œ†๐‘› + ๐‘๐œ†๐‘›+1 + โ‹ฏ + ๐‘’๐œ†๐‘›+4 ⇒
โŸ5 − ๐‘Ž − ๐‘๐œ† − ๐‘๐œ†2 − ๐‘‘๐œ†3 − ๐‘’๐œ†4 = 0
๐œ†
๐‘(๐œ†)
(2) Calculate the roots of the polynomial.
This will give you the eigevalues of A.
(3) Suppose ๐‘(๐œ†) = (๐œ† − ๐œ†1 )3 (๐œ† − ๐œ†2 )2
๐‘ฅ๐‘› = ๐›ผ๐œ†1๐‘› + ๐›ฝ๐‘›๐œ†1๐‘› + ๐›พ๐‘›2 ๐œ†๐‘›1 + ๐›ฟ๐œ†๐‘›2 + ๐œ–๐‘›๐œ†๐‘›2
(4) Obtain ๐›ผ, … , ๐œ– from the given ๐‘ฅ0 , ๐‘ฅ1 , ๐‘ฅ2 , ๐‘ฅ3 , ๐‘ฅ4
Discrete Dynamical Systems
๐‘ฃ1 = ๐ด๐‘ฃ0
๐‘ฃ2 = ๐ด๐‘ฃ1
โ‹ฎ
๐‘ฃ๐‘› = ๐ด๐‘› ๐‘ฃ0
๐ด = ๐‘ˆ๐ฝ๐‘ˆ −1
๐‘ฃ๐‘› = ๐‘ˆ๐ฝ๐‘ˆ −1 ๐‘ฃ0
In such a problem, it is often easier to calculate ๐‘ˆ −1 ๐‘ฃ0 in one step, as apposed to calculating
๐‘ˆ −1 then ๐‘ˆ −1 ๐‘ฃ0
๐‘ค0 = ๐‘ˆ −1 ๐‘ฃ0
๐‘ˆ๐‘ค0 = ๐‘ฃ0
Solve this!
๐ต
6
[0
0
๐‘ค = ๐ต−1 ๐‘ข0
6 2 2
6
= [0 3 1] ,
๐‘ข0 = [ 0 ]
0 0 1
0
๐ต๐‘ค = ๐‘ข0
2 2 ๐‘ค1
6
5
3 1] [๐‘ค2 ] = [0] , ⇒ ๐‘ค = [0]
0 1 ๐‘ค3
0
0
๐œ†1
๐ฝ = [0
0
๐‘ฃ๐‘› = ๐‘ˆ
๐œ†1๐‘›
๐‘›๐œ†1๐‘›−1
0
[0
๐œ†1๐‘›
0
1 0
๐œ†1 1 ]
0 ๐œ†1
๐‘›(๐‘› − 1) ๐‘›−2
๐œ†1
2
๐‘ˆ −1 ๐ฝ
๐‘›๐œ†1๐‘›−1
๐œ†1๐‘›
]
Evaluate:
1
๐‘›2
๐‘ฃ๐‘›
๐‘› 2
๐œ†1 ๐‘›
=๐‘ˆ 0
[0
1
๐‘›๐œ†1
1
๐‘›2
0
๐‘›3 − ๐‘› 1
โˆ™
2๐‘›2 ๐œ†12
1
๐‘‰ −1 ๐‘ฃ0
๐‘›๐œ†1
1
]
๐‘›2
Differential Equations
๐‘ฅ ′′ (๐‘ก) = ๐‘Ž๐‘ฅ ′ (๐‘ก) + ๐‘๐‘ฅ(๐‘ก)
Let
๐‘ฅ1 (๐‘ก) = ๐‘ฅ(๐‘ก)
๐‘ฅ2 (๐‘ก) = ๐‘ฅ ′ (๐‘ก)
[
๐‘ฅ1′ (๐‘ก)
]=
๐‘ฅ2′ (๐‘ก)
๐‘ฅ1 ′
0
[๐‘ฅ ] = [
๐‘
2
1 ๐‘ฅ1
][ ]
๐‘Ž ๐‘ฅ2
๐‘ฅ ′ (๐‘ก) = ๐ด๐‘ฅ(๐‘ก)
๐ด2 ๐ด3
๐‘’ =๐ผ+๐ด+
+
+โ‹ฏ
2! 3!
2 2
๐‘ก ๐ด
๐‘’ ๐‘ก๐ด = ๐ผ + ๐‘ก๐ด +
+โ‹ฏ
2!
๐‘’ ๐ด ๐‘’ ๐ต = ๐‘’ ๐ด+๐ต if ๐ด๐ต = ๐ต๐ด!
๐ด
∞
๐‘’
๐ด+๐ต
=∑
๐‘˜=0
(๐ด + ๐ต)๐‘˜
๐ด๐‘—
๐ต๐‘™
=∑ ∑ =โ‹ฏ
๐‘˜!
๐‘—!
๐‘™!
(๐‘’ ๐‘ก๐ด ๐‘’ ๐œ–๐ด − ๐‘’ ๐‘ก๐ด ) ๐‘๐‘ฆ ๐‘’๐‘ฅ๐‘๐‘™๐‘’๐‘›๐‘Ž๐‘ก๐‘–๐‘œ๐‘› ๐‘๐‘’๐‘™๐‘œ๐‘ค
(๐‘’ ๐œ–๐ด − ๐ผ)
(๐‘’ (๐‘ก+๐œ–)๐ด − ๐‘’ ๐‘ก๐ด )
= lim
=
lim ๐‘’ ๐‘ก๐ด
= ๐‘’ ๐‘ก๐ด ๐ด
ฯต→∞
ฯต→∞
ฯต→∞
๐œ–
๐œ–
๐œ–
= ๐ด๐‘’ ๐‘ก๐ด
lim
Explanation:
๐œ– 2 ๐ด2
+โ‹ฏ
2!
(๐‘’ ๐œ–๐ด − ๐ผ)
๐œ– 2 ๐ด2
= ๐œ–๐ด +
+โ‹ฏ
๐œ–
2!
๐‘ฅ(๐‘ก) = ๐‘’ ๐‘ก๐ด ๐‘ข
๐‘ข is constant
๐‘’ ๐œ–๐ด = ๐ผ + ๐œ–๐ด +
๐‘ฅ ′ (๐‘ก) = ๐ด๐‘’ ๐‘ก๐ด ๐‘ข = ๐ด๐‘ฅ(๐‘ก)
๐‘ฅ(๐‘ก) = ๐‘’ ๐‘ก๐ด ๐‘ฅ(0)
If ๐ด = ๐‘ˆ๐ฝ๐‘ˆ −1
Claim:
๐‘’
๐‘ก๐ด
∞
∞
∞
๐‘˜=0
๐‘˜=0
๐‘˜=0
(๐‘ก๐ด)๐‘˜
(๐‘ก๐‘ˆ๐ฝ๐‘ˆ −1 )๐‘˜
(๐‘ก๐ฝ)๐‘˜
๐‘ˆ(๐‘ก๐ฝ)๐‘˜ ๐‘ˆ −1
=∑
=∑
=∑
= ๐‘ˆ (∑
) ๐‘ˆ −1 = ๐‘ˆ๐‘’ ๐‘ก๐ฝ ๐‘ˆ −1
๐‘˜!
๐‘˜!
๐‘˜!
๐‘˜!
๐‘ฅ1 (๐‘ก)
] = ๐‘ˆ๐‘’ ๐‘ก๐ฝ ๐‘ˆ −1 ๐‘ฅ(0)
๐‘ฅ2 (๐‘ก)
๐‘ฅ(๐‘ก) = [1 0]๐‘ฅ(๐‘ก) = [1 0]๐‘ˆ๐‘’ ๐‘ก๐ฝ ๐‘ˆ −1 ๐‘ฅ(0)
๐‘ฅ(๐‘ก) = [
Case 1: det(๐œ†๐ผ2 − ๐ด) = (๐œ† − ๐œ†1 )(๐œ† − ๐œ†2 ), ๐œ†1 ≠ ๐œ†2 , ๐ฝ = [
๐œ†1
0
0
]
๐œ†2
๐‘ก๐œ†1
0 ] ⇒ ๐‘ฅ(๐‘ก) = ๐›ผ๐‘’ ๐œ†1 ๐‘ก + ๐›ฝ๐‘’ ๐œ†2 ๐‘ก
๐‘’ ๐‘ก๐ฝ = [๐‘’
0
๐‘’ ๐‘ก๐œ†2
Case 2: det(๐œ†๐ผ2 − ๐ด) = (๐œ† − ๐œ†1 )2
๐œ†
0
๐œ†
1
๐ฝ=[ 1
] or [ 1
]
0 ๐œ†1
0 ๐œ†1
But the first form doesn’t exist!!! So we only have the second form
๐‘’ ๐‘ก๐ฝ = ๐‘’ ๐‘ก(๐œ†1 ๐ผ+๐‘) = ๐‘’ ๐œ†1 ๐‘ก๐ผ โˆ™ ๐‘’ ๐‘ก๐‘
But ๐‘ 2 is zero! So ๐‘’ ๐‘ก๐‘ = ๐ผ + ๐‘ก๐‘
Therefore: ๐‘’ ๐‘ก๐ฝ = ๐‘’ ๐œ†1 ๐‘ก๐ผ โˆ™ (๐ผ + ๐‘ก๐‘)
--- end of lesson
∞
๐ด
๐‘’ =∑
๐‘˜=0
๐ด๐‘˜
๐‘˜!
๐ด ๐ต
๐‘’ ๐‘’ = ๐‘’ ๐ด+๐ต if ๐ด๐ต = ๐ต๐ด
๐‘“(๐‘ก) = ๐‘’ ๐‘ก๐ด ๐‘ฃ ๐‘“ ′ (๐‘ก) = ๐ด๐‘’ ๐‘ก๐ด ๐‘ฃ = ๐ด๐‘“(๐‘ก)
Final solution of:
๐‘Ž0 ๐‘ฅ + ๐‘Ž1 ๐‘ฅ ′ (๐‘ก) + ๐‘Ž2 ๐‘ฅ ′′ (๐‘ก) + ๐‘Ž3 ๐‘ฅ ′′′ (๐‘ก) + ๐‘ฅ ๐ผ๐‘‰ (๐‘‡) = 0,
๐‘ฅ(0) = ๐‘1
๐‘ฅ ′ (0) = ๐‘2
๐‘ฅ ′′ (0) = ๐‘3
๐‘ฅ ′′′ (0) = ๐‘4
0≤๐‘ก≤๐‘‘
Strategy:
Set
๐‘ฅ1 (๐‘ก) = ๐‘ฅ(๐‘ก)
๐‘ฅ2 (๐‘ก) = ๐‘ฅ ′ (๐‘ก)
๐‘ฅ3 (๐‘ก) = ๐‘ฅ ′′ (๐‘ก)
๐‘ฅ4 (๐‘ก) = ๐‘ฅ ′′′ (๐‘ก)
๐‘ฅ1′ (๐‘ก)
๐‘ฅ2
๐‘ฅ1
′ (๐‘ก)
๐‘ฅ3
๐‘ฅ2
๐‘ฅ
๐‘ฅ(๐‘ก) = [๐‘ฅ ] ⇒ ๐‘ฅ ′ (๐‘ก) = 2′
=[
]=
๐‘ฅ4
๐‘ฅ3 (๐‘ก)
3
๐‘ฅ4
−๐‘Ž0 ๐‘ฅ − ๐‘Ž1 ๐‘ฅ ′ − ๐‘Ž2 ๐‘ฅ ′′ − ๐‘Ž3 ๐‘ฅ ′′′
[๐‘ฅ4′ (๐‘ก)]
๐‘ฅ2
๐‘ฅ3
[
]
๐‘ฅ4
−๐‘Ž0 ๐‘ฅ1 − ๐‘Ž1 ๐‘ฅ2 − ๐‘Ž2 ๐‘ฅ3 − ๐‘Ž3 ๐‘ฅ4
๐‘ฅ1 = ๐‘ฅ,
๐‘ฅ1′ = ๐‘ฅ ′ ,
๐‘ฅ2′ = ๐‘ฅ1′′ = ๐‘ฅ ′′
So in matrix notation:
๐‘ฅ1
0
1
0
0
๐‘ฅ2
0
0
1
0
[
] [๐‘ฅ ]
0
0
0
1
3
−๐‘Ž0 −๐‘Ž1 −๐‘Ž2 −๐‘Ž3 ๐‘ฅ4
โƒ—โƒ—โƒ—
๐‘ฅ ′ (๐‘ก) = ๐ด๐‘ฅ(๐‘ก)
๐‘ฅ (๐‘ก) = ๐‘’ ๐‘ก๐ด ๐‘ฅ(๐‘ก)
๐‘ฅ(๐‘ก) = [1 0 0
0]๐‘’ ๐‘ก๐ด ๐‘ฅ(0)
๐ด = ๐‘ˆ๐ฝ๐‘ˆ −1 where ๐ฝ is of Jordan form.
(1) Calculate Eigenvalues of ๐ด
These are the roots of the polynomial det(๐œ†๐ผ − ๐ด) = ๐‘Ž0 + ๐‘Ž1 ๐œ† + ๐‘Ž2 ๐œ†2 + ๐‘Ž3 ๐œ†3 + ๐œ†4
If all are different, then ๐ฝ is diagonal!
๐ด = ๐‘ˆ๐ฝ๐‘ˆ −1 ,
๐‘’ ๐‘ก๐ด = ๐‘ˆ๐‘’ ๐‘ก๐ด ๐‘ˆ−1
๐‘’ ๐œ†1 ๐‘ก
โ‹ฑ
๐‘ฅ(๐‘ก) = [1 0 0 0]๐‘ˆ [
] ๐‘ˆ −1 ๐‘ฅ(0)
โ‹ฑ
๐‘’ ๐œ†4 ๐‘ก
๐ด = ๐‘ˆ๐ฝ๐‘ˆ
๐‘’ ๐œ†1 ๐‘ก
๐‘ˆ[
−1
,
๐ด๐‘ˆ = ๐‘ˆ๐ฝ
๐‘’ ๐œ†1 ๐‘ก
โ‹ฑ
โ‹ฑ
] = [๐‘ข1
๐‘ข2
๐‘ข3
๐‘ข4 ] [
๐‘’ ๐œ†4 ๐‘ก
๐‘’ ๐‘ข1 + ๐‘’ ๐‘ข2 + ๐‘’ ๐œ†3 ๐‘ก ๐‘ข3 + ๐‘’ ๐œ†4 ๐‘ก ๐‘ข4
๐‘ฅ(๐‘ก) = ๐‘Ž๐‘’ ๐œ†1 ๐‘ก + ๐‘๐‘’ ๐œ†2 ๐‘ก + ๐‘๐‘’ ๐œ†3 ๐‘ก + ๐‘‘๐‘’ ๐œ†4 ๐‘ก
๐œ†1 ๐‘ก
โ‹ฑ
]=
โ‹ฑ
๐‘’ ๐œ†4 ๐‘ก
๐œ†2 ๐‘ก
1. Find the “eigenvalues” by substituting ๐‘’ ๐œ†๐‘ก into the given differential and then dividing
through by ๐‘’ ๐œ†๐‘ก
๐‘Ž0 ๐‘’ ๐œ†๐‘ก + ๐‘Ž1 ๐œ†๐‘’ ๐œ†๐‘ก + ๐‘Ž2 ๐œ†2 ๐‘’ ๐œ†๐‘ก + ๐‘Ž3 ๐œ†3 ๐‘’ ๐œ†๐‘ก + ๐œ†4 ๐‘’ ๐œ†๐‘ก = 0
After we cross out ๐‘’ ๐œ†๐‘ก we get the same polynomial as we got from the determinant!
2. Find the roots.
Case (1) – 4 distinct roots: ๐œ†1 , ๐œ†2 , ๐œ†3 , ๐œ†4
๐‘ฅ(๐‘ก) = ๐›ผ๐‘’ ๐œ†1 ๐‘ก + ๐›ฝ๐‘’ ๐œ†2 ๐‘ก + ๐›พ๐‘’ ๐œ†3 ๐‘ก + ๐›ฟ๐‘’ ๐œ†4 ๐‘ก
3. Calculate ๐›ผ, ๐›ฝ, ๐›พ, ๐›ฟ from the given initial conditions
2. Case (2) – ๐‘Ž0 + ๐‘Ž1 ๐œ† + โ‹ฏ + ๐œ†4 = (๐œ† − ๐œ†1 )4
๐œ†1
0
๐ฝ=[
0
0
1
๐œ†1
0
0
0
1
๐œ†1
0
0
0
]
1
๐œ†1
๐‘’ ๐‘ก๐ฝ = ๐‘’ ๐‘ก(๐œ†1 ๐ผ+๐‘)
0
0
Where ๐‘ = [
0
0
0 0
0 0
So ๐‘ 2 = [
0 0
0 0
1
0
0
0
1
0
0
0
0 0
1 0
]
0 1
0 0
0
0 0
1
0 0
] , ๐‘3 = [
0
0 0
0
0 0
0
0
0
0
1
0
] , ๐‘4 = 0
0
0
So:
1
๐‘ก
๐‘ก2
2
๐‘ก2๐‘2 ๐‘ก3๐‘3
+
) = ๐‘’ ๐‘ก๐œ†1 0 1 ๐‘ก
2!
3!
0 0 1
[0 0 0
0]๐‘ˆ๐ธ๐‘ก๐ฝ ๐‘ˆ −1 ๐‘ฅ(0) = ๐›ผ๐‘’ ๐‘ก๐œ†1 + ๐›ฝ๐‘ก๐‘’ ๐‘ก๐œ†2 + ๐›พ๐‘ก 2 ๐‘’ ๐‘ก๐œ†3 + ๐›ฟ๐‘ก 3 ๐‘’ ๐‘ก๐œ†4
๐‘’ ๐‘ก๐ฝ = ๐‘’ ๐‘ก(๐œ†1 ๐ผ+๐‘) = ๐‘’ ๐‘ก๐œ†1 ๐ผ ๐‘’ ๐‘ก๐‘ = ๐‘’ ๐‘ก๐œ†1 โˆ™ (๐ผ + ๐‘ก๐‘ +
๐‘ฅ(๐‘ก) = [1
0 0
2. case (3) ๐‘Ž0 + โ‹ฏ + ๐œ†4 = (๐œ† − ๐œ†1 )3 (๐œ† − ๐œ†2 ), ๐œ†1 ≠ ๐œ†2
๐›ผ๐‘’ ๐‘ก๐œ†1 + ๐›ฝ๐‘ก๐‘’ ๐‘ก๐œ†1 + ๐›พ๐‘ก 2 ๐‘’ ๐‘ก๐œ†1 + ๐›ฟ๐‘ก 3 ๐‘’ ๐‘ก๐œ†2
So far everything was homogeneous. But what happens if not?
๐‘Ž0 ๐‘ฅ(5) + ๐‘Ž1 ๐‘ฅ ′ (๐‘ก) + ๐‘ฅ ′′ (๐‘ก) = ๐‘“(๐‘ก),
๐‘ฅ1 = ๐‘ฅ,
๐‘ฅ2 = ๐‘ฅ1′
๐‘ก≥0
๐‘ฅ2
๐‘ฅ1
0
[๐‘ฅ ] = [๐‘“(๐‘ก) − ๐‘Ž ๐‘ฅ − ๐‘Ž ๐‘ฅ ] = [
−๐‘Ž
2
0 1
1 2
0
๐‘ฅ1
1
0
][ ] + [
]
๐‘“(๐‘ก)
−๐‘Ž1 ๐‘ฅ2
๐‘ฅ ′ (๐‘ก) = ๐ด๐‘ฅ(๐‘ก) + ๐ต๐‘ข(๐‘ก)
(๐‘’ −๐‘ก๐ด ๐‘ฅ)′ = −๐ด๐‘’ −๐‘ก๐ด ๐‘ฅ + ๐‘’ −๐‘ก๐ด ๐‘ฅ ′ = ๐‘’ −๐‘ก๐ด (๐‘ฅ ′ − ๐ด๐‘ฅ) = ๐‘’ −๐‘ก๐ด ๐ต๐‘ข(๐‘ก)
๐‘‘ −๐‘ ๐ด
(๐‘’ ๐‘ฅ(๐‘ )) = ๐‘’ −๐‘ ๐ด ๐ต๐‘ข(๐‘ )
๐‘‘๐‘ 
We can integrate both sides!
๐‘ก
๐‘ก
๐‘‘
∫ (๐‘’ −๐‘ ๐ด ๐‘ฅ(๐‘ )) โˆ™ ๐‘‘๐‘  = ∫ ๐‘’ −๐‘ ๐ด ๐ต๐‘ข(๐‘ ) โˆ™ ๐‘‘๐‘ 
๐‘‘๐‘ 
0
0
๐‘ก
๐‘ก
๐‘’ −๐‘ ๐ด ๐‘ฅ(๐‘ )| = ∫ ๐‘’ −๐‘ ๐ด ๐ต๐‘ข(๐‘ ) โˆ™ ๐‘‘๐‘ 
0
0
๐‘ก
๐‘ก
๐‘’ −๐‘ก๐ด ๐‘ฅ(๐‘ก) − ๐‘ฅ(0) = ∫ ๐‘’ −๐‘ ๐ด ๐ต๐‘ข(๐‘ ) โˆ™ ๐‘‘๐‘  ⇒ ๐‘ฅ(๐‘ก) = ๐‘’ ๐‘ก๐ด ๐‘ฅ(0) + ๐‘’ ๐‘ก๐ด (∫ ๐‘’ −๐‘ ๐ด ๐ต๐‘ข(๐‘ ) โˆ™ ๐‘‘๐‘ )
0
0
Norm Linear Spaces
Before we actually go into norm linear spaces, we need to prove some inequalities.
๐‘ก3
3!
๐‘ก2
2
๐‘ก
1]
Inequalities
(1)
๐‘Ž > 0,
๐‘ > 0,
1 1
๐‘  > 1, + = 1
๐‘ก ๐‘ 
๐‘ก > 1,
๐‘Ž๐‘  ๐‘๐‘ก
+
๐‘ 
๐‘ก
Proof: ๐‘  + ๐‘ก = ๐‘ก๐‘  ⇒ ๐‘ก๐‘  − ๐‘  − ๐‘ก + 1 = 1 ⇒ (๐‘ก − 1)(๐‘  − 1) = 1
Following are equivalent for a pair of numbers ๐‘ก, ๐‘  with ๐‘ก > 1, ๐‘  > 1
1 1
+ =1
๐‘ก ๐‘ 
(๐‘ก − 1)(๐‘  − 1) = 1
(๐‘ก − 1)๐‘  = ๐‘ก
(๐‘  − 1)๐‘ก = ๐‘ 
⇒ ๐‘Ž๐‘ <
TODO: Draw graph of ๐‘ฆ = ๐‘ฅ ๐‘ −1
2. (Holder’s Inequality)
1
1
If ๐‘  > 1, ๐‘ก > 1 and ๐‘  + ๐‘ก = 1
๐‘˜
1
๐‘ 
๐‘˜
1
๐‘ก
๐‘˜
๐‘ 
๐‘ก
∑|๐‘Ž๐‘— ๐‘๐‘— | ≤ (∑|๐‘Ž๐‘— | ) โˆ™ (∑|๐‘๐‘— | )
๐‘—=1
๐‘—=1
๐‘—=1
Inequality is obvious if right hand side is 0.
It suffices to focus on a case where the right hand side is strictly positive.
๐‘Ž๐‘—
๐›ผ๐‘— =
๐‘˜
1,
๐‘  ๐‘ 
(∑๐‘˜๐‘—=1|๐‘Ž๐‘— | )
๐‘˜
๐›ฝ๐‘— =
๐‘ 
๐‘˜
๐‘๐‘—
1
๐‘  ๐‘ก
๐‘˜
(∑๐‘—=1|๐‘๐‘— | )
๐‘ก
|๐›ผ๐‘— |
|๐›ฝ๐‘— |
1 1
∑|๐›ผ๐‘— โˆ™ ๐›ฝ๐‘— | ≤ ∑
+∑
= + =1
๐‘ 
๐‘ก
๐‘  ๐‘ก
๐‘—=1
๐‘˜
๐‘—=1
๐‘—=1
๐‘˜
๐‘ 
๐‘ 
∑๐‘˜๐‘—=1|๐‘Ž๐‘— |
|๐›ผ๐‘— |
∑|๐›ผ๐‘— | = ∑ ๐‘˜
=
=1
∑๐‘–=1|๐‘Ž๐‘– |๐‘  ∑๐‘˜๐‘–=1|๐‘Ž๐‘– |๐‘ 
๐‘ 
๐‘—=1
Similarly
๐‘˜
∑(
๐‘—=1
๐‘—=1
๐‘ก
∑|๐‘๐‘— |
=1
|๐‘Ž๐‘— |
(∑๐‘˜๐‘–=1|๐‘Ž๐‘–
1
|๐‘  )๐‘ 
โˆ™
|๐‘๐‘— |
1)
๐‘˜ |๐‘ |๐‘ก ๐‘ก
(∑๐‘™=1 ๐‘™ )
≤1⇒
๐‘˜
1
๐‘ 
๐‘˜
1
๐‘ก
๐‘˜
๐‘ 
๐‘ก
∑|๐‘Ž๐‘— ๐‘๐‘— | ≤ (∑|๐‘Ž๐‘— | ) โˆ™ (∑|๐‘๐‘— | )
๐‘—=1
๐‘—=1
๐‘—=1
So it’s proven.
Let’s prove something extra:
๐‘˜
๐‘˜
∑|๐‘Ž๐‘— ๐‘๐‘— | = (∑|๐‘Ž๐‘– |) โˆ™ (∑|๐‘๐‘– |)
๐‘–=1
๐‘–=1
3. Shwartz inequality:
๐‘˜
๐‘˜
๐‘˜
2
2
∑|๐‘Ž๐‘— ๐‘๐‘— | ≤ √∑|๐‘Ž๐‘— | โˆ™ √∑|๐‘๐‘— |
๐‘—=1
๐‘—=1
๐‘—=1
4.(Minkowski inequality) If ๐‘  ≥ 1, then
1
๐‘ 
๐‘˜
1
๐‘ 
๐‘˜
๐‘ 
1
๐‘ก
๐‘˜
๐‘ 
๐‘ก
(∑|๐‘Ž๐‘— + ๐‘๐‘— | ) ≤ (∑|๐‘Ž๐‘— | ) + (∑|๐‘๐‘— | )
๐‘—=1
๐‘—=1
๐‘—=1
If ๐‘  = 1, its easy.
Suppose ๐‘  > 1, consider:
๐‘˜
๐‘˜
๐‘˜
๐‘ 
∑|๐‘Ž๐‘— + ๐‘๐‘— | = ∑|๐‘Ž๐‘— + ๐‘๐‘— | โˆ™ |๐‘Ž๐‘— + ๐‘๐‘— |
๐‘—=1
๐‘˜
๐‘ −1
= ∑(|๐‘Ž๐‘— | + |๐‘๐‘— |) โˆ™ |๐‘Ž๐‘— + ๐‘๐‘— |
๐‘—=1
๐‘ −1
=
๐‘—=1
๐‘˜
∑|๐‘Ž๐‘— | โˆ™ |๐‘Ž๐‘— + ๐‘๐‘— |
๐‘—=1
โŸ
๐‘ −1
+ ∑|๐‘๐‘— | โˆ™ |๐‘Ž๐‘— + ๐‘๐‘— |
๐‘—=1
โŸ
(1)
๐‘ −1
(2)
1
1
By Holder’s inequality: ๐‘  + ๐‘ก = 1
1
๐‘  ๐‘ 
1
๐‘ก
๐‘˜
(๐‘ −1)๐‘ก
(1) ≤ (∑|๐‘Ž๐‘— | ) โˆ™ (∑|๐‘Ž๐‘— + ๐‘๐‘— |
1
๐‘  ๐‘ 
๐‘ 
๐‘—=1
๐‘ 
1
๐‘ก
(2) ≤ (∑|๐‘๐‘— | ) โˆ™ (∑๐‘˜๐‘—=1|๐‘Ž๐‘— + ๐‘๐‘— | )
๐‘˜
๐‘ 
1
๐‘  ๐‘ 
1
๐‘  ๐‘ 
1
๐‘ก
๐‘˜
๐‘ 
∑|๐‘Ž๐‘— + ๐‘๐‘— | ≤ {(∑|๐‘Ž๐‘— | ) + (∑|๐‘๐‘— | ) } (∑|๐‘Ž๐‘— + ๐‘๐‘— | )
๐‘—=1
--- end of lesson
๐‘ 
) = (∑|๐‘Ž๐‘— | ) โˆ™ (∑|๐‘Ž๐‘— + ๐‘๐‘— | )
๐‘—=1
1
๐‘ 
1
๐‘ก
๐‘˜
๐‘—=1
Normed Linear Spaces (NLS)
Holder’s inequality:
1
1
If ๐‘  > 1, ๐‘ก > 1 and ๐‘  + ๐‘ก = 1
๐‘˜
1
๐‘ 
๐‘˜
1
๐‘ก
๐‘˜
๐‘ 
๐‘ก
∑|๐‘Ž๐‘— ๐‘๐‘— | ≤ (∑|๐‘Ž๐‘— | ) โˆ™ (∑|๐‘๐‘— | )
๐‘—=1
๐‘—=1
๐‘—=1
Cauchy-Shwartz inequality:
๐‘˜
๐‘˜
๐‘˜
2
2
∑|๐‘Ž๐‘— ๐‘๐‘— | ≤ √∑|๐‘Ž๐‘— | โˆ™ √∑|๐‘๐‘— |
๐‘—=1
๐‘—=1
๐‘—=1
(Minkowski inequality) If ๐‘  ≥ 1, then
1
๐‘ 
๐‘˜
1
๐‘ 
๐‘˜
๐‘ 
1
๐‘ก
๐‘˜
๐‘ 
๐‘ก
(∑|๐‘Ž๐‘— + ๐‘๐‘— | ) ≤ (∑|๐‘Ž๐‘— | ) + (∑|๐‘๐‘— | )
๐‘—=1
๐‘—=1
๐‘—=1
A vector space ๐’ฑ is said to be a normed-linear space over ๐”ฝ if for each vector ๐‘ฃ ∈ ๐’ฑ there is
๐œ‘(๐‘ฃ) with the following properties:
1) ๐œ‘(๐‘ฃ) ≥ 0
2) ๐œ‘(๐‘ฃ) = 0 ⇔ ๐‘ฃ = 0
3) ๐œ‘(๐›ผ๐‘ฃ) = |๐›ผ| โˆ™ ๐œ‘(๐‘ฃ)
4) ๐œ‘(๐‘ข + ๐‘ฃ) ≤ ๐œ‘(๐‘ข) + ๐œ‘(๐‘ฃ) (triangle inequality)
Example 1:
๐’ฑ = โ„‚๐‘› , fix ๐‘  ≥ 1
1
Define ๐œ‘(๐‘ฅ) = (∑๐‘›๐‘–=1|๐‘ฅ๐‘– |๐‘  )๐‘ 
Disclaimer: I don’t have the mental forces to copy the proof why ๐œ‘ has all properties.
Example 2: ๐’ฑ = โ„‚๐‘›
Define ๐œ‘(๐‘ฅ) = max{|๐‘ฅ1 |, … , |๐‘ฅ๐‘› |}
Disclaimer 2: See disclaimer 1.
A fundamental fact that on a finite dimension normed-linear space all norms are equivalent.
i.e. If ๐œ‘(๐‘ฅ) is a norm, and ๐œ“(๐‘ฅ) is a norm, then can find constants ๐›ผ1 > 0, ๐›ผ2 > 0 such that:
๐›ผ1 ๐œ‘(๐‘ฅ) ≤ ๐œ“(๐‘ฅ) ≤ ๐›ผ2 ๐œ‘(๐‘ฅ)
It might seem not symmetric. But in fact, the statement implies:
๐œ“(๐‘ฅ)
๐œ“(๐‘ฅ)
๐œ‘(๐‘ฅ) =
,
๐œ‘(๐‘ฅ) ≥
๐›ผ
๐›ผ2
So:
๐œ“(๐‘ฅ)
๐›ผ2
๐œ“(๐‘ฅ)
๐›ผ1
≤ ๐œ‘(๐‘‹) ≤
The three most important norms are: โ€–๐‘ฅโ€–1 , โ€–๐‘ฅโ€–2 , โ€–๐‘ฅโ€–∞
โ€–๐‘ฅโ€–∞ ≤ โ€–๐‘ฅโ€–2 ≤ โ€–๐‘ฅโ€–1 ≤ ๐›พโ€–๐‘ฅโ€–∞
๐‘ฅ1
๐‘ฅ=[ โ‹ฎ ]
๐‘ฅ๐‘›
๐‘›
|๐‘ฅ๐‘– | =
√|๐‘ฅ|2
2
≤ √∑|๐‘ฅ๐‘— | ⇒ |๐‘ฅ๐‘– | ≤ โ€–๐‘ฅโ€–2 ⇒ max|๐‘ฅ๐‘– | ≤ โ€–๐‘ฅโ€–2
๐‘—=1
(โ€–๐‘ฅโ€–2 )2 = |๐‘ฅ1 |2 + โ‹ฏ + |๐‘ฅ๐‘› |2 ≤ |๐‘ฅ1 |(|๐‘ฅ1 | + โ‹ฏ + |๐‘ฅ๐‘› |) + โ‹ฏ + |๐‘ฅ๐‘› |(|๐‘ฅ1 | + โ‹ฏ + |๐‘ฅ๐‘› |) =
(|๐‘ฅ1 | + โ‹ฏ + |๐‘ฅ๐‘› |)2
But since it was normed-negative we are allowed to take the root and therefore:
โ€–๐‘ฅโ€–2 ≤ โ€–๐‘ฅโ€–1
๐‘›
๐‘›
โ€–๐‘ฅโ€–1 = ∑|๐‘ฅ๐‘– | ≤ ∑โ€–๐‘ฅ๐‘– โ€–∞ = ๐‘›โ€–๐‘ฅ๐‘– โ€–∞
๐‘–=1
๐‘–=1
(so in our case ๐‘› = ๐›พ)
๐‘›
1
๐‘ 
โ€–๐‘ฅโ€–๐‘  = (∑|๐‘ฅ๐‘– |) ,
1≤๐‘ <∞
๐‘–=1
โ€–๐‘ฅโ€–∞ = max|๐‘ฅ๐‘– |
If ๐’ฑ is a vector space of functions defined on a finite interval ๐‘Ž ≤ ๐‘ก ≤ ๐‘
๐‘
1
๐‘ 
โ€–๐‘“โ€–๐‘  = (∫|๐‘“(๐‘ก)|๐‘  ๐‘‘๐‘ก)
๐‘Ž
โ€–๐‘ฅโ€–∞ = max|๐‘“(๐‘ก)| ,
๐‘Ž≤๐‘ก≤๐‘
Matrix Norms (Transformation Norms)
Let ๐ด ∈ โ„‚๐‘×๐‘ž , โ€–๐ดโ€– = max
โ€–๐ด๐‘ฅ โ€–2
โ€–๐‘ฅโ€–2
, ๐‘ฅ ∈ โ„‚๐‘ž , ๐‘ฅ ≠ 0
0≤๐‘ฅ<1
What is the maximum value of ๐‘ฅ in this interval? There’s no maximum value.
Lemma: Let ๐ด ∈ โ„‚๐‘×๐‘ž and let
โ€–๐ด๐‘ฅโ€–2
๐›ผ1 = max {
|๐‘ฅ ∈ โ„‚๐‘ž , ๐‘ฅ ≠ 0}
โ€–๐‘ฅโ€–2
๐›ผ2 = max{โ€–๐ด๐‘ฅโ€–2|๐‘ฅ ∈ โ„‚๐‘ž , โ€–๐‘ฅโ€–2 = 1}
๐›ผ3 = max{โ€–๐ด๐‘ฅโ€–2|๐‘ฅ ∈ โ„‚๐‘ž , โ€–๐‘ฅโ€–2 ≤ 1}
Then ๐›ผ1 = ๐›ผ2 = ๐›ผ3
โ€–๐ด๐‘ฅโ€–2
โ€–๐‘ฅโ€–2
Take ๐‘ฅ ∈ โ„‚๐‘ž , ๐‘ฅ ≠ 0 and consider
โ€–๐ด๐‘ฅโ€–2
๐‘ฅ
= โ€–๐ด โˆ™
โ€– ≤ ๐›ผ2 ≤ ๐›ผ3
โ€–๐‘ฅโ€–2
โ€–๐‘ฅโ€–2 2
โ€–๐‘ฅโ€–
๐‘ฅ
Now we know that โ€–โ€–๐‘ฅโ€– โ€– = โ€–๐‘ฅโ€–2 = 1
2
2
So we get: ๐›ผ1 ≥ ๐›ผ2 ≥ ๐›ผ3
Take ๐‘ฅ ∈ โ„‚๐‘ž , โ€–๐‘ฅโ€–2 ≤ 1, ๐‘ฅ ≠ 0
โ€–๐ด๐‘ฅโ€–2 =
โ€–๐ด๐‘ฅโ€–2
,
โ€–๐‘ฅโ€–2
โ€–๐‘ฅโ€–2 ≤
โ€–๐ด๐‘ฅโ€–2
≤ ๐›ผ1
โ€–๐‘ฅโ€–2
And for ๐‘ฅ = 0 it’s a special case where the inequality is true as well.
So we get equality.
โ€–๐ด๐‘ฅโ€–
โ„‚๐‘×๐‘ž is a normed linear space in respect to the norm โ€–๐ดโ€– = max { โ€–๐‘ฅโ€– 2 |๐‘ฅ ∈ โ„‚๐‘ž , ๐‘ฅ ≠ 0}
2
It’s clear ๐œ‘(๐ด) ≥ 0
Is ๐œ‘(๐ด) = 0 ⇒ every vector we put in there equals to zero.
๐œ‘(๐›ผ๐ด) = |๐›ผ|๐œ‘(๐ด)
Lemma: โ€–๐ด๐‘ฅโ€–2 ≤ ๐›ผ1 โ€–๐‘ฅโ€–2
Proof: ๐‘ฅ ≠ 0,
โ€–๐ด๐‘ฅโ€–2
โ€–๐‘ฅโ€–2
≤ ๐›ผ ⇒ โ€–๐ด๐‘ฅโ€–2 ≤ ๐›ผ1 โ€–2โ€–2
If ๐‘ฅ = 0 then it’s also true. So true for all ๐‘ฅ.
๐œ‘(๐ด + ๐ต) ≤
โ€–๐ด + ๐ต๐‘ฅโ€–2 โ€–๐ด๐‘ฅโ€–2 โ€–๐ต๐‘ฅโ€–2
≤
+
≤ ๐œ‘(๐ด) + ๐œ‘(๐ต)
โ€–๐‘ฅโ€–2
โ€–๐‘ฅโ€–2
โ€–๐‘ฅโ€–2
โ„‚๐‘×๐‘ž is a normed-linear space with respect to the norm:
โ€–๐ด๐‘ฅโ€–2
โ€–๐ดโ€– = max {
|๐‘ฅ ≠ 0}
โ€–๐‘ฅโ€–2
Can also show (in a very similar manner)
โ€–๐ด๐‘ฅโ€–๐‘ 
โ€–๐ดโ€–๐‘ ,๐‘  = max {
|๐‘ฅ ≠ 0}
โ€–๐‘ฅโ€–๐‘ 
โ€–๐ดโ€–๐‘ ,๐‘ก = max {
โ€–๐ด๐‘ฅโ€–๐‘ก
|๐‘ฅ ≠ 0}
โ€–๐‘ฅโ€–๐‘ 
Lemma:
Let ๐ด ∈ โ„‚๐‘×๐‘ž , ๐ต ∈ โ„‚ ∈๐‘ž×๐‘Ÿ
โ€–๐ด๐ตโ€– ≤ โ€–๐ดโ€– โˆ™ โ€–๐ตโ€–
Proof: โ€–๐ด๐ต๐‘ฅโ€–2 ≤ โ€–๐ดโ€–โ€–๐ต๐‘ฅโ€–2 ≤ โ€–๐ดโ€–โ€–๐ตโ€–โ€–๐‘ฅโ€–2
If ๐‘ฅ ≠ 0
โ€–๐ด๐ต๐‘ฅโ€–2
โ€–๐‘ฅโ€–2
≤ โ€–๐ดโ€–โ€–๐ตโ€– ⇒ โ€–๐ด๐ตโ€– ≤ โ€–๐ดโ€–โ€–๐ตโ€–
Theorem: ๐ด ∈ โ„‚๐‘›×๐‘› suppose ๐ด invertible.
If ๐ต ∈ โ„‚๐‘›×๐‘› and ๐ต is “close enough” to ๐ด then ๐ต is invertible.
Proof: Let ๐ต = ๐ด − (๐ด − ๐ต) = ๐ด(๐ผ − ๐ด−1 (๐ด − ๐ต)) = ๐ด (๐ผ −
๐‘‹
โŸ
)
๐ด−1 (๐ด−๐ต)
If โ€–๐‘‹โ€– < 1, then ๐ผ − ๐‘‹ is invertible.
Enough to show ๐’ฉ๐ผ−๐‘‹ = {0}
Let ๐‘ข ∈ ๐’ฉ๐ผ−๐‘‹ ⇒ (๐ผ − ๐‘‹)๐‘ข = 0 ⇒ ๐‘ข = ๐‘‹๐‘ข ⇒ โ€–๐‘ขโ€–2 = โ€–๐‘‹๐‘ขโ€–2 ≤ โ€–๐‘‹โ€–โ€–๐‘ขโ€–2
(1 − โ€–๐‘‹โ€–) โˆ™ โ€–๐‘ขโ€–2 ≤ 0 ⇒ โ€–๐‘ขโ€–2 ≤ 0
So โ€–๐‘ขโ€– = 0 ⇒ ๐‘ข = 0.
--- end of lesson
NLS –normed linear spaces
๐ด ∈ โ„‚๐‘×๐‘ž , ๐ต ∈ โ„‚๐‘ž×๐‘Ÿ
โ€–๐ด๐ตโ€–2,2 ≤ โ€–๐ดโ€–2,2 โˆ™ โ€–๐ตโ€–2,2
โ€–๐ด๐‘ฅโ€–2,2 ≤ โ€–๐ดโ€–2,2 โˆ™ โ€–๐‘ฅโ€–2
โ€–๐ด๐‘ฅโ€–๐‘ก
โ€–๐ดโ€–๐‘ ,๐‘ก = max {
|๐‘ฅ ≠ 0}
โ€–๐‘ฅโ€–๐‘ 
โ€–๐ดโ€–2,2 = max{โ€–๐ด๐‘ฅโ€–2,2 : โ€–๐‘ฅโ€–2 = 1} = max{โ€–๐ดโ€–2,2 : โ€–๐‘ฅโ€–2 ≤ 1}
Theorem: If ๐ด ∈ โ„‚๐‘×๐‘ that is invertible and ๐ต ∈ โ„‚๐‘×๐‘ that is close enough to ๐ด, then ๐ต is also
invertible.
First idea: ๐‘ฅ ∈ โ„‚๐‘×๐‘ , โ€–๐‘ฅโ€– < 1, then ๐ผ๐‘ − ๐‘‹ is invertible.
Let ๐‘ข ∈ ๐’ฉ๐ผ๐‘ −๐‘‹ ⇒ (๐ผ๐‘ − ๐‘‹)๐‘ข = 0 ⇒ ๐‘ข = ๐‘‹๐‘ข
⇒ โ€–๐‘ขโ€– = โ€–๐‘‹๐‘ขโ€– ⇒ โ€–๐‘‹๐‘ขโ€– ≤ โ€–๐‘‹โ€– โˆ™ โ€–๐‘ขโ€– ⇒ (1 − โ€–๐‘‹โ€–)โ€–๐‘ขโ€– ≤ 0 ⇒ โ€–๐‘ขโ€– ≤ 0 ⇒ ๐‘ข = 0
๐›ผ∈โ„‚
๐‘†๐‘› = 1 + ๐›ผ + ๐›ผ 2 + โ‹ฏ + ๐›ผ ๐‘›
๐›ผ๐‘†๐‘› = ๐›ผ + ๐›ผ 2 + โ‹ฏ + ๐›ผ ๐‘› + ๐›ผ ๐‘›+1
(1 − ๐›ผ)๐‘†๐‘› = 1 − ๐›ผ ๐‘›+1
1 − ๐›ผ ๐‘›+1
1
๐‘†๐‘› = 1 + ๐›ผ + โ‹ฏ + ๐›ผ ๐‘› =
→
1−๐›ผ
1−๐›ผ
∞
1
|๐›ผ| < 1 ∑ ๐›ผ ๐‘— =
1−๐›ผ
๐‘—=1
So as ๐‘› ↑ ∞
๐‘ฅ ๐‘›+1 → 0
− 0โ€– → 0
โ€–๐‘ฅ ๐‘›+1
So โ€–๐‘ฅ ๐‘›+1 โ€– ≤ โ€–๐‘ฅโ€–๐‘›+1
If โ€–๐‘ฅโ€– < 1 ⇒ (๐ผ − ๐‘‹)๐‘†๐‘› = ๐ผ๐‘ − ๐‘‹ ๐‘›+1 → ๐ผ๐‘ ๐‘Ž๐‘  ๐‘› ↑ ∞
∞
(๐ผ − ๐‘‹)
−1
∞
−1
๐‘—
= ∑ ๐‘ฅ ๐‘–. ๐‘’. โ€–∑ ๐‘ฅ ๐‘— − (๐ผ๐‘ − ๐‘ฅ) โ€– → 0
๐‘—=0
๐‘—=0
Theorem: If ๐ด ∈ โ„‚๐‘×๐‘ž and ๐ด is left invertible, ๐ต ∈ โ„‚๐‘×๐‘ž and ๐ตis close enough to ๐ด, then ๐ต is left
invertible.
Proof: ๐ด is left invertible ⇒ there is a ๐ถ ∈ โ„‚๐‘ž×๐‘ such that ๐ถ๐ด = ๐ผ๐‘ž
๐ต = ๐ด − (๐ด − ๐ต) ⇒ ๐ถ๐ต = ๐ถ๐ด − โŸ
๐ถ(๐ด − ๐ต) = ๐ผ๐‘ž − ๐‘‹
๐‘‹∈โ„‚๐‘ž×๐‘ž
If โ€–๐‘‹โ€– < 1 ⇒ ๐ผ๐‘ž − ๐‘‹ is invertible.
Doing that, we find (๐ผ๐‘ž − ๐‘‹)
−1
๐ถ๐ต = ๐ผ๐‘ž
⇒ (๐ผ๐‘ž − ๐‘‹)๐ถ is left inverse of ๐ต, i.e. ๐ต is left invertible.
๐‘‹ = ๐ถ(๐ด − ๐ต) ⇒ โ€–๐‘‹โ€– ≤ โ€–๐ถโ€– โˆ™ โ€–๐ด − ๐ตโ€–
If โ€–๐ด − ๐ตโ€– <
1
โ€–๐ถโ€–
⇒ โ€–๐‘‹โ€– < 1
We can then prove symmetrically that if it is right invertible then any close enough ๐ต is also
right invertible.
๐ด ∈ โ„‚๐‘×๐‘ž
๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด = ๐‘ or ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด = ๐‘ž
๐ต close to ๐ด it will have the same rank as ๐ด.
Question: ๐‘Ÿ๐‘Ž๐‘›๐‘˜๐ด = ๐‘˜, 1 ≤ ๐‘˜ < min{๐‘, ๐‘ž}
Can we prove that if ๐ต is close enough to ๐ด, then rank ๐ต = ๐‘˜.
2
Observation: ∈ โ„‚๐‘×๐‘ž , ๐ด = [๐‘Ž๐‘–๐‘— ] , then |๐‘Ž๐‘–๐‘— | ≤ โ€–๐ดโ€–2,2 ≤ √∑๐‘–,๐‘—|๐‘Ž๐‘–๐‘— | ⇒
|๐‘Ž๐‘–๐‘— − ๐‘๐‘–๐‘— | ≤ โ€–๐ด − ๐ตโ€–2,2
Proof: โ€–๐ด๐‘ฅโ€–22
๐‘
=
2
∑๐‘๐‘–=1|∑๐‘ž๐‘—=1 ๐‘Ž๐‘–๐‘— ๐‘ฅ๐‘— |
1
2
๐‘ž
1
=
2
∑๐‘๐‘–=1 {(∑๐‘ž๐‘—=1|๐‘Ž๐‘–๐‘— | )2
1
โˆ™
2 2
๐‘ž
(∑๐‘—=1|๐‘ฅ๐‘— | ) }
2
≤
2
2
โ€–๐‘ฅโ€–22 ∑ (∑|๐‘Ž๐‘–๐‘— | )
๐‘–=1
๐‘—=1
{
}
But โ€–๐ดโ€–2,2 = max{โ€–๐ด๐‘ฅโ€–2|โ€–๐‘ฅโ€–2 = 1}
So we get what we wanted to prove…
โ€–๐ด๐‘’๐‘™ โ€–2 ≤ โ€–๐ดโ€–2,2 โˆ™ โ€–๐‘’
โŸ๐‘™ โ€–2,2
=1
๐‘ž
√∑|๐‘Ž๐‘–๐‘™ |2 ≤ โ€–๐ดโ€–2,2 โˆ™ 1 ≤ 0
๐‘–=1
Useful fact: Let ๐œ‘(๐‘ฅ) be a norm on a NLS ๐’ณ over ๐”ฝ. Then |๐œ‘(๐‘‹) − ๐œ‘(๐‘ฆ)| ≤ ๐œ‘(๐‘ฅ − ๐‘ฆ)
Or in other notation: |โ€–๐‘ฅโ€– − โ€–๐‘ฆโ€–| ≤ โ€–๐‘ฅ − ๐‘ฆโ€–
๐œ‘(๐‘ฅ) = ๐œ‘(๐‘ฅ − ๐‘ฆ + ๐‘ฆ) ≤ ๐œ‘(๐‘ฅ − ๐‘ฆ) + ๐œ‘(๐‘ฆ) ⇒
๐œ‘(๐‘ฅ) − ๐œ‘(๐‘ฆ) ≤ ๐œ‘(๐‘ฅ − ๐‘ฆ)
But symmetrical to ๐‘ฆ − ๐‘ฅ so we can also claim:
๐œ‘(๐‘ฆ) − ๐œ‘(๐‘ฅ) ≤ ๐œ‘(๐‘ฆ − ๐‘ฅ)
At least one of them is non-negative and will be selected as the absolute value.
๐ด ∈ โ„‚๐‘×๐‘ž
โ€–๐ดโ€–2,2 = โ‹ฏ
There are other ways of making it a normed linear space.
1
๐‘  ๐‘ 
For instance, can also define: โ€–๐ดโ€–๐‘  = {∑๐‘–,๐‘—|๐‘Ž๐‘–๐‘— | }
This is a valid definition of norm!
But then we lose the fact that โ€–๐ด๐ตโ€–๐‘  ≤ โ€–๐ดโ€–๐‘  โˆ™ โ€–๐ตโ€–๐‘  !
Inner Product Spaces
A vector space ๐’ฐ over ๐”ฝ is said to be an inner product space if there exists a number which we’ll
designate by: ⟨๐‘ข, ๐‘ฃ⟩๐‘ข for every pair of vectors ๐‘ข, ๐‘ฃ ∈ ๐’ฐ with the following properties:
1) ⟨๐‘ข, ๐‘ข⟩๐‘ข ≥ 0
2) ⟨๐‘ข, ๐‘ข⟩๐‘ข = 0 ⇒ ๐‘ข = 0
3) ⟨๐‘ข, ๐‘ฃ⟩ is linear in the first entry – ⟨๐‘ข + ๐‘ค, ๐‘ฃ⟩๐‘ข = ⟨๐‘ข, ๐‘ฃ⟩ + ⟨๐‘ค, ๐‘ฃ⟩ and ⟨๐›ผ๐‘ข, ๐‘ฃ⟩ = ๐›ผ⟨๐‘ข, ๐‘ฃ⟩
⟨๐‘ข, ๐‘ฃ⟩๐‘ข
4) ⟨๐‘ฃ, ๐‘ข⟩ = ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…
Observations:
⟨0, ๐‘ข⟩ = 0 for any ๐‘ข ∈ ๐’ฐ
Proof: ⟨0, ๐‘ข⟩ = ⟨0 + 0, ๐‘ข⟩ = ⟨0, ๐‘ข⟩ + ⟨0, ๐‘ข⟩ ⇒ 0 = ⟨0, ๐‘ข⟩
⟨๐‘ข, ๐‘ฃ⟩ - “conjugate” linear in second entry
⟨๐‘ข, ๐‘ฃ + ๐‘ค⟩ = ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…
⟨๐‘ฃ + ๐‘ค, ๐‘ข⟩ = ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…
⟨๐‘ฃ, ๐‘ข⟩ + ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…
⟨๐‘ค, ๐‘ข⟩ = ⟨๐‘ข, ๐‘ฃ⟩ + ⟨๐‘ข, ๐‘ค⟩
ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…
⟨๐‘ข, ๐›ผ๐‘ฃ⟩ = ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…ฬ…
⟨๐›ผ๐‘ฃ, ๐‘ข⟩ = ๐›ผฬ…⟨๐‘ฃ,
๐‘ข⟩ = ๐›ผฬ…⟨๐‘ข, ๐‘ฃ⟩
Examples:
1) โ„‚๐‘› ⟨๐‘ข, ๐‘ฃ⟩ = ∑๐‘›๐‘—=1 ๐‘ข๐‘— ๐‘ฃฬ…๐‘—
2) โ„‚๐‘› ๐œŒ1 , … , ๐œŒ๐‘› all > 0 ⟨๐‘ข, ๐‘ฃ⟩ = ∑๐‘›๐‘—=1 ๐œŒ๐‘— ๐‘ข๐‘— ๐‘ฃฬ…๐‘—
3) ๐œŒ([๐‘Ž, ๐‘]) = continuous function on [๐‘Ž, ๐‘]
๐‘
ฬ…ฬ…ฬ…ฬ…๐‘‘๐‘ก
⟨๐‘“, ๐‘”⟩ = ∫ ๐‘“(๐‘ก)๐‘”(๐‘ก)
๐‘Ž
Cauchy’s-Shwarz in inner product spaces
Let ๐’ฐ be an inner product space over โ„‚. Then:
1
1
|⟨๐‘ข, ๐‘ฃ⟩| ≤ {⟨๐‘ข, ๐‘ข⟩}2 {⟨๐‘ฃ, ๐‘ฃ⟩2 } with equality if and only if dim ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข, ๐‘ฃ} ≤ 1
Proof: Let ๐œ† ∈ โ„‚ and consider:
⟨๐‘ข + ๐œ†๐‘ฃ, ๐‘ข + ๐œ†๐‘ฃ⟩ = ⟨๐‘ข, ๐‘ข + ๐œ†๐‘ฃ⟩ + ⟨๐œ†๐‘ฃ, ๐‘ข + ๐œ†๐‘ฃ⟩ = ⟨๐‘ข, ๐‘ข⟩ + ๐œ†⟨๐‘ฃ, ๐‘ข⟩ + ๐œ†ฬ…⟨๐‘ฃ, ๐‘ข⟩ + |๐œ†|2 ⟨๐‘ฃ, ๐‘ฃ⟩ +
The inequality is clearly valid if ๐‘ฃ = 0
Can restrict attention to ease ⟨๐‘ฃ, ๐‘ฃ⟩ > 0 and ⟨๐‘ข, ๐‘ฃ⟩ ≠ 0.
Let’s write ⟨๐‘ข, ๐‘ฃ⟩ = ๐‘Ÿ โˆ™ ๐‘’ ๐’พ๐œƒ , ๐‘Ÿ ≥ 0, 0 ≤ ๐œƒ < 2๐œ‹
Choose ๐œ† = ๐‘ฅ๐‘’ ๐’พ๐œƒ . ๐‘ฅ ∈ โ„
⟨๐‘ข + ๐‘ฅ๐‘’ ๐’พ๐œƒ , ๐‘ข + ๐‘’ ๐’พ๐œƒ ⟩ = ⟨๐‘ข, ๐‘ข⟩ + ๐‘ฅ๐‘’ ๐’พ๐œƒ ๐‘Ÿ๐‘ฅ๐‘’ −๐’พ๐œƒ + ๐‘ฅ๐‘’ −๐’พ๐œƒ ๐‘Ÿ๐‘ฅ๐‘’ ๐’พ๐œƒ + ๐‘ฅ 2 ⟨๐‘ฃ, ๐‘ฃ⟩ =
⟨๐‘ข, ๐‘ข⟩ + 2๐‘ฅ๐‘Ÿ + ๐‘ฅ 2 ⟨๐‘ฃ, ๐‘ฃ⟩
๐‘“(๐‘ฅ) = ⟨๐‘ฃ, ๐‘ฃ⟩๐‘ฅ 2 + 2๐‘Ÿ๐‘ฅ + ⟨๐‘ข, ๐‘ข⟩
๐‘“ ′ (๐‘ฅ) = ⟨๐‘ฃ, ๐‘ฃ⟩2๐‘ฅ + 2๐‘Ÿ
๐‘Ÿ
๐‘“ ′ (๐‘ฅ) = 0 ⇒ ๐‘ฅ = −
= ๐‘ฅ0
⟨๐‘ฃ, ๐‘ฃ⟩
⟨๐‘ข + ๐‘ฅ0 ๐‘’ ๐’พ๐œƒ ๐‘ฃ, − − −⟩ = ⟨๐‘ข, ๐‘ข⟩ −
2๐‘Ÿ 2
๐‘Ÿ2
๐‘Ÿ2
⟨๐‘ฃ,
⟨๐‘ข,
+
๐‘ฃ⟩
=
๐‘ข⟩
−
⟨๐‘ฃ, ๐‘ฃ⟩ ⟨๐‘ฃ, ๐‘ฃ⟩2
⟨๐‘ฃ, ๐‘ฃ⟩
๐‘Ÿ2
= ⟨๐‘ข, ๐‘ข⟩ − ⟨๐‘ข + ๐‘ฅ0 ๐‘’ ๐’พ๐œƒ ๐‘ฃ, ๐‘ข + ๐‘ฅ0 ๐‘’ ๐’พ๐œƒ ๐‘ข⟩ ≤ ⟨๐‘ข, ๐‘ข⟩
⟨๐‘ฃ, ๐‘ฃ⟩
๐‘Ÿ 2 ≤ ⟨๐‘ข, ๐‘ข⟩⟨๐‘ฃ, ๐‘ฃ⟩
๐‘Ÿ 2 = ⟨๐‘ฃ, ๐‘ฃ⟩2
And there is an equality if ⟨๐‘ข + ๐‘ฅ0 ๐‘’ ๐’พ๐œƒ ๐‘ฃ, ๐‘ข + ๐‘ฅ0 ๐‘’ ๐’พ๐œƒ ๐‘ข⟩ = 0
But this happens iff ๐‘ข + ๐‘ฅ0 ๐‘’ ๐’พ๐œƒ ๐‘ฃ = 0 ⇒ ๐‘ข and ๐‘ฃ are linearly dependent!
Claim: If ๐’ฐ is an inner product space, then its also a normed-linear-space with respect to ๐œ‘(๐‘ข) =
1
⟨๐‘ข, ๐‘ข⟩2
Check:
1)
2)
3)
4)
๐œ‘(๐‘ข) ≥ 0
๐œ‘(๐‘ข) = 0 ⇒ ๐‘ข = 0
๐œ‘(๐›ผ๐‘ข) = |๐›ผ|๐œ‘(๐‘ข), ๐›ผ ∈ โ„‚
๐œ‘(๐‘ข + ๐‘ฃ) ≤ ๐œ‘(๐‘ข)๐œ‘(๐‘ฃ) ??
๐œ‘(๐‘ข + ๐‘ฃ)2 = ⟨๐‘ข + ๐‘ฃ, ๐‘ข + ๐‘ฃ⟩ = ⟨๐‘ข, ๐‘ข⟩ + ⟨๐‘ข, ๐‘ฃ⟩ + ⟨๐‘ฃ, ๐‘ข⟩ + ⟨๐‘ฃ, ๐‘ฃ⟩ ≤
1
1
⟨๐‘ข, ๐‘ข⟩ + 2⟨๐‘ข, ๐‘ข⟩2 ⟨๐‘ฃ, ๐‘ฃ⟩2 + ⟨๐‘ฃ, ๐‘ฃ⟩ = ๐œ‘(๐‘ข)2 + 2๐œ‘(๐‘ข)2 + ๐œ‘(๐‘ฃ)2 = {๐œ‘(๐‘ข) + ๐œ‘(๐‘ฃ)}2
๐œ‘(๐‘ข + ๐‘ฃ) ≤ ๐œ‘(๐‘ข) + ๐œ‘(๐‘ฃ)
๐œ‘(๐‘ข + ๐‘ฃ)2 = ⟨๐‘ข + ๐‘ฃ, ๐‘ข + ๐‘ฃ⟩ = ⟨๐‘ข, ๐‘ข⟩ + ⟨๐‘ข, ๐‘ฃ⟩ + ⟨๐‘ฃ, ๐‘ข⟩ + ⟨๐‘ฃ, ๐‘ฃ⟩
๐œ‘(๐‘ข − ๐‘ฃ)2 = ⟨๐‘ข − ๐‘ฃ, ๐‘ข − ๐‘ฃ⟩ = ⟨๐‘ข, ๐‘ข⟩ − ⟨๐‘ข, ๐‘ฃ⟩ − ⟨๐‘ฃ, ๐‘ข⟩ − ⟨๐‘ฃ, 0⟩
๐œ‘(๐‘ข + ๐‘ฃ)2 + ๐œ‘(๐‘ข − ๐‘ฃ)2 = 2๐œ‘(๐‘ข)2 + 2๐œ‘(๐‘ฃ)2
So:
โ€–๐‘ข + ๐‘ฃโ€–2 + โ€–๐‘ข − ๐‘ฃโ€–2 = 2โ€–๐‘ขโ€–2 + 2โ€–๐‘ฃโ€–2
Parallelogram law!
If the norm in a NLS satisfies the parallelogram law, then we can define an inner product on that
space by SeHing:
4
1
2
⟨๐‘ข, ๐‘ฃ⟩ = ∑ ๐’พ ๐‘— โ€–๐‘ข + ๐’พ ๐‘— ๐‘ฃโ€–
4
๐‘—=1
โ€–๐‘ฅโ€–๐‘  satisfies parallelogram law ⇔ ๐‘  = 2
--- end of lesson
Inner Product Sace over ๐”ฝ is a vector spac over ๐”ฝ plus a rule that associates a number ⟨๐‘ข, ๐‘ฃ⟩ ∈ ๐”ฝ
for each pair of vectors ๐‘ข, ๐‘ฃ ∈ ๐’ฐ
1) ⟨๐‘ข, ๐‘ข⟩ ≥ 0
2) If ⟨๐‘ข, ๐‘ข⟩ = 0 ⇒ ๐‘ข = 0
3) Linear in first entry
4) ⟨๐‘ข, ๐‘ฃ⟩ = ⟨๐‘ฃ, ๐‘ข⟩
1
1
Checked that ⟨๐‘ข, ๐‘ข⟩2 defines a norm on ๐’ฐ. i.e. ๐’ฐ is a normed-linear space with respect to ⟨๐‘ข, ๐‘ข⟩2
i.e. we shoed that:
1
1) ⟨๐‘ข, ๐‘ข⟩2 ≥ 0
1
2) ⟨๐‘ข, ๐‘ข⟩2 = 0 ⇒ ๐‘ข = 0
1
1
3) ⟨๐›ผ๐‘ข, ๐›ผ๐‘ฃ⟩2 = |๐›ผ|⟨๐‘ข, ๐‘ข⟩2
1
1
1
4) ⟨๐‘ข + ๐‘ฃ, ๐‘ข + ๐‘ฃ⟩2 ≤ ⟨๐‘ข, ๐‘ข⟩2 + ⟨๐‘ฃ, ๐‘ฃ⟩2
Question: Is every NLS an inner product space?
No!
1
The norm that we introduced above โ€–๐‘ขโ€– = ⟨๐‘ข, ๐‘ข⟩2 satisfies the parallelogram law, i.e.:
โ€–๐‘ข + ๐‘ฃโ€–2 + โ€–๐‘ข − ๐‘ฃโ€–2 = 2โ€–๐‘ขโ€–2 + 2โ€–๐‘ฃโ€–2
FACT: If ๐’ฐ is a normed-linear space such that its norm satisfies this extra condition, then can
define an inner product on this space.
Orthogonality
Let ๐’ฐ be an inner product space over ๐”ฝ, then a pair of vectors ๐‘ข ∈ ๐’ฐ is said to be orthogonal to
a vector ๐‘ฃ ∈ ๐’ฐ if ⟨๐‘ข, ๐‘ฃ⟩ = 0
TODO: Draw cosines
โ€–๐‘Žโ€–2 + โ€–๐‘โ€–2 + โ€–๐‘ − ๐‘Žโ€–2 โ€–๐‘Žโ€–2 + โ€–๐‘โ€–2 + ∑3๐‘—=1 ๐‘๐‘—2 + 2๐‘๐‘— ๐‘Ž๐‘— + ๐‘Ž2
cos ๐œƒ =
=
2โ€–๐‘Žโ€– โˆ™ โ€–๐‘โ€–
2โ€–๐‘Žโ€– โˆ™ โ€–๐‘โ€–
Since:
3
โ€–๐‘Žโ€–2
= ∑ ๐‘Ž๐‘–2
๐‘–=1
3
2
2
โ€–๐‘ − ๐‘Žโ€– = ⟨๐‘ − ๐‘Ž, ๐‘ − ๐‘Ž⟩ = ∑(๐‘๐‘— − ๐‘Ž๐‘— )
๐‘—=1
So:
cos ๐œƒ =
⟨๐‘Ž, ๐‘⟩
โ€–๐‘Žโ€– โˆ™ โ€–๐‘โ€–
A family {๐‘ข1 , … , ๐‘ข๐‘˜ } of non-zero vectors is said to be an orthogonal family if ⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = 0 for ๐‘– ≠
๐‘—.
A family {๐‘ข1 , … , ๐‘ข๐‘˜ } of nonzero vectors is said to be orthonormal family if
⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = 0 , ๐‘– ≠ ๐‘—
{
⟨๐‘ข๐‘– , ๐‘ข๐‘– ⟩ = 1, ๐‘– = 1 … ๐‘˜
If ๐’ฑ is a subspace of ๐’ฐ then the orthogonal complement of ๐’ฑ which we write as ๐’ฑ ⊥ of ๐’ฑ (in ๐’ฐ)
is equal to {๐‘ข ∈ ๐’ฐ|⟨๐‘ข, ๐‘ฃ⟩ = 0 ∀๐‘ฃ ∈ ๐’ฑ}
Claim: ๐’ฑ ⊥ is a vector space.
To check: Let ๐‘ข, ๐‘ค ∈ ๐’ฑ ⊥
⟨๐‘ข + ๐‘ค, ๐‘ฃ⟩ = ⟨๐‘ข, ๐‘ฃ⟩ + ⟨๐‘ค, ๐‘ฃ⟩ = 0 + 0 = 0
⟨๐›ผ๐‘ข, ๐‘ฃ⟩ = ๐›ผ⟨๐‘ข, ๐‘ฃ⟩ = ๐›ผ0 = 0
Claim: ๐’ฑ ∩ ๐‘‰ ⊥ = {0}
Let ๐‘ฅ ∈ ๐’ฑ ∩ ๐’ฑ ⊥ , but ๐‘ฅ ∈ ๐’ฑ ⊥ means it’s orthogonal to every vector in ๐’ฑ ⇒ ⟨๐‘ฅ, ๐‘ฅ⟩ = 0 ⇒ ๐‘ฅ = 0.
Observation: If {๐‘ข1 , … , ๐‘ข๐‘˜ } is an orthogonal family, then it’s a linearly independent set of
vectors.
Proof: Suppose you could find coefficients ๐‘1 , … , ๐‘๐‘˜ such that:
๐‘˜
๐‘˜
∑ ๐‘๐‘— ๐‘ข๐‘— = 0 ⇒ ⟨∑ ๐‘๐‘— ๐‘ข๐‘— , ๐‘ข๐‘– ⟩ = ⟨0, ๐‘ข๐‘– ⟩ = 0
๐‘—=1
1
It’s true for all ๐‘–. So suppose ๐‘– = 1:
๐‘˜
∑ ๐‘๐‘— ⟨๐‘ข๐‘— , ๐‘ข1 ⟩
๐ผ๐‘ก ′ ๐‘  ๐‘Ž๐‘› ๐‘œ๐‘Ÿ๐‘กโ„Ž๐‘œ๐‘”๐‘œ๐‘›๐‘Ž๐‘™
๐‘“๐‘Ž๐‘š๐‘–๐‘™๐‘ฆ!
=
๐‘ข1 ≠0
๐‘1 ⟨๐‘ข1 , ๐‘ข1 ⟩ ⇒
๐‘1 = 0
1
We can do the same for ๐‘ข2 and so on…
Thus proving ∀๐‘–. ๐‘๐‘– = 0 which means they are linearly independent.
Let ๐’ฐ be an inner product space over ๐”ฝ and let {๐‘ข1 , … , ๐‘ข๐‘˜ } be a set of ๐‘˜ vectors in ๐’ฐ then:
๐’ข = [๐‘”๐‘–๐‘— ], with ๐‘”๐‘–๐‘— = ⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ called Gram matrix.
Lemma: {๐‘ข1 , … , ๐‘ข๐‘˜ } will be linearly independent ⇔ ๐’ข is invertible.
Suppose first ๐’ข is invertible, claim that if ∑๐‘˜๐‘—=1 ๐‘๐‘— ๐‘ข๐‘— = 0 then ๐‘1 = โ‹ฏ = ๐‘๐‘˜ = 0.
๐‘1
Consider ๐‘ = [ โ‹ฎ ] (๐’ข๐‘)๐‘– = ∑๐‘˜๐‘—=1 ๐‘”๐‘–๐‘— ๐‘๐‘— = ∑๐‘˜๐‘—=1 ๐‘๐‘— ⟨๐‘ข๐‘— , ๐‘ข๐‘– ⟩ = ⟨∑๐‘˜๐‘—=1 ๐‘๐‘— ๐‘ข๐‘— , ๐‘ข๐‘– ⟩ = ⟨0, ๐‘ข๐‘– ⟩ = 0
๐‘๐‘˜
But this is true for all ๐‘–!
So ๐’ข๐‘ = 0
๐’ข is invertible ⇒ ๐‘ = 0
Assume ๐’ข is invertible ⇒ ๐‘1 , … , ๐‘๐‘˜ are such that ∑ ๐‘๐‘— ๐‘ข๐‘— = 0 ⇒ ๐‘1 = โ‹ฏ = ๐‘๐‘˜ = 0 and linearly
independent.
Suppose now ๐‘ข1 , … , ๐‘ข๐‘˜ are linearly independent.
Claim: ๐’ข is invertible.
It’s enough to show that ๐’ฉ๐’ข = {0}
Let ๐‘ ∈ ๐’ฉ๐’ข ⇒ ๐’ข๐‘ = 0 ⇒ ∑๐‘˜๐‘—=1 ๐‘”๐‘–๐‘— ๐‘๐‘— = 0 for ๐‘– = 1, … , ๐‘˜ ⇒ ∑⟨๐‘ข๐‘— , ๐‘ข๐‘– ⟩๐‘๐‘— = 0 ⇒
๐‘˜
๐‘˜
๐‘˜
๐‘˜
๐‘˜
⟨∑ ๐‘๐‘— ๐‘ข๐‘— , ๐‘ข๐‘– ⟩ = 0 ๐‘“๐‘œ๐‘Ÿ ๐‘– = 1, … , ๐‘˜ ⇒ ∑ ๐‘ฬ…๐‘– (⟨∑ ๐‘๐‘— ๐‘ข๐‘— , ๐‘ข๐‘– ⟩) = 0 ⇒ ⟨∑ ๐‘๐‘— ๐‘ข๐‘— , ∑ ๐‘๐‘– ๐‘ข๐‘– ⟩
๐‘—=1
๐‘–=1
๐‘—=1
๐‘—=1
But these are the same vector! Denote it as ๐‘ค so: ⟨๐‘ค, ๐‘ค⟩ = 0 ⇒ ๐‘ค = 0 ⇒
{๐‘ข1 , … , ๐‘ข๐‘˜ } linearly independent ⇒ ๐‘1 = โ‹ฏ = ๐‘๐‘˜ = 0
๐‘–=1
๐‘˜
∑๐‘—=1 ๐‘๐‘— ๐‘ข๐‘—
=0
Thus ๐’ข๐‘ = 0 ⇒ ๐‘ = 0 ⇒ ๐’ฉ๐’ข = {0}. ๐’ข is invertible.
Adjoints
Let ๐‘‡ be a linear transformation from a finite dimensional inner product space ๐’ฐ into a finite
dimensional inner product space ๐’ฑ. Then there exists exactly one linear transformation ๐‘† from
๐’ฑ to ๐’ฐ such that for ๐‘ข ∈ ๐’ฐ, ๐‘ฃ ∈ ๐’ฑ - ⟨๐‘‡๐‘ข, ๐‘ฃ⟩๐’ฑ = ⟨๐‘ข, ๐‘†๐‘ฃ⟩๐’ฐ
๐‘† is called the adjoint of ๐‘‡. And the usual notation is ๐‘‡ ∗.
Proof: At the end of the lesson if we have time.
It’s easy to show that there is most one linear transformation ๐‘† from ๐’ฑ to ๐’ฐ such that
⟨๐‘‡, ๐‘ข, ๐‘ฃ⟩๐’ฑ = ⟨๐‘ข, ๐‘†๐‘ฃ⟩๐’ฐ all ๐‘ข ∈ ๐’ฐ, ๐‘ฃ ∈ ๐’ฑ.
Suppose could find ๐‘†1 , ๐‘†2 that met this condition.
⟨๐‘‡๐‘ข, ๐‘ข⟩๐’ฑ = ⟨๐‘ข, ๐‘†1 ๐‘ฃ⟩๐’ฐ = ⟨๐‘ข, ๐‘†2 ๐‘ฃ⟩๐’ฐ
⟨๐‘ข, ๐‘†1 ๐‘ฃ − ๐‘†2 ๐‘ฃ⟩ = 0 for all ๐‘ข ∈ ๐’ฐ, ๐‘ฃ ∈ ๐’ฑ.
This is true for all ๐‘ข! We can choose ๐‘ข = ๐‘†1 ๐‘ฃ − ๐‘†2 ๐‘ฃ.
But then: ⟨๐‘†1 ๐‘ฃ − ๐‘†2 ๐‘ฃ, ๐‘†1 ๐‘ฃ − ๐‘†2 ๐‘ฃ⟩ = 0
So the vector ๐‘†1 ๐‘ฃ − ๐‘†2 ๐‘ฃ = 0 ⇒ ๐‘†1 ๐‘ฃ = ๐‘†2 ๐‘ฃ ⇒ ๐‘†1 = ๐‘†2 .
Let ๐‘‡ be a linear transformation from a finite dimensional inner product space ๐’ฐ into ๐’ฐ.
Then ๐‘‡ is said to be:
Normal if ๐‘‡ ∗ ๐‘‡ = ๐‘‡๐‘‡ ∗
Selfadjoint if ๐‘‡ ∗ = ๐‘‡
Unitary if ๐‘‡ ∗ ๐‘‡ = ๐‘‡๐‘‡ ∗ = ๐ผ
๐‘‡ = ๐‘‡ ∗ ⇒ ๐‘‡ is normal
๐‘‡๐‘‡ ∗ = ๐‘‡๐‘‡ ∗ ⇒ is normal
Theorem: Let ๐‘‡ be a linear transformation from an ๐‘› dimensional inner product space ๐’ฐ into
itself and suppose ๐‘‡ is normal. Then:
(1) There exists an orthonormal basis {๐‘ข1 , … , ๐‘ข๐‘› } of eigenvectors of ๐‘‡
(2) If ๐‘‡ = ๐‘‡ ∗ ⇒ all the eigenvalues are real.
(3) If ๐‘‡ is unitary⇒ all eigenvalues have magnitude 1.
1) ๐‘‡๐œ† = ๐œ†๐ผ − ๐‘‡, (๐‘‡๐œ† )∗ = ๐œ†ฬ…๐ผ − ๐‘‡
⟨๐‘‡๐œ† ๐‘ข, ๐‘ฃ⟩๐’ฐ = ⟨๐‘ข, (๐‘‡๐œ† )∗ ๐‘ฃ⟩
But the left side equals:
⟨(๐œ†๐ผ − ๐‘‡)๐‘ข, ๐‘ฃ⟩ = ๐œ†⟨๐‘ข, ๐‘ฃ⟩ − ⟨๐‘‡๐‘ข, ๐‘ฃ⟩ = ⟨๐‘ข, ๐œ†ฬ…๐‘ฃ⟩ − ⟨๐‘ข, ๐‘‡ ∗ ๐‘ฃ⟩ = ⟨๐‘ข, (๐œ†ฬ…๐ผ − ๐‘‡ ∗ )๐‘ฃ⟩
But there is only one adjoint matrix! So (๐‘‡๐œ† )∗ = (๐œ†ฬ…๐ผ − ๐‘‡ ∗ )
2) T normal ⇒ ๐‘‡๐œ† is normal.
๐‘‡๐œ† (๐‘‡๐œ† )∗ = (๐œ†๐ผ − ๐‘‡)(๐œ†ฬ…๐ผ − ๐‘‡) = ๐œ†๐ผ(๐œ†ฬ…๐ผ − ๐‘‡ ∗ ) − ๐‘‡(๐œ†ฬ…๐ผ − ๐‘‡ ∗ ) = (๐œ†ฬ…๐ผ − ๐‘‡ ∗ )๐œ†๐ผ − (๐œ†ฬ…๐ผ − ๐‘‡ ∗ )๐‘‡ =
(๐œ†ฬ…๐ผ − ๐‘‡)(๐œ†๐ผ − ๐‘‡) = (๐‘‡๐œ† )∗ ๐‘‡๐œ†
3) ๐‘‡๐‘ข = ๐œ†๐‘ข ⇒ ๐‘‡ ∗ ๐‘ข = ๐œ†ฬ…๐‘ข
Let ๐‘ข ∈ ๐’ฉ๐‘‡๐œ† ⇒ ๐‘‡๐œ† ๐‘ข = 0 ⇒ (๐‘‡๐œ† )∗ ๐‘‡๐œ† ๐‘ข = 0 ⇒ ๐‘‡๐œ† (๐‘‡๐œ† )∗ = 0 ⇒ ⟨๐‘‡๐œ† (๐‘‡๐œ† )∗ ๐‘ข, ๐‘ข⟩ = 0 ⇒
⟨(๐‘‡๐œ† )∗ ๐‘ข, (๐‘‡๐œ† )∗ ๐‘ข⟩ = 0 ⇒ (๐‘‡๐œ† )∗ ๐‘ข = 0 ⇒ (๐œ†ฬ…๐ผ − ๐‘‡ ∗ )๐‘ข = 0
4) Establish orthonormal basis of eigenvectors of ๐‘‡.
๐’ฐ is invariant under ๐‘‡ ⇒ ๐‘‡ has an eigenvector ๐‘ข1 ≠ 0 i.e. there is ๐‘‡๐‘ข1 = ๐œ†1 ๐‘ข1 , ๐‘ข1 ≠ 0
๐‘ข1
๐‘ข1
๐‘‡(
) = ๐œ†1 (
)
โ€–๐‘ข1 โ€–
โ€–๐‘ข1 โ€–
So can assume ๐‘ข1 has norm 1.
Suppose we could find {๐‘ข1 , … , ๐‘ข๐‘˜ } such that ๐‘‡๐‘ข๐‘— = ๐œ†๐‘ข๐‘— , ๐‘— = 1, … , ๐‘˜
⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = 0 if ๐‘– ≠ ๐‘— and 1 otherwise.
โ„’๐‘˜ = ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข1 , … , ๐‘ข๐‘˜ }
โ„’๐‘˜⊥ = vectors in ๐’ฐ that are orthogonal to โ„’๐‘˜ .
๐‘ฃ ∈ โ„’๐‘˜⊥ ⇔ ⟨๐‘ฃ, ๐‘ข๐‘— ⟩ = 0,
๐‘— = 1, … , ๐‘˜
โ„’๐‘˜⊥ is invariant under ๐‘‡, i.e. ๐‘ฃ ∈ โ„’๐‘˜⊥ ⇒ ๐‘‡๐‘ฃ ∈ โ„’๐‘˜⊥
⟨๐‘‡๐‘ฃ, ๐‘ข๐‘— ⟩ = ⟨๐‘ฃ, ๐‘‡ ∗ ๐‘ข๐‘— ⟩ = ⟨๐‘ฃ, ๐œ†ฬ…๐‘— ๐‘ข๐‘— ⟩ = ๐œ†๐‘— ⟨๐‘ฃ, ๐‘ข๐‘— ⟩ = 0,
๐‘— = 1, … , ๐‘˜
Because:
๐‘‡๐‘ข๐‘— = ๐œ†๐‘— ๐‘ข๐‘—
๐‘‡ ∗ ๐‘ข๐‘— = ๐œ†ฬ…๐‘— ๐‘ข๐‘—
So can find an eigenvector of ๐‘‡ in โ„’๐‘˜⊥ . Call it ๐‘ข๐‘˜+1 = ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข1 , … , ๐‘ข๐‘˜ }.
So we can increase our โ„’๐‘˜⊥ until it is big enough…
6) ๐‘‡ unitary ⇒ |๐œ†๐‘— | = 1
If ๐‘‡ unitary:
2
๐‘‡๐‘ข๐‘— = ๐œ†๐‘— ๐‘ข๐‘— ⇒ ๐‘‡ ∗ ๐‘‡๐‘ข๐‘— = ๐œ†๐‘— ๐‘‡ ∗ ๐‘ข๐‘— = ๐œ†๐‘— ๐œ†ฬ…๐‘— ๐‘ข๐‘— ⇒ ๐‘ข๐‘— = |๐œ†๐‘— | ๐‘ข๐‘— .
--- end of lesson
Recall: IF ๐‘‡ is a linear transformation from a finite dimension inner product space ๐’ฐ over ๐”ฝ into
a finite dimension inner product space ๐’ฑ, then there exists exactly one linear transformation ๐‘†
from ๐’ฑ to ๐’ฐ such that ⟨๐‘‡๐‘ข, ๐‘ฃ⟩๐’ฑ = ⟨๐‘ข, ๐‘†๐‘ฃ⟩๐’ฐ for every ๐‘ข ∈ ๐’ฐ, ๐‘ฃ ∈ ๐’ฑ
That ๐‘† is called the adjoint of ๐‘‡ and is denoted by ๐‘‡ ∗ .
So in the future we’ll write ⟨๐‘‡๐‘ข, ๐‘ฃ⟩๐’ฑ = ⟨๐‘ข, ๐‘‡ ∗ ๐‘ฃ⟩๐’ฐ .
Existence: Suppose {๐‘ข1 , … , ๐‘ข๐‘˜ } is a basis for ๐’ฐ and {๐‘ฃ1 , … , ๐‘ฃ๐‘™ } is a basis of ๐’ฑ.
Can reduce the existence to showing that exists ๐‘† such that
⟨๐‘‡๐‘ข๐‘– , ๐‘ฃ๐‘— ⟩๐’ฑ = ⟨๐‘ข๐‘– , ๐‘†๐‘ฃ๐‘— ⟩๐’ฐ ,
๐‘– = 1, … , ๐‘˜, ๐‘— = 1, … , ๐‘™
๐‘™
๐‘™
๐‘™
๐‘‡๐‘ข๐‘– ∈ ๐’ฑ ⇒ ๐‘‡๐‘ข๐‘– = ∑ ๐‘Ž๐‘ก๐‘– ๐‘ฃ๐‘ก ๐‘“๐‘œ๐‘Ÿ ๐‘– = 1, … , ๐‘˜ ⇒ ⟨๐‘‡๐‘ข๐‘– , ๐‘ฃ๐‘— ⟩ = ⟨∑ ๐‘Ž๐‘ก๐‘– ๐‘ฃ๐‘ก , ๐‘ฃ๐‘— ⟩ = ∑ ๐‘Ž๐‘ก๐‘– ⟨๐‘ฃ๐‘ก , ๐‘ฃ๐‘— ⟩๐’ฑ =
๐‘ก=1
๐‘ก=1
๐‘ก=1
๐‘™
∑ ๐‘Ž๐‘ก๐‘– (๐บ๐’ฑ )๐‘—๐‘ก = (๐บ๐’ฑ ๐ด)๐‘—๐‘–
๐‘ก=1
๐‘˜
๐‘†๐‘ฃ๐‘— = ∑ ๐‘๐‘–๐‘— ๐‘ข๐‘Ÿ
๐‘–=1
Want to choose ๐ต = [๐‘๐‘–๐‘— ] so that equiality (1) holds.
๐‘˜
๐‘˜
⟨๐‘ข๐‘– , ๐‘†๐‘ฃ๐‘— ⟩๐’ฐ = ⟨๐‘ข๐‘– , ∑ ๐‘๐‘–๐‘— ๐‘ข๐‘Ÿ ⟩๐’ฐ = ∑ ฬ…ฬ…ฬ…
๐‘๐‘–๐‘—ฬ…⟨๐‘ข๐‘– , ๐‘ข๐‘Ÿ ⟩๐’ฐ = (๐บ๐’ฐ ๐ด)
๐‘–=1
๐‘–=1
๐‘ฬ…๐‘Ÿ๐‘— = (๐ต๐ป )๐‘—๐‘Ÿ
๐ต๐ป ๐’ข๐’ฐ = ๐’ข๐’ฑ ๐ด
๐‘‡: ๐’ฐ → ๐’ฐ
๐‘‡ is said to be normal if ๐‘‡ ∗ ๐‘‡ = ๐‘‡๐‘‡ ∗
๐‘‡ is said to be self-adjoint if ๐‘‡ = ๐‘‡ ∗
๐‘‡ is said to be unitary if ๐‘‡ ∗ ๐‘‡ = ๐‘‡๐‘‡ ∗ = ๐ผ
Let ๐ด ∈ โ„‚๐‘×๐‘ž , ๐’ฐ = โ„‚๐‘ , ⟨, ⟩๐’ฐ , ๐’ฑ = โ„‚๐‘ , ⟨, ⟩๐’ฑ
Then there exists exactly one matrix ๐ด ∈ โ„‚๐‘ž×๐‘ such that ⟨๐ด๐‘ฅ, ๐‘ฆ⟩๐’ฑ = ⟨๐‘ฅ, ๐ด∗ ๐‘ฆ⟩๐’ฐ
Example: ⟨๐‘ข1 , ๐‘ข2 ⟩๐’ฐ = ⟨Δ1 ๐‘ข1 , Δ1 ๐‘ข2 ⟩๐‘ ๐‘ก = (Δ1 ๐‘ข2 )๐ป Δ1 ๐‘ข1
๐‘ž×๐‘ž
⟨๐‘Ž, ๐‘⟩๐‘ ๐‘ก = ∑๐‘˜๐‘—=1 ๐›ผ๐‘— ๐‘ฬ…๐‘— = ๐‘ ๐ป ๐‘Ž = ๐‘ข2๐ป Δ๐ป
invertible.
1 Δ1 ๐‘ข1 , Δ1 ∈ โ„‚
โ„‚๐‘ , ๐ต ∈ ๐ถ ๐‘ž×๐‘ž invertible.
๐‘‘๐‘’๐‘“
1)
2)
3)
4)
⟨๐‘ฅ, ๐‘ฆ⟩๐ต = ⟨๐ต๐‘ฅ, ๐ต๐‘ฆ⟩๐‘ ๐‘ก = (๐ต๐‘ฆ)๐ป ๐ต๐‘ฅ = ๐‘ฆ โ„Ž ๐ต๐ป ๐ต๐‘ฅ
⟨๐‘ฅ, ๐‘ฅ⟩๐ต ≥ 0? ⟨๐‘ฅ, ๐‘ฅ⟩๐ต = ๐‘ฅ ๐ป ๐ต๐ป ๐ต๐‘ฅ = (๐ต๐‘ฅ)๐ป ๐ต๐‘ฅ = ⟨๐ต๐‘ฅ, ๐ต๐‘ฅ⟩๐‘ ๐‘ก = ∑|๐ต๐‘ฅ|2
⟨๐‘ฅ, ๐‘ฅ⟩๐ต = 0 ⇒ ๐‘ฅ = 0 …
…
…
Δ2 ๐‘ × ๐‘ invertible.
๐‘‘๐‘’๐‘“
⟨๐‘ฃ1 , ๐‘ฃ2 ⟩๐’ฑ = ⟨Δ2 ๐‘ฃ1 , Δ2 ๐‘ฃ2 ⟩๐‘ ๐‘ก = (Δ2 ๐‘ฃ2 )๐ป Δ2 ๐‘ฃ1 = ๐‘‰2๐ป (Δ๐ป
2 Δ2 )๐‘ฃ1
๐ป
๐ป
⟨๐ด๐‘ฅ, ๐‘ฆ⟩๐’ฑ = ๐‘ฆ Δ2 Δ2 ๐ด๐‘ฅ
๐ป
∗ ๐ป ๐ป
⟨๐‘ฅ, ๐ด∗ ๐‘ฆ⟩๐’ฐ = (๐ด∗ ๐‘ฆ)๐ป Δ๐ป
1 Δ1 ๐‘ฅ = ๐‘ฆ (๐ด ) Δ1 Δ1 ๐‘ฅ
Since it holds for all ๐‘ฅ ∈ โ„‚๐‘ž and ๐‘ฆ ∈ โ„‚๐‘ :
๐ป
๐ป
∗ ๐ป ๐ป
∗ ๐ป
−1
Δ๐ป
2 Δ2 ๐ด = (๐ด ) Δ1 Δ1 ⇒ (๐ด ) = Δ2 Δ2 ๐ด(Δ1 Δ1 )
−1 ๐ป ๐ป
๐ด∗ = (Δ๐ป
1 Δ1 ) ๐ด Δ2 Δ2
If Δ1 = ๐ผ๐‘ž , Δ2 = ๐ผ๐‘
i.e. if using standard inner product, then ๐ด∗ = ๐ด๐ป
Theorem: Let ๐ด ∈ โ„‚๐‘×๐‘ , โ„‚๐‘ have inner product ⟨, ⟩๐’ฐ
Then:
(1) ๐ด∗ ๐ด = ๐ด๐ด∗ ⇒ ๐ด is diagonalizable and there exists an orthonormal set of eigen vectors i.e.
๐ด๐‘ˆ = ๐‘ˆ๐ท, ๐ท diagonal and columns of ๐‘ˆ are orthonormal. ๐ท ∈ โ„๐‘×๐‘
(2) ๐ด = ๐ด∗
(3) ๐ด∗ ๐ด = ๐ด๐ด∗ = ๐ผ๐‘ , |๐‘‘๐‘–๐‘— | = 1
(1) If ๐ด∗ ๐ด = ๐ด๐ด∗ , then ๐ด๐‘ข = ๐œ†๐‘ข ⇒ ๐ด∗ ๐‘ข = ๐œ†ฬ…๐‘ข
๐ด๐ด∗ =๐ด∗ ๐ด
(๐ด − ๐œ†๐ผ)(๐ด∗ − ๐œ†ฬ…๐ผ)๐‘ข = 0 ⇒
๐ด๐‘ข = ๐œ†๐‘ข ⇒ (๐ด − ๐œ†๐ผ)๐‘ข = 0 ⇒ (๐ด∗ − ๐œ†ฬ…๐ผ)(๐ด − ๐œ†๐ผ)๐‘ข = 0 ⇒
⟨(๐ด − ๐œ†๐ผ)(๐ด∗ − ๐œ†ฬ…๐ผ)๐‘ข, ๐‘ข⟩ = ⟨(๐ด∗ − ๐œ†ฬ…๐ผ)๐‘ข, (๐ด∗ − ๐œ†ฬ…๐ผ)๐‘ข⟩ = 0 ⇒ (๐ด∗ − ๐œ†ฬ…๐ผ)๐‘ข = 0 ⇒ ๐‘ข is an
eigenvector of ๐ด∗ with Eigenvalue ๐œ†ฬ…๐ผ.
Let ๐œ† be an eigenvalue of ๐ด.
It’s enough to show ๐’ฉ(๐ด−๐œ†๐ผ )2 = ๐’ฉ(๐ด−๐œ†๐ผ )
๐‘ฅ ∈ ๐’ฉ(๐ด−๐œ†๐ผ)2 ⇒ (๐ด − ๐œ†๐ผ)2 ๐‘ฅ = 0 ⇒ (๐ด − ๐œ†๐ผ)((๐ด − ๐œ†๐ผ)๐‘ฅ) = 0
Denote ๐‘ฆ = (๐ด − ๐œ†๐ผ)๐‘ฅ. Then ๐ด๐‘ฆ = ๐œ†๐‘ฆ
By (1), ๐ด∗ ๐‘ฆ = ๐œ†ฬ…๐‘ฆ ⇒ (๐ด∗ − ๐œ†ฬ…๐ผ)(๐ด − ๐œ†๐ผ)๐‘ฅ = 0 ⇒ ⟨(๐ด∗ − ๐œ†ฬ…๐ผ)(๐ด − ๐œ†๐ผ)๐‘ฅ, ๐‘ฅ⟩ = 0 ⇒
∗
⟨(๐ด − ๐œ†๐ผ)๐‘ฅ, (๐ด∗ − ๐œ†ฬ…๐ผ) ๐‘ฅ⟩ = 0 ⇒ ⟨(๐ด − ๐œ†๐ผ)๐‘ฅ, (๐ด − ๐œ†๐ผ)๐‘ฅ⟩ = 0 ⇒ ๐‘ฅ ∈ ๐’ฉ๐ด−๐œ†๐ผ
∗
∗
Since: (๐ด∗ − ๐œ†ฬ…๐ผ) = (๐ด∗ )∗ − (๐œ†ฬ…๐ผ) = ๐ด − ๐œ†๐ผ
Have shown:
๐’ฉ(๐ด−๐œ†๐ผ)2 ⊆ ๐’ฉ(๐ด−๐œ†๐ผ)
But this is always true:
๐’ฉ(๐ด−๐œ†๐ผ) ⊆ ๐’ฉ(๐ด−๐œ†๐ผ)2
So we must have equality: ๐’ฉ(๐ด−๐œ†๐ผ) = ๐’ฉ(๐ด−๐œ†๐ผ)2
(3) ๐ด๐‘ข๐‘– = ๐œ†๐‘– ๐‘ข2 , ๐ด๐‘ข๐‘— = ๐œ†๐‘— ๐‘ข๐‘— , ๐œ†๐‘– ≠ ๐œ†๐‘— ⇒ ⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = 0
๐œ†๐‘– ⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = ⟨๐œ†๐‘– ๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = ⟨๐ด๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = ⟨๐‘ข๐‘– , ๐ด∗ ๐‘ข๐‘— ⟩ = ⟨๐‘ข๐‘– , ๐œ†๐‘—ฬ… ๐‘ข๐‘— ⟩ = ๐œ†๐‘— ⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ ⇒
(๐œ†๐‘– − ๐œ†๐‘— )⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = 0. So if ๐œ†๐‘– ≠ ๐œ†๐‘— then it must be that ⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩ = 0 and thus they are
orthogonal.
(4) ๐ด∗ ๐ด = ๐ด๐ด∗ ⇒ can choose an orthonormal basis of eigenvectors of ๐ด.
det(๐œ†๐ผ3 − ๐ด) = (๐œ† − ๐œ†1 )2 (๐œ† − ๐œ†2 ) ⇒ dim ๐’ฉ(๐ด−๐œ†1 ๐ผ3 ) = 2, dim ๐’ฉ(๐ด−๐œ†2 ๐ผ3 ) = 1
⟨๐‘ข1 , ๐‘ข3 ⟩ = 0, ⟨๐‘ข2 , ๐‘ข3 ⟩ = 0
⟨๐‘ข1 , ๐‘ข2 ⟩ =?
We will later use the grahm shmidt method to construct an orthonormal basis for the same
eigenvalue.
๐œ†1
But since 4 is true: ๐ด[๐‘ข1 , … , ๐‘ข๐‘› ] = [๐œ†1 ๐‘ข1 … ๐œ†๐‘› ๐‘ข๐‘› ] = [๐‘ข1 … ๐‘ข๐‘› ] [ 0
0
๐ด๐‘ˆ = ๐‘ˆ๐ท
0
โ‹ฑ
0
0
0]
๐œ†๐‘›
(5) If ๐ด = ๐ด∗ ⇒ ๐œ†๐‘– ∈ โ„
๐œ†๐‘– ⟨๐‘ข๐‘– , ๐‘ฃ๐‘– ⟩ = ⟨๐ด๐‘ข๐‘– , ๐‘ข๐‘– ⟩ = ⟨๐‘ข๐‘– , ๐ด∗ ๐‘ข๐‘– ⟩ = ⟨๐‘ข๐‘– , ๐ด๐‘ข๐‘– ⟩ = ⟨๐‘ข๐‘– , ๐œ†๐‘– ๐‘ข๐‘– ⟩ = ๐œ†ฬ…๐‘– ⟨๐‘ข๐‘– , ๐‘ข๐‘— ⟩
(๐œ†๐‘– − ๐œ†ฬ…๐‘– )⟨๐‘ข๐‘– , ๐‘ข๐‘– ⟩ = 0 ⇒ ๐œ†๐‘– = ๐œ†ฬ…๐‘–
๐ด∗ ๐ด = ๐ด๐ด∗ = ๐ผ ⇒ |๐œ†๐‘– | = 1
⟨๐‘ข๐‘– , ๐‘ข๐‘– ⟩ = ⟨๐ด∗ ๐ด๐‘ข๐‘– , ๐‘ข๐‘– ⟩ = ⟨๐ด๐‘ข๐‘– , (๐ด∗ )∗ ๐‘ข๐‘– ⟩ = ⟨๐ด๐‘ข๐‘– , ๐ด๐‘ข๐‘– ⟩ = ⟨๐œ†๐‘– ๐‘ข๐‘– , ๐œ†๐‘– ๐‘ข๐‘– ⟩ = |๐œ†๐‘– |2 ⟨๐‘ข๐‘– , ๐‘ข๐‘– ⟩ ⇒
(1 − |๐œ†๐‘– |2 )⟨๐‘ข๐‘– , ๐‘ข๐‘– ⟩ = 0 ⇒ |๐œ†๐‘– |2 = 1
If ⟨๐‘ฅ, ๐‘ฆ⟩๐’ฐ = ⟨๐‘ฅ, ๐‘ฆ⟩๐‘ ๐‘ก = ๐‘ฆ ๐ป ๐‘ฅ
๐ด∗ = (… )๐ด๐ป (… )
If Δ1 = ๐ผ๐‘ž , Δ2 = ๐ผ๐‘ then ๐ด∗ = ๐ด๐ป
If ๐ด ∈ โ„‚๐‘×๐‘ and ๐ด = ๐ด๐ป , then:
1) ๐ด is diagonalizable
2) It’s Eigenvalues are real
3) There exists an orthonormal basis (w.r.t. standard inner product) of eigenvectors
๐‘ข1 , … , ๐‘ข๐‘› of ๐ด.
๐ด๐‘ˆ = ๐‘ˆ๐ท, where ๐ท is diagonal with real entries
๐‘ˆ = [๐‘ข1
๐‘ข2
๐‘ข3 ]
๐‘ข1๐ป
๐‘ˆ ๐‘ˆ = [๐‘ข2๐ป ] [๐‘ข1
๐‘ข3๐ป
๐ป
๐‘ข2
๐‘ข3 ]
1, ๐‘–๐‘“ ๐‘– = ๐‘—
(๐‘ˆ ๐ป ๐‘ˆ)๐‘–๐‘— = ๐‘ข๐‘–๐ป ๐‘ข๐‘— = ⟨๐‘ข๐‘— , ๐‘ข๐‘– ⟩๐‘ ๐‘ก = {
0 ๐‘–๐‘“ ๐‘– ≠ ๐‘—
Projections And Direct Sums
A linear transformation ๐’ซ from a vector space ๐’ฐ into ๐’ฐ is called a projection if ๐’ซ 2 = ๐‘ƒ
If ๐’ฐ is an inner product space, ๐’ซ is called an orthogonal projection if ๐’ซ 2 = ๐‘ƒ and ๐‘ƒ = ๐‘ƒ∗ or
๐’ซ 2 = ๐‘ and ⟨๐‘ƒ๐‘ฅ, ๐‘ฆ⟩๐’ฐ = ⟨๐‘ฅ, ๐‘ƒ๐‘ฆ⟩๐’ฐ for all ๐‘ฅ, ๐‘ฆ ∈ ๐’ฐ
๐’ฐ = โ„‚2
1 ๐‘Ž
๐‘ƒ=[
]
0 0
๐‘ฅ ∈ โ„‚2 ,
๐‘ƒ๐‘ฅ = โ„‚2
๐‘2 = ๐‘ƒ
โ€–๐‘ƒโ€– can be large. โ€–๐‘ƒโ€– = 1 will be denoted as an orthogonal projection later…
Let ๐’ฐ be a finite dimension vector space over ๐”ฝ, let ๐‘ƒ a projection. Then ๐’ฐ = ๐’ฉ๐‘ƒ โˆ” โ„›๐‘
๐’ฉ๐‘ = {๐‘ข|๐‘ƒ๐‘ข = 0},
โ„›๐‘ = {๐‘ƒ๐‘ข|๐‘ข ∈ ๐’ฐ}
Let ๐‘ข ∈ ๐’ฐ. Then: ๐‘ข = ๐‘ƒ๐‘ข
โŸ + (๐ผ − ๐‘ƒ)๐‘ข
∈โ„›๐‘ƒ
Claim: (๐ผ − ๐‘ƒ)๐‘ข ∈ ๐’ฉ๐‘ƒ
๐‘ƒ=๐‘ƒ2
๐‘ƒ(๐ผ − ๐‘ƒ)๐‘ข = (๐‘ƒ − ๐‘ƒ2 )๐‘ข = 0
Remains to show that ๐’ฉ๐‘ƒ ∩ โ„›๐‘ƒ = {0}
Let ๐‘ฅ ∈ ๐’ฉ๐‘ƒ ∩ โ„›๐‘ƒ
๐‘ฅ ∈ โ„›๐‘ ⇒ ๐‘ฅ = ๐‘ƒ๐‘ฆ
๐‘ฅ ∈ ๐’ฉ๐‘ƒ ⇒ ๐‘ƒ๐‘ฅ = 0
But then:
0 = ๐‘ƒ๐‘ฅ = ๐‘ƒ2 ๐‘ฆ = ๐‘ƒ๐‘ฆ = ๐‘ฅ
Conversely, let ๐’ฐ = ๐’ฑ โˆ” ๐’ฒ is a direct sum decomposition of vector space ๐’ฐ in terms of the two
subspaces ๐’ฑ and ๐’ฒ.
Then each ๐‘ข ∈ ๐’ฐ have a unique decomposition: ๐‘ข = ๐‘ฃ + ๐‘ค where ๐‘ฃ ∈ ๐’ฑ and ๐‘ค ∈ ๐’ฒ.
Claim: There’s only 1 such decomposition.
Proof: if ๐‘ฃ1 + ๐‘ค1 = ๐‘ฃ2 + ๐‘ค2 ⇒ (๐‘ฃ1 − ๐‘ฃ2 ) = (๐‘ค2 − ๐‘ค1 )
But the left hand side is in ๐’ฑ and the right hand side is in ๐’ฒ. So in their intersection – which
means both sides are zero and therefore ๐‘ฃ1 = ๐‘ฃ2 and ๐‘ค1 = ๐‘ค2 .
Start with ๐‘ข. There is exactly one ๐‘ฃ ∈ ๐’ฑ such that ๐‘ข − ๐‘ฃ ∈ ๐’ฒ.
Call this ๐‘ฃ ๐‘ƒ๐’ฑ ๐‘ข
Claim: ๐‘ƒ๐’ฑ 2 = ๐‘ƒ๐’ฑ
Proof: ๐‘ƒ๐’ฑ ๐‘ข = ๐‘ฃ ∈ ๐’ฑ
๐‘ฃ =๐‘ฃ+0
๐‘ƒ๐’ฑ ๐‘ฃ = ๐‘ฃ
Example: โ„‚2
1
๐’ฑ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]}
0
1
๐’ฒ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]}
2
๐‘Ž
1
1
[ ] = ๐›ผ[ ]+๐›ฝ[ ]
๐‘
0
2
๐‘ = 2๐›ฝ
๐‘Ž =๐›ผ+๐›ฝ
๐‘ 1
๐‘ 1
๐‘Ž
[ ] = (๐‘Ž − ) [ ] + [ ]
๐‘
2 0
2 2
๐‘ 1
๐‘Ž
๐‘ƒ๐’ฑ [ ] = (๐‘Ž − ) [ ]
๐‘
2 0
1
๐’ฑ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]}
0
1
๐’ฒ = ๐‘ ๐‘๐‘Ž๐‘› {[ ]}
3
๐‘ 1
๐‘Ž
๐‘ƒ๐’ฑ [ ] = (๐‘Ž − ) [ ]
๐‘
3 0
--- end of lesson
๐’ฐ = ๐’ฑ โˆ” ๐’ฒ,
๐’ฑ ∩ ๐’ฒ = {0}
๐‘ข ∈ ๐’ฐ there is exactly 1 decomposition ๐‘ข = ๐‘ฃ + ๐‘ค with ๐‘ฃ ∈ ๐’ฑ, ๐‘ค ∈ ๐’ฒ
Let ๐‘ฃ1 , … , ๐‘ฃ๐‘˜ be a basis for ๐’ฑ, ๐‘ค1 , … , ๐‘ค๐‘™ be a basis for ๐’ฒ then
{๐‘ฃ1 , … , ๐‘ฃ๐‘˜ , ๐‘ค1 , … , ๐‘ค๐‘™ } is a basis for ๐’ฐ ⇒
๐‘˜
๐‘™
๐‘ข = ∑ ๐‘Ž๐‘– ๐‘ฃ๐‘– + ∑ ๐‘๐‘— ๐‘ค๐‘— ,
๐‘–=1
๐‘Ž๐‘– , ๐‘๐‘— ∈ ๐”ฝ
๐‘—=1
๐‘˜
๐‘ƒ๐’ฑ ๐‘ข = ∑ ๐‘Ž๐‘– ๐‘ฃ๐‘–
๐‘–=1
This recipe depends upon ๐’ฒ as well as ๐’ฑ.
(1) Start with direct sum decomposition, can define ๐‘ƒ๐’ฑ and observe that
๐‘ƒ๐’ฑ2 ๐‘ข = ๐‘ƒ๐’ฑ2 (๐‘ฃ + ๐‘ค) = ๐‘ƒ๐’ฑ (๐‘ฃ + ๐‘ค)
(2) Can start with ๐‘ƒ, a linear transformation from ๐’ฐ into ๐’ฐ such that ๐‘ƒ2 = ๐‘ƒ, then ๐’ฐ =
โ„›๐‘ƒ โˆ” ๐’ฉ๐‘ƒ
If ๐’ฐ is an inner product space and ๐‘ƒ is linear transformation from ๐’ฐ into ๐’ฐ such that:
1) ๐‘ƒ2 = ๐‘ƒ
2) ๐‘ƒ∗ = ๐‘ƒ
Then ๐’ฐ = โ„›๐‘ƒ โˆ” ๐’ฉ๐‘ƒ
Now ๐’ฉ๐‘ƒ is orthogonal to โ„›๐‘ƒ .
๐‘ฃ ∈ โ„›๐‘ƒ ,
๐‘ค ∈ ๐’ฉ๐‘ƒ
๐‘ฃ ∈ โ„›๐‘ƒ ⇒ ๐‘ฃ = ๐‘ƒ๐‘ฆ
๐‘Ž๐‘™๐‘ค๐‘Ž๐‘ฆ๐‘  ๐‘ก๐‘Ÿ๐‘ข๐‘’
⟨๐‘ฃ, ๐‘ค⟩๐’ฐ = ⟨๐‘ƒ๐‘ฆ, ๐‘ค⟩๐’ฐ
=
Now ๐’ฉ๐‘ƒ is orthogonal to โ„›๐‘ƒ
๐‘ƒ=๐‘ƒ∗
⟨๐‘ฆ, ๐‘ƒ∗ ๐‘ค⟩๐’ฐ = ⟨๐‘ฆ, ๐‘ƒ๐‘ค⟩๐’ฐ = ⟨๐‘ฆ, 0⟩๐’ฐ = 0
If ๐’ฑ is a subspace of ๐’ฐ
๐’ฑ ⊥ - the orthogonal complement of ๐’ฑ in ๐’ฐ = {๐‘ข ∈ ๐’ฐ|⟨๐‘ข, ๐‘ฃ⟩๐’ฐ = 0 ∀๐‘ฃ ∈ ๐’ฑ}
Claim: โ„›๐‘ƒ⊥ ⊆ ๐’ฉ๐‘ƒ
Proof: Suppose that ๐‘ค ∈ โ„›๐‘ƒ⊥ ⇒ ⟨๐‘ƒ๐‘ฆ, ๐‘ค⟩ = 0 for all ๐‘ฆ ∈ ๐’ฐ
This is the same as: ⟨๐‘ฆ, ๐‘ƒ ∗ ๐‘ค⟩ = ⟨๐‘ฆ, ๐‘ƒ๐‘ค⟩
๐‘–๐‘›
๐‘Ž๐‘Ÿ๐‘ก๐‘–๐‘๐‘ข๐‘™๐‘Ž๐‘Ÿ
So we know that 0 = ⟨๐‘ฆ, ๐‘ƒ๐‘ค⟩ ∀๐‘ฆ ∈ ๐’ฐ ⇒
⟨๐‘ƒ๐‘ค, ๐‘ƒ๐‘ค⟩ = 0 ⇒ ๐‘ƒ๐‘ค = 0
Orthogonal Decomposition
๐’ฐ =๐‘‰⊕๐’ฒ
Means that ๐‘ข = ๐‘ฃ + ๐‘ค, ๐‘ฃ ∈ ๐’ฑ, ๐‘ค ∈ ๐’ฒ
Moreover, ⟨๐‘ฃ, ๐‘ค⟩๐’ฐ = 0 ∀๐‘ฃ ∈ ๐’ฑ, ๐‘ค ∈ ๐’ฒ
We really have 3 symbols!
๐’ฑ + ๐’ฒ = {๐‘ฃ + ๐‘ค|๐‘ฃ ∈ ๐’ฑ, ๐‘ค ∈ ๐’ฒ}
๐’ฑ โˆ” ๐’ฒ when ๐’ฑ ∩ ๐’ฒ = {0}
๐’ฑ ⊕ ๐’ฒ when ๐’ฑ ⊥ ๐’ฒ (๐’ฒ = ๐’ฑ ⊥ ) (only defined in an inner product space)
Let ๐’ฐ be an inner product space over ๐”ฝ, let ๐’ฑ be a subspace of ๐’ฐ with basis {๐‘ฃ1 , … , ๐‘ฃ๐‘˜ }
Then ๐’ฐ = ๐’ฑ ⊕ ๐’ฒ
Let ๐‘ข = ๐‘ฃ + ๐‘ค, ๐‘ฃ ∈ ๐’ฑ, ๐‘ค ∈ ๐’ฒ, ⟨๐‘ฃ, ๐‘ค⟩ = 0
Question: Given ๐‘ข find ๐‘ฃ
๐‘˜
๐‘˜
๐‘ข = ∑ ๐‘๐‘— ๐‘ฃ๐‘— + ๐‘ข − ∑ ๐‘๐‘— ๐‘ฃ๐‘—
Try to choose ๐‘๐‘— such that ⟨๐‘ข −
๐‘—=1
๐‘˜
∑๐‘—=1 ๐‘๐‘— ๐‘ฃ๐‘— , ๐‘ฃ๐‘– ⟩๐’ฐ
๐‘—=1
= 0, ๐‘– = 1, … , ๐‘˜
๐‘˜
⟨๐‘ข, ๐‘ฃ๐‘– ⟩๐’ฐ = ๐‘ข − ∑ ๐‘๐‘— ⟨๐‘ฃ
โŸ๐‘— , ๐‘ฃ๐‘– ⟩ = 0
๐‘—=1
๐‘”๐‘–๐‘—
We would like to achieve the equality:
๐‘˜
∑ ๐‘”๐‘–๐‘— ๐‘๐‘— = ⟨๐‘ข, ๐‘ฃ๐‘– ⟩๐’ฐ ,
๐‘– = 1, … , ๐‘˜
๐‘—=1
๐‘1
๐‘1
โ‹ฎ
⟨๐‘ข,
⟩
Let’s choose a vector ๐‘ = [ ] , ๐‘๐‘– =
๐‘ฃ๐‘– ๐’ฐ , ๐‘ = [ โ‹ฎ ]
๐‘๐‘˜
๐‘๐‘˜
So we got that: ๐’ข๐‘ = ๐‘
Since all vectors are linearly independent, ๐’ข is invertible. So: ๐‘ = ๐’ข −1 ๐‘
๐‘ƒ๐’ฑ ๐‘ข = ∑๐‘˜๐‘—=1 ๐‘๐‘— ๐‘ฃ๐‘— where ๐‘ = ๐’ข −1 ๐‘
Suppose ๐’ฐ = โ„‚๐‘› with the standard inner product.
⟨๐‘ข, ๐‘ค⟩ = ∑ ๐‘ข๐‘— ฬ…ฬ…ฬ…
๐‘ค๐‘— = ๐‘ค ๐ป ๐‘ข
⟨๐‘ข, ๐‘ฃ1 ⟩
๐‘ฃ1๐ป ๐‘ข
๐‘ = [ โ‹ฎ ] = [ โ‹ฎ ],
⟨๐‘ข, ๐‘ฃ๐‘˜ ⟩
๐‘ฃ๐‘˜๐ป ๐‘ข
๐‘‰ = [๐‘ฃ1
๐‘ฃ1๐ป
… ๐‘ฃ๐‘˜ ] ⇒ ๐‘‰ = [ โ‹ฎ ] ⇒ ๐‘ = ๐‘‰ ๐ป ๐‘ข
๐‘ฃ๐‘˜๐ป
๐ป
๐‘”๐‘–๐‘— = ⟨๐‘ฃ๐‘— , ๐‘ฃ๐‘– ⟩ = ๐‘ฃ๐‘–๐ป ๐‘ฃ๐‘— = ๐‘’๐‘–๐ป ๐‘‰ ๐ป ๐‘‰๐‘’๐‘—
So ๐’ข = ๐‘‰ ๐ป ๐‘‰
๐‘˜
๐‘ƒ๐’ฑ ๐‘ข = ∑ ๐‘๐‘— ๐‘ฃ๐‘— = ๐‘‰๐‘
๐‘—=1
Because ๐‘ = ๐’ข −1 ๐‘ = (๐‘‰ ๐ป ๐‘‰)−1 ๐‘‰ ๐ป ๐‘ข
So ๐‘ƒ๐’ฑ ๐‘ข = ๐‘‰(๐‘‰ ๐ป ๐‘‰)−1 ๐‘‰ ๐ป ๐‘ข
๐‘‰ = [๐‘ฃ1 … ๐‘ฃ๐‘˜ ]
TODO: Draw area of triangle…
โ„Ž โˆ™ โ€–๐‘Žโ€– = area.
๐’ฑ = ๐‘ ๐‘๐‘Ž๐‘›{๐‘Ž}
๐‘ƒ๐’ฑ ๐‘ = ๐‘Ž(๐‘Ž๐‘‡ ๐‘Ž)−1 ๐‘Ž๐‘‡ ๐‘
๐‘ = ๐‘Ž(๐‘Ž๐‘‡ ๐‘Ž)−1 ๐‘Ž๐‘‡ ๐‘ + ๐‘ − ๐‘Ž(๐‘Ž๐‘‡ ๐‘Ž)−1 ๐‘Ž๐‘‡ ๐‘
So area = โ€–๐‘ − ๐‘Ž(๐‘Ž๐‘‡ ๐‘Ž)−1 ๐‘Ž๐‘‡ ๐‘โ€– โˆ™ โ€–๐‘โ€–
๐‘Ž๐‘Ÿ๐‘’๐‘Ž2 = ⟨๐‘ − ๐‘Ž(๐‘Ž๐‘‡ ๐‘Ž)−1 ๐‘Ž๐‘‡ ๐‘, ๐‘ − ๐‘Ž(๐‘Ž๐‘‡ ๐‘Ž)−1 ๐‘Ž๐‘‡ ๐‘⟩ โˆ™ โ€–๐‘Žโ€–2
(⟨๐‘, ๐‘⟩ − ⟨๐‘ƒ๐’ฑ ๐‘, ๐‘⟩) โˆ™ โ€–๐‘Žโ€–2 = (๐‘ ๐‘‡ ๐‘ − ๐‘ ๐‘‡ ๐‘Ž(๐‘Ž๐‘‡ ๐‘Ž)−1 ๐‘Ž๐‘‡ ๐‘)๐‘Ž๐‘‡ ๐‘Ž = (๐‘ ๐‘‡ ๐‘)(๐‘Ž๐‘‡ ๐‘Ž) − (๐‘ ๐‘‡ ๐‘Ž)(๐‘Ž๐‘‡ ๐‘)
Since: ⟨๐‘ − ๐‘ƒ๐’ฑ ๐‘, ๐‘ƒ๐’ฑ ๐‘⟩ = ⟨๐‘ƒ๐’ฑ ๐‘ − ๐‘๐’ฑ2 ๐‘, ๐‘⟩ = ⟨0, ๐‘⟩ = 0
๐‘‡
๐‘‡
(๐‘Ž๐‘Ÿ๐‘’๐‘Ž)2 = [๐‘Ž๐‘‡ ๐‘Ž ๐‘Ž๐‘‡ ๐‘] = det ๐ถ โˆ™ ๐ถ ๐‘‡
๐‘ ๐‘Ž ๐‘ ๐‘
Where ๐ถ = [๐‘Ž ๐‘]
๐‘ ๐‘–๐‘›๐‘๐‘’ ๐‘–๐‘› โ„2
=
๐ถ=๐ถ ๐‘‡
det ๐ถ โˆ™ det ๐ถ ๐‘‡ = (det ๐ถ)2
If ๐‘Ž, ๐‘ ∈ โ„2
Correction
When we said triangle area earlier, we meant the area of a parallelogram!
๐‘Ž, ๐‘ ∈ โ„2 , then the area is = det[๐‘Ž
๐‘]
Back to the lesson
๐’ฐ inner product space, ๐’ฑ subspace of ๐’ฐ with basis ๐‘ฃ1 , … , ๐‘ฃ๐‘˜ . Let ๐‘ƒ๐’ฑ ๐‘ข denote the orthogonal
projection of ๐‘ข onto ๐’ฑ. Have a recipe.
In addition to what we’ve already done,
Fact (n): min{โ€–๐‘ข − ๐‘ฃโ€–|๐‘ฃ ∈ ๐’ฑ} is achieved by choosing ๐‘ฃ = ๐‘ƒ๐’ฑ ๐‘ข.
2
โ€–๐‘ข − ๐‘ฃโ€–2 = โ€–
๐‘ƒ๐’ฑ ๐‘ข − ๐‘ฃ โ€–
โ€–๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข + โŸ
โ€–
∈๐’ฑ
๐ท๐‘’๐‘›๐‘œ๐‘ก๐‘’
๐‘Ž๐‘  ๐‘ค
Claim: it equals to โ€–๐‘ข − ๐‘ƒ๐’ฑ ๐‘ขโ€–2 + โ€–๐‘ƒ๐’ฑ ๐‘ข − ๐‘ฃโ€–
๐‘๐‘ฆ ๐‘‘๐‘’๐‘“
Proof: โ€–๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข + ๐‘คโ€–2 = ⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข + ๐‘ค, ๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข + ๐‘ค⟩ =
⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข, ๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข⟩ + ⟨๐‘ข − ๐‘ƒ๐’ฑ , ๐‘ค⟩ + ⟨๐‘ค, ๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข⟩ + ⟨๐‘ค, ๐‘ค⟩
Harry claims the second and third piece is zero. Why?
If ๐‘ค ∈ ๐’ฑ then
⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข, ๐‘ค⟩ = ⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข, ๐‘ƒ๐’ฑ ๐‘ค⟩ = ⟨๐‘ƒ๐’ฑ∗ (๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข), ๐‘ค⟩ = ⟨๐‘ƒ๐’ฑ (๐‘ข − ๐‘ƒ๐’ฑ ), ๐‘ค⟩ = ⟨0, ๐‘ค⟩ = 0
So we are left with โ€–๐‘ข − ๐‘ƒ๐’ฑ ๐‘ขโ€–2 + โ€–๐‘ƒ๐’ฑ ๐‘ข − ๐‘ฃโ€–
Since we want the expression as small as possible, we can only play with the right hand part
(since the first hand part is determined!)
So we want to choose the ๐‘ฃ such that โ€–๐‘ƒ๐’ฑ ๐‘ข − ๐‘ฃโ€– = 0
We can do so if we choose ๐‘ฃ = ๐‘ƒ๐’ฑ ๐‘ข
So minโ€–๐‘ข − ๐‘ฃโ€–2 ๐‘ฃ ∈ ๐’ฑ is in fact
โ€–๐‘ข − ๐‘ƒ๐’ฑ ๐‘ขโ€–2 = ⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข, ๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข⟩ = ⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข, ๐‘ข⟩ − ⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข, ๐‘ƒ๐’ฑ ๐‘ข⟩
But ⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข, ๐‘ƒ๐’ฑ ๐‘ข⟩ = 0!
So it actually equals ⟨๐‘ข − ๐‘ƒ๐’ฑ ๐‘ข, ๐‘ข⟩ = ⟨๐‘ข, ๐‘ข⟩ − ⟨๐‘ƒ๐’ฑ ๐‘ข, ๐‘ข⟩ = ⟨๐‘ข, ๐‘ข⟩ − ⟨๐‘ƒ๐’ฑ2 ๐‘ข, ๐‘ข⟩
But ๐‘ƒ๐’ฑ∗ = ๐‘ƒ๐’ฑ !
So equals: ⟨๐‘ข, ๐‘ข⟩ − ⟨๐‘ƒ๐’ฑ ๐‘ข, ๐‘ƒ๐’ฑ∗ ๐‘ข⟩ = โ€–๐‘ขโ€–2 − โ€–๐‘ƒ๐’ฑ ๐‘ขโ€–2
๐‘˜
๐‘ƒ๐’ฑ ๐‘ข = ∑ ๐‘๐‘— ๐‘ฃ๐‘— ,
๐‘ = ๐’ข −1 ๐‘,
1
๐‘=[
⟨๐‘ข, ๐‘ฃ๐‘– ⟩
โ‹ฎ ]
⟨๐‘ข, ๐‘ฃ๐‘˜ ⟩
Another special case, when ๐‘ฃ1 , … , ๐‘ฃ๐‘˜ are orthonormal to the given inner product.
๐‘”๐‘–๐‘— = ⟨๐‘ฃ๐‘— , ๐‘ฃ๐‘– ⟩ = 1 if ๐‘– = ๐‘— and 0 if ๐‘– ≠ ๐‘—
In this case, ๐’ข = ๐ผ ⇒ ๐‘ = ๐‘ ⇒
๐‘˜
๐‘ƒ๐’ฑ ๐‘ข = ∑⟨๐‘ข, ๐‘ฃ๐‘— ⟩๐’ฐ ๐‘ฃ๐‘—
๐‘—=1
Suppose ๐’ฐ is an inner product space, and suppose {๐‘ข1 , … , ๐‘ข๐‘™ } is an orthonormal basis for ๐’ฐ.
Given ๐‘ข ∈ ๐’ฐ ⇒ ๐‘ข = ∑๐‘™๐‘—=1 ๐‘๐‘— ๐‘ข๐‘—
๐‘™
⟨๐‘ข, ๐‘ข๐‘– ⟩ = ∑ ๐‘๐‘— ⟨๐‘ข๐‘— , ๐‘ข๐‘– ⟩ = ๐‘๐‘–
๐‘—=1
2
So ⟨๐‘ข, ๐‘ข⟩ = ∑๐‘™๐‘—=1|๐‘๐‘— |
Gram-Schimdt
Given a linearly independent set of vectors ๐‘ข1 , … , ๐‘ข๐‘˜ in ๐’ฐ.
Claim: There exists an orthonormal set ๐‘ฃ1 , … , ๐‘ฃ๐‘˜ such that
๐‘ ๐‘๐‘Ž๐‘›{๐‘ฃ1 , … , ๐‘ฃ๐‘— } = ๐‘ ๐‘๐‘Ž๐‘›{๐‘ข1 , … , ๐‘ข๐‘— } for ๐‘— = 1, … , ๐‘˜
Proof:
๐‘ข
Define ๐‘ฃ1 = โ€–๐‘ข1 โ€– ⇒ ⟨๐‘ฃ1 , ๐‘ฃ1 ⟩ −
1
⟨๐‘ข1 ,๐‘ข1 ⟩
โ€–๐‘ข1 โ€–2
=1
Introduce the notation: ๐’ฑ๐‘— = ๐‘ ๐‘๐‘Ž๐‘›{๐‘ฃ1 , … , ๐‘ฃ๐‘— }
Define ๐‘ค2 = ๐‘ข2 − ๐‘ƒ๐’ฑ1 ๐‘ข2
We are doing it since this vector will be orthogonal to ๐‘ฃ1 .
Also claim ๐‘ค2 ≠ 0.
๐‘ค
And define: ๐‘ฃ2 = โ€–๐‘ค2 โ€–
2
Keep on doing that.
Define ๐‘ค3 = ๐‘ข3 − ๐‘ƒ๐’ฑ2 ๐‘ข3
So the vector is also orthogonal!
๐‘ค
And then ๐‘ฃ3 = โ€–๐‘ค3 โ€–
3
Now ๐‘ค4 = ๐‘ข4 − ๐‘ƒ๐’ฑ3 ๐‘ข4 = ๐‘ข4 − {⟨๐‘ข4 , ๐‘ฃ1 ⟩๐‘ฃ1 + ⟨๐‘ข4 , ๐‘ฃ2 ⟩๐‘ฃ2 + ⟨๐‘ข4 , ๐‘ฃ3 ⟩๐‘ฃ3 } such that
๐‘ค4
๐‘ฃ4 =
โ€–๐‘ค4 โ€–
Download