Chapter 6: Inner Product Spaces Section 6.1: Inner products

advertisement
Chapter 6: Inner Product Spaces
Section 6.1: Inner products
Definition: An inner product on a real vector space V is a function that
associates a real number h~u, ~vi with each pair of vectors ~u, ~v in V such that
the following axioms are satisfied for all ~u, ~v , ~z in V and all scalars k
1. h~u, ~vi = h~v , ~ui (symmetry axiom)
2. h~u + ~v , ~zi = h~u, ~z i + h~v, ~z i (additivity axiom)
3. hk~u, ~vi = kh~v , ~ui (homogeneity axiom)
4. h~v, ~v i ≥ 0, h~v, ~vi = 0 if and only if ~v = 0 (positivity axiom)
A real vector space with an inner product is called a real inner product space.
Example:
1
Definition: If ~u = (u1 , u2 , . . . , un ), ~v = (v1 , v2 , . . . , vn ) are vectors in Rn
and w1 , w2 , . . . , wn are positive real numbers, then the formula
h~u, ~vi = w1 u1 v1 + w2 u2 v2 + · · · + wn un vn
defines an inner product on Rn called the weighted inner product with weights
w1 , w2 , . . . , wn .
Note: Unless specified otherwise, Rn is assumed to have the usual Euclidean inner product which we refer to as Euclidean n-space.
Example:
2
Definition: If V is an inner product space, then the norm (or length) of
a vector ~u in V is denoted by k~uk and is defined by
k~uk = h~u, ~ui1/2
The distance between two points (vectors) ~u and ~v is denoted by d(u, v) and
is defined by
d(~u, ~v ) = k~u − ~v k
If a vector has norm 1, it is called a unit vector.
Example:
3
Definition: If A is an invertible nxn matrix, ~u, ~v are vectors in Rn , then
the inner product on Rn generated by A is defined as
h~u, ~v i = A~u · A~v = (A~v )T A~u = ~v T AT A~u
Example:
Theorem 6.1.1 Properties of inner products
If ~u, ~v , w
~ are vectors in real inner product space, k is a scalar, then
1. h0, ~ui = h~u, 0i = ~0
2. h~u, ~v + wi
~ = h~u, ~v i + h~u, wi
~
3. h~u, k~vi = kh~u, ~vi
4. h~u − ~v, wi
~ = h~u, wi
~ − h~v , wi
~
5. h~u, ~v − wi
~ = h~u, ~vi − h~u, wi
~
Note: These relations hold for all real inner product spaces.
4
Section 6.2: Angle and orthogonality in inner product spaces
Theorem 6.2.1 Cauchy-Schwarz inequality
If ~u, ~v are vectors in real inner product space, then
|h~u, ~v i| ≤ k~ukk~vk
Equivalently,
h~u, ~vi2 ≤ h~u, ~uih~v, ~vi
or
h~u, ~vi2 ≤ k~uk2 k~v k2
Theorem 6.2.2: Properties of length
If ~u, ~v are vectors in inner product space V , k is a scalar, then
1. k~uk ≥ 0
2. k~uk = 0 if and only if k~uk = 0
3. kk~uk = |k|k~uk
4. k~u + ~v k ≤ k~uk + k~vk (triangle inequality)
Theorem 6.2.3: Properties of distance
If ~u, ~v , w
~ are vectors in inner product space V , k is a scalar, then
1. d(~u, ~v) ≥ 0
2. d(~u, ~v) = 0 if and only if ~u = ~v
3. d(~u, ~v) = d(~v , ~u)
4. d(~u, ~v) ≤ d(~u, w)
~ + d(w,
~ ~v ) (triangle inequality)
Definition: If ~u, ~v are vectors in a real inner product space, the angle between ~u and ~v
is defined by
h~u, ~v i
cos(θ) =
,0≤θ≤π
k~ukk~v k
Definition: Two vectors ~u, ~v in an inner product space are orthogonal
if h~u, ~v i = 0
5
Theorem 6.2.4: Generalized Pythagorean Theorem
If ~u, ~v are orthogonal vectors in a real inner product space, then
k~u + ~v k = k~uk2 + k~vk2
Note: All of these theorems and definitions hold for ANY vector space
with inner product!
Example:
6
Definition: Let W be a subspace of an inner product space V . A vector ~u in V is orthogonal to W if it is orthogonal to every vector in W ,
and the set of all vectors in V that are orthogonal to W is called the
orthogonal complement of W , denoted W ⊥ .
Example:
7
Terminology: We call W and W ⊥ orthogonal complements.
Theorem 6.2.5: Properties of orthogonal complements
If W is a subspace of a finite dimensional inner product space V , then
1. W ⊥ is a subspace of V
2. The only vector common to W and W ⊥ is ~0
3. The orthogonal complement of W ⊥ is W . That is, (W ⊥ )⊥ = W
Theorem 6.2.6: If A is an mxn matrix then
1. the nullspace of A and the row space of A are orthogonal complements
in Rn with respect to the Euclidean inner product. That is,
(Row space of A)⊥ = nullspace of A
2. The nullspace of AT and the column space of A are orthogonal complements in Rm with respect to the Euclidean inner product. That
is,
(Column space of A)⊥ = nullspace of AT
Example:
8
Theorem 6.2.7: If TA : Rn → Rn is multiplication by an n × n matrix
A, then the following are equivalent
1. A is invertible
2. A~x = 0 has only the trivial solution
3. The reduced row echelon form of A is In
4. A can be expressed as a product of elementary matrices
5. A~x = ~b is consistent for every n × 1 matrix ~b
6. A~x = ~b has exactly one solution for every n × 1 matrix ~b
7. det(A) 6= 0
8. The range of TA is Rn
9. TA is one-to-one
10. The column vectors of A are linearly independent
11. The row vectors of A are linearly independent
12. The column vectors of A span Rn
13. The row vectors of A span Rn
14. The column vectors of A form a basis for Rn
15. The row vectors of A form a basis for Rn
16. A has rank n
17. A has nullity 0
18. The orthogonal complement of the nullspace of A is Rn
19. The orthogonal complement of the row space of A is ~0
9
Section 6.3: Orthonormal bases; Gram-Schmidt process; QR-decomposition
Definition: A set of vectors in an inner product space is called an
orthogonal set if all pairs of distinct vectors in the set are orthogonal. An
orthogonal set in which each vector has norm 1 is called orthonormal. A
basis of an inner product space consisting of orthogonal vectors is called an
]underlineorthogonal basis while a basis consisting of orthonormal vectors is
called an orthonormal basis.
Theorem 6.3.1: If S = {v~1 , v~2 , . . . , v~n } is an orthonormal basis for an
inner product space V , and ~u is any vector in V , then
~u = h~u, v~1 iv~1 + h~u, v~2 iv~2 + · · · + h~u, v~n iv~n
Example:
10
Theorem 6.3.2: If S is an orthonormal basis for an n-dimensional inner
product space, and if
(~u)S = (u1 , u2 , . . . , un ), (~v)S = (v1 , v2 , . . . , vn )
then
p
u21 + u22 + · · · + u2n
p
2. d(~u, ~v) = (u1 − v1 )2 + (u2 − v2 )2 + · · · + (un − vn )2
1. k~uk =
3. h~u, ~vi = u1 v1 + u2 v2 + · · · + un vn
Example:
Theorem 6.3.3: If S = {v~1 , v~2 , . . . , v~n } is an orthogonal set of nonzero
vectors in an inner product space, then S is linearly independent.
Example:
11
Theorem 6.3.4: Projection theorem
If W is a finite dimensional subspace of an inner product space V , then every
vector ~u in V can be expressed in exactly one way as
~u = w~1 + w~2
where w~1 is in W and w~2 is in W ⊥ . w~1 is called the orthogonal projection of ~u on W ,
denoted projW ~u, and w~2 is called the complement of ~u orthogonal to W , denoted projW ⊥ ~u, and we can write
~u = projW ~u + projW ⊥ ~u = projW ~u + (~u − projW ~u)
Theorem 6.3.5: Let W be a finite dimensional subspace of an inner
product space V
1. If {v~1 , . . . , v~r } is an orthonormal basis for W , and ~u any vector in V
then
projW ~u = h~u, v~1 iv~1 + h~u, v~2 iv~2 + · · · + h~u, v~r iv~r
2. If {v~1 , . . . , v~r } is an orthogonal basis for W , and ~u any vector in V then
projW ~u =
h~u, v~r i
h~u, v~1 i
v~1 + · · · +
v~r
2
kv~1 k
kv~r k2
Example:
12
Theorem 6.3.6: Every nonzero finite dimensional inner product space
has an orthonormal basis.
Gram-Schmidt process: method to convert an arbitrary basis into an
orthogonal basis. This basis can then be normalized to produce an orthonormal basis.
Example:
13
Summary of Gram-Schmidt process
Let S = {u~1, . . . , u~n } be an arbitrary basis of an inner product space V
v~1 = u~1
hu~2, v~1 i
v~1
kv~1 k2
hu~3, v~2 i
hu~3, v~1 i
v
~
−
v~2
= u~3 −
1
kv~1 k2
kv~2 k2
..
.
hu~n , v~1 i
hu~n , v~2 i
hu~n , vn−1
~ i
= u~n −
v~1 −
v~2 − · · · −
vn−1
~
2
2
kv~1 k
kv~2 k
kvn−1
~ k2
v~2 = u~2 −
v~3
v~n
Then S ′ = {v~1 , . . . , v~n } is an orthogonal basis for V and S ′′ = { kvv~~11 k , . . . , kvv~~nn k }
is an orthonormal basis for V .
Theorem 6.3.7: QR-decomposition
If A is an mxn matrix with linearly independent column vectors u~i , then A
can be factored as
A = QR
where Q is an mxn matrix with orthonormal column vectors q~i , and R is an
invertible upper triangular matrix where


hu~1 , q~1 i hu~2 , q~1 i · · · hu~n , q~1 i

0
hu~2 , q~2 i · · · hu~n , q~2 i 


R=

..
..


.
.
0
0
Example:
14
· · · hu~n , q~n i
Section 6.5: Change of basis
Recall: If the bases S, S ′ span the same space then the vectors in S can
be written as a linear combination of vectors in S ′ , and the vectors in S ′ can
be written as a linear combination of vectors in S. Also recall the coordinate
vector of ~u with respect to S, written as a column vector, is


k1
 k2 


[~v ]S =  .. 
 . 
kn
Change of basis: If B = {b~1 , b~2 , . . . , b~n } is a basis for inner product
space V and C = {~
c1 , c~2 , . . . , c~n } is another basis for V then there is a unique
nxn matrix PC←B , denoted the transition matrix from B to C, such that
[~v ]C = PC←B [~v ]B
where the columns of PC←B are the coordinate vectors of the basis B relative
to the basis C. That is,
h
i
PC←B = [b~1 ]C , [b~2 ]C , . . . , [b~n ]C
Theorem 6.5.1: If PC←B is the transition matrix from a basis B to
a basis C for a finite dimensional space V , then PC←B is invertible and
(PC←B )−1 is the transition matrix from C to B. That is,
[~v ]C = PC←B [~v ]B
[~v ]B = (PC←B )−1 [~v ]C
Example:
15
Section 6.6: Orthogonal matrices
Definition: A square matrix A with the property
A−1 = AT
is said to be an orthogonal matrix. That is, a square matrix is orthogonal if
and only if
AAT = AT A = I
Example:
16
Theorem 6.6.1: For an nxn matrix A, the following statements are
equivalent
1. A is orthogonal
2. The row vectors of A form an orthonormal set in Rn with Euclidean
inner product
3. The column vectors of A form an orthonormal set in Rn with Euclidean
inner product
Theorem 6.6.2:
1. The inverse of an orthogonal matrix is orthogonal
2. The product of orthogonal matrices is orthogonal
3. If A is orthogonal, then det(A) = ±1
Example:
17
Theorem 6.6.3: For an nxn matrix A, the following statements are
equivalent
1. A is orthogonal
2. kA~xk = k~xk for all ~x in Rn
3. A~x · A~y = ~x · ~y for all ~x, ~y in Rn
Example:
18
Definition: If T : Rn → Rn is multiplication by an orthogonal matrix
A, then T is called an orthogonal operator on Rn .
Theorem 6.6.4: If PC←B is the transformation matrix from one orthonormal basis to another orthonormal basis for an inner product space,
then PC←B is an orthogonal matrix. That is,
(PC←B )−1 = (PC←B )T
19
Download