Orthogonal complement and projection notes

advertisement
Recall that any linearly independent set of Rm has at most m elements. For any subspace X of Rm
(including Rm itself), we can find a basis for that subspace; call it B = {b1 , ..., bn }. We can find an
orthonormal basis Q for X via the Gram-Schmidt algorithm applied to B; since the columns of B span
X, the columns of Q span X. On the other hand, the columns of a matrix with orthonormal columns are
linearly independent: suppose Qa = 0, then a = Ia = QT Qa = 0.
1
Orthogonal complement
Now let W ⊂ Rm . Define W ⊥ by
W ⊥ = {x ∈ Rn , hx, wi = 0 ∀w ∈ W }
Note that even if W is not a subspace of Rn , W ⊥ is: if x, y ∈ W ⊥ , a ∈ R, and w ∈ W ,
hx + y, wi = hx, wi + hy, wi = 0,
and
hax, wi = a hx, wi = 0,
So W ⊥ is closed under addition and scalar multiplication.
If W itself is a subspace, say of dimension n, we can divide up Rn into W and W ⊥ in a very natural
way: every v ∈ Rn can be written uniquely as
v = v W + vW ⊥ ,
(1.1)
where vW ∈ W, and vW ⊥ ∈ W ⊥ . To see this, let Q be an orthonormal basis for W , and let v ∈ Rn . Set
vW = QQT v, and vW ⊥ = v − vW .
It is clear that vW + vW ⊥ ; it is also clear that vW ∈ W , because each column of Q is in W , and vW is a
linear combination of columns of Q, and W is closed under linear combinations. On the other hand, if
w ∈ W , since Q is a basis for W , there exists a such that w = Qa. Then since QT Q = In ,
wT vW ⊥ = (Qa)T (v − QQT v) = aT QT v − aT QT v = 0,
so vW ⊥ ∈ W ⊥ , and we have found a decomposition of v into an element from W and from W ⊥ . These
elements are unique: if v = a + b, where a ∈ W and b ∈ W ⊥ ,
vW + vW ⊥ = v = a + b
so
vW − a = b − vW ⊥ ,
and since the left hand side is in W , and the right hand side in W ⊥ , both vW − a and b − vW ⊥ are in
W ∩ W ⊥ . But any vector u ∈ W ∩ W ⊥ satisfies hu, ui = 0, by the definition of W ⊥ , and so u = 0. Thus
a = vW and b = vW ⊥ .
⊥
= W . If w ∈ W , and x ∈ W ⊥ , by definition,
We can now show that if W is a subspace, W ⊥
hw, xi = 0, so w ∈ W ⊥ . On the other hand, if v ∈
/ W , vW ⊥ 6= 0, and so
hv, vW ⊥ i = hvW + vW ⊥ , vW ⊥ i = hvW ⊥ , vW ⊥ i = ||vW ⊥ ||2 > 0,
and so v ∈
/ W ⊥.
1
2
Orthogonal projection
We can thus define a mapping PW : Rm 7→ Rm via v → vW (the above arguments show that this map is
defined and unambiguous for every v ∈ Rm , and is linear). This is called the orthogonal projection onto
W.
• If w ∈ W , PW w = w, and if x ∈ W ⊥ , PW x = 0.
• for any orthonormal basis Q of W , by the uniqueness of the decomposition (1.1),
PW v = QQT v.
• and
PW ⊥ v = Im − QQT v.
• If Q is any matrix with orthonormal columns, QQT is the orthonormal projection onto the column
space of Q.
In class, many times we have mentioned that if m > n, and Q is an m × n orthonormal matrix, QQT is
not the identity; now we can see it is the identity on the column space of Q, and it is the zero mapping
on the orthogonal complement of the column space of Q...
The fact that the choice of basis in the definition of the projection is irrelevant suggest that there is
a basis-free way of writing the orthogonal projection onto W , and indeed,
PW v = arg min ||v − w||2 .
w∈W
To see this, let w ∈ W . Then
||v − w||2 = ||vW + vW ⊥ − w||2 = ||vW − w||2 + ||vW ⊥ ||2 .
But
||v − vW ||2 = ||vW + vW ⊥ − vW ||2 = ||vW ⊥ ||2 ,
so
||v − w||2 = ||vW − w||2 + ||v − vW ||2 ≥ ||v − vW ||2 ,
with equality if and only if w = vW .
3
Some extra notes and perspectives
This section will not tested on the midterm, but may clarify some things but also may unclarify things,
depending on how you like to think: feel free to ignore it.
I wrote W ⊂ Rm instead of W ⊂ V , where V is a finite dimensional vector space, to emphasize that
we wrote everything in the standard basis. This is unecessary, and you may consider it inelegant or even
confusing- but if this actually bothers you, you can replace each expression of the form QQT v by an
expression of the form
n
X
hQj , vi Qj ,
j=1
where Qj is the jth element in the basis Q of W . The point is that here, Q is not a matrix, but rather a
list of vectors, and is not necessarily expressed in the standard basis of Rm .
Moreover, if you want to avoid bases altogether, you can, by making the definition of the projection
of v onto W to be the point closest to v in W , that is, define
vW = arg min ||v − w||2 .
w∈W
2
Using a bit of calculus that is beyond the scope of this course (but is not too hard!), you can check that
a minimizer exists. Then if w ∈ W , consider the degree 2 polynomial in t
||(v − vW ) + tw||2 = ||v − vW ||2 − 2t hv − vW , wi + t2 ||w||2 .
Since this polynomial takes a minimum at t = 0 (because of the definition of vW ), the term hv − vW , wi = 0
(draw a picture or take a derivative in t to see this). So v − vW is orthogonal to every element in W .
Moreover, if w ∈ W ,
||v − w||2 = ||v − vW + vW − w||2 = ||v − vW ||2 + ||vW − w||2 ≥ ||v − vW ||2 ,
with equality only if w = vW ; that is, the projection is unique. So as soon as we know that for each
v ∈ V there is at least one minimizer for arg min ||v − w||2 , we know that there is exactly one, and so the
w∈W
projection is well defined this way. You can further check that the mapping is linear, that it leads to the
definition of the orthogonal complement, etc.
Again, I will not test on this subsection, but wanted to include this in case you were bothered by the
fact that the basis appeared in the construction of the projection even though the projection doesn’t care
which basis you use.
3
Download