18.06.11: ‘Nullspace’ Lecturer: Barwick Monday 29 February 2016

advertisement
18.06.11: ‘Nullspace’
Lecturer: Barwick
Monday 29 February 2016
18.06.11: ‘Nullspace’
Take some vectors ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— ∈ ๐‘‰ and make them into the columns of an ๐‘› × ๐‘˜
matrix
๐ด = ( ๐‘ฃ1โƒ— ๐‘ฃ2โƒ— โ‹ฏ ๐‘ฃ๐‘˜โƒ— ) .
Multiplication by ๐ด is a map ๐‘‡๐ด โˆถ R๐‘˜
๐‘‰ that carries ๐‘’๐‘–ฬ‚ to ๐‘ฃ๐‘–โƒ— .
(1) The vectors ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— ∈ ๐‘‰ are linearly independent if and only if ๐‘‡๐ด is
injective.
(2) The vectors ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— span ๐‘‰ if and only if ๐‘‡๐ด is surjective.
(3) The vectors ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— are a basis of ๐‘‰ if and only if ๐‘‡๐ด is bijective.
18.06.11: ‘Nullspace’
Let’s see why this works.
First, let’s unpack what the injectivity of ๐‘‡๐ด would mean. It’s this condition for
any ๐‘ฅ,โƒ— ๐‘ฆโƒ— ∈ R๐‘˜ :
if ๐ด๐‘ฅโƒ— = ๐ด๐‘ฆ,โƒ— then ๐‘ฅโƒ— = ๐‘ฆ.โƒ—
Defining ๐‘งโƒ— โ‰” ๐‘ฅโƒ— − ๐‘ฆ,โƒ— we see that we want to show that
if ๐ด๐‘งโƒ— = 0,โƒ— then ๐‘งโƒ— = 0.โƒ—
(In other words, we’re saying that the kernel of ๐ด consists of just the zero
vector!)
18.06.11: ‘Nullspace’
Now ๐ด๐‘งโƒ— is a linear combination of the column vectors ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— with coefficients given by the components of ๐‘ง.โƒ— So the injectivity of ๐‘‡๐ด is equivalent to
the following:
๐‘˜
if ∑ ๐‘ง๐‘– ๐‘ฃ๐‘–โƒ— = 0,โƒ— then ๐‘ง1 = โ‹ฏ = ๐‘ง๐‘˜ = 0.
๐‘–=1
That’s exactly what it means for ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— to be linearly independent.
18.06.11: ‘Nullspace’
Now let’s unpack what the surjectivity of ๐‘‡๐ด would mean. It’s the condition
that for any vector ๐‘คโƒ— ∈ ๐‘‰, there exists a vector ๐‘ฅโƒ— ∈ R๐‘˜ such that ๐‘คโƒ— = ๐ด๐‘ฅ.โƒ— In
other words, for any vector ๐‘คโƒ— ∈ ๐‘‰, there exist numbers ๐‘ฅ1 , … , ๐‘ฅ๐‘˜ such that
๐‘˜
๐‘คโƒ— = ∑ ๐‘ฅ๐‘– ๐‘ฃ๐‘–โƒ— .
๐‘–=1
That’s exactly what it means for ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— to span the subspace ๐‘‰.
18.06.11: ‘Nullspace’
So we’ve proved our result: if
๐ด = ( ๐‘ฃ1โƒ— ๐‘ฃ2โƒ— โ‹ฏ ๐‘ฃ๐‘˜โƒ— ) ,
then
(1) ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— ∈ ๐‘‰ are linearly independent if and only if ๐‘‡๐ด is injective.
(2) ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— span ๐‘‰ if and only if ๐‘‡๐ด is surjective.
(3) ๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— are a basis of ๐‘‰ if and only if ๐‘‡๐ด is bijective.
18.06.11: ‘Nullspace’
Now back to our motivating example: we’ve been given a system of linear
equations
0โƒ— = ๐ด๐‘ฅ,โƒ—
where ๐ด is an ๐‘š × ๐‘› matrix. To solve this equation is to find a basis for the
kernel – aka nullspace – of ๐ด.
In other words, the objective is to find a list of linearly independent solutions
๐‘ฃ1โƒ— , … , ๐‘ฃ๐‘˜โƒ— such that any other solution can be written as a linear combination
of these! The dimension of ker(๐ด) – sometimes called the nullity of ๐ด – is the
number ๐‘˜.
18.06.11: ‘Nullspace’
There are two good ways of extracting a basis of the kernel. There’s a way using
row operations, and a way using column operations. You’ve been using these
for a while already, but here’s the way I think of these …
Suppose we can apply some row operations:
( ๐ด ๐ต )
( ๐ถ ๐ท ).
Here, ๐ด and ๐ถ are ๐‘š × ๐‘› matrices, and ๐ต and ๐ท are ๐‘š × ๐‘ matrices. What this
really means is that there’s an invertible ๐‘š × ๐‘š matrix ๐‘€ such that ๐‘€๐ด = ๐ถ
and ๐‘€๐ต = ๐ท. (And it turns out that any ๐‘€ can be built this way!)
That’s why it works to solve equations: if in the end ๐ถ = ๐ผ, then ๐‘€ = ๐ด−1 , and
๐ท = ๐ด−1 ๐ต.
18.06.11: ‘Nullspace’
So you can use row operations to whittle your favorite matrix down, and then
solve. This is nice because it’s so familiar. Let’s do a few examples together as a
team.
First, how about
−3 6 −1 1 −7
๐ด = ( 1 −2 2 3 −1 ) ?
2 −4 5 8 −4
18.06.11: ‘Nullspace’
One more:
2 2 −1 0
1
−1 −1 2 −3 1
๐ด=(
)
1 1 −2 0 −1
0 0
1
1
1
18.06.11: ‘Nullspace’
Column operations work exactly dual to row operations. (Just think of transposing, doing row operations, and transposing back!) So suppose we can apply
some column operations:
(
๐ด
)
๐ต
(
๐ถ
).
๐ท
Here, ๐ด and ๐ถ are ๐‘š × ๐‘› matrices, and ๐ต and ๐ท are ๐‘ × ๐‘› matrices. What this
really means is that there’s an invertible ๐‘› × ๐‘› matrix ๐‘ such that ๐ด๐‘ = ๐ถ and
๐ต๐‘ = ๐ท. (And it turns out that any ๐‘ can be built this way!)
18.06.11: ‘Nullspace’
Why is that a good idea? Well, we’re looking for vectors such that ๐ด๐‘ฅโƒ— = 0.โƒ— So
if we take
๐ด
(
)
๐ผ
where ๐ผ is the ๐‘› × ๐‘› identity matrix, then we can start using column operations
to get it to some
๐ถ
(
)
๐ท
So ๐ด๐ท = ๐ถ. So if ๐ถ has a column of zeroes, then the corresponding column of
๐ท will be a vector in the kernel. Furthermore, if you get ๐ถ into column echelon
form, then the nonzero column vectors of ๐ท lying under the zero columns of
๐ถ form a basis of ker(๐ด).
18.06.11: ‘Nullspace’
Let’s do this one again:
2 2 −1 0
1
−1 −1 2 −3 1
๐ด=(
)
1 1 −2 0 −1
0 0
1
1
1
18.06.11: ‘Nullspace’
Another:
1
0
๐ด=(
0
0
0
1
0
0
3
5
0
0
0 2 −8
0 −1 4
)
1 7 −9
0 0
0
Download